1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17#include "util/record.h"
18#include <traceevent/event-parse.h>
19#include <api/fs/tracing_path.h>
20#include <bpf/bpf.h>
21#include "util/bpf_map.h"
22#include "util/rlimit.h"
23#include "builtin.h"
24#include "util/cgroup.h"
25#include "util/color.h"
26#include "util/config.h"
27#include "util/debug.h"
28#include "util/dso.h"
29#include "util/env.h"
30#include "util/event.h"
31#include "util/evsel.h"
32#include "util/evsel_fprintf.h"
33#include "util/synthetic-events.h"
34#include "util/evlist.h"
35#include "util/evswitch.h"
36#include "util/mmap.h"
37#include <subcmd/pager.h>
38#include <subcmd/exec-cmd.h>
39#include "util/machine.h"
40#include "util/map.h"
41#include "util/symbol.h"
42#include "util/path.h"
43#include "util/session.h"
44#include "util/thread.h"
45#include <subcmd/parse-options.h>
46#include "util/strlist.h"
47#include "util/intlist.h"
48#include "util/thread_map.h"
49#include "util/stat.h"
50#include "util/tool.h"
51#include "util/util.h"
52#include "trace/beauty/beauty.h"
53#include "trace-event.h"
54#include "util/parse-events.h"
55#include "util/bpf-loader.h"
56#include "callchain.h"
57#include "print_binary.h"
58#include "string2.h"
59#include "syscalltbl.h"
60#include "rb_resort.h"
61#include "../perf.h"
62
63#include <errno.h>
64#include <inttypes.h>
65#include <poll.h>
66#include <signal.h>
67#include <stdlib.h>
68#include <string.h>
69#include <linux/err.h>
70#include <linux/filter.h>
71#include <linux/kernel.h>
72#include <linux/random.h>
73#include <linux/stringify.h>
74#include <linux/time64.h>
75#include <linux/zalloc.h>
76#include <fcntl.h>
77#include <sys/sysmacros.h>
78
79#include <linux/ctype.h>
80#include <perf/mmap.h>
81
82#ifndef O_CLOEXEC
83# define O_CLOEXEC 02000000
84#endif
85
86#ifndef F_LINUX_SPECIFIC_BASE
87# define F_LINUX_SPECIFIC_BASE 1024
88#endif
89
90
91
92
93struct syscall_arg_fmt {
94 size_t (*scnprintf)(char *bf, size_t size, struct syscall_arg *arg);
95 bool (*strtoul)(char *bf, size_t size, struct syscall_arg *arg, u64 *val);
96 unsigned long (*mask_val)(struct syscall_arg *arg, unsigned long val);
97 void *parm;
98 const char *name;
99 u16 nr_entries;
100 bool show_zero;
101};
102
103struct syscall_fmt {
104 const char *name;
105 const char *alias;
106 struct {
107 const char *sys_enter,
108 *sys_exit;
109 } bpf_prog_name;
110 struct syscall_arg_fmt arg[6];
111 u8 nr_args;
112 bool errpid;
113 bool timeout;
114 bool hexret;
115};
116
117struct trace {
118 struct perf_tool tool;
119 struct syscalltbl *sctbl;
120 struct {
121 struct syscall *table;
122 struct bpf_map *map;
123 struct {
124 struct bpf_map *sys_enter,
125 *sys_exit;
126 } prog_array;
127 struct {
128 struct evsel *sys_enter,
129 *sys_exit,
130 *augmented;
131 } events;
132 struct bpf_program *unaugmented_prog;
133 } syscalls;
134 struct {
135 struct bpf_map *map;
136 } dump;
137 struct record_opts opts;
138 struct evlist *evlist;
139 struct machine *host;
140 struct thread *current;
141 struct bpf_object *bpf_obj;
142 struct cgroup *cgroup;
143 u64 base_time;
144 FILE *output;
145 unsigned long nr_events;
146 unsigned long nr_events_printed;
147 unsigned long max_events;
148 struct evswitch evswitch;
149 struct strlist *ev_qualifier;
150 struct {
151 size_t nr;
152 int *entries;
153 } ev_qualifier_ids;
154 struct {
155 size_t nr;
156 pid_t *entries;
157 struct bpf_map *map;
158 } filter_pids;
159 double duration_filter;
160 double runtime_ms;
161 struct {
162 u64 vfs_getname,
163 proc_getname;
164 } stats;
165 unsigned int max_stack;
166 unsigned int min_stack;
167 int raw_augmented_syscalls_args_size;
168 bool raw_augmented_syscalls;
169 bool fd_path_disabled;
170 bool sort_events;
171 bool not_ev_qualifier;
172 bool live;
173 bool full_time;
174 bool sched;
175 bool multiple_threads;
176 bool summary;
177 bool summary_only;
178 bool errno_summary;
179 bool failure_only;
180 bool show_comm;
181 bool print_sample;
182 bool show_tool_stats;
183 bool trace_syscalls;
184 bool libtraceevent_print;
185 bool kernel_syscallchains;
186 s16 args_alignment;
187 bool show_tstamp;
188 bool show_duration;
189 bool show_zeros;
190 bool show_arg_names;
191 bool show_string_prefix;
192 bool force;
193 bool vfs_getname;
194 int trace_pgfaults;
195 char *perfconfig_events;
196 struct {
197 struct ordered_events data;
198 u64 last;
199 } oe;
200};
201
202struct tp_field {
203 int offset;
204 union {
205 u64 (*integer)(struct tp_field *field, struct perf_sample *sample);
206 void *(*pointer)(struct tp_field *field, struct perf_sample *sample);
207 };
208};
209
210#define TP_UINT_FIELD(bits) \
211static u64 tp_field__u##bits(struct tp_field *field, struct perf_sample *sample) \
212{ \
213 u##bits value; \
214 memcpy(&value, sample->raw_data + field->offset, sizeof(value)); \
215 return value; \
216}
217
218TP_UINT_FIELD(8);
219TP_UINT_FIELD(16);
220TP_UINT_FIELD(32);
221TP_UINT_FIELD(64);
222
223#define TP_UINT_FIELD__SWAPPED(bits) \
224static u64 tp_field__swapped_u##bits(struct tp_field *field, struct perf_sample *sample) \
225{ \
226 u##bits value; \
227 memcpy(&value, sample->raw_data + field->offset, sizeof(value)); \
228 return bswap_##bits(value);\
229}
230
231TP_UINT_FIELD__SWAPPED(16);
232TP_UINT_FIELD__SWAPPED(32);
233TP_UINT_FIELD__SWAPPED(64);
234
235static int __tp_field__init_uint(struct tp_field *field, int size, int offset, bool needs_swap)
236{
237 field->offset = offset;
238
239 switch (size) {
240 case 1:
241 field->integer = tp_field__u8;
242 break;
243 case 2:
244 field->integer = needs_swap ? tp_field__swapped_u16 : tp_field__u16;
245 break;
246 case 4:
247 field->integer = needs_swap ? tp_field__swapped_u32 : tp_field__u32;
248 break;
249 case 8:
250 field->integer = needs_swap ? tp_field__swapped_u64 : tp_field__u64;
251 break;
252 default:
253 return -1;
254 }
255
256 return 0;
257}
258
259static int tp_field__init_uint(struct tp_field *field, struct tep_format_field *format_field, bool needs_swap)
260{
261 return __tp_field__init_uint(field, format_field->size, format_field->offset, needs_swap);
262}
263
264static void *tp_field__ptr(struct tp_field *field, struct perf_sample *sample)
265{
266 return sample->raw_data + field->offset;
267}
268
269static int __tp_field__init_ptr(struct tp_field *field, int offset)
270{
271 field->offset = offset;
272 field->pointer = tp_field__ptr;
273 return 0;
274}
275
276static int tp_field__init_ptr(struct tp_field *field, struct tep_format_field *format_field)
277{
278 return __tp_field__init_ptr(field, format_field->offset);
279}
280
281struct syscall_tp {
282 struct tp_field id;
283 union {
284 struct tp_field args, ret;
285 };
286};
287
288
289
290
291
292
293struct evsel_trace {
294 struct syscall_tp sc;
295 struct syscall_arg_fmt *fmt;
296};
297
298static struct evsel_trace *evsel_trace__new(void)
299{
300 return zalloc(sizeof(struct evsel_trace));
301}
302
303static void evsel_trace__delete(struct evsel_trace *et)
304{
305 if (et == NULL)
306 return;
307
308 zfree(&et->fmt);
309 free(et);
310}
311
312
313
314
315
316static inline struct syscall_tp *__evsel__syscall_tp(struct evsel *evsel)
317{
318 struct evsel_trace *et = evsel->priv;
319
320 return &et->sc;
321}
322
323static struct syscall_tp *evsel__syscall_tp(struct evsel *evsel)
324{
325 if (evsel->priv == NULL) {
326 evsel->priv = evsel_trace__new();
327 if (evsel->priv == NULL)
328 return NULL;
329 }
330
331 return __evsel__syscall_tp(evsel);
332}
333
334
335
336
337static inline struct syscall_arg_fmt *__evsel__syscall_arg_fmt(struct evsel *evsel)
338{
339 struct evsel_trace *et = evsel->priv;
340
341 return et->fmt;
342}
343
344static struct syscall_arg_fmt *evsel__syscall_arg_fmt(struct evsel *evsel)
345{
346 struct evsel_trace *et = evsel->priv;
347
348 if (evsel->priv == NULL) {
349 et = evsel->priv = evsel_trace__new();
350
351 if (et == NULL)
352 return NULL;
353 }
354
355 if (et->fmt == NULL) {
356 et->fmt = calloc(evsel->tp_format->format.nr_fields, sizeof(struct syscall_arg_fmt));
357 if (et->fmt == NULL)
358 goto out_delete;
359 }
360
361 return __evsel__syscall_arg_fmt(evsel);
362
363out_delete:
364 evsel_trace__delete(evsel->priv);
365 evsel->priv = NULL;
366 return NULL;
367}
368
369static int evsel__init_tp_uint_field(struct evsel *evsel, struct tp_field *field, const char *name)
370{
371 struct tep_format_field *format_field = evsel__field(evsel, name);
372
373 if (format_field == NULL)
374 return -1;
375
376 return tp_field__init_uint(field, format_field, evsel->needs_swap);
377}
378
379#define perf_evsel__init_sc_tp_uint_field(evsel, name) \
380 ({ struct syscall_tp *sc = __evsel__syscall_tp(evsel);\
381 evsel__init_tp_uint_field(evsel, &sc->name, #name); })
382
383static int evsel__init_tp_ptr_field(struct evsel *evsel, struct tp_field *field, const char *name)
384{
385 struct tep_format_field *format_field = evsel__field(evsel, name);
386
387 if (format_field == NULL)
388 return -1;
389
390 return tp_field__init_ptr(field, format_field);
391}
392
393#define perf_evsel__init_sc_tp_ptr_field(evsel, name) \
394 ({ struct syscall_tp *sc = __evsel__syscall_tp(evsel);\
395 evsel__init_tp_ptr_field(evsel, &sc->name, #name); })
396
397static void evsel__delete_priv(struct evsel *evsel)
398{
399 zfree(&evsel->priv);
400 evsel__delete(evsel);
401}
402
403static int evsel__init_syscall_tp(struct evsel *evsel)
404{
405 struct syscall_tp *sc = evsel__syscall_tp(evsel);
406
407 if (sc != NULL) {
408 if (evsel__init_tp_uint_field(evsel, &sc->id, "__syscall_nr") &&
409 evsel__init_tp_uint_field(evsel, &sc->id, "nr"))
410 return -ENOENT;
411 return 0;
412 }
413
414 return -ENOMEM;
415}
416
417static int evsel__init_augmented_syscall_tp(struct evsel *evsel, struct evsel *tp)
418{
419 struct syscall_tp *sc = evsel__syscall_tp(evsel);
420
421 if (sc != NULL) {
422 struct tep_format_field *syscall_id = evsel__field(tp, "id");
423 if (syscall_id == NULL)
424 syscall_id = evsel__field(tp, "__syscall_nr");
425 if (syscall_id == NULL ||
426 __tp_field__init_uint(&sc->id, syscall_id->size, syscall_id->offset, evsel->needs_swap))
427 return -EINVAL;
428
429 return 0;
430 }
431
432 return -ENOMEM;
433}
434
435static int evsel__init_augmented_syscall_tp_args(struct evsel *evsel)
436{
437 struct syscall_tp *sc = __evsel__syscall_tp(evsel);
438
439 return __tp_field__init_ptr(&sc->args, sc->id.offset + sizeof(u64));
440}
441
442static int evsel__init_augmented_syscall_tp_ret(struct evsel *evsel)
443{
444 struct syscall_tp *sc = __evsel__syscall_tp(evsel);
445
446 return __tp_field__init_uint(&sc->ret, sizeof(u64), sc->id.offset + sizeof(u64), evsel->needs_swap);
447}
448
449static int evsel__init_raw_syscall_tp(struct evsel *evsel, void *handler)
450{
451 if (evsel__syscall_tp(evsel) != NULL) {
452 if (perf_evsel__init_sc_tp_uint_field(evsel, id))
453 return -ENOENT;
454
455 evsel->handler = handler;
456 return 0;
457 }
458
459 return -ENOMEM;
460}
461
462static struct evsel *perf_evsel__raw_syscall_newtp(const char *direction, void *handler)
463{
464 struct evsel *evsel = evsel__newtp("raw_syscalls", direction);
465
466
467 if (IS_ERR(evsel))
468 evsel = evsel__newtp("syscalls", direction);
469
470 if (IS_ERR(evsel))
471 return NULL;
472
473 if (evsel__init_raw_syscall_tp(evsel, handler))
474 goto out_delete;
475
476 return evsel;
477
478out_delete:
479 evsel__delete_priv(evsel);
480 return NULL;
481}
482
483#define perf_evsel__sc_tp_uint(evsel, name, sample) \
484 ({ struct syscall_tp *fields = __evsel__syscall_tp(evsel); \
485 fields->name.integer(&fields->name, sample); })
486
487#define perf_evsel__sc_tp_ptr(evsel, name, sample) \
488 ({ struct syscall_tp *fields = __evsel__syscall_tp(evsel); \
489 fields->name.pointer(&fields->name, sample); })
490
491size_t strarray__scnprintf_suffix(struct strarray *sa, char *bf, size_t size, const char *intfmt, bool show_suffix, int val)
492{
493 int idx = val - sa->offset;
494
495 if (idx < 0 || idx >= sa->nr_entries || sa->entries[idx] == NULL) {
496 size_t printed = scnprintf(bf, size, intfmt, val);
497 if (show_suffix)
498 printed += scnprintf(bf + printed, size - printed, " /* %s??? */", sa->prefix);
499 return printed;
500 }
501
502 return scnprintf(bf, size, "%s%s", sa->entries[idx], show_suffix ? sa->prefix : "");
503}
504
505size_t strarray__scnprintf(struct strarray *sa, char *bf, size_t size, const char *intfmt, bool show_prefix, int val)
506{
507 int idx = val - sa->offset;
508
509 if (idx < 0 || idx >= sa->nr_entries || sa->entries[idx] == NULL) {
510 size_t printed = scnprintf(bf, size, intfmt, val);
511 if (show_prefix)
512 printed += scnprintf(bf + printed, size - printed, " /* %s??? */", sa->prefix);
513 return printed;
514 }
515
516 return scnprintf(bf, size, "%s%s", show_prefix ? sa->prefix : "", sa->entries[idx]);
517}
518
519static size_t __syscall_arg__scnprintf_strarray(char *bf, size_t size,
520 const char *intfmt,
521 struct syscall_arg *arg)
522{
523 return strarray__scnprintf(arg->parm, bf, size, intfmt, arg->show_string_prefix, arg->val);
524}
525
526static size_t syscall_arg__scnprintf_strarray(char *bf, size_t size,
527 struct syscall_arg *arg)
528{
529 return __syscall_arg__scnprintf_strarray(bf, size, "%d", arg);
530}
531
532#define SCA_STRARRAY syscall_arg__scnprintf_strarray
533
534bool syscall_arg__strtoul_strarray(char *bf, size_t size, struct syscall_arg *arg, u64 *ret)
535{
536 return strarray__strtoul(arg->parm, bf, size, ret);
537}
538
539bool syscall_arg__strtoul_strarray_flags(char *bf, size_t size, struct syscall_arg *arg, u64 *ret)
540{
541 return strarray__strtoul_flags(arg->parm, bf, size, ret);
542}
543
544bool syscall_arg__strtoul_strarrays(char *bf, size_t size, struct syscall_arg *arg, u64 *ret)
545{
546 return strarrays__strtoul(arg->parm, bf, size, ret);
547}
548
549size_t syscall_arg__scnprintf_strarray_flags(char *bf, size_t size, struct syscall_arg *arg)
550{
551 return strarray__scnprintf_flags(arg->parm, bf, size, arg->show_string_prefix, arg->val);
552}
553
554size_t strarrays__scnprintf(struct strarrays *sas, char *bf, size_t size, const char *intfmt, bool show_prefix, int val)
555{
556 size_t printed;
557 int i;
558
559 for (i = 0; i < sas->nr_entries; ++i) {
560 struct strarray *sa = sas->entries[i];
561 int idx = val - sa->offset;
562
563 if (idx >= 0 && idx < sa->nr_entries) {
564 if (sa->entries[idx] == NULL)
565 break;
566 return scnprintf(bf, size, "%s%s", show_prefix ? sa->prefix : "", sa->entries[idx]);
567 }
568 }
569
570 printed = scnprintf(bf, size, intfmt, val);
571 if (show_prefix)
572 printed += scnprintf(bf + printed, size - printed, " /* %s??? */", sas->entries[0]->prefix);
573 return printed;
574}
575
576bool strarray__strtoul(struct strarray *sa, char *bf, size_t size, u64 *ret)
577{
578 int i;
579
580 for (i = 0; i < sa->nr_entries; ++i) {
581 if (sa->entries[i] && strncmp(sa->entries[i], bf, size) == 0 && sa->entries[i][size] == '\0') {
582 *ret = sa->offset + i;
583 return true;
584 }
585 }
586
587 return false;
588}
589
590bool strarray__strtoul_flags(struct strarray *sa, char *bf, size_t size, u64 *ret)
591{
592 u64 val = 0;
593 char *tok = bf, *sep, *end;
594
595 *ret = 0;
596
597 while (size != 0) {
598 int toklen = size;
599
600 sep = memchr(tok, '|', size);
601 if (sep != NULL) {
602 size -= sep - tok + 1;
603
604 end = sep - 1;
605 while (end > tok && isspace(*end))
606 --end;
607
608 toklen = end - tok + 1;
609 }
610
611 while (isspace(*tok))
612 ++tok;
613
614 if (isalpha(*tok) || *tok == '_') {
615 if (!strarray__strtoul(sa, tok, toklen, &val))
616 return false;
617 } else {
618 bool is_hexa = tok[0] == 0 && (tok[1] = 'x' || tok[1] == 'X');
619
620 val = strtoul(tok, NULL, is_hexa ? 16 : 0);
621 }
622
623 *ret |= (1 << (val - 1));
624
625 if (sep == NULL)
626 break;
627 tok = sep + 1;
628 }
629
630 return true;
631}
632
633bool strarrays__strtoul(struct strarrays *sas, char *bf, size_t size, u64 *ret)
634{
635 int i;
636
637 for (i = 0; i < sas->nr_entries; ++i) {
638 struct strarray *sa = sas->entries[i];
639
640 if (strarray__strtoul(sa, bf, size, ret))
641 return true;
642 }
643
644 return false;
645}
646
647size_t syscall_arg__scnprintf_strarrays(char *bf, size_t size,
648 struct syscall_arg *arg)
649{
650 return strarrays__scnprintf(arg->parm, bf, size, "%d", arg->show_string_prefix, arg->val);
651}
652
653#ifndef AT_FDCWD
654#define AT_FDCWD -100
655#endif
656
657static size_t syscall_arg__scnprintf_fd_at(char *bf, size_t size,
658 struct syscall_arg *arg)
659{
660 int fd = arg->val;
661 const char *prefix = "AT_FD";
662
663 if (fd == AT_FDCWD)
664 return scnprintf(bf, size, "%s%s", arg->show_string_prefix ? prefix : "", "CWD");
665
666 return syscall_arg__scnprintf_fd(bf, size, arg);
667}
668
669#define SCA_FDAT syscall_arg__scnprintf_fd_at
670
671static size_t syscall_arg__scnprintf_close_fd(char *bf, size_t size,
672 struct syscall_arg *arg);
673
674#define SCA_CLOSE_FD syscall_arg__scnprintf_close_fd
675
676size_t syscall_arg__scnprintf_hex(char *bf, size_t size, struct syscall_arg *arg)
677{
678 return scnprintf(bf, size, "%#lx", arg->val);
679}
680
681size_t syscall_arg__scnprintf_ptr(char *bf, size_t size, struct syscall_arg *arg)
682{
683 if (arg->val == 0)
684 return scnprintf(bf, size, "NULL");
685 return syscall_arg__scnprintf_hex(bf, size, arg);
686}
687
688size_t syscall_arg__scnprintf_int(char *bf, size_t size, struct syscall_arg *arg)
689{
690 return scnprintf(bf, size, "%d", arg->val);
691}
692
693size_t syscall_arg__scnprintf_long(char *bf, size_t size, struct syscall_arg *arg)
694{
695 return scnprintf(bf, size, "%ld", arg->val);
696}
697
698static size_t syscall_arg__scnprintf_char_array(char *bf, size_t size, struct syscall_arg *arg)
699{
700
701
702
703 return scnprintf(bf, size, "\"%-.*s\"", arg->fmt->nr_entries ?: arg->len, arg->val);
704}
705
706#define SCA_CHAR_ARRAY syscall_arg__scnprintf_char_array
707
708static const char *bpf_cmd[] = {
709 "MAP_CREATE", "MAP_LOOKUP_ELEM", "MAP_UPDATE_ELEM", "MAP_DELETE_ELEM",
710 "MAP_GET_NEXT_KEY", "PROG_LOAD",
711};
712static DEFINE_STRARRAY(bpf_cmd, "BPF_");
713
714static const char *fsmount_flags[] = {
715 [1] = "CLOEXEC",
716};
717static DEFINE_STRARRAY(fsmount_flags, "FSMOUNT_");
718
719#include "trace/beauty/generated/fsconfig_arrays.c"
720
721static DEFINE_STRARRAY(fsconfig_cmds, "FSCONFIG_");
722
723static const char *epoll_ctl_ops[] = { "ADD", "DEL", "MOD", };
724static DEFINE_STRARRAY_OFFSET(epoll_ctl_ops, "EPOLL_CTL_", 1);
725
726static const char *itimers[] = { "REAL", "VIRTUAL", "PROF", };
727static DEFINE_STRARRAY(itimers, "ITIMER_");
728
729static const char *keyctl_options[] = {
730 "GET_KEYRING_ID", "JOIN_SESSION_KEYRING", "UPDATE", "REVOKE", "CHOWN",
731 "SETPERM", "DESCRIBE", "CLEAR", "LINK", "UNLINK", "SEARCH", "READ",
732 "INSTANTIATE", "NEGATE", "SET_REQKEY_KEYRING", "SET_TIMEOUT",
733 "ASSUME_AUTHORITY", "GET_SECURITY", "SESSION_TO_PARENT", "REJECT",
734 "INSTANTIATE_IOV", "INVALIDATE", "GET_PERSISTENT",
735};
736static DEFINE_STRARRAY(keyctl_options, "KEYCTL_");
737
738static const char *whences[] = { "SET", "CUR", "END",
739#ifdef SEEK_DATA
740"DATA",
741#endif
742#ifdef SEEK_HOLE
743"HOLE",
744#endif
745};
746static DEFINE_STRARRAY(whences, "SEEK_");
747
748static const char *fcntl_cmds[] = {
749 "DUPFD", "GETFD", "SETFD", "GETFL", "SETFL", "GETLK", "SETLK",
750 "SETLKW", "SETOWN", "GETOWN", "SETSIG", "GETSIG", "GETLK64",
751 "SETLK64", "SETLKW64", "SETOWN_EX", "GETOWN_EX",
752 "GETOWNER_UIDS",
753};
754static DEFINE_STRARRAY(fcntl_cmds, "F_");
755
756static const char *fcntl_linux_specific_cmds[] = {
757 "SETLEASE", "GETLEASE", "NOTIFY", [5] = "CANCELLK", "DUPFD_CLOEXEC",
758 "SETPIPE_SZ", "GETPIPE_SZ", "ADD_SEALS", "GET_SEALS",
759 "GET_RW_HINT", "SET_RW_HINT", "GET_FILE_RW_HINT", "SET_FILE_RW_HINT",
760};
761
762static DEFINE_STRARRAY_OFFSET(fcntl_linux_specific_cmds, "F_", F_LINUX_SPECIFIC_BASE);
763
764static struct strarray *fcntl_cmds_arrays[] = {
765 &strarray__fcntl_cmds,
766 &strarray__fcntl_linux_specific_cmds,
767};
768
769static DEFINE_STRARRAYS(fcntl_cmds_arrays);
770
771static const char *rlimit_resources[] = {
772 "CPU", "FSIZE", "DATA", "STACK", "CORE", "RSS", "NPROC", "NOFILE",
773 "MEMLOCK", "AS", "LOCKS", "SIGPENDING", "MSGQUEUE", "NICE", "RTPRIO",
774 "RTTIME",
775};
776static DEFINE_STRARRAY(rlimit_resources, "RLIMIT_");
777
778static const char *sighow[] = { "BLOCK", "UNBLOCK", "SETMASK", };
779static DEFINE_STRARRAY(sighow, "SIG_");
780
781static const char *clockid[] = {
782 "REALTIME", "MONOTONIC", "PROCESS_CPUTIME_ID", "THREAD_CPUTIME_ID",
783 "MONOTONIC_RAW", "REALTIME_COARSE", "MONOTONIC_COARSE", "BOOTTIME",
784 "REALTIME_ALARM", "BOOTTIME_ALARM", "SGI_CYCLE", "TAI"
785};
786static DEFINE_STRARRAY(clockid, "CLOCK_");
787
788static size_t syscall_arg__scnprintf_access_mode(char *bf, size_t size,
789 struct syscall_arg *arg)
790{
791 bool show_prefix = arg->show_string_prefix;
792 const char *suffix = "_OK";
793 size_t printed = 0;
794 int mode = arg->val;
795
796 if (mode == F_OK)
797 return scnprintf(bf, size, "F%s", show_prefix ? suffix : "");
798#define P_MODE(n) \
799 if (mode & n##_OK) { \
800 printed += scnprintf(bf + printed, size - printed, "%s%s", #n, show_prefix ? suffix : ""); \
801 mode &= ~n##_OK; \
802 }
803
804 P_MODE(R);
805 P_MODE(W);
806 P_MODE(X);
807#undef P_MODE
808
809 if (mode)
810 printed += scnprintf(bf + printed, size - printed, "|%#x", mode);
811
812 return printed;
813}
814
815#define SCA_ACCMODE syscall_arg__scnprintf_access_mode
816
817static size_t syscall_arg__scnprintf_filename(char *bf, size_t size,
818 struct syscall_arg *arg);
819
820#define SCA_FILENAME syscall_arg__scnprintf_filename
821
822static size_t syscall_arg__scnprintf_pipe_flags(char *bf, size_t size,
823 struct syscall_arg *arg)
824{
825 bool show_prefix = arg->show_string_prefix;
826 const char *prefix = "O_";
827 int printed = 0, flags = arg->val;
828
829#define P_FLAG(n) \
830 if (flags & O_##n) { \
831 printed += scnprintf(bf + printed, size - printed, "%s%s%s", printed ? "|" : "", show_prefix ? prefix : "", #n); \
832 flags &= ~O_##n; \
833 }
834
835 P_FLAG(CLOEXEC);
836 P_FLAG(NONBLOCK);
837#undef P_FLAG
838
839 if (flags)
840 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags);
841
842 return printed;
843}
844
845#define SCA_PIPE_FLAGS syscall_arg__scnprintf_pipe_flags
846
847#ifndef GRND_NONBLOCK
848#define GRND_NONBLOCK 0x0001
849#endif
850#ifndef GRND_RANDOM
851#define GRND_RANDOM 0x0002
852#endif
853
854static size_t syscall_arg__scnprintf_getrandom_flags(char *bf, size_t size,
855 struct syscall_arg *arg)
856{
857 bool show_prefix = arg->show_string_prefix;
858 const char *prefix = "GRND_";
859 int printed = 0, flags = arg->val;
860
861#define P_FLAG(n) \
862 if (flags & GRND_##n) { \
863 printed += scnprintf(bf + printed, size - printed, "%s%s%s", printed ? "|" : "", show_prefix ? prefix : "", #n); \
864 flags &= ~GRND_##n; \
865 }
866
867 P_FLAG(RANDOM);
868 P_FLAG(NONBLOCK);
869#undef P_FLAG
870
871 if (flags)
872 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags);
873
874 return printed;
875}
876
877#define SCA_GETRANDOM_FLAGS syscall_arg__scnprintf_getrandom_flags
878
879#define STRARRAY(name, array) \
880 { .scnprintf = SCA_STRARRAY, \
881 .strtoul = STUL_STRARRAY, \
882 .parm = &strarray__##array, }
883
884#define STRARRAY_FLAGS(name, array) \
885 { .scnprintf = SCA_STRARRAY_FLAGS, \
886 .strtoul = STUL_STRARRAY_FLAGS, \
887 .parm = &strarray__##array, }
888
889#include "trace/beauty/arch_errno_names.c"
890#include "trace/beauty/eventfd.c"
891#include "trace/beauty/futex_op.c"
892#include "trace/beauty/futex_val3.c"
893#include "trace/beauty/mmap.c"
894#include "trace/beauty/mode_t.c"
895#include "trace/beauty/msg_flags.c"
896#include "trace/beauty/open_flags.c"
897#include "trace/beauty/perf_event_open.c"
898#include "trace/beauty/pid.c"
899#include "trace/beauty/sched_policy.c"
900#include "trace/beauty/seccomp.c"
901#include "trace/beauty/signum.c"
902#include "trace/beauty/socket_type.c"
903#include "trace/beauty/waitid_options.c"
904
905static struct syscall_fmt syscall_fmts[] = {
906 { .name = "access",
907 .arg = { [1] = { .scnprintf = SCA_ACCMODE, }, }, },
908 { .name = "arch_prctl",
909 .arg = { [0] = { .scnprintf = SCA_X86_ARCH_PRCTL_CODE, },
910 [1] = { .scnprintf = SCA_PTR, }, }, },
911 { .name = "bind",
912 .arg = { [0] = { .scnprintf = SCA_INT, },
913 [1] = { .scnprintf = SCA_SOCKADDR, },
914 [2] = { .scnprintf = SCA_INT, }, }, },
915 { .name = "bpf",
916 .arg = { [0] = STRARRAY(cmd, bpf_cmd), }, },
917 { .name = "brk", .hexret = true,
918 .arg = { [0] = { .scnprintf = SCA_PTR, }, }, },
919 { .name = "clock_gettime",
920 .arg = { [0] = STRARRAY(clk_id, clockid), }, },
921 { .name = "clone", .errpid = true, .nr_args = 5,
922 .arg = { [0] = { .name = "flags", .scnprintf = SCA_CLONE_FLAGS, },
923 [1] = { .name = "child_stack", .scnprintf = SCA_HEX, },
924 [2] = { .name = "parent_tidptr", .scnprintf = SCA_HEX, },
925 [3] = { .name = "child_tidptr", .scnprintf = SCA_HEX, },
926 [4] = { .name = "tls", .scnprintf = SCA_HEX, }, }, },
927 { .name = "close",
928 .arg = { [0] = { .scnprintf = SCA_CLOSE_FD, }, }, },
929 { .name = "connect",
930 .arg = { [0] = { .scnprintf = SCA_INT, },
931 [1] = { .scnprintf = SCA_SOCKADDR, },
932 [2] = { .scnprintf = SCA_INT, }, }, },
933 { .name = "epoll_ctl",
934 .arg = { [1] = STRARRAY(op, epoll_ctl_ops), }, },
935 { .name = "eventfd2",
936 .arg = { [1] = { .scnprintf = SCA_EFD_FLAGS, }, }, },
937 { .name = "fchmodat",
938 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
939 { .name = "fchownat",
940 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
941 { .name = "fcntl",
942 .arg = { [1] = { .scnprintf = SCA_FCNTL_CMD,
943 .strtoul = STUL_STRARRAYS,
944 .parm = &strarrays__fcntl_cmds_arrays,
945 .show_zero = true, },
946 [2] = { .scnprintf = SCA_FCNTL_ARG, }, }, },
947 { .name = "flock",
948 .arg = { [1] = { .scnprintf = SCA_FLOCK, }, }, },
949 { .name = "fsconfig",
950 .arg = { [1] = STRARRAY(cmd, fsconfig_cmds), }, },
951 { .name = "fsmount",
952 .arg = { [1] = STRARRAY_FLAGS(flags, fsmount_flags),
953 [2] = { .scnprintf = SCA_FSMOUNT_ATTR_FLAGS, }, }, },
954 { .name = "fspick",
955 .arg = { [0] = { .scnprintf = SCA_FDAT, },
956 [1] = { .scnprintf = SCA_FILENAME, },
957 [2] = { .scnprintf = SCA_FSPICK_FLAGS, }, }, },
958 { .name = "fstat", .alias = "newfstat", },
959 { .name = "fstatat", .alias = "newfstatat", },
960 { .name = "futex",
961 .arg = { [1] = { .scnprintf = SCA_FUTEX_OP, },
962 [5] = { .scnprintf = SCA_FUTEX_VAL3, }, }, },
963 { .name = "futimesat",
964 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
965 { .name = "getitimer",
966 .arg = { [0] = STRARRAY(which, itimers), }, },
967 { .name = "getpid", .errpid = true, },
968 { .name = "getpgid", .errpid = true, },
969 { .name = "getppid", .errpid = true, },
970 { .name = "getrandom",
971 .arg = { [2] = { .scnprintf = SCA_GETRANDOM_FLAGS, }, }, },
972 { .name = "getrlimit",
973 .arg = { [0] = STRARRAY(resource, rlimit_resources), }, },
974 { .name = "gettid", .errpid = true, },
975 { .name = "ioctl",
976 .arg = {
977#if defined(__i386__) || defined(__x86_64__)
978
979
980
981 [1] = { .scnprintf = SCA_IOCTL_CMD, },
982 [2] = { .scnprintf = SCA_HEX, }, }, },
983#else
984 [2] = { .scnprintf = SCA_HEX, }, }, },
985#endif
986 { .name = "kcmp", .nr_args = 5,
987 .arg = { [0] = { .name = "pid1", .scnprintf = SCA_PID, },
988 [1] = { .name = "pid2", .scnprintf = SCA_PID, },
989 [2] = { .name = "type", .scnprintf = SCA_KCMP_TYPE, },
990 [3] = { .name = "idx1", .scnprintf = SCA_KCMP_IDX, },
991 [4] = { .name = "idx2", .scnprintf = SCA_KCMP_IDX, }, }, },
992 { .name = "keyctl",
993 .arg = { [0] = STRARRAY(option, keyctl_options), }, },
994 { .name = "kill",
995 .arg = { [1] = { .scnprintf = SCA_SIGNUM, }, }, },
996 { .name = "linkat",
997 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
998 { .name = "lseek",
999 .arg = { [2] = STRARRAY(whence, whences), }, },
1000 { .name = "lstat", .alias = "newlstat", },
1001 { .name = "madvise",
1002 .arg = { [0] = { .scnprintf = SCA_HEX, },
1003 [2] = { .scnprintf = SCA_MADV_BHV, }, }, },
1004 { .name = "mkdirat",
1005 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1006 { .name = "mknodat",
1007 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1008 { .name = "mmap", .hexret = true,
1009
1010#if defined(__s390x__)
1011 .alias = "old_mmap",
1012#endif
1013 .arg = { [2] = { .scnprintf = SCA_MMAP_PROT, },
1014 [3] = { .scnprintf = SCA_MMAP_FLAGS,
1015 .strtoul = STUL_STRARRAY_FLAGS,
1016 .parm = &strarray__mmap_flags, },
1017 [5] = { .scnprintf = SCA_HEX, }, }, },
1018 { .name = "mount",
1019 .arg = { [0] = { .scnprintf = SCA_FILENAME, },
1020 [3] = { .scnprintf = SCA_MOUNT_FLAGS,
1021 .mask_val = SCAMV_MOUNT_FLAGS, }, }, },
1022 { .name = "move_mount",
1023 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1024 [1] = { .scnprintf = SCA_FILENAME, },
1025 [2] = { .scnprintf = SCA_FDAT, },
1026 [3] = { .scnprintf = SCA_FILENAME, },
1027 [4] = { .scnprintf = SCA_MOVE_MOUNT_FLAGS, }, }, },
1028 { .name = "mprotect",
1029 .arg = { [0] = { .scnprintf = SCA_HEX, },
1030 [2] = { .scnprintf = SCA_MMAP_PROT, }, }, },
1031 { .name = "mq_unlink",
1032 .arg = { [0] = { .scnprintf = SCA_FILENAME, }, }, },
1033 { .name = "mremap", .hexret = true,
1034 .arg = { [3] = { .scnprintf = SCA_MREMAP_FLAGS, }, }, },
1035 { .name = "name_to_handle_at",
1036 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1037 { .name = "newfstatat",
1038 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1039 { .name = "open",
1040 .arg = { [1] = { .scnprintf = SCA_OPEN_FLAGS, }, }, },
1041 { .name = "open_by_handle_at",
1042 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1043 [2] = { .scnprintf = SCA_OPEN_FLAGS, }, }, },
1044 { .name = "openat",
1045 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1046 [2] = { .scnprintf = SCA_OPEN_FLAGS, }, }, },
1047 { .name = "perf_event_open",
1048 .arg = { [2] = { .scnprintf = SCA_INT, },
1049 [3] = { .scnprintf = SCA_FD, },
1050 [4] = { .scnprintf = SCA_PERF_FLAGS, }, }, },
1051 { .name = "pipe2",
1052 .arg = { [1] = { .scnprintf = SCA_PIPE_FLAGS, }, }, },
1053 { .name = "pkey_alloc",
1054 .arg = { [1] = { .scnprintf = SCA_PKEY_ALLOC_ACCESS_RIGHTS, }, }, },
1055 { .name = "pkey_free",
1056 .arg = { [0] = { .scnprintf = SCA_INT, }, }, },
1057 { .name = "pkey_mprotect",
1058 .arg = { [0] = { .scnprintf = SCA_HEX, },
1059 [2] = { .scnprintf = SCA_MMAP_PROT, },
1060 [3] = { .scnprintf = SCA_INT, }, }, },
1061 { .name = "poll", .timeout = true, },
1062 { .name = "ppoll", .timeout = true, },
1063 { .name = "prctl",
1064 .arg = { [0] = { .scnprintf = SCA_PRCTL_OPTION,
1065 .strtoul = STUL_STRARRAY,
1066 .parm = &strarray__prctl_options, },
1067 [1] = { .scnprintf = SCA_PRCTL_ARG2, },
1068 [2] = { .scnprintf = SCA_PRCTL_ARG3, }, }, },
1069 { .name = "pread", .alias = "pread64", },
1070 { .name = "preadv", .alias = "pread", },
1071 { .name = "prlimit64",
1072 .arg = { [1] = STRARRAY(resource, rlimit_resources), }, },
1073 { .name = "pwrite", .alias = "pwrite64", },
1074 { .name = "readlinkat",
1075 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1076 { .name = "recvfrom",
1077 .arg = { [3] = { .scnprintf = SCA_MSG_FLAGS, }, }, },
1078 { .name = "recvmmsg",
1079 .arg = { [3] = { .scnprintf = SCA_MSG_FLAGS, }, }, },
1080 { .name = "recvmsg",
1081 .arg = { [2] = { .scnprintf = SCA_MSG_FLAGS, }, }, },
1082 { .name = "renameat",
1083 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1084 [2] = { .scnprintf = SCA_FDAT, }, }, },
1085 { .name = "renameat2",
1086 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1087 [2] = { .scnprintf = SCA_FDAT, },
1088 [4] = { .scnprintf = SCA_RENAMEAT2_FLAGS, }, }, },
1089 { .name = "rt_sigaction",
1090 .arg = { [0] = { .scnprintf = SCA_SIGNUM, }, }, },
1091 { .name = "rt_sigprocmask",
1092 .arg = { [0] = STRARRAY(how, sighow), }, },
1093 { .name = "rt_sigqueueinfo",
1094 .arg = { [1] = { .scnprintf = SCA_SIGNUM, }, }, },
1095 { .name = "rt_tgsigqueueinfo",
1096 .arg = { [2] = { .scnprintf = SCA_SIGNUM, }, }, },
1097 { .name = "sched_setscheduler",
1098 .arg = { [1] = { .scnprintf = SCA_SCHED_POLICY, }, }, },
1099 { .name = "seccomp",
1100 .arg = { [0] = { .scnprintf = SCA_SECCOMP_OP, },
1101 [1] = { .scnprintf = SCA_SECCOMP_FLAGS, }, }, },
1102 { .name = "select", .timeout = true, },
1103 { .name = "sendfile", .alias = "sendfile64", },
1104 { .name = "sendmmsg",
1105 .arg = { [3] = { .scnprintf = SCA_MSG_FLAGS, }, }, },
1106 { .name = "sendmsg",
1107 .arg = { [2] = { .scnprintf = SCA_MSG_FLAGS, }, }, },
1108 { .name = "sendto",
1109 .arg = { [3] = { .scnprintf = SCA_MSG_FLAGS, },
1110 [4] = { .scnprintf = SCA_SOCKADDR, }, }, },
1111 { .name = "set_tid_address", .errpid = true, },
1112 { .name = "setitimer",
1113 .arg = { [0] = STRARRAY(which, itimers), }, },
1114 { .name = "setrlimit",
1115 .arg = { [0] = STRARRAY(resource, rlimit_resources), }, },
1116 { .name = "socket",
1117 .arg = { [0] = STRARRAY(family, socket_families),
1118 [1] = { .scnprintf = SCA_SK_TYPE, },
1119 [2] = { .scnprintf = SCA_SK_PROTO, }, }, },
1120 { .name = "socketpair",
1121 .arg = { [0] = STRARRAY(family, socket_families),
1122 [1] = { .scnprintf = SCA_SK_TYPE, },
1123 [2] = { .scnprintf = SCA_SK_PROTO, }, }, },
1124 { .name = "stat", .alias = "newstat", },
1125 { .name = "statx",
1126 .arg = { [0] = { .scnprintf = SCA_FDAT, },
1127 [2] = { .scnprintf = SCA_STATX_FLAGS, } ,
1128 [3] = { .scnprintf = SCA_STATX_MASK, }, }, },
1129 { .name = "swapoff",
1130 .arg = { [0] = { .scnprintf = SCA_FILENAME, }, }, },
1131 { .name = "swapon",
1132 .arg = { [0] = { .scnprintf = SCA_FILENAME, }, }, },
1133 { .name = "symlinkat",
1134 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1135 { .name = "sync_file_range",
1136 .arg = { [3] = { .scnprintf = SCA_SYNC_FILE_RANGE_FLAGS, }, }, },
1137 { .name = "tgkill",
1138 .arg = { [2] = { .scnprintf = SCA_SIGNUM, }, }, },
1139 { .name = "tkill",
1140 .arg = { [1] = { .scnprintf = SCA_SIGNUM, }, }, },
1141 { .name = "umount2", .alias = "umount",
1142 .arg = { [0] = { .scnprintf = SCA_FILENAME, }, }, },
1143 { .name = "uname", .alias = "newuname", },
1144 { .name = "unlinkat",
1145 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1146 { .name = "utimensat",
1147 .arg = { [0] = { .scnprintf = SCA_FDAT, }, }, },
1148 { .name = "wait4", .errpid = true,
1149 .arg = { [2] = { .scnprintf = SCA_WAITID_OPTIONS, }, }, },
1150 { .name = "waitid", .errpid = true,
1151 .arg = { [3] = { .scnprintf = SCA_WAITID_OPTIONS, }, }, },
1152};
1153
1154static int syscall_fmt__cmp(const void *name, const void *fmtp)
1155{
1156 const struct syscall_fmt *fmt = fmtp;
1157 return strcmp(name, fmt->name);
1158}
1159
1160static struct syscall_fmt *__syscall_fmt__find(struct syscall_fmt *fmts, const int nmemb, const char *name)
1161{
1162 return bsearch(name, fmts, nmemb, sizeof(struct syscall_fmt), syscall_fmt__cmp);
1163}
1164
1165static struct syscall_fmt *syscall_fmt__find(const char *name)
1166{
1167 const int nmemb = ARRAY_SIZE(syscall_fmts);
1168 return __syscall_fmt__find(syscall_fmts, nmemb, name);
1169}
1170
1171static struct syscall_fmt *__syscall_fmt__find_by_alias(struct syscall_fmt *fmts, const int nmemb, const char *alias)
1172{
1173 int i;
1174
1175 for (i = 0; i < nmemb; ++i) {
1176 if (fmts[i].alias && strcmp(fmts[i].alias, alias) == 0)
1177 return &fmts[i];
1178 }
1179
1180 return NULL;
1181}
1182
1183static struct syscall_fmt *syscall_fmt__find_by_alias(const char *alias)
1184{
1185 const int nmemb = ARRAY_SIZE(syscall_fmts);
1186 return __syscall_fmt__find_by_alias(syscall_fmts, nmemb, alias);
1187}
1188
1189
1190
1191
1192
1193
1194
1195struct syscall {
1196 struct tep_event *tp_format;
1197 int nr_args;
1198 int args_size;
1199 struct {
1200 struct bpf_program *sys_enter,
1201 *sys_exit;
1202 } bpf_prog;
1203 bool is_exit;
1204 bool is_open;
1205 bool nonexistent;
1206 struct tep_format_field *args;
1207 const char *name;
1208 struct syscall_fmt *fmt;
1209 struct syscall_arg_fmt *arg_fmt;
1210};
1211
1212
1213
1214
1215
1216
1217struct bpf_map_syscall_entry {
1218 bool enabled;
1219 u16 string_args_len[6];
1220};
1221
1222
1223
1224
1225
1226
1227
1228
1229static size_t fprintf_duration(unsigned long t, bool calculated, FILE *fp)
1230{
1231 double duration = (double)t / NSEC_PER_MSEC;
1232 size_t printed = fprintf(fp, "(");
1233
1234 if (!calculated)
1235 printed += fprintf(fp, " ");
1236 else if (duration >= 1.0)
1237 printed += color_fprintf(fp, PERF_COLOR_RED, "%6.3f ms", duration);
1238 else if (duration >= 0.01)
1239 printed += color_fprintf(fp, PERF_COLOR_YELLOW, "%6.3f ms", duration);
1240 else
1241 printed += color_fprintf(fp, PERF_COLOR_NORMAL, "%6.3f ms", duration);
1242 return printed + fprintf(fp, "): ");
1243}
1244
1245
1246
1247
1248
1249
1250
1251
1252struct thread_trace {
1253 u64 entry_time;
1254 bool entry_pending;
1255 unsigned long nr_events;
1256 unsigned long pfmaj, pfmin;
1257 char *entry_str;
1258 double runtime_ms;
1259 size_t (*ret_scnprintf)(char *bf, size_t size, struct syscall_arg *arg);
1260 struct {
1261 unsigned long ptr;
1262 short int entry_str_pos;
1263 bool pending_open;
1264 unsigned int namelen;
1265 char *name;
1266 } filename;
1267 struct {
1268 int max;
1269 struct file *table;
1270 } files;
1271
1272 struct intlist *syscall_stats;
1273};
1274
1275static struct thread_trace *thread_trace__new(void)
1276{
1277 struct thread_trace *ttrace = zalloc(sizeof(struct thread_trace));
1278
1279 if (ttrace) {
1280 ttrace->files.max = -1;
1281 ttrace->syscall_stats = intlist__new(NULL);
1282 }
1283
1284 return ttrace;
1285}
1286
1287static struct thread_trace *thread__trace(struct thread *thread, FILE *fp)
1288{
1289 struct thread_trace *ttrace;
1290
1291 if (thread == NULL)
1292 goto fail;
1293
1294 if (thread__priv(thread) == NULL)
1295 thread__set_priv(thread, thread_trace__new());
1296
1297 if (thread__priv(thread) == NULL)
1298 goto fail;
1299
1300 ttrace = thread__priv(thread);
1301 ++ttrace->nr_events;
1302
1303 return ttrace;
1304fail:
1305 color_fprintf(fp, PERF_COLOR_RED,
1306 "WARNING: not enough memory, dropping samples!\n");
1307 return NULL;
1308}
1309
1310
1311void syscall_arg__set_ret_scnprintf(struct syscall_arg *arg,
1312 size_t (*ret_scnprintf)(char *bf, size_t size, struct syscall_arg *arg))
1313{
1314 struct thread_trace *ttrace = thread__priv(arg->thread);
1315
1316 ttrace->ret_scnprintf = ret_scnprintf;
1317}
1318
1319#define TRACE_PFMAJ (1 << 0)
1320#define TRACE_PFMIN (1 << 1)
1321
1322static const size_t trace__entry_str_size = 2048;
1323
1324static struct file *thread_trace__files_entry(struct thread_trace *ttrace, int fd)
1325{
1326 if (fd < 0)
1327 return NULL;
1328
1329 if (fd > ttrace->files.max) {
1330 struct file *nfiles = realloc(ttrace->files.table, (fd + 1) * sizeof(struct file));
1331
1332 if (nfiles == NULL)
1333 return NULL;
1334
1335 if (ttrace->files.max != -1) {
1336 memset(nfiles + ttrace->files.max + 1, 0,
1337 (fd - ttrace->files.max) * sizeof(struct file));
1338 } else {
1339 memset(nfiles, 0, (fd + 1) * sizeof(struct file));
1340 }
1341
1342 ttrace->files.table = nfiles;
1343 ttrace->files.max = fd;
1344 }
1345
1346 return ttrace->files.table + fd;
1347}
1348
1349struct file *thread__files_entry(struct thread *thread, int fd)
1350{
1351 return thread_trace__files_entry(thread__priv(thread), fd);
1352}
1353
1354static int trace__set_fd_pathname(struct thread *thread, int fd, const char *pathname)
1355{
1356 struct thread_trace *ttrace = thread__priv(thread);
1357 struct file *file = thread_trace__files_entry(ttrace, fd);
1358
1359 if (file != NULL) {
1360 struct stat st;
1361 if (stat(pathname, &st) == 0)
1362 file->dev_maj = major(st.st_rdev);
1363 file->pathname = strdup(pathname);
1364 if (file->pathname)
1365 return 0;
1366 }
1367
1368 return -1;
1369}
1370
1371static int thread__read_fd_path(struct thread *thread, int fd)
1372{
1373 char linkname[PATH_MAX], pathname[PATH_MAX];
1374 struct stat st;
1375 int ret;
1376
1377 if (thread->pid_ == thread->tid) {
1378 scnprintf(linkname, sizeof(linkname),
1379 "/proc/%d/fd/%d", thread->pid_, fd);
1380 } else {
1381 scnprintf(linkname, sizeof(linkname),
1382 "/proc/%d/task/%d/fd/%d", thread->pid_, thread->tid, fd);
1383 }
1384
1385 if (lstat(linkname, &st) < 0 || st.st_size + 1 > (off_t)sizeof(pathname))
1386 return -1;
1387
1388 ret = readlink(linkname, pathname, sizeof(pathname));
1389
1390 if (ret < 0 || ret > st.st_size)
1391 return -1;
1392
1393 pathname[ret] = '\0';
1394 return trace__set_fd_pathname(thread, fd, pathname);
1395}
1396
1397static const char *thread__fd_path(struct thread *thread, int fd,
1398 struct trace *trace)
1399{
1400 struct thread_trace *ttrace = thread__priv(thread);
1401
1402 if (ttrace == NULL || trace->fd_path_disabled)
1403 return NULL;
1404
1405 if (fd < 0)
1406 return NULL;
1407
1408 if ((fd > ttrace->files.max || ttrace->files.table[fd].pathname == NULL)) {
1409 if (!trace->live)
1410 return NULL;
1411 ++trace->stats.proc_getname;
1412 if (thread__read_fd_path(thread, fd))
1413 return NULL;
1414 }
1415
1416 return ttrace->files.table[fd].pathname;
1417}
1418
1419size_t syscall_arg__scnprintf_fd(char *bf, size_t size, struct syscall_arg *arg)
1420{
1421 int fd = arg->val;
1422 size_t printed = scnprintf(bf, size, "%d", fd);
1423 const char *path = thread__fd_path(arg->thread, fd, arg->trace);
1424
1425 if (path)
1426 printed += scnprintf(bf + printed, size - printed, "<%s>", path);
1427
1428 return printed;
1429}
1430
1431size_t pid__scnprintf_fd(struct trace *trace, pid_t pid, int fd, char *bf, size_t size)
1432{
1433 size_t printed = scnprintf(bf, size, "%d", fd);
1434 struct thread *thread = machine__find_thread(trace->host, pid, pid);
1435
1436 if (thread) {
1437 const char *path = thread__fd_path(thread, fd, trace);
1438
1439 if (path)
1440 printed += scnprintf(bf + printed, size - printed, "<%s>", path);
1441
1442 thread__put(thread);
1443 }
1444
1445 return printed;
1446}
1447
1448static size_t syscall_arg__scnprintf_close_fd(char *bf, size_t size,
1449 struct syscall_arg *arg)
1450{
1451 int fd = arg->val;
1452 size_t printed = syscall_arg__scnprintf_fd(bf, size, arg);
1453 struct thread_trace *ttrace = thread__priv(arg->thread);
1454
1455 if (ttrace && fd >= 0 && fd <= ttrace->files.max)
1456 zfree(&ttrace->files.table[fd].pathname);
1457
1458 return printed;
1459}
1460
1461static void thread__set_filename_pos(struct thread *thread, const char *bf,
1462 unsigned long ptr)
1463{
1464 struct thread_trace *ttrace = thread__priv(thread);
1465
1466 ttrace->filename.ptr = ptr;
1467 ttrace->filename.entry_str_pos = bf - ttrace->entry_str;
1468}
1469
1470static size_t syscall_arg__scnprintf_augmented_string(struct syscall_arg *arg, char *bf, size_t size)
1471{
1472 struct augmented_arg *augmented_arg = arg->augmented.args;
1473 size_t printed = scnprintf(bf, size, "\"%.*s\"", augmented_arg->size, augmented_arg->value);
1474
1475
1476
1477
1478 int consumed = sizeof(*augmented_arg) + augmented_arg->size;
1479
1480 arg->augmented.args = ((void *)arg->augmented.args) + consumed;
1481 arg->augmented.size -= consumed;
1482
1483 return printed;
1484}
1485
1486static size_t syscall_arg__scnprintf_filename(char *bf, size_t size,
1487 struct syscall_arg *arg)
1488{
1489 unsigned long ptr = arg->val;
1490
1491 if (arg->augmented.args)
1492 return syscall_arg__scnprintf_augmented_string(arg, bf, size);
1493
1494 if (!arg->trace->vfs_getname)
1495 return scnprintf(bf, size, "%#x", ptr);
1496
1497 thread__set_filename_pos(arg->thread, bf, ptr);
1498 return 0;
1499}
1500
1501static bool trace__filter_duration(struct trace *trace, double t)
1502{
1503 return t < (trace->duration_filter * NSEC_PER_MSEC);
1504}
1505
1506static size_t __trace__fprintf_tstamp(struct trace *trace, u64 tstamp, FILE *fp)
1507{
1508 double ts = (double)(tstamp - trace->base_time) / NSEC_PER_MSEC;
1509
1510 return fprintf(fp, "%10.3f ", ts);
1511}
1512
1513
1514
1515
1516
1517
1518
1519static size_t trace__fprintf_tstamp(struct trace *trace, u64 tstamp, FILE *fp)
1520{
1521 if (tstamp > 0)
1522 return __trace__fprintf_tstamp(trace, tstamp, fp);
1523
1524 return fprintf(fp, " ? ");
1525}
1526
1527static bool done = false;
1528static bool interrupted = false;
1529
1530static void sig_handler(int sig)
1531{
1532 done = true;
1533 interrupted = sig == SIGINT;
1534}
1535
1536static size_t trace__fprintf_comm_tid(struct trace *trace, struct thread *thread, FILE *fp)
1537{
1538 size_t printed = 0;
1539
1540 if (trace->multiple_threads) {
1541 if (trace->show_comm)
1542 printed += fprintf(fp, "%.14s/", thread__comm_str(thread));
1543 printed += fprintf(fp, "%d ", thread->tid);
1544 }
1545
1546 return printed;
1547}
1548
1549static size_t trace__fprintf_entry_head(struct trace *trace, struct thread *thread,
1550 u64 duration, bool duration_calculated, u64 tstamp, FILE *fp)
1551{
1552 size_t printed = 0;
1553
1554 if (trace->show_tstamp)
1555 printed = trace__fprintf_tstamp(trace, tstamp, fp);
1556 if (trace->show_duration)
1557 printed += fprintf_duration(duration, duration_calculated, fp);
1558 return printed + trace__fprintf_comm_tid(trace, thread, fp);
1559}
1560
1561static int trace__process_event(struct trace *trace, struct machine *machine,
1562 union perf_event *event, struct perf_sample *sample)
1563{
1564 int ret = 0;
1565
1566 switch (event->header.type) {
1567 case PERF_RECORD_LOST:
1568 color_fprintf(trace->output, PERF_COLOR_RED,
1569 "LOST %" PRIu64 " events!\n", event->lost.lost);
1570 ret = machine__process_lost_event(machine, event, sample);
1571 break;
1572 default:
1573 ret = machine__process_event(machine, event, sample);
1574 break;
1575 }
1576
1577 return ret;
1578}
1579
1580static int trace__tool_process(struct perf_tool *tool,
1581 union perf_event *event,
1582 struct perf_sample *sample,
1583 struct machine *machine)
1584{
1585 struct trace *trace = container_of(tool, struct trace, tool);
1586 return trace__process_event(trace, machine, event, sample);
1587}
1588
1589static char *trace__machine__resolve_kernel_addr(void *vmachine, unsigned long long *addrp, char **modp)
1590{
1591 struct machine *machine = vmachine;
1592
1593 if (machine->kptr_restrict_warned)
1594 return NULL;
1595
1596 if (symbol_conf.kptr_restrict) {
1597 pr_warning("Kernel address maps (/proc/{kallsyms,modules}) are restricted.\n\n"
1598 "Check /proc/sys/kernel/kptr_restrict and /proc/sys/kernel/perf_event_paranoid.\n\n"
1599 "Kernel samples will not be resolved.\n");
1600 machine->kptr_restrict_warned = true;
1601 return NULL;
1602 }
1603
1604 return machine__resolve_kernel_addr(vmachine, addrp, modp);
1605}
1606
1607static int trace__symbols_init(struct trace *trace, struct evlist *evlist)
1608{
1609 int err = symbol__init(NULL);
1610
1611 if (err)
1612 return err;
1613
1614 trace->host = machine__new_host();
1615 if (trace->host == NULL)
1616 return -ENOMEM;
1617
1618 err = trace_event__register_resolver(trace->host, trace__machine__resolve_kernel_addr);
1619 if (err < 0)
1620 goto out;
1621
1622 err = __machine__synthesize_threads(trace->host, &trace->tool, &trace->opts.target,
1623 evlist->core.threads, trace__tool_process, false,
1624 1);
1625out:
1626 if (err)
1627 symbol__exit();
1628
1629 return err;
1630}
1631
1632static void trace__symbols__exit(struct trace *trace)
1633{
1634 machine__exit(trace->host);
1635 trace->host = NULL;
1636
1637 symbol__exit();
1638}
1639
1640static int syscall__alloc_arg_fmts(struct syscall *sc, int nr_args)
1641{
1642 int idx;
1643
1644 if (nr_args == 6 && sc->fmt && sc->fmt->nr_args != 0)
1645 nr_args = sc->fmt->nr_args;
1646
1647 sc->arg_fmt = calloc(nr_args, sizeof(*sc->arg_fmt));
1648 if (sc->arg_fmt == NULL)
1649 return -1;
1650
1651 for (idx = 0; idx < nr_args; ++idx) {
1652 if (sc->fmt)
1653 sc->arg_fmt[idx] = sc->fmt->arg[idx];
1654 }
1655
1656 sc->nr_args = nr_args;
1657 return 0;
1658}
1659
1660static struct syscall_arg_fmt syscall_arg_fmts__by_name[] = {
1661 { .name = "msr", .scnprintf = SCA_X86_MSR, .strtoul = STUL_X86_MSR, },
1662 { .name = "vector", .scnprintf = SCA_X86_IRQ_VECTORS, .strtoul = STUL_X86_IRQ_VECTORS, },
1663};
1664
1665static int syscall_arg_fmt__cmp(const void *name, const void *fmtp)
1666{
1667 const struct syscall_arg_fmt *fmt = fmtp;
1668 return strcmp(name, fmt->name);
1669}
1670
1671static struct syscall_arg_fmt *
1672__syscall_arg_fmt__find_by_name(struct syscall_arg_fmt *fmts, const int nmemb, const char *name)
1673{
1674 return bsearch(name, fmts, nmemb, sizeof(struct syscall_arg_fmt), syscall_arg_fmt__cmp);
1675}
1676
1677static struct syscall_arg_fmt *syscall_arg_fmt__find_by_name(const char *name)
1678{
1679 const int nmemb = ARRAY_SIZE(syscall_arg_fmts__by_name);
1680 return __syscall_arg_fmt__find_by_name(syscall_arg_fmts__by_name, nmemb, name);
1681}
1682
1683static struct tep_format_field *
1684syscall_arg_fmt__init_array(struct syscall_arg_fmt *arg, struct tep_format_field *field)
1685{
1686 struct tep_format_field *last_field = NULL;
1687 int len;
1688
1689 for (; field; field = field->next, ++arg) {
1690 last_field = field;
1691
1692 if (arg->scnprintf)
1693 continue;
1694
1695 len = strlen(field->name);
1696
1697 if (strcmp(field->type, "const char *") == 0 &&
1698 ((len >= 4 && strcmp(field->name + len - 4, "name") == 0) ||
1699 strstr(field->name, "path") != NULL))
1700 arg->scnprintf = SCA_FILENAME;
1701 else if ((field->flags & TEP_FIELD_IS_POINTER) || strstr(field->name, "addr"))
1702 arg->scnprintf = SCA_PTR;
1703 else if (strcmp(field->type, "pid_t") == 0)
1704 arg->scnprintf = SCA_PID;
1705 else if (strcmp(field->type, "umode_t") == 0)
1706 arg->scnprintf = SCA_MODE_T;
1707 else if ((field->flags & TEP_FIELD_IS_ARRAY) && strstr(field->type, "char")) {
1708 arg->scnprintf = SCA_CHAR_ARRAY;
1709 arg->nr_entries = field->arraylen;
1710 } else if ((strcmp(field->type, "int") == 0 ||
1711 strcmp(field->type, "unsigned int") == 0 ||
1712 strcmp(field->type, "long") == 0) &&
1713 len >= 2 && strcmp(field->name + len - 2, "fd") == 0) {
1714
1715
1716
1717
1718
1719
1720
1721 arg->scnprintf = SCA_FD;
1722 } else {
1723 struct syscall_arg_fmt *fmt = syscall_arg_fmt__find_by_name(field->name);
1724
1725 if (fmt) {
1726 arg->scnprintf = fmt->scnprintf;
1727 arg->strtoul = fmt->strtoul;
1728 }
1729 }
1730 }
1731
1732 return last_field;
1733}
1734
1735static int syscall__set_arg_fmts(struct syscall *sc)
1736{
1737 struct tep_format_field *last_field = syscall_arg_fmt__init_array(sc->arg_fmt, sc->args);
1738
1739 if (last_field)
1740 sc->args_size = last_field->offset + last_field->size;
1741
1742 return 0;
1743}
1744
1745static int trace__read_syscall_info(struct trace *trace, int id)
1746{
1747 char tp_name[128];
1748 struct syscall *sc;
1749 const char *name = syscalltbl__name(trace->sctbl, id);
1750
1751#ifdef HAVE_SYSCALL_TABLE_SUPPORT
1752 if (trace->syscalls.table == NULL) {
1753 trace->syscalls.table = calloc(trace->sctbl->syscalls.max_id + 1, sizeof(*sc));
1754 if (trace->syscalls.table == NULL)
1755 return -ENOMEM;
1756 }
1757#else
1758 if (id > trace->sctbl->syscalls.max_id || (id == 0 && trace->syscalls.table == NULL)) {
1759
1760 struct syscall *table = realloc(trace->syscalls.table, (id + 1) * sizeof(*sc));
1761
1762 if (table == NULL)
1763 return -ENOMEM;
1764
1765
1766 if (trace->syscalls.table == NULL)
1767 memset(table, 0, (id + 1) * sizeof(*sc));
1768 else
1769 memset(table + trace->sctbl->syscalls.max_id + 1, 0, (id - trace->sctbl->syscalls.max_id) * sizeof(*sc));
1770
1771 trace->syscalls.table = table;
1772 trace->sctbl->syscalls.max_id = id;
1773 }
1774#endif
1775 sc = trace->syscalls.table + id;
1776 if (sc->nonexistent)
1777 return 0;
1778
1779 if (name == NULL) {
1780 sc->nonexistent = true;
1781 return 0;
1782 }
1783
1784 sc->name = name;
1785 sc->fmt = syscall_fmt__find(sc->name);
1786
1787 snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->name);
1788 sc->tp_format = trace_event__tp_format("syscalls", tp_name);
1789
1790 if (IS_ERR(sc->tp_format) && sc->fmt && sc->fmt->alias) {
1791 snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->fmt->alias);
1792 sc->tp_format = trace_event__tp_format("syscalls", tp_name);
1793 }
1794
1795 if (syscall__alloc_arg_fmts(sc, IS_ERR(sc->tp_format) ? 6 : sc->tp_format->format.nr_fields))
1796 return -ENOMEM;
1797
1798 if (IS_ERR(sc->tp_format))
1799 return PTR_ERR(sc->tp_format);
1800
1801 sc->args = sc->tp_format->format.fields;
1802
1803
1804
1805
1806
1807 if (sc->args && (!strcmp(sc->args->name, "__syscall_nr") || !strcmp(sc->args->name, "nr"))) {
1808 sc->args = sc->args->next;
1809 --sc->nr_args;
1810 }
1811
1812 sc->is_exit = !strcmp(name, "exit_group") || !strcmp(name, "exit");
1813 sc->is_open = !strcmp(name, "open") || !strcmp(name, "openat");
1814
1815 return syscall__set_arg_fmts(sc);
1816}
1817
1818static int evsel__init_tp_arg_scnprintf(struct evsel *evsel)
1819{
1820 struct syscall_arg_fmt *fmt = evsel__syscall_arg_fmt(evsel);
1821
1822 if (fmt != NULL) {
1823 syscall_arg_fmt__init_array(fmt, evsel->tp_format->format.fields);
1824 return 0;
1825 }
1826
1827 return -ENOMEM;
1828}
1829
1830static int intcmp(const void *a, const void *b)
1831{
1832 const int *one = a, *another = b;
1833
1834 return *one - *another;
1835}
1836
1837static int trace__validate_ev_qualifier(struct trace *trace)
1838{
1839 int err = 0;
1840 bool printed_invalid_prefix = false;
1841 struct str_node *pos;
1842 size_t nr_used = 0, nr_allocated = strlist__nr_entries(trace->ev_qualifier);
1843
1844 trace->ev_qualifier_ids.entries = malloc(nr_allocated *
1845 sizeof(trace->ev_qualifier_ids.entries[0]));
1846
1847 if (trace->ev_qualifier_ids.entries == NULL) {
1848 fputs("Error:\tNot enough memory for allocating events qualifier ids\n",
1849 trace->output);
1850 err = -EINVAL;
1851 goto out;
1852 }
1853
1854 strlist__for_each_entry(pos, trace->ev_qualifier) {
1855 const char *sc = pos->s;
1856 int id = syscalltbl__id(trace->sctbl, sc), match_next = -1;
1857
1858 if (id < 0) {
1859 id = syscalltbl__strglobmatch_first(trace->sctbl, sc, &match_next);
1860 if (id >= 0)
1861 goto matches;
1862
1863 if (!printed_invalid_prefix) {
1864 pr_debug("Skipping unknown syscalls: ");
1865 printed_invalid_prefix = true;
1866 } else {
1867 pr_debug(", ");
1868 }
1869
1870 pr_debug("%s", sc);
1871 continue;
1872 }
1873matches:
1874 trace->ev_qualifier_ids.entries[nr_used++] = id;
1875 if (match_next == -1)
1876 continue;
1877
1878 while (1) {
1879 id = syscalltbl__strglobmatch_next(trace->sctbl, sc, &match_next);
1880 if (id < 0)
1881 break;
1882 if (nr_allocated == nr_used) {
1883 void *entries;
1884
1885 nr_allocated += 8;
1886 entries = realloc(trace->ev_qualifier_ids.entries,
1887 nr_allocated * sizeof(trace->ev_qualifier_ids.entries[0]));
1888 if (entries == NULL) {
1889 err = -ENOMEM;
1890 fputs("\nError:\t Not enough memory for parsing\n", trace->output);
1891 goto out_free;
1892 }
1893 trace->ev_qualifier_ids.entries = entries;
1894 }
1895 trace->ev_qualifier_ids.entries[nr_used++] = id;
1896 }
1897 }
1898
1899 trace->ev_qualifier_ids.nr = nr_used;
1900 qsort(trace->ev_qualifier_ids.entries, nr_used, sizeof(int), intcmp);
1901out:
1902 if (printed_invalid_prefix)
1903 pr_debug("\n");
1904 return err;
1905out_free:
1906 zfree(&trace->ev_qualifier_ids.entries);
1907 trace->ev_qualifier_ids.nr = 0;
1908 goto out;
1909}
1910
1911static __maybe_unused bool trace__syscall_enabled(struct trace *trace, int id)
1912{
1913 bool in_ev_qualifier;
1914
1915 if (trace->ev_qualifier_ids.nr == 0)
1916 return true;
1917
1918 in_ev_qualifier = bsearch(&id, trace->ev_qualifier_ids.entries,
1919 trace->ev_qualifier_ids.nr, sizeof(int), intcmp) != NULL;
1920
1921 if (in_ev_qualifier)
1922 return !trace->not_ev_qualifier;
1923
1924 return trace->not_ev_qualifier;
1925}
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935unsigned long syscall_arg__val(struct syscall_arg *arg, u8 idx)
1936{
1937 unsigned long val;
1938 unsigned char *p = arg->args + sizeof(unsigned long) * idx;
1939
1940 memcpy(&val, p, sizeof(val));
1941 return val;
1942}
1943
1944static size_t syscall__scnprintf_name(struct syscall *sc, char *bf, size_t size,
1945 struct syscall_arg *arg)
1946{
1947 if (sc->arg_fmt && sc->arg_fmt[arg->idx].name)
1948 return scnprintf(bf, size, "%s: ", sc->arg_fmt[arg->idx].name);
1949
1950 return scnprintf(bf, size, "arg%d: ", arg->idx);
1951}
1952
1953
1954
1955
1956
1957
1958static unsigned long syscall_arg_fmt__mask_val(struct syscall_arg_fmt *fmt, struct syscall_arg *arg, unsigned long val)
1959{
1960 if (fmt && fmt->mask_val)
1961 return fmt->mask_val(arg, val);
1962
1963 return val;
1964}
1965
1966static size_t syscall_arg_fmt__scnprintf_val(struct syscall_arg_fmt *fmt, char *bf, size_t size,
1967 struct syscall_arg *arg, unsigned long val)
1968{
1969 if (fmt && fmt->scnprintf) {
1970 arg->val = val;
1971 if (fmt->parm)
1972 arg->parm = fmt->parm;
1973 return fmt->scnprintf(bf, size, arg);
1974 }
1975 return scnprintf(bf, size, "%ld", val);
1976}
1977
1978static size_t syscall__scnprintf_args(struct syscall *sc, char *bf, size_t size,
1979 unsigned char *args, void *augmented_args, int augmented_args_size,
1980 struct trace *trace, struct thread *thread)
1981{
1982 size_t printed = 0;
1983 unsigned long val;
1984 u8 bit = 1;
1985 struct syscall_arg arg = {
1986 .args = args,
1987 .augmented = {
1988 .size = augmented_args_size,
1989 .args = augmented_args,
1990 },
1991 .idx = 0,
1992 .mask = 0,
1993 .trace = trace,
1994 .thread = thread,
1995 .show_string_prefix = trace->show_string_prefix,
1996 };
1997 struct thread_trace *ttrace = thread__priv(thread);
1998
1999
2000
2001
2002
2003
2004 ttrace->ret_scnprintf = NULL;
2005
2006 if (sc->args != NULL) {
2007 struct tep_format_field *field;
2008
2009 for (field = sc->args; field;
2010 field = field->next, ++arg.idx, bit <<= 1) {
2011 if (arg.mask & bit)
2012 continue;
2013
2014 arg.fmt = &sc->arg_fmt[arg.idx];
2015 val = syscall_arg__val(&arg, arg.idx);
2016
2017
2018
2019
2020 val = syscall_arg_fmt__mask_val(&sc->arg_fmt[arg.idx], &arg, val);
2021
2022
2023
2024
2025
2026
2027 if (val == 0 &&
2028 !trace->show_zeros &&
2029 !(sc->arg_fmt &&
2030 (sc->arg_fmt[arg.idx].show_zero ||
2031 sc->arg_fmt[arg.idx].scnprintf == SCA_STRARRAY ||
2032 sc->arg_fmt[arg.idx].scnprintf == SCA_STRARRAYS) &&
2033 sc->arg_fmt[arg.idx].parm))
2034 continue;
2035
2036 printed += scnprintf(bf + printed, size - printed, "%s", printed ? ", " : "");
2037
2038 if (trace->show_arg_names)
2039 printed += scnprintf(bf + printed, size - printed, "%s: ", field->name);
2040
2041 printed += syscall_arg_fmt__scnprintf_val(&sc->arg_fmt[arg.idx],
2042 bf + printed, size - printed, &arg, val);
2043 }
2044 } else if (IS_ERR(sc->tp_format)) {
2045
2046
2047
2048
2049
2050 while (arg.idx < sc->nr_args) {
2051 if (arg.mask & bit)
2052 goto next_arg;
2053 val = syscall_arg__val(&arg, arg.idx);
2054 if (printed)
2055 printed += scnprintf(bf + printed, size - printed, ", ");
2056 printed += syscall__scnprintf_name(sc, bf + printed, size - printed, &arg);
2057 printed += syscall_arg_fmt__scnprintf_val(&sc->arg_fmt[arg.idx], bf + printed, size - printed, &arg, val);
2058next_arg:
2059 ++arg.idx;
2060 bit <<= 1;
2061 }
2062 }
2063
2064 return printed;
2065}
2066
2067typedef int (*tracepoint_handler)(struct trace *trace, struct evsel *evsel,
2068 union perf_event *event,
2069 struct perf_sample *sample);
2070
2071static struct syscall *trace__syscall_info(struct trace *trace,
2072 struct evsel *evsel, int id)
2073{
2074 int err = 0;
2075
2076 if (id < 0) {
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088 if (verbose > 1) {
2089 static u64 n;
2090 fprintf(trace->output, "Invalid syscall %d id, skipping (%s, %" PRIu64 ") ...\n",
2091 id, evsel__name(evsel), ++n);
2092 }
2093 return NULL;
2094 }
2095
2096 err = -EINVAL;
2097
2098#ifdef HAVE_SYSCALL_TABLE_SUPPORT
2099 if (id > trace->sctbl->syscalls.max_id) {
2100#else
2101 if (id >= trace->sctbl->syscalls.max_id) {
2102
2103
2104
2105
2106
2107 err = trace__read_syscall_info(trace, id);
2108 if (err)
2109#endif
2110 goto out_cant_read;
2111 }
2112
2113 if ((trace->syscalls.table == NULL || trace->syscalls.table[id].name == NULL) &&
2114 (err = trace__read_syscall_info(trace, id)) != 0)
2115 goto out_cant_read;
2116
2117 if (trace->syscalls.table[id].name == NULL) {
2118 if (trace->syscalls.table[id].nonexistent)
2119 return NULL;
2120 goto out_cant_read;
2121 }
2122
2123 return &trace->syscalls.table[id];
2124
2125out_cant_read:
2126 if (verbose > 0) {
2127 char sbuf[STRERR_BUFSIZE];
2128 fprintf(trace->output, "Problems reading syscall %d: %d (%s)", id, -err, str_error_r(-err, sbuf, sizeof(sbuf)));
2129 if (id <= trace->sctbl->syscalls.max_id && trace->syscalls.table[id].name != NULL)
2130 fprintf(trace->output, "(%s)", trace->syscalls.table[id].name);
2131 fputs(" information\n", trace->output);
2132 }
2133 return NULL;
2134}
2135
2136struct syscall_stats {
2137 struct stats stats;
2138 u64 nr_failures;
2139 int max_errno;
2140 u32 *errnos;
2141};
2142
2143static void thread__update_stats(struct thread *thread, struct thread_trace *ttrace,
2144 int id, struct perf_sample *sample, long err, bool errno_summary)
2145{
2146 struct int_node *inode;
2147 struct syscall_stats *stats;
2148 u64 duration = 0;
2149
2150 inode = intlist__findnew(ttrace->syscall_stats, id);
2151 if (inode == NULL)
2152 return;
2153
2154 stats = inode->priv;
2155 if (stats == NULL) {
2156 stats = malloc(sizeof(*stats));
2157 if (stats == NULL)
2158 return;
2159
2160 stats->nr_failures = 0;
2161 stats->max_errno = 0;
2162 stats->errnos = NULL;
2163 init_stats(&stats->stats);
2164 inode->priv = stats;
2165 }
2166
2167 if (ttrace->entry_time && sample->time > ttrace->entry_time)
2168 duration = sample->time - ttrace->entry_time;
2169
2170 update_stats(&stats->stats, duration);
2171
2172 if (err < 0) {
2173 ++stats->nr_failures;
2174
2175 if (!errno_summary)
2176 return;
2177
2178 err = -err;
2179 if (err > stats->max_errno) {
2180 u32 *new_errnos = realloc(stats->errnos, err * sizeof(u32));
2181
2182 if (new_errnos) {
2183 memset(new_errnos + stats->max_errno, 0, (err - stats->max_errno) * sizeof(u32));
2184 } else {
2185 pr_debug("Not enough memory for errno stats for thread \"%s\"(%d/%d), results will be incomplete\n",
2186 thread__comm_str(thread), thread->pid_, thread->tid);
2187 return;
2188 }
2189
2190 stats->errnos = new_errnos;
2191 stats->max_errno = err;
2192 }
2193
2194 ++stats->errnos[err - 1];
2195 }
2196}
2197
2198static int trace__printf_interrupted_entry(struct trace *trace)
2199{
2200 struct thread_trace *ttrace;
2201 size_t printed;
2202 int len;
2203
2204 if (trace->failure_only || trace->current == NULL)
2205 return 0;
2206
2207 ttrace = thread__priv(trace->current);
2208
2209 if (!ttrace->entry_pending)
2210 return 0;
2211
2212 printed = trace__fprintf_entry_head(trace, trace->current, 0, false, ttrace->entry_time, trace->output);
2213 printed += len = fprintf(trace->output, "%s)", ttrace->entry_str);
2214
2215 if (len < trace->args_alignment - 4)
2216 printed += fprintf(trace->output, "%-*s", trace->args_alignment - 4 - len, " ");
2217
2218 printed += fprintf(trace->output, " ...\n");
2219
2220 ttrace->entry_pending = false;
2221 ++trace->nr_events_printed;
2222
2223 return printed;
2224}
2225
2226static int trace__fprintf_sample(struct trace *trace, struct evsel *evsel,
2227 struct perf_sample *sample, struct thread *thread)
2228{
2229 int printed = 0;
2230
2231 if (trace->print_sample) {
2232 double ts = (double)sample->time / NSEC_PER_MSEC;
2233
2234 printed += fprintf(trace->output, "%22s %10.3f %s %d/%d [%d]\n",
2235 evsel__name(evsel), ts,
2236 thread__comm_str(thread),
2237 sample->pid, sample->tid, sample->cpu);
2238 }
2239
2240 return printed;
2241}
2242
2243static void *syscall__augmented_args(struct syscall *sc, struct perf_sample *sample, int *augmented_args_size, int raw_augmented_args_size)
2244{
2245 void *augmented_args = NULL;
2246
2247
2248
2249
2250
2251
2252
2253
2254
2255
2256
2257
2258
2259
2260 int args_size = raw_augmented_args_size ?: sc->args_size;
2261
2262 *augmented_args_size = sample->raw_size - args_size;
2263 if (*augmented_args_size > 0)
2264 augmented_args = sample->raw_data + args_size;
2265
2266 return augmented_args;
2267}
2268
2269static int trace__sys_enter(struct trace *trace, struct evsel *evsel,
2270 union perf_event *event __maybe_unused,
2271 struct perf_sample *sample)
2272{
2273 char *msg;
2274 void *args;
2275 int printed = 0;
2276 struct thread *thread;
2277 int id = perf_evsel__sc_tp_uint(evsel, id, sample), err = -1;
2278 int augmented_args_size = 0;
2279 void *augmented_args = NULL;
2280 struct syscall *sc = trace__syscall_info(trace, evsel, id);
2281 struct thread_trace *ttrace;
2282
2283 if (sc == NULL)
2284 return -1;
2285
2286 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2287 ttrace = thread__trace(thread, trace->output);
2288 if (ttrace == NULL)
2289 goto out_put;
2290
2291 trace__fprintf_sample(trace, evsel, sample, thread);
2292
2293 args = perf_evsel__sc_tp_ptr(evsel, args, sample);
2294
2295 if (ttrace->entry_str == NULL) {
2296 ttrace->entry_str = malloc(trace__entry_str_size);
2297 if (!ttrace->entry_str)
2298 goto out_put;
2299 }
2300
2301 if (!(trace->duration_filter || trace->summary_only || trace->min_stack))
2302 trace__printf_interrupted_entry(trace);
2303
2304
2305
2306
2307
2308
2309
2310
2311
2312
2313 if (evsel != trace->syscalls.events.sys_enter)
2314 augmented_args = syscall__augmented_args(sc, sample, &augmented_args_size, trace->raw_augmented_syscalls_args_size);
2315 ttrace->entry_time = sample->time;
2316 msg = ttrace->entry_str;
2317 printed += scnprintf(msg + printed, trace__entry_str_size - printed, "%s(", sc->name);
2318
2319 printed += syscall__scnprintf_args(sc, msg + printed, trace__entry_str_size - printed,
2320 args, augmented_args, augmented_args_size, trace, thread);
2321
2322 if (sc->is_exit) {
2323 if (!(trace->duration_filter || trace->summary_only || trace->failure_only || trace->min_stack)) {
2324 int alignment = 0;
2325
2326 trace__fprintf_entry_head(trace, thread, 0, false, ttrace->entry_time, trace->output);
2327 printed = fprintf(trace->output, "%s)", ttrace->entry_str);
2328 if (trace->args_alignment > printed)
2329 alignment = trace->args_alignment - printed;
2330 fprintf(trace->output, "%*s= ?\n", alignment, " ");
2331 }
2332 } else {
2333 ttrace->entry_pending = true;
2334
2335 ttrace->filename.pending_open = false;
2336 }
2337
2338 if (trace->current != thread) {
2339 thread__put(trace->current);
2340 trace->current = thread__get(thread);
2341 }
2342 err = 0;
2343out_put:
2344 thread__put(thread);
2345 return err;
2346}
2347
2348static int trace__fprintf_sys_enter(struct trace *trace, struct evsel *evsel,
2349 struct perf_sample *sample)
2350{
2351 struct thread_trace *ttrace;
2352 struct thread *thread;
2353 int id = perf_evsel__sc_tp_uint(evsel, id, sample), err = -1;
2354 struct syscall *sc = trace__syscall_info(trace, evsel, id);
2355 char msg[1024];
2356 void *args, *augmented_args = NULL;
2357 int augmented_args_size;
2358
2359 if (sc == NULL)
2360 return -1;
2361
2362 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2363 ttrace = thread__trace(thread, trace->output);
2364
2365
2366
2367
2368 if (ttrace == NULL)
2369 goto out_put;
2370
2371 args = perf_evsel__sc_tp_ptr(evsel, args, sample);
2372 augmented_args = syscall__augmented_args(sc, sample, &augmented_args_size, trace->raw_augmented_syscalls_args_size);
2373 syscall__scnprintf_args(sc, msg, sizeof(msg), args, augmented_args, augmented_args_size, trace, thread);
2374 fprintf(trace->output, "%s", msg);
2375 err = 0;
2376out_put:
2377 thread__put(thread);
2378 return err;
2379}
2380
2381static int trace__resolve_callchain(struct trace *trace, struct evsel *evsel,
2382 struct perf_sample *sample,
2383 struct callchain_cursor *cursor)
2384{
2385 struct addr_location al;
2386 int max_stack = evsel->core.attr.sample_max_stack ?
2387 evsel->core.attr.sample_max_stack :
2388 trace->max_stack;
2389 int err;
2390
2391 if (machine__resolve(trace->host, &al, sample) < 0)
2392 return -1;
2393
2394 err = thread__resolve_callchain(al.thread, cursor, evsel, sample, NULL, NULL, max_stack);
2395 addr_location__put(&al);
2396 return err;
2397}
2398
2399static int trace__fprintf_callchain(struct trace *trace, struct perf_sample *sample)
2400{
2401
2402 const unsigned int print_opts = EVSEL__PRINT_SYM |
2403 EVSEL__PRINT_DSO |
2404 EVSEL__PRINT_UNKNOWN_AS_ADDR;
2405
2406 return sample__fprintf_callchain(sample, 38, print_opts, &callchain_cursor, symbol_conf.bt_stop_list, trace->output);
2407}
2408
2409static const char *errno_to_name(struct evsel *evsel, int err)
2410{
2411 struct perf_env *env = evsel__env(evsel);
2412 const char *arch_name = perf_env__arch(env);
2413
2414 return arch_syscalls__strerrno(arch_name, err);
2415}
2416
2417static int trace__sys_exit(struct trace *trace, struct evsel *evsel,
2418 union perf_event *event __maybe_unused,
2419 struct perf_sample *sample)
2420{
2421 long ret;
2422 u64 duration = 0;
2423 bool duration_calculated = false;
2424 struct thread *thread;
2425 int id = perf_evsel__sc_tp_uint(evsel, id, sample), err = -1, callchain_ret = 0, printed = 0;
2426 int alignment = trace->args_alignment;
2427 struct syscall *sc = trace__syscall_info(trace, evsel, id);
2428 struct thread_trace *ttrace;
2429
2430 if (sc == NULL)
2431 return -1;
2432
2433 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2434 ttrace = thread__trace(thread, trace->output);
2435 if (ttrace == NULL)
2436 goto out_put;
2437
2438 trace__fprintf_sample(trace, evsel, sample, thread);
2439
2440 ret = perf_evsel__sc_tp_uint(evsel, ret, sample);
2441
2442 if (trace->summary)
2443 thread__update_stats(thread, ttrace, id, sample, ret, trace->errno_summary);
2444
2445 if (!trace->fd_path_disabled && sc->is_open && ret >= 0 && ttrace->filename.pending_open) {
2446 trace__set_fd_pathname(thread, ret, ttrace->filename.name);
2447 ttrace->filename.pending_open = false;
2448 ++trace->stats.vfs_getname;
2449 }
2450
2451 if (ttrace->entry_time) {
2452 duration = sample->time - ttrace->entry_time;
2453 if (trace__filter_duration(trace, duration))
2454 goto out;
2455 duration_calculated = true;
2456 } else if (trace->duration_filter)
2457 goto out;
2458
2459 if (sample->callchain) {
2460 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
2461 if (callchain_ret == 0) {
2462 if (callchain_cursor.nr < trace->min_stack)
2463 goto out;
2464 callchain_ret = 1;
2465 }
2466 }
2467
2468 if (trace->summary_only || (ret >= 0 && trace->failure_only))
2469 goto out;
2470
2471 trace__fprintf_entry_head(trace, thread, duration, duration_calculated, ttrace->entry_time, trace->output);
2472
2473 if (ttrace->entry_pending) {
2474 printed = fprintf(trace->output, "%s", ttrace->entry_str);
2475 } else {
2476 printed += fprintf(trace->output, " ... [");
2477 color_fprintf(trace->output, PERF_COLOR_YELLOW, "continued");
2478 printed += 9;
2479 printed += fprintf(trace->output, "]: %s()", sc->name);
2480 }
2481
2482 printed++;
2483
2484 if (alignment > printed)
2485 alignment -= printed;
2486 else
2487 alignment = 0;
2488
2489 fprintf(trace->output, ")%*s= ", alignment, " ");
2490
2491 if (sc->fmt == NULL) {
2492 if (ret < 0)
2493 goto errno_print;
2494signed_print:
2495 fprintf(trace->output, "%ld", ret);
2496 } else if (ret < 0) {
2497errno_print: {
2498 char bf[STRERR_BUFSIZE];
2499 const char *emsg = str_error_r(-ret, bf, sizeof(bf)),
2500 *e = errno_to_name(evsel, -ret);
2501
2502 fprintf(trace->output, "-1 %s (%s)", e, emsg);
2503 }
2504 } else if (ret == 0 && sc->fmt->timeout)
2505 fprintf(trace->output, "0 (Timeout)");
2506 else if (ttrace->ret_scnprintf) {
2507 char bf[1024];
2508 struct syscall_arg arg = {
2509 .val = ret,
2510 .thread = thread,
2511 .trace = trace,
2512 };
2513 ttrace->ret_scnprintf(bf, sizeof(bf), &arg);
2514 ttrace->ret_scnprintf = NULL;
2515 fprintf(trace->output, "%s", bf);
2516 } else if (sc->fmt->hexret)
2517 fprintf(trace->output, "%#lx", ret);
2518 else if (sc->fmt->errpid) {
2519 struct thread *child = machine__find_thread(trace->host, ret, ret);
2520
2521 if (child != NULL) {
2522 fprintf(trace->output, "%ld", ret);
2523 if (child->comm_set)
2524 fprintf(trace->output, " (%s)", thread__comm_str(child));
2525 thread__put(child);
2526 }
2527 } else
2528 goto signed_print;
2529
2530 fputc('\n', trace->output);
2531
2532
2533
2534
2535
2536 if (++trace->nr_events_printed == trace->max_events && trace->max_events != ULONG_MAX)
2537 interrupted = true;
2538
2539 if (callchain_ret > 0)
2540 trace__fprintf_callchain(trace, sample);
2541 else if (callchain_ret < 0)
2542 pr_err("Problem processing %s callchain, skipping...\n", evsel__name(evsel));
2543out:
2544 ttrace->entry_pending = false;
2545 err = 0;
2546out_put:
2547 thread__put(thread);
2548 return err;
2549}
2550
2551static int trace__vfs_getname(struct trace *trace, struct evsel *evsel,
2552 union perf_event *event __maybe_unused,
2553 struct perf_sample *sample)
2554{
2555 struct thread *thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2556 struct thread_trace *ttrace;
2557 size_t filename_len, entry_str_len, to_move;
2558 ssize_t remaining_space;
2559 char *pos;
2560 const char *filename = evsel__rawptr(evsel, sample, "pathname");
2561
2562 if (!thread)
2563 goto out;
2564
2565 ttrace = thread__priv(thread);
2566 if (!ttrace)
2567 goto out_put;
2568
2569 filename_len = strlen(filename);
2570 if (filename_len == 0)
2571 goto out_put;
2572
2573 if (ttrace->filename.namelen < filename_len) {
2574 char *f = realloc(ttrace->filename.name, filename_len + 1);
2575
2576 if (f == NULL)
2577 goto out_put;
2578
2579 ttrace->filename.namelen = filename_len;
2580 ttrace->filename.name = f;
2581 }
2582
2583 strcpy(ttrace->filename.name, filename);
2584 ttrace->filename.pending_open = true;
2585
2586 if (!ttrace->filename.ptr)
2587 goto out_put;
2588
2589 entry_str_len = strlen(ttrace->entry_str);
2590 remaining_space = trace__entry_str_size - entry_str_len - 1;
2591 if (remaining_space <= 0)
2592 goto out_put;
2593
2594 if (filename_len > (size_t)remaining_space) {
2595 filename += filename_len - remaining_space;
2596 filename_len = remaining_space;
2597 }
2598
2599 to_move = entry_str_len - ttrace->filename.entry_str_pos + 1;
2600 pos = ttrace->entry_str + ttrace->filename.entry_str_pos;
2601 memmove(pos + filename_len, pos, to_move);
2602 memcpy(pos, filename, filename_len);
2603
2604 ttrace->filename.ptr = 0;
2605 ttrace->filename.entry_str_pos = 0;
2606out_put:
2607 thread__put(thread);
2608out:
2609 return 0;
2610}
2611
2612static int trace__sched_stat_runtime(struct trace *trace, struct evsel *evsel,
2613 union perf_event *event __maybe_unused,
2614 struct perf_sample *sample)
2615{
2616 u64 runtime = evsel__intval(evsel, sample, "runtime");
2617 double runtime_ms = (double)runtime / NSEC_PER_MSEC;
2618 struct thread *thread = machine__findnew_thread(trace->host,
2619 sample->pid,
2620 sample->tid);
2621 struct thread_trace *ttrace = thread__trace(thread, trace->output);
2622
2623 if (ttrace == NULL)
2624 goto out_dump;
2625
2626 ttrace->runtime_ms += runtime_ms;
2627 trace->runtime_ms += runtime_ms;
2628out_put:
2629 thread__put(thread);
2630 return 0;
2631
2632out_dump:
2633 fprintf(trace->output, "%s: comm=%s,pid=%u,runtime=%" PRIu64 ",vruntime=%" PRIu64 ")\n",
2634 evsel->name,
2635 evsel__strval(evsel, sample, "comm"),
2636 (pid_t)evsel__intval(evsel, sample, "pid"),
2637 runtime,
2638 evsel__intval(evsel, sample, "vruntime"));
2639 goto out_put;
2640}
2641
2642static int bpf_output__printer(enum binary_printer_ops op,
2643 unsigned int val, void *extra __maybe_unused, FILE *fp)
2644{
2645 unsigned char ch = (unsigned char)val;
2646
2647 switch (op) {
2648 case BINARY_PRINT_CHAR_DATA:
2649 return fprintf(fp, "%c", isprint(ch) ? ch : '.');
2650 case BINARY_PRINT_DATA_BEGIN:
2651 case BINARY_PRINT_LINE_BEGIN:
2652 case BINARY_PRINT_ADDR:
2653 case BINARY_PRINT_NUM_DATA:
2654 case BINARY_PRINT_NUM_PAD:
2655 case BINARY_PRINT_SEP:
2656 case BINARY_PRINT_CHAR_PAD:
2657 case BINARY_PRINT_LINE_END:
2658 case BINARY_PRINT_DATA_END:
2659 default:
2660 break;
2661 }
2662
2663 return 0;
2664}
2665
2666static void bpf_output__fprintf(struct trace *trace,
2667 struct perf_sample *sample)
2668{
2669 binary__fprintf(sample->raw_data, sample->raw_size, 8,
2670 bpf_output__printer, NULL, trace->output);
2671 ++trace->nr_events_printed;
2672}
2673
2674static size_t trace__fprintf_tp_fields(struct trace *trace, struct evsel *evsel, struct perf_sample *sample,
2675 struct thread *thread, void *augmented_args, int augmented_args_size)
2676{
2677 char bf[2048];
2678 size_t size = sizeof(bf);
2679 struct tep_format_field *field = evsel->tp_format->format.fields;
2680 struct syscall_arg_fmt *arg = __evsel__syscall_arg_fmt(evsel);
2681 size_t printed = 0;
2682 unsigned long val;
2683 u8 bit = 1;
2684 struct syscall_arg syscall_arg = {
2685 .augmented = {
2686 .size = augmented_args_size,
2687 .args = augmented_args,
2688 },
2689 .idx = 0,
2690 .mask = 0,
2691 .trace = trace,
2692 .thread = thread,
2693 .show_string_prefix = trace->show_string_prefix,
2694 };
2695
2696 for (; field && arg; field = field->next, ++syscall_arg.idx, bit <<= 1, ++arg) {
2697 if (syscall_arg.mask & bit)
2698 continue;
2699
2700 syscall_arg.len = 0;
2701 syscall_arg.fmt = arg;
2702 if (field->flags & TEP_FIELD_IS_ARRAY) {
2703 int offset = field->offset;
2704
2705 if (field->flags & TEP_FIELD_IS_DYNAMIC) {
2706 offset = format_field__intval(field, sample, evsel->needs_swap);
2707 syscall_arg.len = offset >> 16;
2708 offset &= 0xffff;
2709 }
2710
2711 val = (uintptr_t)(sample->raw_data + offset);
2712 } else
2713 val = format_field__intval(field, sample, evsel->needs_swap);
2714
2715
2716
2717
2718 val = syscall_arg_fmt__mask_val(arg, &syscall_arg, val);
2719
2720
2721
2722
2723
2724
2725 if (val == 0 &&
2726 !trace->show_zeros &&
2727 !((arg->show_zero ||
2728 arg->scnprintf == SCA_STRARRAY ||
2729 arg->scnprintf == SCA_STRARRAYS) &&
2730 arg->parm))
2731 continue;
2732
2733 printed += scnprintf(bf + printed, size - printed, "%s", printed ? ", " : "");
2734
2735
2736
2737
2738
2739 if (1 || trace->show_arg_names)
2740 printed += scnprintf(bf + printed, size - printed, "%s: ", field->name);
2741
2742 printed += syscall_arg_fmt__scnprintf_val(arg, bf + printed, size - printed, &syscall_arg, val);
2743 }
2744
2745 return printed + fprintf(trace->output, "%s", bf);
2746}
2747
2748static int trace__event_handler(struct trace *trace, struct evsel *evsel,
2749 union perf_event *event __maybe_unused,
2750 struct perf_sample *sample)
2751{
2752 struct thread *thread;
2753 int callchain_ret = 0;
2754
2755
2756
2757
2758
2759
2760 if (evsel->disabled)
2761 return 0;
2762
2763 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2764
2765 if (sample->callchain) {
2766 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
2767 if (callchain_ret == 0) {
2768 if (callchain_cursor.nr < trace->min_stack)
2769 goto out;
2770 callchain_ret = 1;
2771 }
2772 }
2773
2774 trace__printf_interrupted_entry(trace);
2775 trace__fprintf_tstamp(trace, sample->time, trace->output);
2776
2777 if (trace->trace_syscalls && trace->show_duration)
2778 fprintf(trace->output, "( ): ");
2779
2780 if (thread)
2781 trace__fprintf_comm_tid(trace, thread, trace->output);
2782
2783 if (evsel == trace->syscalls.events.augmented) {
2784 int id = perf_evsel__sc_tp_uint(evsel, id, sample);
2785 struct syscall *sc = trace__syscall_info(trace, evsel, id);
2786
2787 if (sc) {
2788 fprintf(trace->output, "%s(", sc->name);
2789 trace__fprintf_sys_enter(trace, evsel, sample);
2790 fputc(')', trace->output);
2791 goto newline;
2792 }
2793
2794
2795
2796
2797
2798
2799 }
2800
2801 fprintf(trace->output, "%s(", evsel->name);
2802
2803 if (evsel__is_bpf_output(evsel)) {
2804 bpf_output__fprintf(trace, sample);
2805 } else if (evsel->tp_format) {
2806 if (strncmp(evsel->tp_format->name, "sys_enter_", 10) ||
2807 trace__fprintf_sys_enter(trace, evsel, sample)) {
2808 if (trace->libtraceevent_print) {
2809 event_format__fprintf(evsel->tp_format, sample->cpu,
2810 sample->raw_data, sample->raw_size,
2811 trace->output);
2812 } else {
2813 trace__fprintf_tp_fields(trace, evsel, sample, thread, NULL, 0);
2814 }
2815 }
2816 }
2817
2818newline:
2819 fprintf(trace->output, ")\n");
2820
2821 if (callchain_ret > 0)
2822 trace__fprintf_callchain(trace, sample);
2823 else if (callchain_ret < 0)
2824 pr_err("Problem processing %s callchain, skipping...\n", evsel__name(evsel));
2825
2826 ++trace->nr_events_printed;
2827
2828 if (evsel->max_events != ULONG_MAX && ++evsel->nr_events_printed == evsel->max_events) {
2829 evsel__disable(evsel);
2830 evsel__close(evsel);
2831 }
2832out:
2833 thread__put(thread);
2834 return 0;
2835}
2836
2837static void print_location(FILE *f, struct perf_sample *sample,
2838 struct addr_location *al,
2839 bool print_dso, bool print_sym)
2840{
2841
2842 if ((verbose > 0 || print_dso) && al->map)
2843 fprintf(f, "%s@", al->map->dso->long_name);
2844
2845 if ((verbose > 0 || print_sym) && al->sym)
2846 fprintf(f, "%s+0x%" PRIx64, al->sym->name,
2847 al->addr - al->sym->start);
2848 else if (al->map)
2849 fprintf(f, "0x%" PRIx64, al->addr);
2850 else
2851 fprintf(f, "0x%" PRIx64, sample->addr);
2852}
2853
2854static int trace__pgfault(struct trace *trace,
2855 struct evsel *evsel,
2856 union perf_event *event __maybe_unused,
2857 struct perf_sample *sample)
2858{
2859 struct thread *thread;
2860 struct addr_location al;
2861 char map_type = 'd';
2862 struct thread_trace *ttrace;
2863 int err = -1;
2864 int callchain_ret = 0;
2865
2866 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2867
2868 if (sample->callchain) {
2869 callchain_ret = trace__resolve_callchain(trace, evsel, sample, &callchain_cursor);
2870 if (callchain_ret == 0) {
2871 if (callchain_cursor.nr < trace->min_stack)
2872 goto out_put;
2873 callchain_ret = 1;
2874 }
2875 }
2876
2877 ttrace = thread__trace(thread, trace->output);
2878 if (ttrace == NULL)
2879 goto out_put;
2880
2881 if (evsel->core.attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ)
2882 ttrace->pfmaj++;
2883 else
2884 ttrace->pfmin++;
2885
2886 if (trace->summary_only)
2887 goto out;
2888
2889 thread__find_symbol(thread, sample->cpumode, sample->ip, &al);
2890
2891 trace__fprintf_entry_head(trace, thread, 0, true, sample->time, trace->output);
2892
2893 fprintf(trace->output, "%sfault [",
2894 evsel->core.attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ ?
2895 "maj" : "min");
2896
2897 print_location(trace->output, sample, &al, false, true);
2898
2899 fprintf(trace->output, "] => ");
2900
2901 thread__find_symbol(thread, sample->cpumode, sample->addr, &al);
2902
2903 if (!al.map) {
2904 thread__find_symbol(thread, sample->cpumode, sample->addr, &al);
2905
2906 if (al.map)
2907 map_type = 'x';
2908 else
2909 map_type = '?';
2910 }
2911
2912 print_location(trace->output, sample, &al, true, false);
2913
2914 fprintf(trace->output, " (%c%c)\n", map_type, al.level);
2915
2916 if (callchain_ret > 0)
2917 trace__fprintf_callchain(trace, sample);
2918 else if (callchain_ret < 0)
2919 pr_err("Problem processing %s callchain, skipping...\n", evsel__name(evsel));
2920
2921 ++trace->nr_events_printed;
2922out:
2923 err = 0;
2924out_put:
2925 thread__put(thread);
2926 return err;
2927}
2928
2929static void trace__set_base_time(struct trace *trace,
2930 struct evsel *evsel,
2931 struct perf_sample *sample)
2932{
2933
2934
2935
2936
2937
2938
2939
2940
2941 if (trace->base_time == 0 && !trace->full_time &&
2942 (evsel->core.attr.sample_type & PERF_SAMPLE_TIME))
2943 trace->base_time = sample->time;
2944}
2945
2946static int trace__process_sample(struct perf_tool *tool,
2947 union perf_event *event,
2948 struct perf_sample *sample,
2949 struct evsel *evsel,
2950 struct machine *machine __maybe_unused)
2951{
2952 struct trace *trace = container_of(tool, struct trace, tool);
2953 struct thread *thread;
2954 int err = 0;
2955
2956 tracepoint_handler handler = evsel->handler;
2957
2958 thread = machine__findnew_thread(trace->host, sample->pid, sample->tid);
2959 if (thread && thread__is_filtered(thread))
2960 goto out;
2961
2962 trace__set_base_time(trace, evsel, sample);
2963
2964 if (handler) {
2965 ++trace->nr_events;
2966 handler(trace, evsel, event, sample);
2967 }
2968out:
2969 thread__put(thread);
2970 return err;
2971}
2972
2973static int trace__record(struct trace *trace, int argc, const char **argv)
2974{
2975 unsigned int rec_argc, i, j;
2976 const char **rec_argv;
2977 const char * const record_args[] = {
2978 "record",
2979 "-R",
2980 "-m", "1024",
2981 "-c", "1",
2982 };
2983 pid_t pid = getpid();
2984 char *filter = asprintf__tp_filter_pids(1, &pid);
2985 const char * const sc_args[] = { "-e", };
2986 unsigned int sc_args_nr = ARRAY_SIZE(sc_args);
2987 const char * const majpf_args[] = { "-e", "major-faults" };
2988 unsigned int majpf_args_nr = ARRAY_SIZE(majpf_args);
2989 const char * const minpf_args[] = { "-e", "minor-faults" };
2990 unsigned int minpf_args_nr = ARRAY_SIZE(minpf_args);
2991 int err = -1;
2992
2993
2994 rec_argc = ARRAY_SIZE(record_args) + sc_args_nr + 3 +
2995 majpf_args_nr + minpf_args_nr + argc;
2996 rec_argv = calloc(rec_argc + 1, sizeof(char *));
2997
2998 if (rec_argv == NULL || filter == NULL)
2999 goto out_free;
3000
3001 j = 0;
3002 for (i = 0; i < ARRAY_SIZE(record_args); i++)
3003 rec_argv[j++] = record_args[i];
3004
3005 if (trace->trace_syscalls) {
3006 for (i = 0; i < sc_args_nr; i++)
3007 rec_argv[j++] = sc_args[i];
3008
3009
3010 if (is_valid_tracepoint("raw_syscalls:sys_enter"))
3011 rec_argv[j++] = "raw_syscalls:sys_enter,raw_syscalls:sys_exit";
3012 else if (is_valid_tracepoint("syscalls:sys_enter"))
3013 rec_argv[j++] = "syscalls:sys_enter,syscalls:sys_exit";
3014 else {
3015 pr_err("Neither raw_syscalls nor syscalls events exist.\n");
3016 goto out_free;
3017 }
3018 }
3019
3020 rec_argv[j++] = "--filter";
3021 rec_argv[j++] = filter;
3022
3023 if (trace->trace_pgfaults & TRACE_PFMAJ)
3024 for (i = 0; i < majpf_args_nr; i++)
3025 rec_argv[j++] = majpf_args[i];
3026
3027 if (trace->trace_pgfaults & TRACE_PFMIN)
3028 for (i = 0; i < minpf_args_nr; i++)
3029 rec_argv[j++] = minpf_args[i];
3030
3031 for (i = 0; i < (unsigned int)argc; i++)
3032 rec_argv[j++] = argv[i];
3033
3034 err = cmd_record(j, rec_argv);
3035out_free:
3036 free(filter);
3037 free(rec_argv);
3038 return err;
3039}
3040
3041static size_t trace__fprintf_thread_summary(struct trace *trace, FILE *fp);
3042
3043static bool evlist__add_vfs_getname(struct evlist *evlist)
3044{
3045 bool found = false;
3046 struct evsel *evsel, *tmp;
3047 struct parse_events_error err;
3048 int ret;
3049
3050 bzero(&err, sizeof(err));
3051 ret = parse_events(evlist, "probe:vfs_getname*", &err);
3052 if (ret) {
3053 free(err.str);
3054 free(err.help);
3055 free(err.first_str);
3056 free(err.first_help);
3057 return false;
3058 }
3059
3060 evlist__for_each_entry_safe(evlist, evsel, tmp) {
3061 if (!strstarts(evsel__name(evsel), "probe:vfs_getname"))
3062 continue;
3063
3064 if (evsel__field(evsel, "pathname")) {
3065 evsel->handler = trace__vfs_getname;
3066 found = true;
3067 continue;
3068 }
3069
3070 list_del_init(&evsel->core.node);
3071 evsel->evlist = NULL;
3072 evsel__delete(evsel);
3073 }
3074
3075 return found;
3076}
3077
3078static struct evsel *evsel__new_pgfault(u64 config)
3079{
3080 struct evsel *evsel;
3081 struct perf_event_attr attr = {
3082 .type = PERF_TYPE_SOFTWARE,
3083 .mmap_data = 1,
3084 };
3085
3086 attr.config = config;
3087 attr.sample_period = 1;
3088
3089 event_attr_init(&attr);
3090
3091 evsel = evsel__new(&attr);
3092 if (evsel)
3093 evsel->handler = trace__pgfault;
3094
3095 return evsel;
3096}
3097
3098static void trace__handle_event(struct trace *trace, union perf_event *event, struct perf_sample *sample)
3099{
3100 const u32 type = event->header.type;
3101 struct evsel *evsel;
3102
3103 if (type != PERF_RECORD_SAMPLE) {
3104 trace__process_event(trace, trace->host, event, sample);
3105 return;
3106 }
3107
3108 evsel = evlist__id2evsel(trace->evlist, sample->id);
3109 if (evsel == NULL) {
3110 fprintf(trace->output, "Unknown tp ID %" PRIu64 ", skipping...\n", sample->id);
3111 return;
3112 }
3113
3114 if (evswitch__discard(&trace->evswitch, evsel))
3115 return;
3116
3117 trace__set_base_time(trace, evsel, sample);
3118
3119 if (evsel->core.attr.type == PERF_TYPE_TRACEPOINT &&
3120 sample->raw_data == NULL) {
3121 fprintf(trace->output, "%s sample with no payload for tid: %d, cpu %d, raw_size=%d, skipping...\n",
3122 evsel__name(evsel), sample->tid,
3123 sample->cpu, sample->raw_size);
3124 } else {
3125 tracepoint_handler handler = evsel->handler;
3126 handler(trace, evsel, event, sample);
3127 }
3128
3129 if (trace->nr_events_printed >= trace->max_events && trace->max_events != ULONG_MAX)
3130 interrupted = true;
3131}
3132
3133static int trace__add_syscall_newtp(struct trace *trace)
3134{
3135 int ret = -1;
3136 struct evlist *evlist = trace->evlist;
3137 struct evsel *sys_enter, *sys_exit;
3138
3139 sys_enter = perf_evsel__raw_syscall_newtp("sys_enter", trace__sys_enter);
3140 if (sys_enter == NULL)
3141 goto out;
3142
3143 if (perf_evsel__init_sc_tp_ptr_field(sys_enter, args))
3144 goto out_delete_sys_enter;
3145
3146 sys_exit = perf_evsel__raw_syscall_newtp("sys_exit", trace__sys_exit);
3147 if (sys_exit == NULL)
3148 goto out_delete_sys_enter;
3149
3150 if (perf_evsel__init_sc_tp_uint_field(sys_exit, ret))
3151 goto out_delete_sys_exit;
3152
3153 evsel__config_callchain(sys_enter, &trace->opts, &callchain_param);
3154 evsel__config_callchain(sys_exit, &trace->opts, &callchain_param);
3155
3156 evlist__add(evlist, sys_enter);
3157 evlist__add(evlist, sys_exit);
3158
3159 if (callchain_param.enabled && !trace->kernel_syscallchains) {
3160
3161
3162
3163
3164
3165 sys_exit->core.attr.exclude_callchain_kernel = 1;
3166 }
3167
3168 trace->syscalls.events.sys_enter = sys_enter;
3169 trace->syscalls.events.sys_exit = sys_exit;
3170
3171 ret = 0;
3172out:
3173 return ret;
3174
3175out_delete_sys_exit:
3176 evsel__delete_priv(sys_exit);
3177out_delete_sys_enter:
3178 evsel__delete_priv(sys_enter);
3179 goto out;
3180}
3181
3182static int trace__set_ev_qualifier_tp_filter(struct trace *trace)
3183{
3184 int err = -1;
3185 struct evsel *sys_exit;
3186 char *filter = asprintf_expr_inout_ints("id", !trace->not_ev_qualifier,
3187 trace->ev_qualifier_ids.nr,
3188 trace->ev_qualifier_ids.entries);
3189
3190 if (filter == NULL)
3191 goto out_enomem;
3192
3193 if (!evsel__append_tp_filter(trace->syscalls.events.sys_enter, filter)) {
3194 sys_exit = trace->syscalls.events.sys_exit;
3195 err = evsel__append_tp_filter(sys_exit, filter);
3196 }
3197
3198 free(filter);
3199out:
3200 return err;
3201out_enomem:
3202 errno = ENOMEM;
3203 goto out;
3204}
3205
3206#ifdef HAVE_LIBBPF_SUPPORT
3207static struct bpf_map *trace__find_bpf_map_by_name(struct trace *trace, const char *name)
3208{
3209 if (trace->bpf_obj == NULL)
3210 return NULL;
3211
3212 return bpf_object__find_map_by_name(trace->bpf_obj, name);
3213}
3214
3215static void trace__set_bpf_map_filtered_pids(struct trace *trace)
3216{
3217 trace->filter_pids.map = trace__find_bpf_map_by_name(trace, "pids_filtered");
3218}
3219
3220static void trace__set_bpf_map_syscalls(struct trace *trace)
3221{
3222 trace->syscalls.map = trace__find_bpf_map_by_name(trace, "syscalls");
3223 trace->syscalls.prog_array.sys_enter = trace__find_bpf_map_by_name(trace, "syscalls_sys_enter");
3224 trace->syscalls.prog_array.sys_exit = trace__find_bpf_map_by_name(trace, "syscalls_sys_exit");
3225}
3226
3227static struct bpf_program *trace__find_bpf_program_by_title(struct trace *trace, const char *name)
3228{
3229 if (trace->bpf_obj == NULL)
3230 return NULL;
3231
3232 return bpf_object__find_program_by_title(trace->bpf_obj, name);
3233}
3234
3235static struct bpf_program *trace__find_syscall_bpf_prog(struct trace *trace, struct syscall *sc,
3236 const char *prog_name, const char *type)
3237{
3238 struct bpf_program *prog;
3239
3240 if (prog_name == NULL) {
3241 char default_prog_name[256];
3242 scnprintf(default_prog_name, sizeof(default_prog_name), "!syscalls:sys_%s_%s", type, sc->name);
3243 prog = trace__find_bpf_program_by_title(trace, default_prog_name);
3244 if (prog != NULL)
3245 goto out_found;
3246 if (sc->fmt && sc->fmt->alias) {
3247 scnprintf(default_prog_name, sizeof(default_prog_name), "!syscalls:sys_%s_%s", type, sc->fmt->alias);
3248 prog = trace__find_bpf_program_by_title(trace, default_prog_name);
3249 if (prog != NULL)
3250 goto out_found;
3251 }
3252 goto out_unaugmented;
3253 }
3254
3255 prog = trace__find_bpf_program_by_title(trace, prog_name);
3256
3257 if (prog != NULL) {
3258out_found:
3259 return prog;
3260 }
3261
3262 pr_debug("Couldn't find BPF prog \"%s\" to associate with syscalls:sys_%s_%s, not augmenting it\n",
3263 prog_name, type, sc->name);
3264out_unaugmented:
3265 return trace->syscalls.unaugmented_prog;
3266}
3267
3268static void trace__init_syscall_bpf_progs(struct trace *trace, int id)
3269{
3270 struct syscall *sc = trace__syscall_info(trace, NULL, id);
3271
3272 if (sc == NULL)
3273 return;
3274
3275 sc->bpf_prog.sys_enter = trace__find_syscall_bpf_prog(trace, sc, sc->fmt ? sc->fmt->bpf_prog_name.sys_enter : NULL, "enter");
3276 sc->bpf_prog.sys_exit = trace__find_syscall_bpf_prog(trace, sc, sc->fmt ? sc->fmt->bpf_prog_name.sys_exit : NULL, "exit");
3277}
3278
3279static int trace__bpf_prog_sys_enter_fd(struct trace *trace, int id)
3280{
3281 struct syscall *sc = trace__syscall_info(trace, NULL, id);
3282 return sc ? bpf_program__fd(sc->bpf_prog.sys_enter) : bpf_program__fd(trace->syscalls.unaugmented_prog);
3283}
3284
3285static int trace__bpf_prog_sys_exit_fd(struct trace *trace, int id)
3286{
3287 struct syscall *sc = trace__syscall_info(trace, NULL, id);
3288 return sc ? bpf_program__fd(sc->bpf_prog.sys_exit) : bpf_program__fd(trace->syscalls.unaugmented_prog);
3289}
3290
3291static void trace__init_bpf_map_syscall_args(struct trace *trace, int id, struct bpf_map_syscall_entry *entry)
3292{
3293 struct syscall *sc = trace__syscall_info(trace, NULL, id);
3294 int arg = 0;
3295
3296 if (sc == NULL)
3297 goto out;
3298
3299 for (; arg < sc->nr_args; ++arg) {
3300 entry->string_args_len[arg] = 0;
3301 if (sc->arg_fmt[arg].scnprintf == SCA_FILENAME) {
3302
3303 entry->string_args_len[arg] = PATH_MAX;
3304 }
3305 }
3306out:
3307 for (; arg < 6; ++arg)
3308 entry->string_args_len[arg] = 0;
3309}
3310static int trace__set_ev_qualifier_bpf_filter(struct trace *trace)
3311{
3312 int fd = bpf_map__fd(trace->syscalls.map);
3313 struct bpf_map_syscall_entry value = {
3314 .enabled = !trace->not_ev_qualifier,
3315 };
3316 int err = 0;
3317 size_t i;
3318
3319 for (i = 0; i < trace->ev_qualifier_ids.nr; ++i) {
3320 int key = trace->ev_qualifier_ids.entries[i];
3321
3322 if (value.enabled) {
3323 trace__init_bpf_map_syscall_args(trace, key, &value);
3324 trace__init_syscall_bpf_progs(trace, key);
3325 }
3326
3327 err = bpf_map_update_elem(fd, &key, &value, BPF_EXIST);
3328 if (err)
3329 break;
3330 }
3331
3332 return err;
3333}
3334
3335static int __trace__init_syscalls_bpf_map(struct trace *trace, bool enabled)
3336{
3337 int fd = bpf_map__fd(trace->syscalls.map);
3338 struct bpf_map_syscall_entry value = {
3339 .enabled = enabled,
3340 };
3341 int err = 0, key;
3342
3343 for (key = 0; key < trace->sctbl->syscalls.nr_entries; ++key) {
3344 if (enabled)
3345 trace__init_bpf_map_syscall_args(trace, key, &value);
3346
3347 err = bpf_map_update_elem(fd, &key, &value, BPF_ANY);
3348 if (err)
3349 break;
3350 }
3351
3352 return err;
3353}
3354
3355static int trace__init_syscalls_bpf_map(struct trace *trace)
3356{
3357 bool enabled = true;
3358
3359 if (trace->ev_qualifier_ids.nr)
3360 enabled = trace->not_ev_qualifier;
3361
3362 return __trace__init_syscalls_bpf_map(trace, enabled);
3363}
3364
3365static struct bpf_program *trace__find_usable_bpf_prog_entry(struct trace *trace, struct syscall *sc)
3366{
3367 struct tep_format_field *field, *candidate_field;
3368 int id;
3369
3370
3371
3372
3373 for (field = sc->args; field; field = field->next) {
3374 if (field->flags & TEP_FIELD_IS_POINTER)
3375 goto try_to_find_pair;
3376 }
3377
3378 return NULL;
3379
3380try_to_find_pair:
3381 for (id = 0; id < trace->sctbl->syscalls.nr_entries; ++id) {
3382 struct syscall *pair = trace__syscall_info(trace, NULL, id);
3383 struct bpf_program *pair_prog;
3384 bool is_candidate = false;
3385
3386 if (pair == NULL || pair == sc ||
3387 pair->bpf_prog.sys_enter == trace->syscalls.unaugmented_prog)
3388 continue;
3389
3390 for (field = sc->args, candidate_field = pair->args;
3391 field && candidate_field; field = field->next, candidate_field = candidate_field->next) {
3392 bool is_pointer = field->flags & TEP_FIELD_IS_POINTER,
3393 candidate_is_pointer = candidate_field->flags & TEP_FIELD_IS_POINTER;
3394
3395 if (is_pointer) {
3396 if (!candidate_is_pointer) {
3397
3398 continue;
3399 }
3400 } else {
3401 if (candidate_is_pointer) {
3402
3403 goto next_candidate;
3404 }
3405 continue;
3406 }
3407
3408 if (strcmp(field->type, candidate_field->type))
3409 goto next_candidate;
3410
3411 is_candidate = true;
3412 }
3413
3414 if (!is_candidate)
3415 goto next_candidate;
3416
3417
3418
3419
3420
3421
3422 if (candidate_field) {
3423 for (candidate_field = candidate_field->next; candidate_field; candidate_field = candidate_field->next)
3424 if (candidate_field->flags & TEP_FIELD_IS_POINTER)
3425 goto next_candidate;
3426 }
3427
3428 pair_prog = pair->bpf_prog.sys_enter;
3429
3430
3431
3432
3433
3434
3435
3436
3437
3438 if (pair_prog == NULL) {
3439 pair_prog = trace__find_syscall_bpf_prog(trace, pair, pair->fmt ? pair->fmt->bpf_prog_name.sys_enter : NULL, "enter");
3440 if (pair_prog == trace->syscalls.unaugmented_prog)
3441 goto next_candidate;
3442 }
3443
3444 pr_debug("Reusing \"%s\" BPF sys_enter augmenter for \"%s\"\n", pair->name, sc->name);
3445 return pair_prog;
3446 next_candidate:
3447 continue;
3448 }
3449
3450 return NULL;
3451}
3452
3453static int trace__init_syscalls_bpf_prog_array_maps(struct trace *trace)
3454{
3455 int map_enter_fd = bpf_map__fd(trace->syscalls.prog_array.sys_enter),
3456 map_exit_fd = bpf_map__fd(trace->syscalls.prog_array.sys_exit);
3457 int err = 0, key;
3458
3459 for (key = 0; key < trace->sctbl->syscalls.nr_entries; ++key) {
3460 int prog_fd;
3461
3462 if (!trace__syscall_enabled(trace, key))
3463 continue;
3464
3465 trace__init_syscall_bpf_progs(trace, key);
3466
3467
3468 prog_fd = trace__bpf_prog_sys_enter_fd(trace, key);
3469 err = bpf_map_update_elem(map_enter_fd, &key, &prog_fd, BPF_ANY);
3470 if (err)
3471 break;
3472 prog_fd = trace__bpf_prog_sys_exit_fd(trace, key);
3473 err = bpf_map_update_elem(map_exit_fd, &key, &prog_fd, BPF_ANY);
3474 if (err)
3475 break;
3476 }
3477
3478
3479
3480
3481
3482
3483
3484
3485
3486
3487
3488
3489
3490
3491
3492
3493
3494
3495
3496
3497
3498
3499
3500
3501
3502
3503
3504
3505
3506 for (key = 0; key < trace->sctbl->syscalls.nr_entries; ++key) {
3507 struct syscall *sc = trace__syscall_info(trace, NULL, key);
3508 struct bpf_program *pair_prog;
3509 int prog_fd;
3510
3511 if (sc == NULL || sc->bpf_prog.sys_enter == NULL)
3512 continue;
3513
3514
3515
3516
3517
3518 if (sc->bpf_prog.sys_enter != trace->syscalls.unaugmented_prog)
3519 continue;
3520
3521
3522
3523
3524
3525 pair_prog = trace__find_usable_bpf_prog_entry(trace, sc);
3526 if (pair_prog == NULL)
3527 continue;
3528
3529 sc->bpf_prog.sys_enter = pair_prog;
3530
3531
3532
3533
3534
3535 prog_fd = bpf_program__fd(sc->bpf_prog.sys_enter);
3536 err = bpf_map_update_elem(map_enter_fd, &key, &prog_fd, BPF_ANY);
3537 if (err)
3538 break;
3539 }
3540
3541
3542 return err;
3543}
3544
3545static void trace__delete_augmented_syscalls(struct trace *trace)
3546{
3547 struct evsel *evsel, *tmp;
3548
3549 evlist__remove(trace->evlist, trace->syscalls.events.augmented);
3550 evsel__delete(trace->syscalls.events.augmented);
3551 trace->syscalls.events.augmented = NULL;
3552
3553 evlist__for_each_entry_safe(trace->evlist, tmp, evsel) {
3554 if (evsel->bpf_obj == trace->bpf_obj) {
3555 evlist__remove(trace->evlist, evsel);
3556 evsel__delete(evsel);
3557 }
3558
3559 }
3560
3561 bpf_object__close(trace->bpf_obj);
3562 trace->bpf_obj = NULL;
3563}
3564#else
3565static struct bpf_map *trace__find_bpf_map_by_name(struct trace *trace __maybe_unused,
3566 const char *name __maybe_unused)
3567{
3568 return NULL;
3569}
3570
3571static void trace__set_bpf_map_filtered_pids(struct trace *trace __maybe_unused)
3572{
3573}
3574
3575static void trace__set_bpf_map_syscalls(struct trace *trace __maybe_unused)
3576{
3577}
3578
3579static int trace__set_ev_qualifier_bpf_filter(struct trace *trace __maybe_unused)
3580{
3581 return 0;
3582}
3583
3584static int trace__init_syscalls_bpf_map(struct trace *trace __maybe_unused)
3585{
3586 return 0;
3587}
3588
3589static struct bpf_program *trace__find_bpf_program_by_title(struct trace *trace __maybe_unused,
3590 const char *name __maybe_unused)
3591{
3592 return NULL;
3593}
3594
3595static int trace__init_syscalls_bpf_prog_array_maps(struct trace *trace __maybe_unused)
3596{
3597 return 0;
3598}
3599
3600static void trace__delete_augmented_syscalls(struct trace *trace __maybe_unused)
3601{
3602}
3603#endif
3604
3605static bool trace__only_augmented_syscalls_evsels(struct trace *trace)
3606{
3607 struct evsel *evsel;
3608
3609 evlist__for_each_entry(trace->evlist, evsel) {
3610 if (evsel == trace->syscalls.events.augmented ||
3611 evsel->bpf_obj == trace->bpf_obj)
3612 continue;
3613
3614 return false;
3615 }
3616
3617 return true;
3618}
3619
3620static int trace__set_ev_qualifier_filter(struct trace *trace)
3621{
3622 if (trace->syscalls.map)
3623 return trace__set_ev_qualifier_bpf_filter(trace);
3624 if (trace->syscalls.events.sys_enter)
3625 return trace__set_ev_qualifier_tp_filter(trace);
3626 return 0;
3627}
3628
3629static int bpf_map__set_filter_pids(struct bpf_map *map __maybe_unused,
3630 size_t npids __maybe_unused, pid_t *pids __maybe_unused)
3631{
3632 int err = 0;
3633#ifdef HAVE_LIBBPF_SUPPORT
3634 bool value = true;
3635 int map_fd = bpf_map__fd(map);
3636 size_t i;
3637
3638 for (i = 0; i < npids; ++i) {
3639 err = bpf_map_update_elem(map_fd, &pids[i], &value, BPF_ANY);
3640 if (err)
3641 break;
3642 }
3643#endif
3644 return err;
3645}
3646
3647static int trace__set_filter_loop_pids(struct trace *trace)
3648{
3649 unsigned int nr = 1, err;
3650 pid_t pids[32] = {
3651 getpid(),
3652 };
3653 struct thread *thread = machine__find_thread(trace->host, pids[0], pids[0]);
3654
3655 while (thread && nr < ARRAY_SIZE(pids)) {
3656 struct thread *parent = machine__find_thread(trace->host, thread->ppid, thread->ppid);
3657
3658 if (parent == NULL)
3659 break;
3660
3661 if (!strcmp(thread__comm_str(parent), "sshd") ||
3662 strstarts(thread__comm_str(parent), "gnome-terminal")) {
3663 pids[nr++] = parent->tid;
3664 break;
3665 }
3666 thread = parent;
3667 }
3668
3669 err = evlist__append_tp_filter_pids(trace->evlist, nr, pids);
3670 if (!err && trace->filter_pids.map)
3671 err = bpf_map__set_filter_pids(trace->filter_pids.map, nr, pids);
3672
3673 return err;
3674}
3675
3676static int trace__set_filter_pids(struct trace *trace)
3677{
3678 int err = 0;
3679
3680
3681
3682
3683
3684
3685 if (trace->filter_pids.nr > 0) {
3686 err = evlist__append_tp_filter_pids(trace->evlist, trace->filter_pids.nr,
3687 trace->filter_pids.entries);
3688 if (!err && trace->filter_pids.map) {
3689 err = bpf_map__set_filter_pids(trace->filter_pids.map, trace->filter_pids.nr,
3690 trace->filter_pids.entries);
3691 }
3692 } else if (perf_thread_map__pid(trace->evlist->core.threads, 0) == -1) {
3693 err = trace__set_filter_loop_pids(trace);
3694 }
3695
3696 return err;
3697}
3698
3699static int __trace__deliver_event(struct trace *trace, union perf_event *event)
3700{
3701 struct evlist *evlist = trace->evlist;
3702 struct perf_sample sample;
3703 int err = evlist__parse_sample(evlist, event, &sample);
3704
3705 if (err)
3706 fprintf(trace->output, "Can't parse sample, err = %d, skipping...\n", err);
3707 else
3708 trace__handle_event(trace, event, &sample);
3709
3710 return 0;
3711}
3712
3713static int __trace__flush_events(struct trace *trace)
3714{
3715 u64 first = ordered_events__first_time(&trace->oe.data);
3716 u64 flush = trace->oe.last - NSEC_PER_SEC;
3717
3718
3719 if (first && first < flush)
3720 return ordered_events__flush_time(&trace->oe.data, flush);
3721
3722 return 0;
3723}
3724
3725static int trace__flush_events(struct trace *trace)
3726{
3727 return !trace->sort_events ? 0 : __trace__flush_events(trace);
3728}
3729
3730static int trace__deliver_event(struct trace *trace, union perf_event *event)
3731{
3732 int err;
3733
3734 if (!trace->sort_events)
3735 return __trace__deliver_event(trace, event);
3736
3737 err = evlist__parse_sample_timestamp(trace->evlist, event, &trace->oe.last);
3738 if (err && err != -1)
3739 return err;
3740
3741 err = ordered_events__queue(&trace->oe.data, event, trace->oe.last, 0);
3742 if (err)
3743 return err;
3744
3745 return trace__flush_events(trace);
3746}
3747
3748static int ordered_events__deliver_event(struct ordered_events *oe,
3749 struct ordered_event *event)
3750{
3751 struct trace *trace = container_of(oe, struct trace, oe.data);
3752
3753 return __trace__deliver_event(trace, event->event);
3754}
3755
3756static struct syscall_arg_fmt *evsel__find_syscall_arg_fmt_by_name(struct evsel *evsel, char *arg)
3757{
3758 struct tep_format_field *field;
3759 struct syscall_arg_fmt *fmt = __evsel__syscall_arg_fmt(evsel);
3760
3761 if (evsel->tp_format == NULL || fmt == NULL)
3762 return NULL;
3763
3764 for (field = evsel->tp_format->format.fields; field; field = field->next, ++fmt)
3765 if (strcmp(field->name, arg) == 0)
3766 return fmt;
3767
3768 return NULL;
3769}
3770
3771static int trace__expand_filter(struct trace *trace __maybe_unused, struct evsel *evsel)
3772{
3773 char *tok, *left = evsel->filter, *new_filter = evsel->filter;
3774
3775 while ((tok = strpbrk(left, "=<>!")) != NULL) {
3776 char *right = tok + 1, *right_end;
3777
3778 if (*right == '=')
3779 ++right;
3780
3781 while (isspace(*right))
3782 ++right;
3783
3784 if (*right == '\0')
3785 break;
3786
3787 while (!isalpha(*left))
3788 if (++left == tok) {
3789
3790
3791
3792
3793 return 0;
3794 }
3795
3796 right_end = right + 1;
3797 while (isalnum(*right_end) || *right_end == '_' || *right_end == '|')
3798 ++right_end;
3799
3800 if (isalpha(*right)) {
3801 struct syscall_arg_fmt *fmt;
3802 int left_size = tok - left,
3803 right_size = right_end - right;
3804 char arg[128];
3805
3806 while (isspace(left[left_size - 1]))
3807 --left_size;
3808
3809 scnprintf(arg, sizeof(arg), "%.*s", left_size, left);
3810
3811 fmt = evsel__find_syscall_arg_fmt_by_name(evsel, arg);
3812 if (fmt == NULL) {
3813 pr_err("\"%s\" not found in \"%s\", can't set filter \"%s\"\n",
3814 arg, evsel->name, evsel->filter);
3815 return -1;
3816 }
3817
3818 pr_debug2("trying to expand \"%s\" \"%.*s\" \"%.*s\" -> ",
3819 arg, (int)(right - tok), tok, right_size, right);
3820
3821 if (fmt->strtoul) {
3822 u64 val;
3823 struct syscall_arg syscall_arg = {
3824 .parm = fmt->parm,
3825 };
3826
3827 if (fmt->strtoul(right, right_size, &syscall_arg, &val)) {
3828 char *n, expansion[19];
3829 int expansion_lenght = scnprintf(expansion, sizeof(expansion), "%#" PRIx64, val);
3830 int expansion_offset = right - new_filter;
3831
3832 pr_debug("%s", expansion);
3833
3834 if (asprintf(&n, "%.*s%s%s", expansion_offset, new_filter, expansion, right_end) < 0) {
3835 pr_debug(" out of memory!\n");
3836 free(new_filter);
3837 return -1;
3838 }
3839 if (new_filter != evsel->filter)
3840 free(new_filter);
3841 left = n + expansion_offset + expansion_lenght;
3842 new_filter = n;
3843 } else {
3844 pr_err("\"%.*s\" not found for \"%s\" in \"%s\", can't set filter \"%s\"\n",
3845 right_size, right, arg, evsel->name, evsel->filter);
3846 return -1;
3847 }
3848 } else {
3849 pr_err("No resolver (strtoul) for \"%s\" in \"%s\", can't set filter \"%s\"\n",
3850 arg, evsel->name, evsel->filter);
3851 return -1;
3852 }
3853
3854 pr_debug("\n");
3855 } else {
3856 left = right_end;
3857 }
3858 }
3859
3860 if (new_filter != evsel->filter) {
3861 pr_debug("New filter for %s: %s\n", evsel->name, new_filter);
3862 evsel__set_filter(evsel, new_filter);
3863 free(new_filter);
3864 }
3865
3866 return 0;
3867}
3868
3869static int trace__expand_filters(struct trace *trace, struct evsel **err_evsel)
3870{
3871 struct evlist *evlist = trace->evlist;
3872 struct evsel *evsel;
3873
3874 evlist__for_each_entry(evlist, evsel) {
3875 if (evsel->filter == NULL)
3876 continue;
3877
3878 if (trace__expand_filter(trace, evsel)) {
3879 *err_evsel = evsel;
3880 return -1;
3881 }
3882 }
3883
3884 return 0;
3885}
3886
3887static int trace__run(struct trace *trace, int argc, const char **argv)
3888{
3889 struct evlist *evlist = trace->evlist;
3890 struct evsel *evsel, *pgfault_maj = NULL, *pgfault_min = NULL;
3891 int err = -1, i;
3892 unsigned long before;
3893 const bool forks = argc > 0;
3894 bool draining = false;
3895
3896 trace->live = true;
3897
3898 if (!trace->raw_augmented_syscalls) {
3899 if (trace->trace_syscalls && trace__add_syscall_newtp(trace))
3900 goto out_error_raw_syscalls;
3901
3902 if (trace->trace_syscalls)
3903 trace->vfs_getname = evlist__add_vfs_getname(evlist);
3904 }
3905
3906 if ((trace->trace_pgfaults & TRACE_PFMAJ)) {
3907 pgfault_maj = evsel__new_pgfault(PERF_COUNT_SW_PAGE_FAULTS_MAJ);
3908 if (pgfault_maj == NULL)
3909 goto out_error_mem;
3910 evsel__config_callchain(pgfault_maj, &trace->opts, &callchain_param);
3911 evlist__add(evlist, pgfault_maj);
3912 }
3913
3914 if ((trace->trace_pgfaults & TRACE_PFMIN)) {
3915 pgfault_min = evsel__new_pgfault(PERF_COUNT_SW_PAGE_FAULTS_MIN);
3916 if (pgfault_min == NULL)
3917 goto out_error_mem;
3918 evsel__config_callchain(pgfault_min, &trace->opts, &callchain_param);
3919 evlist__add(evlist, pgfault_min);
3920 }
3921
3922 if (trace->sched &&
3923 evlist__add_newtp(evlist, "sched", "sched_stat_runtime", trace__sched_stat_runtime))
3924 goto out_error_sched_stat_runtime;
3925
3926
3927
3928
3929
3930
3931
3932
3933
3934
3935
3936
3937
3938
3939
3940
3941
3942
3943
3944
3945
3946
3947
3948
3949
3950 if (trace->cgroup)
3951 evlist__set_default_cgroup(trace->evlist, trace->cgroup);
3952
3953 err = evlist__create_maps(evlist, &trace->opts.target);
3954 if (err < 0) {
3955 fprintf(trace->output, "Problems parsing the target to trace, check your options!\n");
3956 goto out_delete_evlist;
3957 }
3958
3959 err = trace__symbols_init(trace, evlist);
3960 if (err < 0) {
3961 fprintf(trace->output, "Problems initializing symbol libraries!\n");
3962 goto out_delete_evlist;
3963 }
3964
3965 evlist__config(evlist, &trace->opts, &callchain_param);
3966
3967 if (forks) {
3968 err = evlist__prepare_workload(evlist, &trace->opts.target, argv, false, NULL);
3969 if (err < 0) {
3970 fprintf(trace->output, "Couldn't run the workload!\n");
3971 goto out_delete_evlist;
3972 }
3973 }
3974
3975 err = evlist__open(evlist);
3976 if (err < 0)
3977 goto out_error_open;
3978
3979 err = bpf__apply_obj_config();
3980 if (err) {
3981 char errbuf[BUFSIZ];
3982
3983 bpf__strerror_apply_obj_config(err, errbuf, sizeof(errbuf));
3984 pr_err("ERROR: Apply config to BPF failed: %s\n",
3985 errbuf);
3986 goto out_error_open;
3987 }
3988
3989 err = trace__set_filter_pids(trace);
3990 if (err < 0)
3991 goto out_error_mem;
3992
3993 if (trace->syscalls.map)
3994 trace__init_syscalls_bpf_map(trace);
3995
3996 if (trace->syscalls.prog_array.sys_enter)
3997 trace__init_syscalls_bpf_prog_array_maps(trace);
3998
3999 if (trace->ev_qualifier_ids.nr > 0) {
4000 err = trace__set_ev_qualifier_filter(trace);
4001 if (err < 0)
4002 goto out_errno;
4003
4004 if (trace->syscalls.events.sys_exit) {
4005 pr_debug("event qualifier tracepoint filter: %s\n",
4006 trace->syscalls.events.sys_exit->filter);
4007 }
4008 }
4009
4010
4011
4012
4013
4014
4015
4016
4017
4018
4019
4020
4021 trace->fd_path_disabled = !trace__syscall_enabled(trace, syscalltbl__id(trace->sctbl, "close"));
4022
4023 err = trace__expand_filters(trace, &evsel);
4024 if (err)
4025 goto out_delete_evlist;
4026 err = evlist__apply_filters(evlist, &evsel);
4027 if (err < 0)
4028 goto out_error_apply_filters;
4029
4030 if (trace->dump.map)
4031 bpf_map__fprintf(trace->dump.map, trace->output);
4032
4033 err = evlist__mmap(evlist, trace->opts.mmap_pages);
4034 if (err < 0)
4035 goto out_error_mmap;
4036
4037 if (!target__none(&trace->opts.target) && !trace->opts.initial_delay)
4038 evlist__enable(evlist);
4039
4040 if (forks)
4041 evlist__start_workload(evlist);
4042
4043 if (trace->opts.initial_delay) {
4044 usleep(trace->opts.initial_delay * 1000);
4045 evlist__enable(evlist);
4046 }
4047
4048 trace->multiple_threads = perf_thread_map__pid(evlist->core.threads, 0) == -1 ||
4049 evlist->core.threads->nr > 1 ||
4050 evlist__first(evlist)->core.attr.inherit;
4051
4052
4053
4054
4055
4056
4057
4058 evlist__for_each_entry(evlist, evsel) {
4059 if (evsel__has_callchain(evsel) &&
4060 evsel->core.attr.sample_max_stack == 0)
4061 evsel->core.attr.sample_max_stack = trace->max_stack;
4062 }
4063again:
4064 before = trace->nr_events;
4065
4066 for (i = 0; i < evlist->core.nr_mmaps; i++) {
4067 union perf_event *event;
4068 struct mmap *md;
4069
4070 md = &evlist->mmap[i];
4071 if (perf_mmap__read_init(&md->core) < 0)
4072 continue;
4073
4074 while ((event = perf_mmap__read_event(&md->core)) != NULL) {
4075 ++trace->nr_events;
4076
4077 err = trace__deliver_event(trace, event);
4078 if (err)
4079 goto out_disable;
4080
4081 perf_mmap__consume(&md->core);
4082
4083 if (interrupted)
4084 goto out_disable;
4085
4086 if (done && !draining) {
4087 evlist__disable(evlist);
4088 draining = true;
4089 }
4090 }
4091 perf_mmap__read_done(&md->core);
4092 }
4093
4094 if (trace->nr_events == before) {
4095 int timeout = done ? 100 : -1;
4096
4097 if (!draining && evlist__poll(evlist, timeout) > 0) {
4098 if (evlist__filter_pollfd(evlist, POLLERR | POLLHUP | POLLNVAL) == 0)
4099 draining = true;
4100
4101 goto again;
4102 } else {
4103 if (trace__flush_events(trace))
4104 goto out_disable;
4105 }
4106 } else {
4107 goto again;
4108 }
4109
4110out_disable:
4111 thread__zput(trace->current);
4112
4113 evlist__disable(evlist);
4114
4115 if (trace->sort_events)
4116 ordered_events__flush(&trace->oe.data, OE_FLUSH__FINAL);
4117
4118 if (!err) {
4119 if (trace->summary)
4120 trace__fprintf_thread_summary(trace, trace->output);
4121
4122 if (trace->show_tool_stats) {
4123 fprintf(trace->output, "Stats:\n "
4124 " vfs_getname : %" PRIu64 "\n"
4125 " proc_getname: %" PRIu64 "\n",
4126 trace->stats.vfs_getname,
4127 trace->stats.proc_getname);
4128 }
4129 }
4130
4131out_delete_evlist:
4132 trace__symbols__exit(trace);
4133
4134 evlist__delete(evlist);
4135 cgroup__put(trace->cgroup);
4136 trace->evlist = NULL;
4137 trace->live = false;
4138 return err;
4139{
4140 char errbuf[BUFSIZ];
4141
4142out_error_sched_stat_runtime:
4143 tracing_path__strerror_open_tp(errno, errbuf, sizeof(errbuf), "sched", "sched_stat_runtime");
4144 goto out_error;
4145
4146out_error_raw_syscalls:
4147 tracing_path__strerror_open_tp(errno, errbuf, sizeof(errbuf), "raw_syscalls", "sys_(enter|exit)");
4148 goto out_error;
4149
4150out_error_mmap:
4151 evlist__strerror_mmap(evlist, errno, errbuf, sizeof(errbuf));
4152 goto out_error;
4153
4154out_error_open:
4155 evlist__strerror_open(evlist, errno, errbuf, sizeof(errbuf));
4156
4157out_error:
4158 fprintf(trace->output, "%s\n", errbuf);
4159 goto out_delete_evlist;
4160
4161out_error_apply_filters:
4162 fprintf(trace->output,
4163 "Failed to set filter \"%s\" on event %s with %d (%s)\n",
4164 evsel->filter, evsel__name(evsel), errno,
4165 str_error_r(errno, errbuf, sizeof(errbuf)));
4166 goto out_delete_evlist;
4167}
4168out_error_mem:
4169 fprintf(trace->output, "Not enough memory to run!\n");
4170 goto out_delete_evlist;
4171
4172out_errno:
4173 fprintf(trace->output, "errno=%d,%s\n", errno, strerror(errno));
4174 goto out_delete_evlist;
4175}
4176
4177static int trace__replay(struct trace *trace)
4178{
4179 const struct evsel_str_handler handlers[] = {
4180 { "probe:vfs_getname", trace__vfs_getname, },
4181 };
4182 struct perf_data data = {
4183 .path = input_name,
4184 .mode = PERF_DATA_MODE_READ,
4185 .force = trace->force,
4186 };
4187 struct perf_session *session;
4188 struct evsel *evsel;
4189 int err = -1;
4190
4191 trace->tool.sample = trace__process_sample;
4192 trace->tool.mmap = perf_event__process_mmap;
4193 trace->tool.mmap2 = perf_event__process_mmap2;
4194 trace->tool.comm = perf_event__process_comm;
4195 trace->tool.exit = perf_event__process_exit;
4196 trace->tool.fork = perf_event__process_fork;
4197 trace->tool.attr = perf_event__process_attr;
4198 trace->tool.tracing_data = perf_event__process_tracing_data;
4199 trace->tool.build_id = perf_event__process_build_id;
4200 trace->tool.namespaces = perf_event__process_namespaces;
4201
4202 trace->tool.ordered_events = true;
4203 trace->tool.ordering_requires_timestamps = true;
4204
4205
4206 trace->multiple_threads = true;
4207
4208 session = perf_session__new(&data, false, &trace->tool);
4209 if (IS_ERR(session))
4210 return PTR_ERR(session);
4211
4212 if (trace->opts.target.pid)
4213 symbol_conf.pid_list_str = strdup(trace->opts.target.pid);
4214
4215 if (trace->opts.target.tid)
4216 symbol_conf.tid_list_str = strdup(trace->opts.target.tid);
4217
4218 if (symbol__init(&session->header.env) < 0)
4219 goto out;
4220
4221 trace->host = &session->machines.host;
4222
4223 err = perf_session__set_tracepoints_handlers(session, handlers);
4224 if (err)
4225 goto out;
4226
4227 evsel = evlist__find_tracepoint_by_name(session->evlist, "raw_syscalls:sys_enter");
4228
4229 if (evsel == NULL)
4230 evsel = evlist__find_tracepoint_by_name(session->evlist, "syscalls:sys_enter");
4231
4232 if (evsel &&
4233 (evsel__init_raw_syscall_tp(evsel, trace__sys_enter) < 0 ||
4234 perf_evsel__init_sc_tp_ptr_field(evsel, args))) {
4235 pr_err("Error during initialize raw_syscalls:sys_enter event\n");
4236 goto out;
4237 }
4238
4239 evsel = evlist__find_tracepoint_by_name(session->evlist, "raw_syscalls:sys_exit");
4240 if (evsel == NULL)
4241 evsel = evlist__find_tracepoint_by_name(session->evlist, "syscalls:sys_exit");
4242 if (evsel &&
4243 (evsel__init_raw_syscall_tp(evsel, trace__sys_exit) < 0 ||
4244 perf_evsel__init_sc_tp_uint_field(evsel, ret))) {
4245 pr_err("Error during initialize raw_syscalls:sys_exit event\n");
4246 goto out;
4247 }
4248
4249 evlist__for_each_entry(session->evlist, evsel) {
4250 if (evsel->core.attr.type == PERF_TYPE_SOFTWARE &&
4251 (evsel->core.attr.config == PERF_COUNT_SW_PAGE_FAULTS_MAJ ||
4252 evsel->core.attr.config == PERF_COUNT_SW_PAGE_FAULTS_MIN ||
4253 evsel->core.attr.config == PERF_COUNT_SW_PAGE_FAULTS))
4254 evsel->handler = trace__pgfault;
4255 }
4256
4257 setup_pager();
4258
4259 err = perf_session__process_events(session);
4260 if (err)
4261 pr_err("Failed to process events, error %d", err);
4262
4263 else if (trace->summary)
4264 trace__fprintf_thread_summary(trace, trace->output);
4265
4266out:
4267 perf_session__delete(session);
4268
4269 return err;
4270}
4271
4272static size_t trace__fprintf_threads_header(FILE *fp)
4273{
4274 size_t printed;
4275
4276 printed = fprintf(fp, "\n Summary of events:\n\n");
4277
4278 return printed;
4279}
4280
4281DEFINE_RESORT_RB(syscall_stats, a->msecs > b->msecs,
4282 struct syscall_stats *stats;
4283 double msecs;
4284 int syscall;
4285)
4286{
4287 struct int_node *source = rb_entry(nd, struct int_node, rb_node);
4288 struct syscall_stats *stats = source->priv;
4289
4290 entry->syscall = source->i;
4291 entry->stats = stats;
4292 entry->msecs = stats ? (u64)stats->stats.n * (avg_stats(&stats->stats) / NSEC_PER_MSEC) : 0;
4293}
4294
4295static size_t thread__dump_stats(struct thread_trace *ttrace,
4296 struct trace *trace, FILE *fp)
4297{
4298 size_t printed = 0;
4299 struct syscall *sc;
4300 struct rb_node *nd;
4301 DECLARE_RESORT_RB_INTLIST(syscall_stats, ttrace->syscall_stats);
4302
4303 if (syscall_stats == NULL)
4304 return 0;
4305
4306 printed += fprintf(fp, "\n");
4307
4308 printed += fprintf(fp, " syscall calls errors total min avg max stddev\n");
4309 printed += fprintf(fp, " (msec) (msec) (msec) (msec) (%%)\n");
4310 printed += fprintf(fp, " --------------- -------- ------ -------- --------- --------- --------- ------\n");
4311
4312 resort_rb__for_each_entry(nd, syscall_stats) {
4313 struct syscall_stats *stats = syscall_stats_entry->stats;
4314 if (stats) {
4315 double min = (double)(stats->stats.min) / NSEC_PER_MSEC;
4316 double max = (double)(stats->stats.max) / NSEC_PER_MSEC;
4317 double avg = avg_stats(&stats->stats);
4318 double pct;
4319 u64 n = (u64)stats->stats.n;
4320
4321 pct = avg ? 100.0 * stddev_stats(&stats->stats) / avg : 0.0;
4322 avg /= NSEC_PER_MSEC;
4323
4324 sc = &trace->syscalls.table[syscall_stats_entry->syscall];
4325 printed += fprintf(fp, " %-15s", sc->name);
4326 printed += fprintf(fp, " %8" PRIu64 " %6" PRIu64 " %9.3f %9.3f %9.3f",
4327 n, stats->nr_failures, syscall_stats_entry->msecs, min, avg);
4328 printed += fprintf(fp, " %9.3f %9.2f%%\n", max, pct);
4329
4330 if (trace->errno_summary && stats->nr_failures) {
4331 const char *arch_name = perf_env__arch(trace->host->env);
4332 int e;
4333
4334 for (e = 0; e < stats->max_errno; ++e) {
4335 if (stats->errnos[e] != 0)
4336 fprintf(fp, "\t\t\t\t%s: %d\n", arch_syscalls__strerrno(arch_name, e + 1), stats->errnos[e]);
4337 }
4338 }
4339 }
4340 }
4341
4342 resort_rb__delete(syscall_stats);
4343 printed += fprintf(fp, "\n\n");
4344
4345 return printed;
4346}
4347
4348static size_t trace__fprintf_thread(FILE *fp, struct thread *thread, struct trace *trace)
4349{
4350 size_t printed = 0;
4351 struct thread_trace *ttrace = thread__priv(thread);
4352 double ratio;
4353
4354 if (ttrace == NULL)
4355 return 0;
4356
4357 ratio = (double)ttrace->nr_events / trace->nr_events * 100.0;
4358
4359 printed += fprintf(fp, " %s (%d), ", thread__comm_str(thread), thread->tid);
4360 printed += fprintf(fp, "%lu events, ", ttrace->nr_events);
4361 printed += fprintf(fp, "%.1f%%", ratio);
4362 if (ttrace->pfmaj)
4363 printed += fprintf(fp, ", %lu majfaults", ttrace->pfmaj);
4364 if (ttrace->pfmin)
4365 printed += fprintf(fp, ", %lu minfaults", ttrace->pfmin);
4366 if (trace->sched)
4367 printed += fprintf(fp, ", %.3f msec\n", ttrace->runtime_ms);
4368 else if (fputc('\n', fp) != EOF)
4369 ++printed;
4370
4371 printed += thread__dump_stats(ttrace, trace, fp);
4372
4373 return printed;
4374}
4375
4376static unsigned long thread__nr_events(struct thread_trace *ttrace)
4377{
4378 return ttrace ? ttrace->nr_events : 0;
4379}
4380
4381DEFINE_RESORT_RB(threads, (thread__nr_events(a->thread->priv) < thread__nr_events(b->thread->priv)),
4382 struct thread *thread;
4383)
4384{
4385 entry->thread = rb_entry(nd, struct thread, rb_node);
4386}
4387
4388static size_t trace__fprintf_thread_summary(struct trace *trace, FILE *fp)
4389{
4390 size_t printed = trace__fprintf_threads_header(fp);
4391 struct rb_node *nd;
4392 int i;
4393
4394 for (i = 0; i < THREADS__TABLE_SIZE; i++) {
4395 DECLARE_RESORT_RB_MACHINE_THREADS(threads, trace->host, i);
4396
4397 if (threads == NULL) {
4398 fprintf(fp, "%s", "Error sorting output by nr_events!\n");
4399 return 0;
4400 }
4401
4402 resort_rb__for_each_entry(nd, threads)
4403 printed += trace__fprintf_thread(fp, threads_entry->thread, trace);
4404
4405 resort_rb__delete(threads);
4406 }
4407 return printed;
4408}
4409
4410static int trace__set_duration(const struct option *opt, const char *str,
4411 int unset __maybe_unused)
4412{
4413 struct trace *trace = opt->value;
4414
4415 trace->duration_filter = atof(str);
4416 return 0;
4417}
4418
4419static int trace__set_filter_pids_from_option(const struct option *opt, const char *str,
4420 int unset __maybe_unused)
4421{
4422 int ret = -1;
4423 size_t i;
4424 struct trace *trace = opt->value;
4425
4426
4427
4428
4429 struct intlist *list = intlist__new(str);
4430
4431 if (list == NULL)
4432 return -1;
4433
4434 i = trace->filter_pids.nr = intlist__nr_entries(list) + 1;
4435 trace->filter_pids.entries = calloc(i, sizeof(pid_t));
4436
4437 if (trace->filter_pids.entries == NULL)
4438 goto out;
4439
4440 trace->filter_pids.entries[0] = getpid();
4441
4442 for (i = 1; i < trace->filter_pids.nr; ++i)
4443 trace->filter_pids.entries[i] = intlist__entry(list, i - 1)->i;
4444
4445 intlist__delete(list);
4446 ret = 0;
4447out:
4448 return ret;
4449}
4450
4451static int trace__open_output(struct trace *trace, const char *filename)
4452{
4453 struct stat st;
4454
4455 if (!stat(filename, &st) && st.st_size) {
4456 char oldname[PATH_MAX];
4457
4458 scnprintf(oldname, sizeof(oldname), "%s.old", filename);
4459 unlink(oldname);
4460 rename(filename, oldname);
4461 }
4462
4463 trace->output = fopen(filename, "w");
4464
4465 return trace->output == NULL ? -errno : 0;
4466}
4467
4468static int parse_pagefaults(const struct option *opt, const char *str,
4469 int unset __maybe_unused)
4470{
4471 int *trace_pgfaults = opt->value;
4472
4473 if (strcmp(str, "all") == 0)
4474 *trace_pgfaults |= TRACE_PFMAJ | TRACE_PFMIN;
4475 else if (strcmp(str, "maj") == 0)
4476 *trace_pgfaults |= TRACE_PFMAJ;
4477 else if (strcmp(str, "min") == 0)
4478 *trace_pgfaults |= TRACE_PFMIN;
4479 else
4480 return -1;
4481
4482 return 0;
4483}
4484
4485static void evlist__set_default_evsel_handler(struct evlist *evlist, void *handler)
4486{
4487 struct evsel *evsel;
4488
4489 evlist__for_each_entry(evlist, evsel) {
4490 if (evsel->handler == NULL)
4491 evsel->handler = handler;
4492 }
4493}
4494
4495static void evsel__set_syscall_arg_fmt(struct evsel *evsel, const char *name)
4496{
4497 struct syscall_arg_fmt *fmt = evsel__syscall_arg_fmt(evsel);
4498
4499 if (fmt) {
4500 struct syscall_fmt *scfmt = syscall_fmt__find(name);
4501
4502 if (scfmt) {
4503 int skip = 0;
4504
4505 if (strcmp(evsel->tp_format->format.fields->name, "__syscall_nr") == 0 ||
4506 strcmp(evsel->tp_format->format.fields->name, "nr") == 0)
4507 ++skip;
4508
4509 memcpy(fmt + skip, scfmt->arg, (evsel->tp_format->format.nr_fields - skip) * sizeof(*fmt));
4510 }
4511 }
4512}
4513
4514static int evlist__set_syscall_tp_fields(struct evlist *evlist)
4515{
4516 struct evsel *evsel;
4517
4518 evlist__for_each_entry(evlist, evsel) {
4519 if (evsel->priv || !evsel->tp_format)
4520 continue;
4521
4522 if (strcmp(evsel->tp_format->system, "syscalls")) {
4523 evsel__init_tp_arg_scnprintf(evsel);
4524 continue;
4525 }
4526
4527 if (evsel__init_syscall_tp(evsel))
4528 return -1;
4529
4530 if (!strncmp(evsel->tp_format->name, "sys_enter_", 10)) {
4531 struct syscall_tp *sc = __evsel__syscall_tp(evsel);
4532
4533 if (__tp_field__init_ptr(&sc->args, sc->id.offset + sizeof(u64)))
4534 return -1;
4535
4536 evsel__set_syscall_arg_fmt(evsel, evsel->tp_format->name + sizeof("sys_enter_") - 1);
4537 } else if (!strncmp(evsel->tp_format->name, "sys_exit_", 9)) {
4538 struct syscall_tp *sc = __evsel__syscall_tp(evsel);
4539
4540 if (__tp_field__init_uint(&sc->ret, sizeof(u64), sc->id.offset + sizeof(u64), evsel->needs_swap))
4541 return -1;
4542
4543 evsel__set_syscall_arg_fmt(evsel, evsel->tp_format->name + sizeof("sys_exit_") - 1);
4544 }
4545 }
4546
4547 return 0;
4548}
4549
4550
4551
4552
4553
4554
4555
4556
4557
4558static int trace__parse_events_option(const struct option *opt, const char *str,
4559 int unset __maybe_unused)
4560{
4561 struct trace *trace = (struct trace *)opt->value;
4562 const char *s = str;
4563 char *sep = NULL, *lists[2] = { NULL, NULL, };
4564 int len = strlen(str) + 1, err = -1, list, idx;
4565 char *strace_groups_dir = system_path(STRACE_GROUPS_DIR);
4566 char group_name[PATH_MAX];
4567 struct syscall_fmt *fmt;
4568
4569 if (strace_groups_dir == NULL)
4570 return -1;
4571
4572 if (*s == '!') {
4573 ++s;
4574 trace->not_ev_qualifier = true;
4575 }
4576
4577 while (1) {
4578 if ((sep = strchr(s, ',')) != NULL)
4579 *sep = '\0';
4580
4581 list = 0;
4582 if (syscalltbl__id(trace->sctbl, s) >= 0 ||
4583 syscalltbl__strglobmatch_first(trace->sctbl, s, &idx) >= 0) {
4584 list = 1;
4585 goto do_concat;
4586 }
4587
4588 fmt = syscall_fmt__find_by_alias(s);
4589 if (fmt != NULL) {
4590 list = 1;
4591 s = fmt->name;
4592 } else {
4593 path__join(group_name, sizeof(group_name), strace_groups_dir, s);
4594 if (access(group_name, R_OK) == 0)
4595 list = 1;
4596 }
4597do_concat:
4598 if (lists[list]) {
4599 sprintf(lists[list] + strlen(lists[list]), ",%s", s);
4600 } else {
4601 lists[list] = malloc(len);
4602 if (lists[list] == NULL)
4603 goto out;
4604 strcpy(lists[list], s);
4605 }
4606
4607 if (!sep)
4608 break;
4609
4610 *sep = ',';
4611 s = sep + 1;
4612 }
4613
4614 if (lists[1] != NULL) {
4615 struct strlist_config slist_config = {
4616 .dirname = strace_groups_dir,
4617 };
4618
4619 trace->ev_qualifier = strlist__new(lists[1], &slist_config);
4620 if (trace->ev_qualifier == NULL) {
4621 fputs("Not enough memory to parse event qualifier", trace->output);
4622 goto out;
4623 }
4624
4625 if (trace__validate_ev_qualifier(trace))
4626 goto out;
4627 trace->trace_syscalls = true;
4628 }
4629
4630 err = 0;
4631
4632 if (lists[0]) {
4633 struct option o = {
4634 .value = &trace->evlist,
4635 };
4636 err = parse_events_option(&o, lists[0], 0);
4637 }
4638out:
4639 if (sep)
4640 *sep = ',';
4641
4642 return err;
4643}
4644
4645static int trace__parse_cgroups(const struct option *opt, const char *str, int unset)
4646{
4647 struct trace *trace = opt->value;
4648
4649 if (!list_empty(&trace->evlist->core.entries)) {
4650 struct option o = {
4651 .value = &trace->evlist,
4652 };
4653 return parse_cgroups(&o, str, unset);
4654 }
4655 trace->cgroup = evlist__findnew_cgroup(trace->evlist, str);
4656
4657 return 0;
4658}
4659
4660static int trace__config(const char *var, const char *value, void *arg)
4661{
4662 struct trace *trace = arg;
4663 int err = 0;
4664
4665 if (!strcmp(var, "trace.add_events")) {
4666 trace->perfconfig_events = strdup(value);
4667 if (trace->perfconfig_events == NULL) {
4668 pr_err("Not enough memory for %s\n", "trace.add_events");
4669 return -1;
4670 }
4671 } else if (!strcmp(var, "trace.show_timestamp")) {
4672 trace->show_tstamp = perf_config_bool(var, value);
4673 } else if (!strcmp(var, "trace.show_duration")) {
4674 trace->show_duration = perf_config_bool(var, value);
4675 } else if (!strcmp(var, "trace.show_arg_names")) {
4676 trace->show_arg_names = perf_config_bool(var, value);
4677 if (!trace->show_arg_names)
4678 trace->show_zeros = true;
4679 } else if (!strcmp(var, "trace.show_zeros")) {
4680 bool new_show_zeros = perf_config_bool(var, value);
4681 if (!trace->show_arg_names && !new_show_zeros) {
4682 pr_warning("trace.show_zeros has to be set when trace.show_arg_names=no\n");
4683 goto out;
4684 }
4685 trace->show_zeros = new_show_zeros;
4686 } else if (!strcmp(var, "trace.show_prefix")) {
4687 trace->show_string_prefix = perf_config_bool(var, value);
4688 } else if (!strcmp(var, "trace.no_inherit")) {
4689 trace->opts.no_inherit = perf_config_bool(var, value);
4690 } else if (!strcmp(var, "trace.args_alignment")) {
4691 int args_alignment = 0;
4692 if (perf_config_int(&args_alignment, var, value) == 0)
4693 trace->args_alignment = args_alignment;
4694 } else if (!strcmp(var, "trace.tracepoint_beautifiers")) {
4695 if (strcasecmp(value, "libtraceevent") == 0)
4696 trace->libtraceevent_print = true;
4697 else if (strcasecmp(value, "libbeauty") == 0)
4698 trace->libtraceevent_print = false;
4699 }
4700out:
4701 return err;
4702}
4703
4704int cmd_trace(int argc, const char **argv)
4705{
4706 const char *trace_usage[] = {
4707 "perf trace [<options>] [<command>]",
4708 "perf trace [<options>] -- <command> [<options>]",
4709 "perf trace record [<options>] [<command>]",
4710 "perf trace record [<options>] -- <command> [<options>]",
4711 NULL
4712 };
4713 struct trace trace = {
4714 .opts = {
4715 .target = {
4716 .uid = UINT_MAX,
4717 .uses_mmap = true,
4718 },
4719 .user_freq = UINT_MAX,
4720 .user_interval = ULLONG_MAX,
4721 .no_buffering = true,
4722 .mmap_pages = UINT_MAX,
4723 },
4724 .output = stderr,
4725 .show_comm = true,
4726 .show_tstamp = true,
4727 .show_duration = true,
4728 .show_arg_names = true,
4729 .args_alignment = 70,
4730 .trace_syscalls = false,
4731 .kernel_syscallchains = false,
4732 .max_stack = UINT_MAX,
4733 .max_events = ULONG_MAX,
4734 };
4735 const char *map_dump_str = NULL;
4736 const char *output_name = NULL;
4737 const struct option trace_options[] = {
4738 OPT_CALLBACK('e', "event", &trace, "event",
4739 "event/syscall selector. use 'perf list' to list available events",
4740 trace__parse_events_option),
4741 OPT_CALLBACK(0, "filter", &trace.evlist, "filter",
4742 "event filter", parse_filter),
4743 OPT_BOOLEAN(0, "comm", &trace.show_comm,
4744 "show the thread COMM next to its id"),
4745 OPT_BOOLEAN(0, "tool_stats", &trace.show_tool_stats, "show tool stats"),
4746 OPT_CALLBACK(0, "expr", &trace, "expr", "list of syscalls/events to trace",
4747 trace__parse_events_option),
4748 OPT_STRING('o', "output", &output_name, "file", "output file name"),
4749 OPT_STRING('i', "input", &input_name, "file", "Analyze events in file"),
4750 OPT_STRING('p', "pid", &trace.opts.target.pid, "pid",
4751 "trace events on existing process id"),
4752 OPT_STRING('t', "tid", &trace.opts.target.tid, "tid",
4753 "trace events on existing thread id"),
4754 OPT_CALLBACK(0, "filter-pids", &trace, "CSV list of pids",
4755 "pids to filter (by the kernel)", trace__set_filter_pids_from_option),
4756 OPT_BOOLEAN('a', "all-cpus", &trace.opts.target.system_wide,
4757 "system-wide collection from all CPUs"),
4758 OPT_STRING('C', "cpu", &trace.opts.target.cpu_list, "cpu",
4759 "list of cpus to monitor"),
4760 OPT_BOOLEAN(0, "no-inherit", &trace.opts.no_inherit,
4761 "child tasks do not inherit counters"),
4762 OPT_CALLBACK('m', "mmap-pages", &trace.opts.mmap_pages, "pages",
4763 "number of mmap data pages", evlist__parse_mmap_pages),
4764 OPT_STRING('u', "uid", &trace.opts.target.uid_str, "user",
4765 "user to profile"),
4766 OPT_CALLBACK(0, "duration", &trace, "float",
4767 "show only events with duration > N.M ms",
4768 trace__set_duration),
4769#ifdef HAVE_LIBBPF_SUPPORT
4770 OPT_STRING(0, "map-dump", &map_dump_str, "BPF map", "BPF map to periodically dump"),
4771#endif
4772 OPT_BOOLEAN(0, "sched", &trace.sched, "show blocking scheduler events"),
4773 OPT_INCR('v', "verbose", &verbose, "be more verbose"),
4774 OPT_BOOLEAN('T', "time", &trace.full_time,
4775 "Show full timestamp, not time relative to first start"),
4776 OPT_BOOLEAN(0, "failure", &trace.failure_only,
4777 "Show only syscalls that failed"),
4778 OPT_BOOLEAN('s', "summary", &trace.summary_only,
4779 "Show only syscall summary with statistics"),
4780 OPT_BOOLEAN('S', "with-summary", &trace.summary,
4781 "Show all syscalls and summary with statistics"),
4782 OPT_BOOLEAN(0, "errno-summary", &trace.errno_summary,
4783 "Show errno stats per syscall, use with -s or -S"),
4784 OPT_CALLBACK_DEFAULT('F', "pf", &trace.trace_pgfaults, "all|maj|min",
4785 "Trace pagefaults", parse_pagefaults, "maj"),
4786 OPT_BOOLEAN(0, "syscalls", &trace.trace_syscalls, "Trace syscalls"),
4787 OPT_BOOLEAN('f', "force", &trace.force, "don't complain, do it"),
4788 OPT_CALLBACK(0, "call-graph", &trace.opts,
4789 "record_mode[,record_size]", record_callchain_help,
4790 &record_parse_callchain_opt),
4791 OPT_BOOLEAN(0, "libtraceevent_print", &trace.libtraceevent_print,
4792 "Use libtraceevent to print the tracepoint arguments."),
4793 OPT_BOOLEAN(0, "kernel-syscall-graph", &trace.kernel_syscallchains,
4794 "Show the kernel callchains on the syscall exit path"),
4795 OPT_ULONG(0, "max-events", &trace.max_events,
4796 "Set the maximum number of events to print, exit after that is reached. "),
4797 OPT_UINTEGER(0, "min-stack", &trace.min_stack,
4798 "Set the minimum stack depth when parsing the callchain, "
4799 "anything below the specified depth will be ignored."),
4800 OPT_UINTEGER(0, "max-stack", &trace.max_stack,
4801 "Set the maximum stack depth when parsing the callchain, "
4802 "anything beyond the specified depth will be ignored. "
4803 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)),
4804 OPT_BOOLEAN(0, "sort-events", &trace.sort_events,
4805 "Sort batch of events before processing, use if getting out of order events"),
4806 OPT_BOOLEAN(0, "print-sample", &trace.print_sample,
4807 "print the PERF_RECORD_SAMPLE PERF_SAMPLE_ info, for debugging"),
4808 OPT_UINTEGER(0, "proc-map-timeout", &proc_map_timeout,
4809 "per thread proc mmap processing timeout in ms"),
4810 OPT_CALLBACK('G', "cgroup", &trace, "name", "monitor event in cgroup name only",
4811 trace__parse_cgroups),
4812 OPT_INTEGER('D', "delay", &trace.opts.initial_delay,
4813 "ms to wait before starting measurement after program "
4814 "start"),
4815 OPTS_EVSWITCH(&trace.evswitch),
4816 OPT_END()
4817 };
4818 bool __maybe_unused max_stack_user_set = true;
4819 bool mmap_pages_user_set = true;
4820 struct evsel *evsel;
4821 const char * const trace_subcommands[] = { "record", NULL };
4822 int err = -1;
4823 char bf[BUFSIZ];
4824
4825 signal(SIGSEGV, sighandler_dump_stack);
4826 signal(SIGFPE, sighandler_dump_stack);
4827 signal(SIGCHLD, sig_handler);
4828 signal(SIGINT, sig_handler);
4829
4830 trace.evlist = evlist__new();
4831 trace.sctbl = syscalltbl__new();
4832
4833 if (trace.evlist == NULL || trace.sctbl == NULL) {
4834 pr_err("Not enough memory to run!\n");
4835 err = -ENOMEM;
4836 goto out;
4837 }
4838
4839
4840
4841
4842
4843
4844
4845
4846 rlimit__bump_memlock();
4847
4848 err = perf_config(trace__config, &trace);
4849 if (err)
4850 goto out;
4851
4852 argc = parse_options_subcommand(argc, argv, trace_options, trace_subcommands,
4853 trace_usage, PARSE_OPT_STOP_AT_NON_OPTION);
4854
4855
4856
4857
4858
4859
4860
4861
4862
4863
4864
4865
4866 if (!trace.trace_syscalls && !trace.trace_pgfaults &&
4867 trace.evlist->core.nr_entries == 0 ) {
4868 trace.trace_syscalls = true;
4869 }
4870
4871
4872
4873
4874
4875
4876 if (trace.perfconfig_events != NULL) {
4877 struct parse_events_error parse_err;
4878
4879 bzero(&parse_err, sizeof(parse_err));
4880 err = parse_events(trace.evlist, trace.perfconfig_events, &parse_err);
4881 if (err) {
4882 parse_events_print_error(&parse_err, trace.perfconfig_events);
4883 goto out;
4884 }
4885 }
4886
4887 if ((nr_cgroups || trace.cgroup) && !trace.opts.target.system_wide) {
4888 usage_with_options_msg(trace_usage, trace_options,
4889 "cgroup monitoring only available in system-wide mode");
4890 }
4891
4892 evsel = bpf__setup_output_event(trace.evlist, "__augmented_syscalls__");
4893 if (IS_ERR(evsel)) {
4894 bpf__strerror_setup_output_event(trace.evlist, PTR_ERR(evsel), bf, sizeof(bf));
4895 pr_err("ERROR: Setup trace syscalls enter failed: %s\n", bf);
4896 goto out;
4897 }
4898
4899 if (evsel) {
4900 trace.syscalls.events.augmented = evsel;
4901
4902 evsel = evlist__find_tracepoint_by_name(trace.evlist, "raw_syscalls:sys_enter");
4903 if (evsel == NULL) {
4904 pr_err("ERROR: raw_syscalls:sys_enter not found in the augmented BPF object\n");
4905 goto out;
4906 }
4907
4908 if (evsel->bpf_obj == NULL) {
4909 pr_err("ERROR: raw_syscalls:sys_enter not associated to a BPF object\n");
4910 goto out;
4911 }
4912
4913 trace.bpf_obj = evsel->bpf_obj;
4914
4915
4916
4917
4918
4919
4920 if (!trace.trace_syscalls && trace__only_augmented_syscalls_evsels(&trace))
4921 trace.trace_syscalls = true;
4922
4923
4924
4925
4926
4927
4928
4929
4930
4931
4932
4933
4934
4935
4936
4937
4938
4939
4940
4941
4942
4943
4944
4945
4946
4947 if (!trace.trace_syscalls) {
4948 trace__delete_augmented_syscalls(&trace);
4949 } else {
4950 trace__set_bpf_map_filtered_pids(&trace);
4951 trace__set_bpf_map_syscalls(&trace);
4952 trace.syscalls.unaugmented_prog = trace__find_bpf_program_by_title(&trace, "!raw_syscalls:unaugmented");
4953 }
4954 }
4955
4956 err = bpf__setup_stdout(trace.evlist);
4957 if (err) {
4958 bpf__strerror_setup_stdout(trace.evlist, err, bf, sizeof(bf));
4959 pr_err("ERROR: Setup BPF stdout failed: %s\n", bf);
4960 goto out;
4961 }
4962
4963 err = -1;
4964
4965 if (map_dump_str) {
4966 trace.dump.map = trace__find_bpf_map_by_name(&trace, map_dump_str);
4967 if (trace.dump.map == NULL) {
4968 pr_err("ERROR: BPF map \"%s\" not found\n", map_dump_str);
4969 goto out;
4970 }
4971 }
4972
4973 if (trace.trace_pgfaults) {
4974 trace.opts.sample_address = true;
4975 trace.opts.sample_time = true;
4976 }
4977
4978 if (trace.opts.mmap_pages == UINT_MAX)
4979 mmap_pages_user_set = false;
4980
4981 if (trace.max_stack == UINT_MAX) {
4982 trace.max_stack = input_name ? PERF_MAX_STACK_DEPTH : sysctl__max_stack();
4983 max_stack_user_set = false;
4984 }
4985
4986#ifdef HAVE_DWARF_UNWIND_SUPPORT
4987 if ((trace.min_stack || max_stack_user_set) && !callchain_param.enabled) {
4988 record_opts__parse_callchain(&trace.opts, &callchain_param, "dwarf", false);
4989 }
4990#endif
4991
4992 if (callchain_param.enabled) {
4993 if (!mmap_pages_user_set && geteuid() == 0)
4994 trace.opts.mmap_pages = perf_event_mlock_kb_in_pages() * 4;
4995
4996 symbol_conf.use_callchain = true;
4997 }
4998
4999 if (trace.evlist->core.nr_entries > 0) {
5000 evlist__set_default_evsel_handler(trace.evlist, trace__event_handler);
5001 if (evlist__set_syscall_tp_fields(trace.evlist)) {
5002 perror("failed to set syscalls:* tracepoint fields");
5003 goto out;
5004 }
5005 }
5006
5007 if (trace.sort_events) {
5008 ordered_events__init(&trace.oe.data, ordered_events__deliver_event, &trace);
5009 ordered_events__set_copy_on_queue(&trace.oe.data, true);
5010 }
5011
5012
5013
5014
5015
5016
5017
5018
5019
5020
5021
5022
5023 if (trace.syscalls.events.augmented) {
5024 evlist__for_each_entry(trace.evlist, evsel) {
5025 bool raw_syscalls_sys_exit = strcmp(evsel__name(evsel), "raw_syscalls:sys_exit") == 0;
5026
5027 if (raw_syscalls_sys_exit) {
5028 trace.raw_augmented_syscalls = true;
5029 goto init_augmented_syscall_tp;
5030 }
5031
5032 if (trace.syscalls.events.augmented->priv == NULL &&
5033 strstr(evsel__name(evsel), "syscalls:sys_enter")) {
5034 struct evsel *augmented = trace.syscalls.events.augmented;
5035 if (evsel__init_augmented_syscall_tp(augmented, evsel) ||
5036 evsel__init_augmented_syscall_tp_args(augmented))
5037 goto out;
5038
5039
5040
5041
5042
5043 augmented->handler = trace__sys_enter;
5044
5045
5046
5047
5048
5049
5050 if (evsel__init_augmented_syscall_tp(evsel, evsel) ||
5051 evsel__init_augmented_syscall_tp_args(evsel))
5052 goto out;
5053 evsel->handler = trace__sys_enter;
5054 }
5055
5056 if (strstarts(evsel__name(evsel), "syscalls:sys_exit_")) {
5057 struct syscall_tp *sc;
5058init_augmented_syscall_tp:
5059 if (evsel__init_augmented_syscall_tp(evsel, evsel))
5060 goto out;
5061 sc = __evsel__syscall_tp(evsel);
5062
5063
5064
5065
5066
5067
5068
5069
5070
5071
5072
5073
5074
5075
5076
5077
5078
5079
5080
5081 if (trace.raw_augmented_syscalls)
5082 trace.raw_augmented_syscalls_args_size = (6 + 1) * sizeof(long) + sc->id.offset;
5083 evsel__init_augmented_syscall_tp_ret(evsel);
5084 evsel->handler = trace__sys_exit;
5085 }
5086 }
5087 }
5088
5089 if ((argc >= 1) && (strcmp(argv[0], "record") == 0))
5090 return trace__record(&trace, argc-1, &argv[1]);
5091
5092
5093 if (trace.errno_summary && !trace.summary && !trace.summary_only)
5094 trace.summary_only = true;
5095
5096
5097 if (trace.summary_only)
5098 trace.summary = trace.summary_only;
5099
5100 if (output_name != NULL) {
5101 err = trace__open_output(&trace, output_name);
5102 if (err < 0) {
5103 perror("failed to create output file");
5104 goto out;
5105 }
5106 }
5107
5108 err = evswitch__init(&trace.evswitch, trace.evlist, stderr);
5109 if (err)
5110 goto out_close;
5111
5112 err = target__validate(&trace.opts.target);
5113 if (err) {
5114 target__strerror(&trace.opts.target, err, bf, sizeof(bf));
5115 fprintf(trace.output, "%s", bf);
5116 goto out_close;
5117 }
5118
5119 err = target__parse_uid(&trace.opts.target);
5120 if (err) {
5121 target__strerror(&trace.opts.target, err, bf, sizeof(bf));
5122 fprintf(trace.output, "%s", bf);
5123 goto out_close;
5124 }
5125
5126 if (!argc && target__none(&trace.opts.target))
5127 trace.opts.target.system_wide = true;
5128
5129 if (input_name)
5130 err = trace__replay(&trace);
5131 else
5132 err = trace__run(&trace, argc, argv);
5133
5134out_close:
5135 if (output_name != NULL)
5136 fclose(trace.output);
5137out:
5138 zfree(&trace.perfconfig_events);
5139 return err;
5140}
5141