perf trace: Exclude the kernel part of the callchain leading to a syscall
[linux-2.6-block.git] / tools / perf / builtin-record.c
CommitLineData
abaff32a 1/*
bf9e1876
IM
2 * builtin-record.c
3 *
4 * Builtin record command: Record the profile of a workload
5 * (or a CPU, or a PID) into the perf.data output file - for
6 * later analysis via perf report.
abaff32a 7 */
16f762a2 8#include "builtin.h"
bf9e1876
IM
9
10#include "perf.h"
11
6122e4e4 12#include "util/build-id.h"
6eda5838 13#include "util/util.h"
4b6ab94e 14#include <subcmd/parse-options.h>
8ad8db37 15#include "util/parse-events.h"
6eda5838 16
8f651eae 17#include "util/callchain.h"
f14d5707 18#include "util/cgroup.h"
7c6a1c65 19#include "util/header.h"
66e274f3 20#include "util/event.h"
361c99a6 21#include "util/evlist.h"
69aad6f1 22#include "util/evsel.h"
8f28827a 23#include "util/debug.h"
94c744b6 24#include "util/session.h"
45694aa7 25#include "util/tool.h"
8d06367f 26#include "util/symbol.h"
a12b51c4 27#include "util/cpumap.h"
fd78260b 28#include "util/thread_map.h"
f5fc1412 29#include "util/data.h"
bcc84ec6 30#include "util/perf_regs.h"
ef149c25 31#include "util/auxtrace.h"
46bc29b9 32#include "util/tsc.h"
f00898f4 33#include "util/parse-branch-options.h"
bcc84ec6 34#include "util/parse-regs-options.h"
71dc2326 35#include "util/llvm-utils.h"
8690a2a7 36#include "util/bpf-loader.h"
d8871ea7 37#include "asm/bug.h"
7c6a1c65 38
97124d5e 39#include <unistd.h>
de9ac07b 40#include <sched.h>
a41794cd 41#include <sys/mman.h>
de9ac07b 42
78da39fa 43
8c6f45a7 44struct record {
45694aa7 45 struct perf_tool tool;
b4006796 46 struct record_opts opts;
d20deb64 47 u64 bytes_written;
f5fc1412 48 struct perf_data_file file;
ef149c25 49 struct auxtrace_record *itr;
d20deb64
ACM
50 struct perf_evlist *evlist;
51 struct perf_session *session;
52 const char *progname;
d20deb64 53 int realtime_prio;
d20deb64 54 bool no_buildid;
d2db9a98 55 bool no_buildid_set;
d20deb64 56 bool no_buildid_cache;
d2db9a98 57 bool no_buildid_cache_set;
6156681b 58 bool buildid_all;
9f065194 59 unsigned long long samples;
0f82ebc4 60};
a21ca2ca 61
8c6f45a7 62static int record__write(struct record *rec, void *bf, size_t size)
f5970550 63{
cf8b2e69 64 if (perf_data_file__write(rec->session->file, bf, size) < 0) {
50a9b868
JO
65 pr_err("failed to write perf data, error: %m\n");
66 return -1;
f5970550 67 }
8d3eca20 68
cf8b2e69 69 rec->bytes_written += size;
8d3eca20 70 return 0;
f5970550
PZ
71}
72
45694aa7 73static int process_synthesized_event(struct perf_tool *tool,
d20deb64 74 union perf_event *event,
1d037ca1
IT
75 struct perf_sample *sample __maybe_unused,
76 struct machine *machine __maybe_unused)
234fbbf5 77{
8c6f45a7
ACM
78 struct record *rec = container_of(tool, struct record, tool);
79 return record__write(rec, event, event->header.size);
234fbbf5
ACM
80}
81
e5685730 82static int record__mmap_read(struct record *rec, int idx)
de9ac07b 83{
e5685730 84 struct perf_mmap *md = &rec->evlist->mmap[idx];
7b8283b5
DA
85 u64 head = perf_mmap__read_head(md);
86 u64 old = md->prev;
918512b4 87 unsigned char *data = md->base + page_size;
de9ac07b
PZ
88 unsigned long size;
89 void *buf;
8d3eca20 90 int rc = 0;
de9ac07b 91
dc82009a 92 if (old == head)
8d3eca20 93 return 0;
dc82009a 94
d20deb64 95 rec->samples++;
de9ac07b
PZ
96
97 size = head - old;
98
99 if ((old & md->mask) + size != (head & md->mask)) {
100 buf = &data[old & md->mask];
101 size = md->mask + 1 - (old & md->mask);
102 old += size;
021e9f47 103
8c6f45a7 104 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
105 rc = -1;
106 goto out;
107 }
de9ac07b
PZ
108 }
109
110 buf = &data[old & md->mask];
111 size = head - old;
112 old += size;
021e9f47 113
8c6f45a7 114 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
115 rc = -1;
116 goto out;
117 }
de9ac07b
PZ
118
119 md->prev = old;
e5685730 120 perf_evlist__mmap_consume(rec->evlist, idx);
8d3eca20
DA
121out:
122 return rc;
de9ac07b
PZ
123}
124
2dd6d8a1
AH
125static volatile int done;
126static volatile int signr = -1;
127static volatile int child_finished;
128static volatile int auxtrace_snapshot_enabled;
129static volatile int auxtrace_snapshot_err;
130static volatile int auxtrace_record__snapshot_started;
131
132static void sig_handler(int sig)
133{
134 if (sig == SIGCHLD)
135 child_finished = 1;
136 else
137 signr = sig;
138
139 done = 1;
140}
141
142static void record__sig_exit(void)
143{
144 if (signr == -1)
145 return;
146
147 signal(signr, SIG_DFL);
148 raise(signr);
149}
150
e31f0d01
AH
151#ifdef HAVE_AUXTRACE_SUPPORT
152
ef149c25
AH
153static int record__process_auxtrace(struct perf_tool *tool,
154 union perf_event *event, void *data1,
155 size_t len1, void *data2, size_t len2)
156{
157 struct record *rec = container_of(tool, struct record, tool);
99fa2984 158 struct perf_data_file *file = &rec->file;
ef149c25
AH
159 size_t padding;
160 u8 pad[8] = {0};
161
99fa2984
AH
162 if (!perf_data_file__is_pipe(file)) {
163 off_t file_offset;
164 int fd = perf_data_file__fd(file);
165 int err;
166
167 file_offset = lseek(fd, 0, SEEK_CUR);
168 if (file_offset == -1)
169 return -1;
170 err = auxtrace_index__auxtrace_event(&rec->session->auxtrace_index,
171 event, file_offset);
172 if (err)
173 return err;
174 }
175
ef149c25
AH
176 /* event.auxtrace.size includes padding, see __auxtrace_mmap__read() */
177 padding = (len1 + len2) & 7;
178 if (padding)
179 padding = 8 - padding;
180
181 record__write(rec, event, event->header.size);
182 record__write(rec, data1, len1);
183 if (len2)
184 record__write(rec, data2, len2);
185 record__write(rec, &pad, padding);
186
187 return 0;
188}
189
190static int record__auxtrace_mmap_read(struct record *rec,
191 struct auxtrace_mmap *mm)
192{
193 int ret;
194
195 ret = auxtrace_mmap__read(mm, rec->itr, &rec->tool,
196 record__process_auxtrace);
197 if (ret < 0)
198 return ret;
199
200 if (ret)
201 rec->samples++;
202
203 return 0;
204}
205
2dd6d8a1
AH
206static int record__auxtrace_mmap_read_snapshot(struct record *rec,
207 struct auxtrace_mmap *mm)
208{
209 int ret;
210
211 ret = auxtrace_mmap__read_snapshot(mm, rec->itr, &rec->tool,
212 record__process_auxtrace,
213 rec->opts.auxtrace_snapshot_size);
214 if (ret < 0)
215 return ret;
216
217 if (ret)
218 rec->samples++;
219
220 return 0;
221}
222
223static int record__auxtrace_read_snapshot_all(struct record *rec)
224{
225 int i;
226 int rc = 0;
227
228 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
229 struct auxtrace_mmap *mm =
230 &rec->evlist->mmap[i].auxtrace_mmap;
231
232 if (!mm->base)
233 continue;
234
235 if (record__auxtrace_mmap_read_snapshot(rec, mm) != 0) {
236 rc = -1;
237 goto out;
238 }
239 }
240out:
241 return rc;
242}
243
244static void record__read_auxtrace_snapshot(struct record *rec)
245{
246 pr_debug("Recording AUX area tracing snapshot\n");
247 if (record__auxtrace_read_snapshot_all(rec) < 0) {
248 auxtrace_snapshot_err = -1;
249 } else {
250 auxtrace_snapshot_err = auxtrace_record__snapshot_finish(rec->itr);
251 if (!auxtrace_snapshot_err)
252 auxtrace_snapshot_enabled = 1;
253 }
254}
255
e31f0d01
AH
256#else
257
258static inline
259int record__auxtrace_mmap_read(struct record *rec __maybe_unused,
260 struct auxtrace_mmap *mm __maybe_unused)
261{
262 return 0;
263}
264
2dd6d8a1
AH
265static inline
266void record__read_auxtrace_snapshot(struct record *rec __maybe_unused)
de9ac07b 267{
f7b7c26e
PZ
268}
269
2dd6d8a1
AH
270static inline
271int auxtrace_record__snapshot_start(struct auxtrace_record *itr __maybe_unused)
f7b7c26e 272{
2dd6d8a1 273 return 0;
de9ac07b
PZ
274}
275
2dd6d8a1
AH
276#endif
277
8c6f45a7 278static int record__open(struct record *rec)
dd7927f4 279{
56e52e85 280 char msg[512];
6a4bb04c 281 struct perf_evsel *pos;
d20deb64
ACM
282 struct perf_evlist *evlist = rec->evlist;
283 struct perf_session *session = rec->session;
b4006796 284 struct record_opts *opts = &rec->opts;
8d3eca20 285 int rc = 0;
dd7927f4 286
f77a9518 287 perf_evlist__config(evlist, opts);
cac21425 288
0050f7aa 289 evlist__for_each(evlist, pos) {
dd7927f4 290try_again:
d988d5ee 291 if (perf_evsel__open(pos, pos->cpus, pos->threads) < 0) {
56e52e85 292 if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
d6d901c2 293 if (verbose)
c0a54341 294 ui__warning("%s\n", msg);
d6d901c2
ZY
295 goto try_again;
296 }
ca6a4258 297
56e52e85
ACM
298 rc = -errno;
299 perf_evsel__open_strerror(pos, &opts->target,
300 errno, msg, sizeof(msg));
301 ui__error("%s\n", msg);
8d3eca20 302 goto out;
c171b552
LZ
303 }
304 }
a43d3f08 305
23d4aad4
ACM
306 if (perf_evlist__apply_filters(evlist, &pos)) {
307 error("failed to set filter \"%s\" on event %s with %d (%s)\n",
308 pos->filter, perf_evsel__name(pos), errno,
35550da3 309 strerror_r(errno, msg, sizeof(msg)));
8d3eca20
DA
310 rc = -1;
311 goto out;
0a102479
FW
312 }
313
ef149c25 314 if (perf_evlist__mmap_ex(evlist, opts->mmap_pages, false,
2dd6d8a1
AH
315 opts->auxtrace_mmap_pages,
316 opts->auxtrace_snapshot_mode) < 0) {
8d3eca20
DA
317 if (errno == EPERM) {
318 pr_err("Permission error mapping pages.\n"
319 "Consider increasing "
320 "/proc/sys/kernel/perf_event_mlock_kb,\n"
321 "or try again with a smaller value of -m/--mmap_pages.\n"
ef149c25
AH
322 "(current value: %u,%u)\n",
323 opts->mmap_pages, opts->auxtrace_mmap_pages);
8d3eca20 324 rc = -errno;
8d3eca20 325 } else {
35550da3
MH
326 pr_err("failed to mmap with %d (%s)\n", errno,
327 strerror_r(errno, msg, sizeof(msg)));
95c36561
WN
328 if (errno)
329 rc = -errno;
330 else
331 rc = -EINVAL;
8d3eca20
DA
332 }
333 goto out;
18e60939 334 }
0a27d7f9 335
563aecb2 336 session->evlist = evlist;
7b56cce2 337 perf_session__set_id_hdr_size(session);
8d3eca20
DA
338out:
339 return rc;
16c8a109
PZ
340}
341
e3d59112
NK
342static int process_sample_event(struct perf_tool *tool,
343 union perf_event *event,
344 struct perf_sample *sample,
345 struct perf_evsel *evsel,
346 struct machine *machine)
347{
348 struct record *rec = container_of(tool, struct record, tool);
349
350 rec->samples++;
351
352 return build_id__mark_dso_hit(tool, event, sample, evsel, machine);
353}
354
8c6f45a7 355static int process_buildids(struct record *rec)
6122e4e4 356{
f5fc1412
JO
357 struct perf_data_file *file = &rec->file;
358 struct perf_session *session = rec->session;
6122e4e4 359
457ae94a 360 if (file->size == 0)
9f591fd7
ACM
361 return 0;
362
00dc8657
NK
363 /*
364 * During this process, it'll load kernel map and replace the
365 * dso->long_name to a real pathname it found. In this case
366 * we prefer the vmlinux path like
367 * /lib/modules/3.16.4/build/vmlinux
368 *
369 * rather than build-id path (in debug directory).
370 * $HOME/.debug/.build-id/f0/6e17aa50adf4d00b88925e03775de107611551
371 */
372 symbol_conf.ignore_vmlinux_buildid = true;
373
6156681b
NK
374 /*
375 * If --buildid-all is given, it marks all DSO regardless of hits,
376 * so no need to process samples.
377 */
378 if (rec->buildid_all)
379 rec->tool.sample = NULL;
380
b7b61cbe 381 return perf_session__process_events(session);
6122e4e4
ACM
382}
383
8115d60c 384static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
a1645ce1
ZY
385{
386 int err;
45694aa7 387 struct perf_tool *tool = data;
a1645ce1
ZY
388 /*
389 *As for guest kernel when processing subcommand record&report,
390 *we arrange module mmap prior to guest kernel mmap and trigger
391 *a preload dso because default guest module symbols are loaded
392 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
393 *method is used to avoid symbol missing when the first addr is
394 *in module instead of in guest kernel.
395 */
45694aa7 396 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 397 machine);
a1645ce1
ZY
398 if (err < 0)
399 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 400 " relocation symbol.\n", machine->pid);
a1645ce1 401
a1645ce1
ZY
402 /*
403 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
404 * have no _text sometimes.
405 */
45694aa7 406 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
0ae617be 407 machine);
a1645ce1
ZY
408 if (err < 0)
409 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 410 " relocation symbol.\n", machine->pid);
a1645ce1
ZY
411}
412
98402807
FW
413static struct perf_event_header finished_round_event = {
414 .size = sizeof(struct perf_event_header),
415 .type = PERF_RECORD_FINISHED_ROUND,
416};
417
8c6f45a7 418static int record__mmap_read_all(struct record *rec)
98402807 419{
dcabb507 420 u64 bytes_written = rec->bytes_written;
0e2e63dd 421 int i;
8d3eca20 422 int rc = 0;
98402807 423
d20deb64 424 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
ef149c25
AH
425 struct auxtrace_mmap *mm = &rec->evlist->mmap[i].auxtrace_mmap;
426
8d3eca20 427 if (rec->evlist->mmap[i].base) {
e5685730 428 if (record__mmap_read(rec, i) != 0) {
8d3eca20
DA
429 rc = -1;
430 goto out;
431 }
432 }
ef149c25 433
2dd6d8a1 434 if (mm->base && !rec->opts.auxtrace_snapshot_mode &&
ef149c25
AH
435 record__auxtrace_mmap_read(rec, mm) != 0) {
436 rc = -1;
437 goto out;
438 }
98402807
FW
439 }
440
dcabb507
JO
441 /*
442 * Mark the round finished in case we wrote
443 * at least one event.
444 */
445 if (bytes_written != rec->bytes_written)
446 rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
8d3eca20
DA
447
448out:
449 return rc;
98402807
FW
450}
451
8c6f45a7 452static void record__init_features(struct record *rec)
57706abc 453{
57706abc
DA
454 struct perf_session *session = rec->session;
455 int feat;
456
457 for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
458 perf_header__set_feat(&session->header, feat);
459
460 if (rec->no_buildid)
461 perf_header__clear_feat(&session->header, HEADER_BUILD_ID);
462
3e2be2da 463 if (!have_tracepoints(&rec->evlist->entries))
57706abc
DA
464 perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
465
466 if (!rec->opts.branch_stack)
467 perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
ef149c25
AH
468
469 if (!rec->opts.full_auxtrace)
470 perf_header__clear_feat(&session->header, HEADER_AUXTRACE);
ffa517ad
JO
471
472 perf_header__clear_feat(&session->header, HEADER_STAT);
57706abc
DA
473}
474
e1ab48ba
WN
475static void
476record__finish_output(struct record *rec)
477{
478 struct perf_data_file *file = &rec->file;
479 int fd = perf_data_file__fd(file);
480
481 if (file->is_pipe)
482 return;
483
484 rec->session->header.data_size += rec->bytes_written;
485 file->size = lseek(perf_data_file__fd(file), 0, SEEK_CUR);
486
487 if (!rec->no_buildid) {
488 process_buildids(rec);
489
490 if (rec->buildid_all)
491 dsos__hit_all(rec->session);
492 }
493 perf_session__write_header(rec->session, rec->evlist, fd, true);
494
495 return;
496}
497
f33cbe72
ACM
498static volatile int workload_exec_errno;
499
500/*
501 * perf_evlist__prepare_workload will send a SIGUSR1
502 * if the fork fails, since we asked by setting its
503 * want_signal to true.
504 */
45604710
NK
505static void workload_exec_failed_signal(int signo __maybe_unused,
506 siginfo_t *info,
f33cbe72
ACM
507 void *ucontext __maybe_unused)
508{
509 workload_exec_errno = info->si_value.sival_int;
510 done = 1;
f33cbe72
ACM
511 child_finished = 1;
512}
513
2dd6d8a1
AH
514static void snapshot_sig_handler(int sig);
515
46bc29b9
AH
516int __weak
517perf_event__synth_time_conv(const struct perf_event_mmap_page *pc __maybe_unused,
518 struct perf_tool *tool __maybe_unused,
519 perf_event__handler_t process __maybe_unused,
520 struct machine *machine __maybe_unused)
521{
522 return 0;
523}
524
c45c86eb
WN
525static int record__synthesize(struct record *rec)
526{
527 struct perf_session *session = rec->session;
528 struct machine *machine = &session->machines.host;
529 struct perf_data_file *file = &rec->file;
530 struct record_opts *opts = &rec->opts;
531 struct perf_tool *tool = &rec->tool;
532 int fd = perf_data_file__fd(file);
533 int err = 0;
534
535 if (file->is_pipe) {
536 err = perf_event__synthesize_attrs(tool, session,
537 process_synthesized_event);
538 if (err < 0) {
539 pr_err("Couldn't synthesize attrs.\n");
540 goto out;
541 }
542
543 if (have_tracepoints(&rec->evlist->entries)) {
544 /*
545 * FIXME err <= 0 here actually means that
546 * there were no tracepoints so its not really
547 * an error, just that we don't need to
548 * synthesize anything. We really have to
549 * return this more properly and also
550 * propagate errors that now are calling die()
551 */
552 err = perf_event__synthesize_tracing_data(tool, fd, rec->evlist,
553 process_synthesized_event);
554 if (err <= 0) {
555 pr_err("Couldn't record tracing data.\n");
556 goto out;
557 }
558 rec->bytes_written += err;
559 }
560 }
561
46bc29b9
AH
562 err = perf_event__synth_time_conv(rec->evlist->mmap[0].base, tool,
563 process_synthesized_event, machine);
564 if (err)
565 goto out;
566
c45c86eb
WN
567 if (rec->opts.full_auxtrace) {
568 err = perf_event__synthesize_auxtrace_info(rec->itr, tool,
569 session, process_synthesized_event);
570 if (err)
571 goto out;
572 }
573
574 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
575 machine);
576 WARN_ONCE(err < 0, "Couldn't record kernel reference relocation symbol\n"
577 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
578 "Check /proc/kallsyms permission or run as root.\n");
579
580 err = perf_event__synthesize_modules(tool, process_synthesized_event,
581 machine);
582 WARN_ONCE(err < 0, "Couldn't record kernel module information.\n"
583 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
584 "Check /proc/modules permission or run as root.\n");
585
586 if (perf_guest) {
587 machines__process_guests(&session->machines,
588 perf_event__synthesize_guest_os, tool);
589 }
590
591 err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
592 process_synthesized_event, opts->sample_address,
593 opts->proc_map_timeout);
594out:
595 return err;
596}
597
8c6f45a7 598static int __cmd_record(struct record *rec, int argc, const char **argv)
16c8a109 599{
57706abc 600 int err;
45604710 601 int status = 0;
8b412664 602 unsigned long waking = 0;
46be604b 603 const bool forks = argc > 0;
23346f21 604 struct machine *machine;
45694aa7 605 struct perf_tool *tool = &rec->tool;
b4006796 606 struct record_opts *opts = &rec->opts;
f5fc1412 607 struct perf_data_file *file = &rec->file;
d20deb64 608 struct perf_session *session;
6dcf45ef 609 bool disabled = false, draining = false;
42aa276f 610 int fd;
de9ac07b 611
d20deb64 612 rec->progname = argv[0];
33e49ea7 613
45604710 614 atexit(record__sig_exit);
f5970550
PZ
615 signal(SIGCHLD, sig_handler);
616 signal(SIGINT, sig_handler);
804f7ac7 617 signal(SIGTERM, sig_handler);
2dd6d8a1
AH
618 if (rec->opts.auxtrace_snapshot_mode)
619 signal(SIGUSR2, snapshot_sig_handler);
620 else
621 signal(SIGUSR2, SIG_IGN);
f5970550 622
b7b61cbe 623 session = perf_session__new(file, false, tool);
94c744b6 624 if (session == NULL) {
ffa91880 625 pr_err("Perf session creation failed.\n");
a9a70bbc
ACM
626 return -1;
627 }
628
42aa276f 629 fd = perf_data_file__fd(file);
d20deb64
ACM
630 rec->session = session;
631
8c6f45a7 632 record__init_features(rec);
330aa675 633
d4db3f16 634 if (forks) {
3e2be2da 635 err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
f5fc1412 636 argv, file->is_pipe,
735f7e0b 637 workload_exec_failed_signal);
35b9d88e
ACM
638 if (err < 0) {
639 pr_err("Couldn't run the workload!\n");
45604710 640 status = err;
35b9d88e 641 goto out_delete_session;
856e9660 642 }
856e9660
PZ
643 }
644
8c6f45a7 645 if (record__open(rec) != 0) {
8d3eca20 646 err = -1;
45604710 647 goto out_child;
8d3eca20 648 }
de9ac07b 649
8690a2a7
WN
650 err = bpf__apply_obj_config();
651 if (err) {
652 char errbuf[BUFSIZ];
653
654 bpf__strerror_apply_obj_config(err, errbuf, sizeof(errbuf));
655 pr_err("ERROR: Apply config to BPF failed: %s\n",
656 errbuf);
657 goto out_child;
658 }
659
cca8482c
AH
660 /*
661 * Normally perf_session__new would do this, but it doesn't have the
662 * evlist.
663 */
664 if (rec->tool.ordered_events && !perf_evlist__sample_id_all(rec->evlist)) {
665 pr_warning("WARNING: No sample_id_all support, falling back to unordered processing\n");
666 rec->tool.ordered_events = false;
667 }
668
3e2be2da 669 if (!rec->evlist->nr_groups)
a8bb559b
NK
670 perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);
671
f5fc1412 672 if (file->is_pipe) {
42aa276f 673 err = perf_header__write_pipe(fd);
529870e3 674 if (err < 0)
45604710 675 goto out_child;
563aecb2 676 } else {
42aa276f 677 err = perf_session__write_header(session, rec->evlist, fd, false);
d5eed904 678 if (err < 0)
45604710 679 goto out_child;
56b03f3c
ACM
680 }
681
d3665498 682 if (!rec->no_buildid
e20960c0 683 && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
d3665498 684 pr_err("Couldn't generate buildids. "
e20960c0 685 "Use --no-buildid to profile anyway.\n");
8d3eca20 686 err = -1;
45604710 687 goto out_child;
e20960c0
RR
688 }
689
34ba5122 690 machine = &session->machines.host;
743eb868 691
c45c86eb
WN
692 err = record__synthesize(rec);
693 if (err < 0)
45604710 694 goto out_child;
8d3eca20 695
d20deb64 696 if (rec->realtime_prio) {
de9ac07b
PZ
697 struct sched_param param;
698
d20deb64 699 param.sched_priority = rec->realtime_prio;
de9ac07b 700 if (sched_setscheduler(0, SCHED_FIFO, &param)) {
6beba7ad 701 pr_err("Could not set realtime priority.\n");
8d3eca20 702 err = -1;
45604710 703 goto out_child;
de9ac07b
PZ
704 }
705 }
706
774cb499
JO
707 /*
708 * When perf is starting the traced process, all the events
709 * (apart from group members) have enable_on_exec=1 set,
710 * so don't spoil it by prematurely enabling them.
711 */
6619a53e 712 if (!target__none(&opts->target) && !opts->initial_delay)
3e2be2da 713 perf_evlist__enable(rec->evlist);
764e16a3 714
856e9660
PZ
715 /*
716 * Let the child rip
717 */
e803cf97 718 if (forks) {
e5bed564
NK
719 union perf_event *event;
720
721 event = malloc(sizeof(event->comm) + machine->id_hdr_size);
722 if (event == NULL) {
723 err = -ENOMEM;
724 goto out_child;
725 }
726
e803cf97
NK
727 /*
728 * Some H/W events are generated before COMM event
729 * which is emitted during exec(), so perf script
730 * cannot see a correct process name for those events.
731 * Synthesize COMM event to prevent it.
732 */
e5bed564 733 perf_event__synthesize_comm(tool, event,
e803cf97
NK
734 rec->evlist->workload.pid,
735 process_synthesized_event,
736 machine);
e5bed564 737 free(event);
e803cf97 738
3e2be2da 739 perf_evlist__start_workload(rec->evlist);
e803cf97 740 }
856e9660 741
6619a53e
AK
742 if (opts->initial_delay) {
743 usleep(opts->initial_delay * 1000);
744 perf_evlist__enable(rec->evlist);
745 }
746
2dd6d8a1 747 auxtrace_snapshot_enabled = 1;
649c48a9 748 for (;;) {
9f065194 749 unsigned long long hits = rec->samples;
de9ac07b 750
8c6f45a7 751 if (record__mmap_read_all(rec) < 0) {
2dd6d8a1 752 auxtrace_snapshot_enabled = 0;
8d3eca20 753 err = -1;
45604710 754 goto out_child;
8d3eca20 755 }
de9ac07b 756
2dd6d8a1
AH
757 if (auxtrace_record__snapshot_started) {
758 auxtrace_record__snapshot_started = 0;
759 if (!auxtrace_snapshot_err)
760 record__read_auxtrace_snapshot(rec);
761 if (auxtrace_snapshot_err) {
762 pr_err("AUX area tracing snapshot failed\n");
763 err = -1;
764 goto out_child;
765 }
766 }
767
d20deb64 768 if (hits == rec->samples) {
6dcf45ef 769 if (done || draining)
649c48a9 770 break;
f66a889d 771 err = perf_evlist__poll(rec->evlist, -1);
a515114f
JO
772 /*
773 * Propagate error, only if there's any. Ignore positive
774 * number of returned events and interrupt error.
775 */
776 if (err > 0 || (err < 0 && errno == EINTR))
45604710 777 err = 0;
8b412664 778 waking++;
6dcf45ef
ACM
779
780 if (perf_evlist__filter_pollfd(rec->evlist, POLLERR | POLLHUP) == 0)
781 draining = true;
8b412664
PZ
782 }
783
774cb499
JO
784 /*
785 * When perf is starting the traced process, at the end events
786 * die with the process and we wait for that. Thus no need to
787 * disable events in this case.
788 */
602ad878 789 if (done && !disabled && !target__none(&opts->target)) {
2dd6d8a1 790 auxtrace_snapshot_enabled = 0;
3e2be2da 791 perf_evlist__disable(rec->evlist);
2711926a
JO
792 disabled = true;
793 }
de9ac07b 794 }
2dd6d8a1 795 auxtrace_snapshot_enabled = 0;
de9ac07b 796
f33cbe72 797 if (forks && workload_exec_errno) {
35550da3 798 char msg[STRERR_BUFSIZE];
f33cbe72
ACM
799 const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
800 pr_err("Workload failed: %s\n", emsg);
801 err = -1;
45604710 802 goto out_child;
f33cbe72
ACM
803 }
804
e3d59112 805 if (!quiet)
45604710 806 fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);
b44308f5 807
45604710
NK
808out_child:
809 if (forks) {
810 int exit_status;
addc2785 811
45604710
NK
812 if (!child_finished)
813 kill(rec->evlist->workload.pid, SIGTERM);
814
815 wait(&exit_status);
816
817 if (err < 0)
818 status = err;
819 else if (WIFEXITED(exit_status))
820 status = WEXITSTATUS(exit_status);
821 else if (WIFSIGNALED(exit_status))
822 signr = WTERMSIG(exit_status);
823 } else
824 status = err;
825
e3d59112
NK
826 /* this will be recalculated during process_buildids() */
827 rec->samples = 0;
828
e1ab48ba
WN
829 if (!err)
830 record__finish_output(rec);
39d17dac 831
e3d59112
NK
832 if (!err && !quiet) {
833 char samples[128];
834
ef149c25 835 if (rec->samples && !rec->opts.full_auxtrace)
e3d59112
NK
836 scnprintf(samples, sizeof(samples),
837 " (%" PRIu64 " samples)", rec->samples);
838 else
839 samples[0] = '\0';
840
841 fprintf(stderr, "[ perf record: Captured and wrote %.3f MB %s%s ]\n",
842 perf_data_file__size(file) / 1024.0 / 1024.0,
843 file->path, samples);
844 }
845
39d17dac
ACM
846out_delete_session:
847 perf_session__delete(session);
45604710 848 return status;
de9ac07b 849}
0e9b20b8 850
72a128aa 851static void callchain_debug(void)
09b0fd45 852{
aad2b21c 853 static const char *str[CALLCHAIN_MAX] = { "NONE", "FP", "DWARF", "LBR" };
a601fdff 854
72a128aa 855 pr_debug("callchain: type %s\n", str[callchain_param.record_mode]);
26d33022 856
72a128aa 857 if (callchain_param.record_mode == CALLCHAIN_DWARF)
09b0fd45 858 pr_debug("callchain: stack dump size %d\n",
72a128aa 859 callchain_param.dump_size);
09b0fd45
JO
860}
861
c421e80b 862int record_parse_callchain_opt(const struct option *opt,
09b0fd45
JO
863 const char *arg,
864 int unset)
865{
09b0fd45 866 int ret;
c421e80b 867 struct record_opts *record = (struct record_opts *)opt->value;
09b0fd45 868
c421e80b 869 record->callgraph_set = true;
72a128aa 870 callchain_param.enabled = !unset;
eb853e80 871
09b0fd45
JO
872 /* --no-call-graph */
873 if (unset) {
72a128aa 874 callchain_param.record_mode = CALLCHAIN_NONE;
09b0fd45
JO
875 pr_debug("callchain: disabled\n");
876 return 0;
877 }
878
c3a6a8c4 879 ret = parse_callchain_record_opt(arg, &callchain_param);
5c0cf224
JO
880 if (!ret) {
881 /* Enable data address sampling for DWARF unwind. */
882 if (callchain_param.record_mode == CALLCHAIN_DWARF)
883 record->sample_address = true;
72a128aa 884 callchain_debug();
5c0cf224 885 }
26d33022
JO
886
887 return ret;
888}
889
c421e80b 890int record_callchain_opt(const struct option *opt,
09b0fd45
JO
891 const char *arg __maybe_unused,
892 int unset __maybe_unused)
893{
c421e80b
KL
894 struct record_opts *record = (struct record_opts *)opt->value;
895
896 record->callgraph_set = true;
72a128aa 897 callchain_param.enabled = true;
09b0fd45 898
72a128aa
NK
899 if (callchain_param.record_mode == CALLCHAIN_NONE)
900 callchain_param.record_mode = CALLCHAIN_FP;
eb853e80 901
72a128aa 902 callchain_debug();
09b0fd45
JO
903 return 0;
904}
905
eb853e80
JO
906static int perf_record_config(const char *var, const char *value, void *cb)
907{
7a29c087
NK
908 struct record *rec = cb;
909
910 if (!strcmp(var, "record.build-id")) {
911 if (!strcmp(value, "cache"))
912 rec->no_buildid_cache = false;
913 else if (!strcmp(value, "no-cache"))
914 rec->no_buildid_cache = true;
915 else if (!strcmp(value, "skip"))
916 rec->no_buildid = true;
917 else
918 return -1;
919 return 0;
920 }
eb853e80 921 if (!strcmp(var, "record.call-graph"))
5a2e5e85 922 var = "call-graph.record-mode"; /* fall-through */
eb853e80
JO
923
924 return perf_default_config(var, value, cb);
925}
926
814c8c38
PZ
927struct clockid_map {
928 const char *name;
929 int clockid;
930};
931
932#define CLOCKID_MAP(n, c) \
933 { .name = n, .clockid = (c), }
934
935#define CLOCKID_END { .name = NULL, }
936
937
938/*
939 * Add the missing ones, we need to build on many distros...
940 */
941#ifndef CLOCK_MONOTONIC_RAW
942#define CLOCK_MONOTONIC_RAW 4
943#endif
944#ifndef CLOCK_BOOTTIME
945#define CLOCK_BOOTTIME 7
946#endif
947#ifndef CLOCK_TAI
948#define CLOCK_TAI 11
949#endif
950
951static const struct clockid_map clockids[] = {
952 /* available for all events, NMI safe */
953 CLOCKID_MAP("monotonic", CLOCK_MONOTONIC),
954 CLOCKID_MAP("monotonic_raw", CLOCK_MONOTONIC_RAW),
955
956 /* available for some events */
957 CLOCKID_MAP("realtime", CLOCK_REALTIME),
958 CLOCKID_MAP("boottime", CLOCK_BOOTTIME),
959 CLOCKID_MAP("tai", CLOCK_TAI),
960
961 /* available for the lazy */
962 CLOCKID_MAP("mono", CLOCK_MONOTONIC),
963 CLOCKID_MAP("raw", CLOCK_MONOTONIC_RAW),
964 CLOCKID_MAP("real", CLOCK_REALTIME),
965 CLOCKID_MAP("boot", CLOCK_BOOTTIME),
966
967 CLOCKID_END,
968};
969
970static int parse_clockid(const struct option *opt, const char *str, int unset)
971{
972 struct record_opts *opts = (struct record_opts *)opt->value;
973 const struct clockid_map *cm;
974 const char *ostr = str;
975
976 if (unset) {
977 opts->use_clockid = 0;
978 return 0;
979 }
980
981 /* no arg passed */
982 if (!str)
983 return 0;
984
985 /* no setting it twice */
986 if (opts->use_clockid)
987 return -1;
988
989 opts->use_clockid = true;
990
991 /* if its a number, we're done */
992 if (sscanf(str, "%d", &opts->clockid) == 1)
993 return 0;
994
995 /* allow a "CLOCK_" prefix to the name */
996 if (!strncasecmp(str, "CLOCK_", 6))
997 str += 6;
998
999 for (cm = clockids; cm->name; cm++) {
1000 if (!strcasecmp(str, cm->name)) {
1001 opts->clockid = cm->clockid;
1002 return 0;
1003 }
1004 }
1005
1006 opts->use_clockid = false;
1007 ui__warning("unknown clockid %s, check man page\n", ostr);
1008 return -1;
1009}
1010
e9db1310
AH
1011static int record__parse_mmap_pages(const struct option *opt,
1012 const char *str,
1013 int unset __maybe_unused)
1014{
1015 struct record_opts *opts = opt->value;
1016 char *s, *p;
1017 unsigned int mmap_pages;
1018 int ret;
1019
1020 if (!str)
1021 return -EINVAL;
1022
1023 s = strdup(str);
1024 if (!s)
1025 return -ENOMEM;
1026
1027 p = strchr(s, ',');
1028 if (p)
1029 *p = '\0';
1030
1031 if (*s) {
1032 ret = __perf_evlist__parse_mmap_pages(&mmap_pages, s);
1033 if (ret)
1034 goto out_free;
1035 opts->mmap_pages = mmap_pages;
1036 }
1037
1038 if (!p) {
1039 ret = 0;
1040 goto out_free;
1041 }
1042
1043 ret = __perf_evlist__parse_mmap_pages(&mmap_pages, p + 1);
1044 if (ret)
1045 goto out_free;
1046
1047 opts->auxtrace_mmap_pages = mmap_pages;
1048
1049out_free:
1050 free(s);
1051 return ret;
1052}
1053
e5b2c207 1054static const char * const __record_usage[] = {
9e096753
MG
1055 "perf record [<options>] [<command>]",
1056 "perf record [<options>] -- <command> [<options>]",
0e9b20b8
IM
1057 NULL
1058};
e5b2c207 1059const char * const *record_usage = __record_usage;
0e9b20b8 1060
d20deb64 1061/*
8c6f45a7
ACM
1062 * XXX Ideally would be local to cmd_record() and passed to a record__new
1063 * because we need to have access to it in record__exit, that is called
d20deb64
ACM
1064 * after cmd_record() exits, but since record_options need to be accessible to
1065 * builtin-script, leave it here.
1066 *
1067 * At least we don't ouch it in all the other functions here directly.
1068 *
1069 * Just say no to tons of global variables, sigh.
1070 */
8c6f45a7 1071static struct record record = {
d20deb64 1072 .opts = {
8affc2b8 1073 .sample_time = true,
d20deb64
ACM
1074 .mmap_pages = UINT_MAX,
1075 .user_freq = UINT_MAX,
1076 .user_interval = ULLONG_MAX,
447a6013 1077 .freq = 4000,
d1cb9fce
NK
1078 .target = {
1079 .uses_mmap = true,
3aa5939d 1080 .default_per_cpu = true,
d1cb9fce 1081 },
9d9cad76 1082 .proc_map_timeout = 500,
d20deb64 1083 },
e3d59112
NK
1084 .tool = {
1085 .sample = process_sample_event,
1086 .fork = perf_event__process_fork,
cca8482c 1087 .exit = perf_event__process_exit,
e3d59112
NK
1088 .comm = perf_event__process_comm,
1089 .mmap = perf_event__process_mmap,
1090 .mmap2 = perf_event__process_mmap2,
cca8482c 1091 .ordered_events = true,
e3d59112 1092 },
d20deb64 1093};
7865e817 1094
76a26549
NK
1095const char record_callchain_help[] = CALLCHAIN_RECORD_HELP
1096 "\n\t\t\t\tDefault: fp";
61eaa3be 1097
d20deb64
ACM
1098/*
1099 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
1100 * with it and switch to use the library functions in perf_evlist that came
b4006796 1101 * from builtin-record.c, i.e. use record_opts,
d20deb64
ACM
1102 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
1103 * using pipes, etc.
1104 */
e5b2c207 1105struct option __record_options[] = {
d20deb64 1106 OPT_CALLBACK('e', "event", &record.evlist, "event",
86847b62 1107 "event selector. use 'perf list' to list available events",
f120f9d5 1108 parse_events_option),
d20deb64 1109 OPT_CALLBACK(0, "filter", &record.evlist, "filter",
c171b552 1110 "event filter", parse_filter),
4ba1faa1
WN
1111 OPT_CALLBACK_NOOPT(0, "exclude-perf", &record.evlist,
1112 NULL, "don't record events from perf itself",
1113 exclude_perf),
bea03405 1114 OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
d6d901c2 1115 "record events on existing process id"),
bea03405 1116 OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
d6d901c2 1117 "record events on existing thread id"),
d20deb64 1118 OPT_INTEGER('r', "realtime", &record.realtime_prio,
0e9b20b8 1119 "collect data with this RT SCHED_FIFO priority"),
509051ea 1120 OPT_BOOLEAN(0, "no-buffering", &record.opts.no_buffering,
acac03fa 1121 "collect data without buffering"),
d20deb64 1122 OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
daac07b2 1123 "collect raw sample records from all opened counters"),
bea03405 1124 OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
0e9b20b8 1125 "system-wide collection from all CPUs"),
bea03405 1126 OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
c45c6ea2 1127 "list of cpus to monitor"),
d20deb64 1128 OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
f5fc1412 1129 OPT_STRING('o', "output", &record.file.path, "file",
abaff32a 1130 "output file name"),
69e7e5b0
AH
1131 OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
1132 &record.opts.no_inherit_set,
1133 "child tasks do not inherit counters"),
d20deb64 1134 OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
e9db1310
AH
1135 OPT_CALLBACK('m', "mmap-pages", &record.opts, "pages[,pages]",
1136 "number of mmap data pages and AUX area tracing mmap pages",
1137 record__parse_mmap_pages),
d20deb64 1138 OPT_BOOLEAN(0, "group", &record.opts.group,
43bece79 1139 "put the counters into a counter group"),
09b0fd45
JO
1140 OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
1141 NULL, "enables call-graph recording" ,
1142 &record_callchain_opt),
1143 OPT_CALLBACK(0, "call-graph", &record.opts,
76a26549 1144 "record_mode[,record_size]", record_callchain_help,
09b0fd45 1145 &record_parse_callchain_opt),
c0555642 1146 OPT_INCR('v', "verbose", &verbose,
3da297a6 1147 "be more verbose (show counter open errors, etc)"),
b44308f5 1148 OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
d20deb64 1149 OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
649c48a9 1150 "per thread counts"),
56100321 1151 OPT_BOOLEAN('d', "data", &record.opts.sample_address, "Record the sample addresses"),
3abebc55
AH
1152 OPT_BOOLEAN_SET('T', "timestamp", &record.opts.sample_time,
1153 &record.opts.sample_time_set,
1154 "Record the sample timestamps"),
56100321 1155 OPT_BOOLEAN('P', "period", &record.opts.period, "Record the sample period"),
d20deb64 1156 OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
649c48a9 1157 "don't sample"),
d2db9a98
WN
1158 OPT_BOOLEAN_SET('N', "no-buildid-cache", &record.no_buildid_cache,
1159 &record.no_buildid_cache_set,
1160 "do not update the buildid cache"),
1161 OPT_BOOLEAN_SET('B', "no-buildid", &record.no_buildid,
1162 &record.no_buildid_set,
1163 "do not collect buildids in perf.data"),
d20deb64 1164 OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
023695d9
SE
1165 "monitor event in cgroup name only",
1166 parse_cgroups),
a6205a35 1167 OPT_UINTEGER('D', "delay", &record.opts.initial_delay,
6619a53e 1168 "ms to wait before starting measurement after program start"),
bea03405
NK
1169 OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
1170 "user to profile"),
a5aabdac
SE
1171
1172 OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
1173 "branch any", "sample any taken branches",
1174 parse_branch_stack),
1175
1176 OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
1177 "branch filter mask", "branch stack filter modes",
bdfebd84 1178 parse_branch_stack),
05484298
AK
1179 OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
1180 "sample by weight (on special events only)"),
475eeab9
AK
1181 OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
1182 "sample transaction flags (special events only)"),
3aa5939d
AH
1183 OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
1184 "use per-thread mmaps"),
bcc84ec6
SE
1185 OPT_CALLBACK_OPTARG('I', "intr-regs", &record.opts.sample_intr_regs, NULL, "any register",
1186 "sample selected machine registers on interrupt,"
1187 " use -I ? to list register names", parse_regs),
85c273d2
AK
1188 OPT_BOOLEAN(0, "running-time", &record.opts.running_time,
1189 "Record running/enabled time of read (:S) events"),
814c8c38
PZ
1190 OPT_CALLBACK('k', "clockid", &record.opts,
1191 "clockid", "clockid to use for events, see clock_gettime()",
1192 parse_clockid),
2dd6d8a1
AH
1193 OPT_STRING_OPTARG('S', "snapshot", &record.opts.auxtrace_snapshot_opts,
1194 "opts", "AUX area tracing Snapshot Mode", ""),
9d9cad76
KL
1195 OPT_UINTEGER(0, "proc-map-timeout", &record.opts.proc_map_timeout,
1196 "per thread proc mmap processing timeout in ms"),
b757bb09
AH
1197 OPT_BOOLEAN(0, "switch-events", &record.opts.record_switch_events,
1198 "Record context switch events"),
85723885
JO
1199 OPT_BOOLEAN_FLAG(0, "all-kernel", &record.opts.all_kernel,
1200 "Configure all used events to run in kernel space.",
1201 PARSE_OPT_EXCLUSIVE),
1202 OPT_BOOLEAN_FLAG(0, "all-user", &record.opts.all_user,
1203 "Configure all used events to run in user space.",
1204 PARSE_OPT_EXCLUSIVE),
71dc2326
WN
1205 OPT_STRING(0, "clang-path", &llvm_param.clang_path, "clang path",
1206 "clang binary to use for compiling BPF scriptlets"),
1207 OPT_STRING(0, "clang-opt", &llvm_param.clang_opt, "clang options",
1208 "options passed to clang when compiling BPF scriptlets"),
7efe0e03
HK
1209 OPT_STRING(0, "vmlinux", &symbol_conf.vmlinux_name,
1210 "file", "vmlinux pathname"),
6156681b
NK
1211 OPT_BOOLEAN(0, "buildid-all", &record.buildid_all,
1212 "Record build-id of all DSOs regardless of hits"),
0e9b20b8
IM
1213 OPT_END()
1214};
1215
e5b2c207
NK
1216struct option *record_options = __record_options;
1217
1d037ca1 1218int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
0e9b20b8 1219{
ef149c25 1220 int err;
8c6f45a7 1221 struct record *rec = &record;
16ad2ffb 1222 char errbuf[BUFSIZ];
0e9b20b8 1223
48e1cab1
WN
1224#ifndef HAVE_LIBBPF_SUPPORT
1225# define set_nobuild(s, l, c) set_option_nobuild(record_options, s, l, "NO_LIBBPF=1", c)
1226 set_nobuild('\0', "clang-path", true);
1227 set_nobuild('\0', "clang-opt", true);
1228# undef set_nobuild
7efe0e03
HK
1229#endif
1230
1231#ifndef HAVE_BPF_PROLOGUE
1232# if !defined (HAVE_DWARF_SUPPORT)
1233# define REASON "NO_DWARF=1"
1234# elif !defined (HAVE_LIBBPF_SUPPORT)
1235# define REASON "NO_LIBBPF=1"
1236# else
1237# define REASON "this architecture doesn't support BPF prologue"
1238# endif
1239# define set_nobuild(s, l, c) set_option_nobuild(record_options, s, l, REASON, c)
1240 set_nobuild('\0', "vmlinux", true);
1241# undef set_nobuild
1242# undef REASON
48e1cab1
WN
1243#endif
1244
3e2be2da
ACM
1245 rec->evlist = perf_evlist__new();
1246 if (rec->evlist == NULL)
361c99a6
ACM
1247 return -ENOMEM;
1248
eb853e80
JO
1249 perf_config(perf_record_config, rec);
1250
bca647aa 1251 argc = parse_options(argc, argv, record_options, record_usage,
655000e7 1252 PARSE_OPT_STOP_AT_NON_OPTION);
602ad878 1253 if (!argc && target__none(&rec->opts.target))
bca647aa 1254 usage_with_options(record_usage, record_options);
0e9b20b8 1255
bea03405 1256 if (nr_cgroups && !rec->opts.target.system_wide) {
c7118369
NK
1257 usage_with_options_msg(record_usage, record_options,
1258 "cgroup monitoring only available in system-wide mode");
1259
023695d9 1260 }
b757bb09
AH
1261 if (rec->opts.record_switch_events &&
1262 !perf_can_record_switch_events()) {
c7118369
NK
1263 ui__error("kernel does not support recording context switch events\n");
1264 parse_options_usage(record_usage, record_options, "switch-events", 0);
1265 return -EINVAL;
b757bb09 1266 }
023695d9 1267
ef149c25
AH
1268 if (!rec->itr) {
1269 rec->itr = auxtrace_record__init(rec->evlist, &err);
1270 if (err)
1271 return err;
1272 }
1273
2dd6d8a1
AH
1274 err = auxtrace_parse_snapshot_options(rec->itr, &rec->opts,
1275 rec->opts.auxtrace_snapshot_opts);
1276 if (err)
1277 return err;
1278
d7888573
WN
1279 err = bpf__setup_stdout(rec->evlist);
1280 if (err) {
1281 bpf__strerror_setup_stdout(rec->evlist, err, errbuf, sizeof(errbuf));
1282 pr_err("ERROR: Setup BPF stdout failed: %s\n",
1283 errbuf);
1284 return err;
1285 }
1286
ef149c25
AH
1287 err = -ENOMEM;
1288
0a7e6d1b 1289 symbol__init(NULL);
baa2f6ce 1290
ec80fde7 1291 if (symbol_conf.kptr_restrict)
646aaea6
ACM
1292 pr_warning(
1293"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
1294"check /proc/sys/kernel/kptr_restrict.\n\n"
1295"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
1296"file is not found in the buildid cache or in the vmlinux path.\n\n"
1297"Samples in kernel modules won't be resolved at all.\n\n"
1298"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
1299"even with a suitable vmlinux or kallsyms file.\n\n");
ec80fde7 1300
d20deb64 1301 if (rec->no_buildid_cache || rec->no_buildid)
a1ac1d3c 1302 disable_buildid_cache();
655000e7 1303
3e2be2da
ACM
1304 if (rec->evlist->nr_entries == 0 &&
1305 perf_evlist__add_default(rec->evlist) < 0) {
69aad6f1
ACM
1306 pr_err("Not enough memory for event selector list\n");
1307 goto out_symbol_exit;
bbd36e5e 1308 }
0e9b20b8 1309
69e7e5b0
AH
1310 if (rec->opts.target.tid && !rec->opts.no_inherit_set)
1311 rec->opts.no_inherit = true;
1312
602ad878 1313 err = target__validate(&rec->opts.target);
16ad2ffb 1314 if (err) {
602ad878 1315 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
16ad2ffb
NK
1316 ui__warning("%s", errbuf);
1317 }
1318
602ad878 1319 err = target__parse_uid(&rec->opts.target);
16ad2ffb
NK
1320 if (err) {
1321 int saved_errno = errno;
4bd0f2d2 1322
602ad878 1323 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
3780f488 1324 ui__error("%s", errbuf);
16ad2ffb
NK
1325
1326 err = -saved_errno;
8fa60e1f 1327 goto out_symbol_exit;
16ad2ffb 1328 }
0d37aa34 1329
16ad2ffb 1330 err = -ENOMEM;
3e2be2da 1331 if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
dd7927f4 1332 usage_with_options(record_usage, record_options);
69aad6f1 1333
ef149c25
AH
1334 err = auxtrace_record__options(rec->itr, rec->evlist, &rec->opts);
1335 if (err)
1336 goto out_symbol_exit;
1337
6156681b
NK
1338 /*
1339 * We take all buildids when the file contains
1340 * AUX area tracing data because we do not decode the
1341 * trace because it would take too long.
1342 */
1343 if (rec->opts.full_auxtrace)
1344 rec->buildid_all = true;
1345
b4006796 1346 if (record_opts__config(&rec->opts)) {
39d17dac 1347 err = -EINVAL;
03ad9747 1348 goto out_symbol_exit;
7e4ff9e3
MG
1349 }
1350
d20deb64 1351 err = __cmd_record(&record, argc, argv);
d65a458b 1352out_symbol_exit:
45604710 1353 perf_evlist__delete(rec->evlist);
d65a458b 1354 symbol__exit();
ef149c25 1355 auxtrace_record__free(rec->itr);
39d17dac 1356 return err;
0e9b20b8 1357}
2dd6d8a1
AH
1358
1359static void snapshot_sig_handler(int sig __maybe_unused)
1360{
1361 if (!auxtrace_snapshot_enabled)
1362 return;
1363 auxtrace_snapshot_enabled = 0;
1364 auxtrace_snapshot_err = auxtrace_record__snapshot_start(record.itr);
1365 auxtrace_record__snapshot_started = 1;
1366}