perf trace: Introduce --min-stack filter
[linux-2.6-block.git] / tools / perf / builtin-record.c
CommitLineData
abaff32a 1/*
bf9e1876
IM
2 * builtin-record.c
3 *
4 * Builtin record command: Record the profile of a workload
5 * (or a CPU, or a PID) into the perf.data output file - for
6 * later analysis via perf report.
abaff32a 7 */
16f762a2 8#include "builtin.h"
bf9e1876
IM
9
10#include "perf.h"
11
6122e4e4 12#include "util/build-id.h"
6eda5838 13#include "util/util.h"
4b6ab94e 14#include <subcmd/parse-options.h>
8ad8db37 15#include "util/parse-events.h"
6eda5838 16
8f651eae 17#include "util/callchain.h"
f14d5707 18#include "util/cgroup.h"
7c6a1c65 19#include "util/header.h"
66e274f3 20#include "util/event.h"
361c99a6 21#include "util/evlist.h"
69aad6f1 22#include "util/evsel.h"
8f28827a 23#include "util/debug.h"
94c744b6 24#include "util/session.h"
45694aa7 25#include "util/tool.h"
8d06367f 26#include "util/symbol.h"
a12b51c4 27#include "util/cpumap.h"
fd78260b 28#include "util/thread_map.h"
f5fc1412 29#include "util/data.h"
bcc84ec6 30#include "util/perf_regs.h"
ef149c25 31#include "util/auxtrace.h"
46bc29b9 32#include "util/tsc.h"
f00898f4 33#include "util/parse-branch-options.h"
bcc84ec6 34#include "util/parse-regs-options.h"
71dc2326 35#include "util/llvm-utils.h"
8690a2a7 36#include "util/bpf-loader.h"
d8871ea7 37#include "asm/bug.h"
7c6a1c65 38
97124d5e 39#include <unistd.h>
de9ac07b 40#include <sched.h>
a41794cd 41#include <sys/mman.h>
de9ac07b 42
78da39fa 43
8c6f45a7 44struct record {
45694aa7 45 struct perf_tool tool;
b4006796 46 struct record_opts opts;
d20deb64 47 u64 bytes_written;
f5fc1412 48 struct perf_data_file file;
ef149c25 49 struct auxtrace_record *itr;
d20deb64
ACM
50 struct perf_evlist *evlist;
51 struct perf_session *session;
52 const char *progname;
d20deb64 53 int realtime_prio;
d20deb64 54 bool no_buildid;
d2db9a98 55 bool no_buildid_set;
d20deb64 56 bool no_buildid_cache;
d2db9a98 57 bool no_buildid_cache_set;
6156681b 58 bool buildid_all;
ecfd7a9c 59 bool timestamp_filename;
9f065194 60 unsigned long long samples;
0f82ebc4 61};
a21ca2ca 62
8c6f45a7 63static int record__write(struct record *rec, void *bf, size_t size)
f5970550 64{
cf8b2e69 65 if (perf_data_file__write(rec->session->file, bf, size) < 0) {
50a9b868
JO
66 pr_err("failed to write perf data, error: %m\n");
67 return -1;
f5970550 68 }
8d3eca20 69
cf8b2e69 70 rec->bytes_written += size;
8d3eca20 71 return 0;
f5970550
PZ
72}
73
45694aa7 74static int process_synthesized_event(struct perf_tool *tool,
d20deb64 75 union perf_event *event,
1d037ca1
IT
76 struct perf_sample *sample __maybe_unused,
77 struct machine *machine __maybe_unused)
234fbbf5 78{
8c6f45a7
ACM
79 struct record *rec = container_of(tool, struct record, tool);
80 return record__write(rec, event, event->header.size);
234fbbf5
ACM
81}
82
e5685730 83static int record__mmap_read(struct record *rec, int idx)
de9ac07b 84{
e5685730 85 struct perf_mmap *md = &rec->evlist->mmap[idx];
7b8283b5
DA
86 u64 head = perf_mmap__read_head(md);
87 u64 old = md->prev;
918512b4 88 unsigned char *data = md->base + page_size;
de9ac07b
PZ
89 unsigned long size;
90 void *buf;
8d3eca20 91 int rc = 0;
de9ac07b 92
dc82009a 93 if (old == head)
8d3eca20 94 return 0;
dc82009a 95
d20deb64 96 rec->samples++;
de9ac07b
PZ
97
98 size = head - old;
99
100 if ((old & md->mask) + size != (head & md->mask)) {
101 buf = &data[old & md->mask];
102 size = md->mask + 1 - (old & md->mask);
103 old += size;
021e9f47 104
8c6f45a7 105 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
106 rc = -1;
107 goto out;
108 }
de9ac07b
PZ
109 }
110
111 buf = &data[old & md->mask];
112 size = head - old;
113 old += size;
021e9f47 114
8c6f45a7 115 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
116 rc = -1;
117 goto out;
118 }
de9ac07b
PZ
119
120 md->prev = old;
e5685730 121 perf_evlist__mmap_consume(rec->evlist, idx);
8d3eca20
DA
122out:
123 return rc;
de9ac07b
PZ
124}
125
2dd6d8a1
AH
126static volatile int done;
127static volatile int signr = -1;
128static volatile int child_finished;
c0bdc1c4
WN
129
130static volatile enum {
131 AUXTRACE_SNAPSHOT_OFF = -1,
132 AUXTRACE_SNAPSHOT_DISABLED = 0,
133 AUXTRACE_SNAPSHOT_ENABLED = 1,
134} auxtrace_snapshot_state = AUXTRACE_SNAPSHOT_OFF;
135
136static inline void
137auxtrace_snapshot_on(void)
138{
139 auxtrace_snapshot_state = AUXTRACE_SNAPSHOT_DISABLED;
140}
141
142static inline void
143auxtrace_snapshot_enable(void)
144{
145 if (auxtrace_snapshot_state == AUXTRACE_SNAPSHOT_OFF)
146 return;
147 auxtrace_snapshot_state = AUXTRACE_SNAPSHOT_ENABLED;
148}
149
150static inline void
151auxtrace_snapshot_disable(void)
152{
153 if (auxtrace_snapshot_state == AUXTRACE_SNAPSHOT_OFF)
154 return;
155 auxtrace_snapshot_state = AUXTRACE_SNAPSHOT_DISABLED;
156}
157
158static inline bool
159auxtrace_snapshot_is_enabled(void)
160{
161 if (auxtrace_snapshot_state == AUXTRACE_SNAPSHOT_OFF)
162 return false;
163 return auxtrace_snapshot_state == AUXTRACE_SNAPSHOT_ENABLED;
164}
165
2dd6d8a1
AH
166static volatile int auxtrace_snapshot_err;
167static volatile int auxtrace_record__snapshot_started;
168
169static void sig_handler(int sig)
170{
171 if (sig == SIGCHLD)
172 child_finished = 1;
173 else
174 signr = sig;
175
176 done = 1;
177}
178
179static void record__sig_exit(void)
180{
181 if (signr == -1)
182 return;
183
184 signal(signr, SIG_DFL);
185 raise(signr);
186}
187
e31f0d01
AH
188#ifdef HAVE_AUXTRACE_SUPPORT
189
ef149c25
AH
190static int record__process_auxtrace(struct perf_tool *tool,
191 union perf_event *event, void *data1,
192 size_t len1, void *data2, size_t len2)
193{
194 struct record *rec = container_of(tool, struct record, tool);
99fa2984 195 struct perf_data_file *file = &rec->file;
ef149c25
AH
196 size_t padding;
197 u8 pad[8] = {0};
198
99fa2984
AH
199 if (!perf_data_file__is_pipe(file)) {
200 off_t file_offset;
201 int fd = perf_data_file__fd(file);
202 int err;
203
204 file_offset = lseek(fd, 0, SEEK_CUR);
205 if (file_offset == -1)
206 return -1;
207 err = auxtrace_index__auxtrace_event(&rec->session->auxtrace_index,
208 event, file_offset);
209 if (err)
210 return err;
211 }
212
ef149c25
AH
213 /* event.auxtrace.size includes padding, see __auxtrace_mmap__read() */
214 padding = (len1 + len2) & 7;
215 if (padding)
216 padding = 8 - padding;
217
218 record__write(rec, event, event->header.size);
219 record__write(rec, data1, len1);
220 if (len2)
221 record__write(rec, data2, len2);
222 record__write(rec, &pad, padding);
223
224 return 0;
225}
226
227static int record__auxtrace_mmap_read(struct record *rec,
228 struct auxtrace_mmap *mm)
229{
230 int ret;
231
232 ret = auxtrace_mmap__read(mm, rec->itr, &rec->tool,
233 record__process_auxtrace);
234 if (ret < 0)
235 return ret;
236
237 if (ret)
238 rec->samples++;
239
240 return 0;
241}
242
2dd6d8a1
AH
243static int record__auxtrace_mmap_read_snapshot(struct record *rec,
244 struct auxtrace_mmap *mm)
245{
246 int ret;
247
248 ret = auxtrace_mmap__read_snapshot(mm, rec->itr, &rec->tool,
249 record__process_auxtrace,
250 rec->opts.auxtrace_snapshot_size);
251 if (ret < 0)
252 return ret;
253
254 if (ret)
255 rec->samples++;
256
257 return 0;
258}
259
260static int record__auxtrace_read_snapshot_all(struct record *rec)
261{
262 int i;
263 int rc = 0;
264
265 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
266 struct auxtrace_mmap *mm =
267 &rec->evlist->mmap[i].auxtrace_mmap;
268
269 if (!mm->base)
270 continue;
271
272 if (record__auxtrace_mmap_read_snapshot(rec, mm) != 0) {
273 rc = -1;
274 goto out;
275 }
276 }
277out:
278 return rc;
279}
280
281static void record__read_auxtrace_snapshot(struct record *rec)
282{
283 pr_debug("Recording AUX area tracing snapshot\n");
284 if (record__auxtrace_read_snapshot_all(rec) < 0) {
285 auxtrace_snapshot_err = -1;
286 } else {
287 auxtrace_snapshot_err = auxtrace_record__snapshot_finish(rec->itr);
288 if (!auxtrace_snapshot_err)
c0bdc1c4 289 auxtrace_snapshot_enable();
2dd6d8a1
AH
290 }
291}
292
e31f0d01
AH
293#else
294
295static inline
296int record__auxtrace_mmap_read(struct record *rec __maybe_unused,
297 struct auxtrace_mmap *mm __maybe_unused)
298{
299 return 0;
300}
301
2dd6d8a1
AH
302static inline
303void record__read_auxtrace_snapshot(struct record *rec __maybe_unused)
de9ac07b 304{
f7b7c26e
PZ
305}
306
2dd6d8a1
AH
307static inline
308int auxtrace_record__snapshot_start(struct auxtrace_record *itr __maybe_unused)
f7b7c26e 309{
2dd6d8a1 310 return 0;
de9ac07b
PZ
311}
312
2dd6d8a1
AH
313#endif
314
8c6f45a7 315static int record__open(struct record *rec)
dd7927f4 316{
56e52e85 317 char msg[512];
6a4bb04c 318 struct perf_evsel *pos;
d20deb64
ACM
319 struct perf_evlist *evlist = rec->evlist;
320 struct perf_session *session = rec->session;
b4006796 321 struct record_opts *opts = &rec->opts;
8d3eca20 322 int rc = 0;
dd7927f4 323
e68ae9cf 324 perf_evlist__config(evlist, opts, &callchain_param);
cac21425 325
0050f7aa 326 evlist__for_each(evlist, pos) {
dd7927f4 327try_again:
d988d5ee 328 if (perf_evsel__open(pos, pos->cpus, pos->threads) < 0) {
56e52e85 329 if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
d6d901c2 330 if (verbose)
c0a54341 331 ui__warning("%s\n", msg);
d6d901c2
ZY
332 goto try_again;
333 }
ca6a4258 334
56e52e85
ACM
335 rc = -errno;
336 perf_evsel__open_strerror(pos, &opts->target,
337 errno, msg, sizeof(msg));
338 ui__error("%s\n", msg);
8d3eca20 339 goto out;
c171b552
LZ
340 }
341 }
a43d3f08 342
23d4aad4
ACM
343 if (perf_evlist__apply_filters(evlist, &pos)) {
344 error("failed to set filter \"%s\" on event %s with %d (%s)\n",
345 pos->filter, perf_evsel__name(pos), errno,
35550da3 346 strerror_r(errno, msg, sizeof(msg)));
8d3eca20
DA
347 rc = -1;
348 goto out;
0a102479
FW
349 }
350
ef149c25 351 if (perf_evlist__mmap_ex(evlist, opts->mmap_pages, false,
2dd6d8a1
AH
352 opts->auxtrace_mmap_pages,
353 opts->auxtrace_snapshot_mode) < 0) {
8d3eca20
DA
354 if (errno == EPERM) {
355 pr_err("Permission error mapping pages.\n"
356 "Consider increasing "
357 "/proc/sys/kernel/perf_event_mlock_kb,\n"
358 "or try again with a smaller value of -m/--mmap_pages.\n"
ef149c25
AH
359 "(current value: %u,%u)\n",
360 opts->mmap_pages, opts->auxtrace_mmap_pages);
8d3eca20 361 rc = -errno;
8d3eca20 362 } else {
35550da3
MH
363 pr_err("failed to mmap with %d (%s)\n", errno,
364 strerror_r(errno, msg, sizeof(msg)));
95c36561
WN
365 if (errno)
366 rc = -errno;
367 else
368 rc = -EINVAL;
8d3eca20
DA
369 }
370 goto out;
18e60939 371 }
0a27d7f9 372
563aecb2 373 session->evlist = evlist;
7b56cce2 374 perf_session__set_id_hdr_size(session);
8d3eca20
DA
375out:
376 return rc;
16c8a109
PZ
377}
378
e3d59112
NK
379static int process_sample_event(struct perf_tool *tool,
380 union perf_event *event,
381 struct perf_sample *sample,
382 struct perf_evsel *evsel,
383 struct machine *machine)
384{
385 struct record *rec = container_of(tool, struct record, tool);
386
387 rec->samples++;
388
389 return build_id__mark_dso_hit(tool, event, sample, evsel, machine);
390}
391
8c6f45a7 392static int process_buildids(struct record *rec)
6122e4e4 393{
f5fc1412
JO
394 struct perf_data_file *file = &rec->file;
395 struct perf_session *session = rec->session;
6122e4e4 396
457ae94a 397 if (file->size == 0)
9f591fd7
ACM
398 return 0;
399
00dc8657
NK
400 /*
401 * During this process, it'll load kernel map and replace the
402 * dso->long_name to a real pathname it found. In this case
403 * we prefer the vmlinux path like
404 * /lib/modules/3.16.4/build/vmlinux
405 *
406 * rather than build-id path (in debug directory).
407 * $HOME/.debug/.build-id/f0/6e17aa50adf4d00b88925e03775de107611551
408 */
409 symbol_conf.ignore_vmlinux_buildid = true;
410
6156681b
NK
411 /*
412 * If --buildid-all is given, it marks all DSO regardless of hits,
413 * so no need to process samples.
414 */
415 if (rec->buildid_all)
416 rec->tool.sample = NULL;
417
b7b61cbe 418 return perf_session__process_events(session);
6122e4e4
ACM
419}
420
8115d60c 421static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
a1645ce1
ZY
422{
423 int err;
45694aa7 424 struct perf_tool *tool = data;
a1645ce1
ZY
425 /*
426 *As for guest kernel when processing subcommand record&report,
427 *we arrange module mmap prior to guest kernel mmap and trigger
428 *a preload dso because default guest module symbols are loaded
429 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
430 *method is used to avoid symbol missing when the first addr is
431 *in module instead of in guest kernel.
432 */
45694aa7 433 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 434 machine);
a1645ce1
ZY
435 if (err < 0)
436 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 437 " relocation symbol.\n", machine->pid);
a1645ce1 438
a1645ce1
ZY
439 /*
440 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
441 * have no _text sometimes.
442 */
45694aa7 443 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
0ae617be 444 machine);
a1645ce1
ZY
445 if (err < 0)
446 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 447 " relocation symbol.\n", machine->pid);
a1645ce1
ZY
448}
449
98402807
FW
450static struct perf_event_header finished_round_event = {
451 .size = sizeof(struct perf_event_header),
452 .type = PERF_RECORD_FINISHED_ROUND,
453};
454
8c6f45a7 455static int record__mmap_read_all(struct record *rec)
98402807 456{
dcabb507 457 u64 bytes_written = rec->bytes_written;
0e2e63dd 458 int i;
8d3eca20 459 int rc = 0;
98402807 460
d20deb64 461 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
ef149c25
AH
462 struct auxtrace_mmap *mm = &rec->evlist->mmap[i].auxtrace_mmap;
463
8d3eca20 464 if (rec->evlist->mmap[i].base) {
e5685730 465 if (record__mmap_read(rec, i) != 0) {
8d3eca20
DA
466 rc = -1;
467 goto out;
468 }
469 }
ef149c25 470
2dd6d8a1 471 if (mm->base && !rec->opts.auxtrace_snapshot_mode &&
ef149c25
AH
472 record__auxtrace_mmap_read(rec, mm) != 0) {
473 rc = -1;
474 goto out;
475 }
98402807
FW
476 }
477
dcabb507
JO
478 /*
479 * Mark the round finished in case we wrote
480 * at least one event.
481 */
482 if (bytes_written != rec->bytes_written)
483 rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
8d3eca20
DA
484
485out:
486 return rc;
98402807
FW
487}
488
8c6f45a7 489static void record__init_features(struct record *rec)
57706abc 490{
57706abc
DA
491 struct perf_session *session = rec->session;
492 int feat;
493
494 for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
495 perf_header__set_feat(&session->header, feat);
496
497 if (rec->no_buildid)
498 perf_header__clear_feat(&session->header, HEADER_BUILD_ID);
499
3e2be2da 500 if (!have_tracepoints(&rec->evlist->entries))
57706abc
DA
501 perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
502
503 if (!rec->opts.branch_stack)
504 perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
ef149c25
AH
505
506 if (!rec->opts.full_auxtrace)
507 perf_header__clear_feat(&session->header, HEADER_AUXTRACE);
ffa517ad
JO
508
509 perf_header__clear_feat(&session->header, HEADER_STAT);
57706abc
DA
510}
511
e1ab48ba
WN
512static void
513record__finish_output(struct record *rec)
514{
515 struct perf_data_file *file = &rec->file;
516 int fd = perf_data_file__fd(file);
517
518 if (file->is_pipe)
519 return;
520
521 rec->session->header.data_size += rec->bytes_written;
522 file->size = lseek(perf_data_file__fd(file), 0, SEEK_CUR);
523
524 if (!rec->no_buildid) {
525 process_buildids(rec);
526
527 if (rec->buildid_all)
528 dsos__hit_all(rec->session);
529 }
530 perf_session__write_header(rec->session, rec->evlist, fd, true);
531
532 return;
533}
534
ecfd7a9c
WN
535static int
536record__switch_output(struct record *rec, bool at_exit)
537{
538 struct perf_data_file *file = &rec->file;
539 int fd, err;
540
541 /* Same Size: "2015122520103046"*/
542 char timestamp[] = "InvalidTimestamp";
543
544 rec->samples = 0;
545 record__finish_output(rec);
546 err = fetch_current_timestamp(timestamp, sizeof(timestamp));
547 if (err) {
548 pr_err("Failed to get current timestamp\n");
549 return -EINVAL;
550 }
551
552 fd = perf_data_file__switch(file, timestamp,
553 rec->session->header.data_offset,
554 at_exit);
555 if (fd >= 0 && !at_exit) {
556 rec->bytes_written = 0;
557 rec->session->header.data_size = 0;
558 }
559
560 if (!quiet)
561 fprintf(stderr, "[ perf record: Dump %s.%s ]\n",
562 file->path, timestamp);
563 return fd;
564}
565
f33cbe72
ACM
566static volatile int workload_exec_errno;
567
568/*
569 * perf_evlist__prepare_workload will send a SIGUSR1
570 * if the fork fails, since we asked by setting its
571 * want_signal to true.
572 */
45604710
NK
573static void workload_exec_failed_signal(int signo __maybe_unused,
574 siginfo_t *info,
f33cbe72
ACM
575 void *ucontext __maybe_unused)
576{
577 workload_exec_errno = info->si_value.sival_int;
578 done = 1;
f33cbe72
ACM
579 child_finished = 1;
580}
581
2dd6d8a1
AH
582static void snapshot_sig_handler(int sig);
583
46bc29b9
AH
584int __weak
585perf_event__synth_time_conv(const struct perf_event_mmap_page *pc __maybe_unused,
586 struct perf_tool *tool __maybe_unused,
587 perf_event__handler_t process __maybe_unused,
588 struct machine *machine __maybe_unused)
589{
590 return 0;
591}
592
c45c86eb
WN
593static int record__synthesize(struct record *rec)
594{
595 struct perf_session *session = rec->session;
596 struct machine *machine = &session->machines.host;
597 struct perf_data_file *file = &rec->file;
598 struct record_opts *opts = &rec->opts;
599 struct perf_tool *tool = &rec->tool;
600 int fd = perf_data_file__fd(file);
601 int err = 0;
602
603 if (file->is_pipe) {
604 err = perf_event__synthesize_attrs(tool, session,
605 process_synthesized_event);
606 if (err < 0) {
607 pr_err("Couldn't synthesize attrs.\n");
608 goto out;
609 }
610
611 if (have_tracepoints(&rec->evlist->entries)) {
612 /*
613 * FIXME err <= 0 here actually means that
614 * there were no tracepoints so its not really
615 * an error, just that we don't need to
616 * synthesize anything. We really have to
617 * return this more properly and also
618 * propagate errors that now are calling die()
619 */
620 err = perf_event__synthesize_tracing_data(tool, fd, rec->evlist,
621 process_synthesized_event);
622 if (err <= 0) {
623 pr_err("Couldn't record tracing data.\n");
624 goto out;
625 }
626 rec->bytes_written += err;
627 }
628 }
629
46bc29b9
AH
630 err = perf_event__synth_time_conv(rec->evlist->mmap[0].base, tool,
631 process_synthesized_event, machine);
632 if (err)
633 goto out;
634
c45c86eb
WN
635 if (rec->opts.full_auxtrace) {
636 err = perf_event__synthesize_auxtrace_info(rec->itr, tool,
637 session, process_synthesized_event);
638 if (err)
639 goto out;
640 }
641
642 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
643 machine);
644 WARN_ONCE(err < 0, "Couldn't record kernel reference relocation symbol\n"
645 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
646 "Check /proc/kallsyms permission or run as root.\n");
647
648 err = perf_event__synthesize_modules(tool, process_synthesized_event,
649 machine);
650 WARN_ONCE(err < 0, "Couldn't record kernel module information.\n"
651 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
652 "Check /proc/modules permission or run as root.\n");
653
654 if (perf_guest) {
655 machines__process_guests(&session->machines,
656 perf_event__synthesize_guest_os, tool);
657 }
658
659 err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
660 process_synthesized_event, opts->sample_address,
661 opts->proc_map_timeout);
662out:
663 return err;
664}
665
8c6f45a7 666static int __cmd_record(struct record *rec, int argc, const char **argv)
16c8a109 667{
57706abc 668 int err;
45604710 669 int status = 0;
8b412664 670 unsigned long waking = 0;
46be604b 671 const bool forks = argc > 0;
23346f21 672 struct machine *machine;
45694aa7 673 struct perf_tool *tool = &rec->tool;
b4006796 674 struct record_opts *opts = &rec->opts;
f5fc1412 675 struct perf_data_file *file = &rec->file;
d20deb64 676 struct perf_session *session;
6dcf45ef 677 bool disabled = false, draining = false;
42aa276f 678 int fd;
de9ac07b 679
d20deb64 680 rec->progname = argv[0];
33e49ea7 681
45604710 682 atexit(record__sig_exit);
f5970550
PZ
683 signal(SIGCHLD, sig_handler);
684 signal(SIGINT, sig_handler);
804f7ac7 685 signal(SIGTERM, sig_handler);
c0bdc1c4
WN
686
687 if (rec->opts.auxtrace_snapshot_mode) {
2dd6d8a1 688 signal(SIGUSR2, snapshot_sig_handler);
c0bdc1c4
WN
689 auxtrace_snapshot_on();
690 } else {
2dd6d8a1 691 signal(SIGUSR2, SIG_IGN);
c0bdc1c4 692 }
f5970550 693
b7b61cbe 694 session = perf_session__new(file, false, tool);
94c744b6 695 if (session == NULL) {
ffa91880 696 pr_err("Perf session creation failed.\n");
a9a70bbc
ACM
697 return -1;
698 }
699
42aa276f 700 fd = perf_data_file__fd(file);
d20deb64
ACM
701 rec->session = session;
702
8c6f45a7 703 record__init_features(rec);
330aa675 704
d4db3f16 705 if (forks) {
3e2be2da 706 err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
f5fc1412 707 argv, file->is_pipe,
735f7e0b 708 workload_exec_failed_signal);
35b9d88e
ACM
709 if (err < 0) {
710 pr_err("Couldn't run the workload!\n");
45604710 711 status = err;
35b9d88e 712 goto out_delete_session;
856e9660 713 }
856e9660
PZ
714 }
715
8c6f45a7 716 if (record__open(rec) != 0) {
8d3eca20 717 err = -1;
45604710 718 goto out_child;
8d3eca20 719 }
de9ac07b 720
8690a2a7
WN
721 err = bpf__apply_obj_config();
722 if (err) {
723 char errbuf[BUFSIZ];
724
725 bpf__strerror_apply_obj_config(err, errbuf, sizeof(errbuf));
726 pr_err("ERROR: Apply config to BPF failed: %s\n",
727 errbuf);
728 goto out_child;
729 }
730
cca8482c
AH
731 /*
732 * Normally perf_session__new would do this, but it doesn't have the
733 * evlist.
734 */
735 if (rec->tool.ordered_events && !perf_evlist__sample_id_all(rec->evlist)) {
736 pr_warning("WARNING: No sample_id_all support, falling back to unordered processing\n");
737 rec->tool.ordered_events = false;
738 }
739
3e2be2da 740 if (!rec->evlist->nr_groups)
a8bb559b
NK
741 perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);
742
f5fc1412 743 if (file->is_pipe) {
42aa276f 744 err = perf_header__write_pipe(fd);
529870e3 745 if (err < 0)
45604710 746 goto out_child;
563aecb2 747 } else {
42aa276f 748 err = perf_session__write_header(session, rec->evlist, fd, false);
d5eed904 749 if (err < 0)
45604710 750 goto out_child;
56b03f3c
ACM
751 }
752
d3665498 753 if (!rec->no_buildid
e20960c0 754 && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
d3665498 755 pr_err("Couldn't generate buildids. "
e20960c0 756 "Use --no-buildid to profile anyway.\n");
8d3eca20 757 err = -1;
45604710 758 goto out_child;
e20960c0
RR
759 }
760
34ba5122 761 machine = &session->machines.host;
743eb868 762
c45c86eb
WN
763 err = record__synthesize(rec);
764 if (err < 0)
45604710 765 goto out_child;
8d3eca20 766
d20deb64 767 if (rec->realtime_prio) {
de9ac07b
PZ
768 struct sched_param param;
769
d20deb64 770 param.sched_priority = rec->realtime_prio;
de9ac07b 771 if (sched_setscheduler(0, SCHED_FIFO, &param)) {
6beba7ad 772 pr_err("Could not set realtime priority.\n");
8d3eca20 773 err = -1;
45604710 774 goto out_child;
de9ac07b
PZ
775 }
776 }
777
774cb499
JO
778 /*
779 * When perf is starting the traced process, all the events
780 * (apart from group members) have enable_on_exec=1 set,
781 * so don't spoil it by prematurely enabling them.
782 */
6619a53e 783 if (!target__none(&opts->target) && !opts->initial_delay)
3e2be2da 784 perf_evlist__enable(rec->evlist);
764e16a3 785
856e9660
PZ
786 /*
787 * Let the child rip
788 */
e803cf97 789 if (forks) {
e5bed564
NK
790 union perf_event *event;
791
792 event = malloc(sizeof(event->comm) + machine->id_hdr_size);
793 if (event == NULL) {
794 err = -ENOMEM;
795 goto out_child;
796 }
797
e803cf97
NK
798 /*
799 * Some H/W events are generated before COMM event
800 * which is emitted during exec(), so perf script
801 * cannot see a correct process name for those events.
802 * Synthesize COMM event to prevent it.
803 */
e5bed564 804 perf_event__synthesize_comm(tool, event,
e803cf97
NK
805 rec->evlist->workload.pid,
806 process_synthesized_event,
807 machine);
e5bed564 808 free(event);
e803cf97 809
3e2be2da 810 perf_evlist__start_workload(rec->evlist);
e803cf97 811 }
856e9660 812
6619a53e
AK
813 if (opts->initial_delay) {
814 usleep(opts->initial_delay * 1000);
815 perf_evlist__enable(rec->evlist);
816 }
817
c0bdc1c4 818 auxtrace_snapshot_enable();
649c48a9 819 for (;;) {
9f065194 820 unsigned long long hits = rec->samples;
de9ac07b 821
8c6f45a7 822 if (record__mmap_read_all(rec) < 0) {
c0bdc1c4 823 auxtrace_snapshot_disable();
8d3eca20 824 err = -1;
45604710 825 goto out_child;
8d3eca20 826 }
de9ac07b 827
2dd6d8a1
AH
828 if (auxtrace_record__snapshot_started) {
829 auxtrace_record__snapshot_started = 0;
830 if (!auxtrace_snapshot_err)
831 record__read_auxtrace_snapshot(rec);
832 if (auxtrace_snapshot_err) {
833 pr_err("AUX area tracing snapshot failed\n");
834 err = -1;
835 goto out_child;
836 }
837 }
838
d20deb64 839 if (hits == rec->samples) {
6dcf45ef 840 if (done || draining)
649c48a9 841 break;
f66a889d 842 err = perf_evlist__poll(rec->evlist, -1);
a515114f
JO
843 /*
844 * Propagate error, only if there's any. Ignore positive
845 * number of returned events and interrupt error.
846 */
847 if (err > 0 || (err < 0 && errno == EINTR))
45604710 848 err = 0;
8b412664 849 waking++;
6dcf45ef
ACM
850
851 if (perf_evlist__filter_pollfd(rec->evlist, POLLERR | POLLHUP) == 0)
852 draining = true;
8b412664
PZ
853 }
854
774cb499
JO
855 /*
856 * When perf is starting the traced process, at the end events
857 * die with the process and we wait for that. Thus no need to
858 * disable events in this case.
859 */
602ad878 860 if (done && !disabled && !target__none(&opts->target)) {
c0bdc1c4 861 auxtrace_snapshot_disable();
3e2be2da 862 perf_evlist__disable(rec->evlist);
2711926a
JO
863 disabled = true;
864 }
de9ac07b 865 }
c0bdc1c4 866 auxtrace_snapshot_disable();
de9ac07b 867
f33cbe72 868 if (forks && workload_exec_errno) {
35550da3 869 char msg[STRERR_BUFSIZE];
f33cbe72
ACM
870 const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
871 pr_err("Workload failed: %s\n", emsg);
872 err = -1;
45604710 873 goto out_child;
f33cbe72
ACM
874 }
875
e3d59112 876 if (!quiet)
45604710 877 fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);
b44308f5 878
45604710
NK
879out_child:
880 if (forks) {
881 int exit_status;
addc2785 882
45604710
NK
883 if (!child_finished)
884 kill(rec->evlist->workload.pid, SIGTERM);
885
886 wait(&exit_status);
887
888 if (err < 0)
889 status = err;
890 else if (WIFEXITED(exit_status))
891 status = WEXITSTATUS(exit_status);
892 else if (WIFSIGNALED(exit_status))
893 signr = WTERMSIG(exit_status);
894 } else
895 status = err;
896
e3d59112
NK
897 /* this will be recalculated during process_buildids() */
898 rec->samples = 0;
899
ecfd7a9c
WN
900 if (!err) {
901 if (!rec->timestamp_filename) {
902 record__finish_output(rec);
903 } else {
904 fd = record__switch_output(rec, true);
905 if (fd < 0) {
906 status = fd;
907 goto out_delete_session;
908 }
909 }
910 }
39d17dac 911
e3d59112
NK
912 if (!err && !quiet) {
913 char samples[128];
ecfd7a9c
WN
914 const char *postfix = rec->timestamp_filename ?
915 ".<timestamp>" : "";
e3d59112 916
ef149c25 917 if (rec->samples && !rec->opts.full_auxtrace)
e3d59112
NK
918 scnprintf(samples, sizeof(samples),
919 " (%" PRIu64 " samples)", rec->samples);
920 else
921 samples[0] = '\0';
922
ecfd7a9c 923 fprintf(stderr, "[ perf record: Captured and wrote %.3f MB %s%s%s ]\n",
e3d59112 924 perf_data_file__size(file) / 1024.0 / 1024.0,
ecfd7a9c 925 file->path, postfix, samples);
e3d59112
NK
926 }
927
39d17dac
ACM
928out_delete_session:
929 perf_session__delete(session);
45604710 930 return status;
de9ac07b 931}
0e9b20b8 932
72a128aa 933static void callchain_debug(void)
09b0fd45 934{
aad2b21c 935 static const char *str[CALLCHAIN_MAX] = { "NONE", "FP", "DWARF", "LBR" };
a601fdff 936
72a128aa 937 pr_debug("callchain: type %s\n", str[callchain_param.record_mode]);
26d33022 938
72a128aa 939 if (callchain_param.record_mode == CALLCHAIN_DWARF)
09b0fd45 940 pr_debug("callchain: stack dump size %d\n",
72a128aa 941 callchain_param.dump_size);
09b0fd45
JO
942}
943
c421e80b 944int record_parse_callchain_opt(const struct option *opt,
09b0fd45
JO
945 const char *arg,
946 int unset)
947{
09b0fd45 948 int ret;
c421e80b 949 struct record_opts *record = (struct record_opts *)opt->value;
09b0fd45 950
c421e80b 951 record->callgraph_set = true;
72a128aa 952 callchain_param.enabled = !unset;
eb853e80 953
09b0fd45
JO
954 /* --no-call-graph */
955 if (unset) {
72a128aa 956 callchain_param.record_mode = CALLCHAIN_NONE;
09b0fd45
JO
957 pr_debug("callchain: disabled\n");
958 return 0;
959 }
960
c3a6a8c4 961 ret = parse_callchain_record_opt(arg, &callchain_param);
5c0cf224
JO
962 if (!ret) {
963 /* Enable data address sampling for DWARF unwind. */
964 if (callchain_param.record_mode == CALLCHAIN_DWARF)
965 record->sample_address = true;
72a128aa 966 callchain_debug();
5c0cf224 967 }
26d33022
JO
968
969 return ret;
970}
971
c421e80b 972int record_callchain_opt(const struct option *opt,
09b0fd45
JO
973 const char *arg __maybe_unused,
974 int unset __maybe_unused)
975{
c421e80b
KL
976 struct record_opts *record = (struct record_opts *)opt->value;
977
978 record->callgraph_set = true;
72a128aa 979 callchain_param.enabled = true;
09b0fd45 980
72a128aa
NK
981 if (callchain_param.record_mode == CALLCHAIN_NONE)
982 callchain_param.record_mode = CALLCHAIN_FP;
eb853e80 983
72a128aa 984 callchain_debug();
09b0fd45
JO
985 return 0;
986}
987
eb853e80
JO
988static int perf_record_config(const char *var, const char *value, void *cb)
989{
7a29c087
NK
990 struct record *rec = cb;
991
992 if (!strcmp(var, "record.build-id")) {
993 if (!strcmp(value, "cache"))
994 rec->no_buildid_cache = false;
995 else if (!strcmp(value, "no-cache"))
996 rec->no_buildid_cache = true;
997 else if (!strcmp(value, "skip"))
998 rec->no_buildid = true;
999 else
1000 return -1;
1001 return 0;
1002 }
eb853e80 1003 if (!strcmp(var, "record.call-graph"))
5a2e5e85 1004 var = "call-graph.record-mode"; /* fall-through */
eb853e80
JO
1005
1006 return perf_default_config(var, value, cb);
1007}
1008
814c8c38
PZ
1009struct clockid_map {
1010 const char *name;
1011 int clockid;
1012};
1013
1014#define CLOCKID_MAP(n, c) \
1015 { .name = n, .clockid = (c), }
1016
1017#define CLOCKID_END { .name = NULL, }
1018
1019
1020/*
1021 * Add the missing ones, we need to build on many distros...
1022 */
1023#ifndef CLOCK_MONOTONIC_RAW
1024#define CLOCK_MONOTONIC_RAW 4
1025#endif
1026#ifndef CLOCK_BOOTTIME
1027#define CLOCK_BOOTTIME 7
1028#endif
1029#ifndef CLOCK_TAI
1030#define CLOCK_TAI 11
1031#endif
1032
1033static const struct clockid_map clockids[] = {
1034 /* available for all events, NMI safe */
1035 CLOCKID_MAP("monotonic", CLOCK_MONOTONIC),
1036 CLOCKID_MAP("monotonic_raw", CLOCK_MONOTONIC_RAW),
1037
1038 /* available for some events */
1039 CLOCKID_MAP("realtime", CLOCK_REALTIME),
1040 CLOCKID_MAP("boottime", CLOCK_BOOTTIME),
1041 CLOCKID_MAP("tai", CLOCK_TAI),
1042
1043 /* available for the lazy */
1044 CLOCKID_MAP("mono", CLOCK_MONOTONIC),
1045 CLOCKID_MAP("raw", CLOCK_MONOTONIC_RAW),
1046 CLOCKID_MAP("real", CLOCK_REALTIME),
1047 CLOCKID_MAP("boot", CLOCK_BOOTTIME),
1048
1049 CLOCKID_END,
1050};
1051
1052static int parse_clockid(const struct option *opt, const char *str, int unset)
1053{
1054 struct record_opts *opts = (struct record_opts *)opt->value;
1055 const struct clockid_map *cm;
1056 const char *ostr = str;
1057
1058 if (unset) {
1059 opts->use_clockid = 0;
1060 return 0;
1061 }
1062
1063 /* no arg passed */
1064 if (!str)
1065 return 0;
1066
1067 /* no setting it twice */
1068 if (opts->use_clockid)
1069 return -1;
1070
1071 opts->use_clockid = true;
1072
1073 /* if its a number, we're done */
1074 if (sscanf(str, "%d", &opts->clockid) == 1)
1075 return 0;
1076
1077 /* allow a "CLOCK_" prefix to the name */
1078 if (!strncasecmp(str, "CLOCK_", 6))
1079 str += 6;
1080
1081 for (cm = clockids; cm->name; cm++) {
1082 if (!strcasecmp(str, cm->name)) {
1083 opts->clockid = cm->clockid;
1084 return 0;
1085 }
1086 }
1087
1088 opts->use_clockid = false;
1089 ui__warning("unknown clockid %s, check man page\n", ostr);
1090 return -1;
1091}
1092
e9db1310
AH
1093static int record__parse_mmap_pages(const struct option *opt,
1094 const char *str,
1095 int unset __maybe_unused)
1096{
1097 struct record_opts *opts = opt->value;
1098 char *s, *p;
1099 unsigned int mmap_pages;
1100 int ret;
1101
1102 if (!str)
1103 return -EINVAL;
1104
1105 s = strdup(str);
1106 if (!s)
1107 return -ENOMEM;
1108
1109 p = strchr(s, ',');
1110 if (p)
1111 *p = '\0';
1112
1113 if (*s) {
1114 ret = __perf_evlist__parse_mmap_pages(&mmap_pages, s);
1115 if (ret)
1116 goto out_free;
1117 opts->mmap_pages = mmap_pages;
1118 }
1119
1120 if (!p) {
1121 ret = 0;
1122 goto out_free;
1123 }
1124
1125 ret = __perf_evlist__parse_mmap_pages(&mmap_pages, p + 1);
1126 if (ret)
1127 goto out_free;
1128
1129 opts->auxtrace_mmap_pages = mmap_pages;
1130
1131out_free:
1132 free(s);
1133 return ret;
1134}
1135
e5b2c207 1136static const char * const __record_usage[] = {
9e096753
MG
1137 "perf record [<options>] [<command>]",
1138 "perf record [<options>] -- <command> [<options>]",
0e9b20b8
IM
1139 NULL
1140};
e5b2c207 1141const char * const *record_usage = __record_usage;
0e9b20b8 1142
d20deb64 1143/*
8c6f45a7
ACM
1144 * XXX Ideally would be local to cmd_record() and passed to a record__new
1145 * because we need to have access to it in record__exit, that is called
d20deb64
ACM
1146 * after cmd_record() exits, but since record_options need to be accessible to
1147 * builtin-script, leave it here.
1148 *
1149 * At least we don't ouch it in all the other functions here directly.
1150 *
1151 * Just say no to tons of global variables, sigh.
1152 */
8c6f45a7 1153static struct record record = {
d20deb64 1154 .opts = {
8affc2b8 1155 .sample_time = true,
d20deb64
ACM
1156 .mmap_pages = UINT_MAX,
1157 .user_freq = UINT_MAX,
1158 .user_interval = ULLONG_MAX,
447a6013 1159 .freq = 4000,
d1cb9fce
NK
1160 .target = {
1161 .uses_mmap = true,
3aa5939d 1162 .default_per_cpu = true,
d1cb9fce 1163 },
9d9cad76 1164 .proc_map_timeout = 500,
d20deb64 1165 },
e3d59112
NK
1166 .tool = {
1167 .sample = process_sample_event,
1168 .fork = perf_event__process_fork,
cca8482c 1169 .exit = perf_event__process_exit,
e3d59112
NK
1170 .comm = perf_event__process_comm,
1171 .mmap = perf_event__process_mmap,
1172 .mmap2 = perf_event__process_mmap2,
cca8482c 1173 .ordered_events = true,
e3d59112 1174 },
d20deb64 1175};
7865e817 1176
76a26549
NK
1177const char record_callchain_help[] = CALLCHAIN_RECORD_HELP
1178 "\n\t\t\t\tDefault: fp";
61eaa3be 1179
d20deb64
ACM
1180/*
1181 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
1182 * with it and switch to use the library functions in perf_evlist that came
b4006796 1183 * from builtin-record.c, i.e. use record_opts,
d20deb64
ACM
1184 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
1185 * using pipes, etc.
1186 */
e5b2c207 1187struct option __record_options[] = {
d20deb64 1188 OPT_CALLBACK('e', "event", &record.evlist, "event",
86847b62 1189 "event selector. use 'perf list' to list available events",
f120f9d5 1190 parse_events_option),
d20deb64 1191 OPT_CALLBACK(0, "filter", &record.evlist, "filter",
c171b552 1192 "event filter", parse_filter),
4ba1faa1
WN
1193 OPT_CALLBACK_NOOPT(0, "exclude-perf", &record.evlist,
1194 NULL, "don't record events from perf itself",
1195 exclude_perf),
bea03405 1196 OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
d6d901c2 1197 "record events on existing process id"),
bea03405 1198 OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
d6d901c2 1199 "record events on existing thread id"),
d20deb64 1200 OPT_INTEGER('r', "realtime", &record.realtime_prio,
0e9b20b8 1201 "collect data with this RT SCHED_FIFO priority"),
509051ea 1202 OPT_BOOLEAN(0, "no-buffering", &record.opts.no_buffering,
acac03fa 1203 "collect data without buffering"),
d20deb64 1204 OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
daac07b2 1205 "collect raw sample records from all opened counters"),
bea03405 1206 OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
0e9b20b8 1207 "system-wide collection from all CPUs"),
bea03405 1208 OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
c45c6ea2 1209 "list of cpus to monitor"),
d20deb64 1210 OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
f5fc1412 1211 OPT_STRING('o', "output", &record.file.path, "file",
abaff32a 1212 "output file name"),
69e7e5b0
AH
1213 OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
1214 &record.opts.no_inherit_set,
1215 "child tasks do not inherit counters"),
d20deb64 1216 OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
e9db1310
AH
1217 OPT_CALLBACK('m', "mmap-pages", &record.opts, "pages[,pages]",
1218 "number of mmap data pages and AUX area tracing mmap pages",
1219 record__parse_mmap_pages),
d20deb64 1220 OPT_BOOLEAN(0, "group", &record.opts.group,
43bece79 1221 "put the counters into a counter group"),
09b0fd45
JO
1222 OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
1223 NULL, "enables call-graph recording" ,
1224 &record_callchain_opt),
1225 OPT_CALLBACK(0, "call-graph", &record.opts,
76a26549 1226 "record_mode[,record_size]", record_callchain_help,
09b0fd45 1227 &record_parse_callchain_opt),
c0555642 1228 OPT_INCR('v', "verbose", &verbose,
3da297a6 1229 "be more verbose (show counter open errors, etc)"),
b44308f5 1230 OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
d20deb64 1231 OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
649c48a9 1232 "per thread counts"),
56100321 1233 OPT_BOOLEAN('d', "data", &record.opts.sample_address, "Record the sample addresses"),
3abebc55
AH
1234 OPT_BOOLEAN_SET('T', "timestamp", &record.opts.sample_time,
1235 &record.opts.sample_time_set,
1236 "Record the sample timestamps"),
56100321 1237 OPT_BOOLEAN('P', "period", &record.opts.period, "Record the sample period"),
d20deb64 1238 OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
649c48a9 1239 "don't sample"),
d2db9a98
WN
1240 OPT_BOOLEAN_SET('N', "no-buildid-cache", &record.no_buildid_cache,
1241 &record.no_buildid_cache_set,
1242 "do not update the buildid cache"),
1243 OPT_BOOLEAN_SET('B', "no-buildid", &record.no_buildid,
1244 &record.no_buildid_set,
1245 "do not collect buildids in perf.data"),
d20deb64 1246 OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
023695d9
SE
1247 "monitor event in cgroup name only",
1248 parse_cgroups),
a6205a35 1249 OPT_UINTEGER('D', "delay", &record.opts.initial_delay,
6619a53e 1250 "ms to wait before starting measurement after program start"),
bea03405
NK
1251 OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
1252 "user to profile"),
a5aabdac
SE
1253
1254 OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
1255 "branch any", "sample any taken branches",
1256 parse_branch_stack),
1257
1258 OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
1259 "branch filter mask", "branch stack filter modes",
bdfebd84 1260 parse_branch_stack),
05484298
AK
1261 OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
1262 "sample by weight (on special events only)"),
475eeab9
AK
1263 OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
1264 "sample transaction flags (special events only)"),
3aa5939d
AH
1265 OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
1266 "use per-thread mmaps"),
bcc84ec6
SE
1267 OPT_CALLBACK_OPTARG('I', "intr-regs", &record.opts.sample_intr_regs, NULL, "any register",
1268 "sample selected machine registers on interrupt,"
1269 " use -I ? to list register names", parse_regs),
85c273d2
AK
1270 OPT_BOOLEAN(0, "running-time", &record.opts.running_time,
1271 "Record running/enabled time of read (:S) events"),
814c8c38
PZ
1272 OPT_CALLBACK('k', "clockid", &record.opts,
1273 "clockid", "clockid to use for events, see clock_gettime()",
1274 parse_clockid),
2dd6d8a1
AH
1275 OPT_STRING_OPTARG('S', "snapshot", &record.opts.auxtrace_snapshot_opts,
1276 "opts", "AUX area tracing Snapshot Mode", ""),
9d9cad76
KL
1277 OPT_UINTEGER(0, "proc-map-timeout", &record.opts.proc_map_timeout,
1278 "per thread proc mmap processing timeout in ms"),
b757bb09
AH
1279 OPT_BOOLEAN(0, "switch-events", &record.opts.record_switch_events,
1280 "Record context switch events"),
85723885
JO
1281 OPT_BOOLEAN_FLAG(0, "all-kernel", &record.opts.all_kernel,
1282 "Configure all used events to run in kernel space.",
1283 PARSE_OPT_EXCLUSIVE),
1284 OPT_BOOLEAN_FLAG(0, "all-user", &record.opts.all_user,
1285 "Configure all used events to run in user space.",
1286 PARSE_OPT_EXCLUSIVE),
71dc2326
WN
1287 OPT_STRING(0, "clang-path", &llvm_param.clang_path, "clang path",
1288 "clang binary to use for compiling BPF scriptlets"),
1289 OPT_STRING(0, "clang-opt", &llvm_param.clang_opt, "clang options",
1290 "options passed to clang when compiling BPF scriptlets"),
7efe0e03
HK
1291 OPT_STRING(0, "vmlinux", &symbol_conf.vmlinux_name,
1292 "file", "vmlinux pathname"),
6156681b
NK
1293 OPT_BOOLEAN(0, "buildid-all", &record.buildid_all,
1294 "Record build-id of all DSOs regardless of hits"),
ecfd7a9c
WN
1295 OPT_BOOLEAN(0, "timestamp-filename", &record.timestamp_filename,
1296 "append timestamp to output filename"),
0e9b20b8
IM
1297 OPT_END()
1298};
1299
e5b2c207
NK
1300struct option *record_options = __record_options;
1301
1d037ca1 1302int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
0e9b20b8 1303{
ef149c25 1304 int err;
8c6f45a7 1305 struct record *rec = &record;
16ad2ffb 1306 char errbuf[BUFSIZ];
0e9b20b8 1307
48e1cab1
WN
1308#ifndef HAVE_LIBBPF_SUPPORT
1309# define set_nobuild(s, l, c) set_option_nobuild(record_options, s, l, "NO_LIBBPF=1", c)
1310 set_nobuild('\0', "clang-path", true);
1311 set_nobuild('\0', "clang-opt", true);
1312# undef set_nobuild
7efe0e03
HK
1313#endif
1314
1315#ifndef HAVE_BPF_PROLOGUE
1316# if !defined (HAVE_DWARF_SUPPORT)
1317# define REASON "NO_DWARF=1"
1318# elif !defined (HAVE_LIBBPF_SUPPORT)
1319# define REASON "NO_LIBBPF=1"
1320# else
1321# define REASON "this architecture doesn't support BPF prologue"
1322# endif
1323# define set_nobuild(s, l, c) set_option_nobuild(record_options, s, l, REASON, c)
1324 set_nobuild('\0', "vmlinux", true);
1325# undef set_nobuild
1326# undef REASON
48e1cab1
WN
1327#endif
1328
3e2be2da
ACM
1329 rec->evlist = perf_evlist__new();
1330 if (rec->evlist == NULL)
361c99a6
ACM
1331 return -ENOMEM;
1332
eb853e80
JO
1333 perf_config(perf_record_config, rec);
1334
bca647aa 1335 argc = parse_options(argc, argv, record_options, record_usage,
655000e7 1336 PARSE_OPT_STOP_AT_NON_OPTION);
602ad878 1337 if (!argc && target__none(&rec->opts.target))
bca647aa 1338 usage_with_options(record_usage, record_options);
0e9b20b8 1339
bea03405 1340 if (nr_cgroups && !rec->opts.target.system_wide) {
c7118369
NK
1341 usage_with_options_msg(record_usage, record_options,
1342 "cgroup monitoring only available in system-wide mode");
1343
023695d9 1344 }
b757bb09
AH
1345 if (rec->opts.record_switch_events &&
1346 !perf_can_record_switch_events()) {
c7118369
NK
1347 ui__error("kernel does not support recording context switch events\n");
1348 parse_options_usage(record_usage, record_options, "switch-events", 0);
1349 return -EINVAL;
b757bb09 1350 }
023695d9 1351
ef149c25
AH
1352 if (!rec->itr) {
1353 rec->itr = auxtrace_record__init(rec->evlist, &err);
1354 if (err)
1355 return err;
1356 }
1357
2dd6d8a1
AH
1358 err = auxtrace_parse_snapshot_options(rec->itr, &rec->opts,
1359 rec->opts.auxtrace_snapshot_opts);
1360 if (err)
1361 return err;
1362
d7888573
WN
1363 err = bpf__setup_stdout(rec->evlist);
1364 if (err) {
1365 bpf__strerror_setup_stdout(rec->evlist, err, errbuf, sizeof(errbuf));
1366 pr_err("ERROR: Setup BPF stdout failed: %s\n",
1367 errbuf);
1368 return err;
1369 }
1370
ef149c25
AH
1371 err = -ENOMEM;
1372
0a7e6d1b 1373 symbol__init(NULL);
baa2f6ce 1374
ec80fde7 1375 if (symbol_conf.kptr_restrict)
646aaea6
ACM
1376 pr_warning(
1377"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
1378"check /proc/sys/kernel/kptr_restrict.\n\n"
1379"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
1380"file is not found in the buildid cache or in the vmlinux path.\n\n"
1381"Samples in kernel modules won't be resolved at all.\n\n"
1382"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
1383"even with a suitable vmlinux or kallsyms file.\n\n");
ec80fde7 1384
d20deb64 1385 if (rec->no_buildid_cache || rec->no_buildid)
a1ac1d3c 1386 disable_buildid_cache();
655000e7 1387
3e2be2da
ACM
1388 if (rec->evlist->nr_entries == 0 &&
1389 perf_evlist__add_default(rec->evlist) < 0) {
69aad6f1
ACM
1390 pr_err("Not enough memory for event selector list\n");
1391 goto out_symbol_exit;
bbd36e5e 1392 }
0e9b20b8 1393
69e7e5b0
AH
1394 if (rec->opts.target.tid && !rec->opts.no_inherit_set)
1395 rec->opts.no_inherit = true;
1396
602ad878 1397 err = target__validate(&rec->opts.target);
16ad2ffb 1398 if (err) {
602ad878 1399 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
16ad2ffb
NK
1400 ui__warning("%s", errbuf);
1401 }
1402
602ad878 1403 err = target__parse_uid(&rec->opts.target);
16ad2ffb
NK
1404 if (err) {
1405 int saved_errno = errno;
4bd0f2d2 1406
602ad878 1407 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
3780f488 1408 ui__error("%s", errbuf);
16ad2ffb
NK
1409
1410 err = -saved_errno;
8fa60e1f 1411 goto out_symbol_exit;
16ad2ffb 1412 }
0d37aa34 1413
16ad2ffb 1414 err = -ENOMEM;
3e2be2da 1415 if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
dd7927f4 1416 usage_with_options(record_usage, record_options);
69aad6f1 1417
ef149c25
AH
1418 err = auxtrace_record__options(rec->itr, rec->evlist, &rec->opts);
1419 if (err)
1420 goto out_symbol_exit;
1421
6156681b
NK
1422 /*
1423 * We take all buildids when the file contains
1424 * AUX area tracing data because we do not decode the
1425 * trace because it would take too long.
1426 */
1427 if (rec->opts.full_auxtrace)
1428 rec->buildid_all = true;
1429
b4006796 1430 if (record_opts__config(&rec->opts)) {
39d17dac 1431 err = -EINVAL;
03ad9747 1432 goto out_symbol_exit;
7e4ff9e3
MG
1433 }
1434
d20deb64 1435 err = __cmd_record(&record, argc, argv);
d65a458b 1436out_symbol_exit:
45604710 1437 perf_evlist__delete(rec->evlist);
d65a458b 1438 symbol__exit();
ef149c25 1439 auxtrace_record__free(rec->itr);
39d17dac 1440 return err;
0e9b20b8 1441}
2dd6d8a1
AH
1442
1443static void snapshot_sig_handler(int sig __maybe_unused)
1444{
c0bdc1c4 1445 if (!auxtrace_snapshot_is_enabled())
2dd6d8a1 1446 return;
c0bdc1c4 1447 auxtrace_snapshot_disable();
2dd6d8a1
AH
1448 auxtrace_snapshot_err = auxtrace_record__snapshot_start(record.itr);
1449 auxtrace_record__snapshot_started = 1;
1450}