perf tools: Put proper period for for samples without PERIOD sample_type
[linux-2.6-block.git] / tools / perf / builtin-record.c
CommitLineData
abaff32a 1/*
bf9e1876
IM
2 * builtin-record.c
3 *
4 * Builtin record command: Record the profile of a workload
5 * (or a CPU, or a PID) into the perf.data output file - for
6 * later analysis via perf report.
abaff32a 7 */
16f762a2 8#include "builtin.h"
bf9e1876
IM
9
10#include "perf.h"
11
6122e4e4 12#include "util/build-id.h"
6eda5838 13#include "util/util.h"
0e9b20b8 14#include "util/parse-options.h"
8ad8db37 15#include "util/parse-events.h"
6eda5838 16
7c6a1c65 17#include "util/header.h"
66e274f3 18#include "util/event.h"
361c99a6 19#include "util/evlist.h"
69aad6f1 20#include "util/evsel.h"
8f28827a 21#include "util/debug.h"
94c744b6 22#include "util/session.h"
45694aa7 23#include "util/tool.h"
8d06367f 24#include "util/symbol.h"
a12b51c4 25#include "util/cpumap.h"
fd78260b 26#include "util/thread_map.h"
f5fc1412 27#include "util/data.h"
7c6a1c65 28
97124d5e 29#include <unistd.h>
de9ac07b 30#include <sched.h>
a41794cd 31#include <sys/mman.h>
de9ac07b 32
89fe808a 33#ifndef HAVE_ON_EXIT_SUPPORT
78da39fa
BR
34#ifndef ATEXIT_MAX
35#define ATEXIT_MAX 32
36#endif
37static int __on_exit_count = 0;
38typedef void (*on_exit_func_t) (int, void *);
39static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
40static void *__on_exit_args[ATEXIT_MAX];
41static int __exitcode = 0;
42static void __handle_on_exit_funcs(void);
43static int on_exit(on_exit_func_t function, void *arg);
44#define exit(x) (exit)(__exitcode = (x))
45
46static int on_exit(on_exit_func_t function, void *arg)
47{
48 if (__on_exit_count == ATEXIT_MAX)
49 return -ENOMEM;
50 else if (__on_exit_count == 0)
51 atexit(__handle_on_exit_funcs);
52 __on_exit_funcs[__on_exit_count] = function;
53 __on_exit_args[__on_exit_count++] = arg;
54 return 0;
55}
56
57static void __handle_on_exit_funcs(void)
58{
59 int i;
60 for (i = 0; i < __on_exit_count; i++)
61 __on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
62}
63#endif
64
8c6f45a7 65struct record {
45694aa7 66 struct perf_tool tool;
b4006796 67 struct record_opts opts;
d20deb64 68 u64 bytes_written;
f5fc1412 69 struct perf_data_file file;
d20deb64
ACM
70 struct perf_evlist *evlist;
71 struct perf_session *session;
72 const char *progname;
d20deb64 73 int realtime_prio;
d20deb64
ACM
74 bool no_buildid;
75 bool no_buildid_cache;
d20deb64 76 long samples;
0f82ebc4 77};
a21ca2ca 78
8c6f45a7 79static int record__write(struct record *rec, void *bf, size_t size)
f5970550 80{
cf8b2e69 81 if (perf_data_file__write(rec->session->file, bf, size) < 0) {
50a9b868
JO
82 pr_err("failed to write perf data, error: %m\n");
83 return -1;
f5970550 84 }
8d3eca20 85
cf8b2e69 86 rec->bytes_written += size;
8d3eca20 87 return 0;
f5970550
PZ
88}
89
45694aa7 90static int process_synthesized_event(struct perf_tool *tool,
d20deb64 91 union perf_event *event,
1d037ca1
IT
92 struct perf_sample *sample __maybe_unused,
93 struct machine *machine __maybe_unused)
234fbbf5 94{
8c6f45a7
ACM
95 struct record *rec = container_of(tool, struct record, tool);
96 return record__write(rec, event, event->header.size);
234fbbf5
ACM
97}
98
8c6f45a7 99static int record__mmap_read(struct record *rec, struct perf_mmap *md)
de9ac07b 100{
744bd8aa 101 unsigned int head = perf_mmap__read_head(md);
de9ac07b 102 unsigned int old = md->prev;
918512b4 103 unsigned char *data = md->base + page_size;
de9ac07b
PZ
104 unsigned long size;
105 void *buf;
8d3eca20 106 int rc = 0;
de9ac07b 107
dc82009a 108 if (old == head)
8d3eca20 109 return 0;
dc82009a 110
d20deb64 111 rec->samples++;
de9ac07b
PZ
112
113 size = head - old;
114
115 if ((old & md->mask) + size != (head & md->mask)) {
116 buf = &data[old & md->mask];
117 size = md->mask + 1 - (old & md->mask);
118 old += size;
021e9f47 119
8c6f45a7 120 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
121 rc = -1;
122 goto out;
123 }
de9ac07b
PZ
124 }
125
126 buf = &data[old & md->mask];
127 size = head - old;
128 old += size;
021e9f47 129
8c6f45a7 130 if (record__write(rec, buf, size) < 0) {
8d3eca20
DA
131 rc = -1;
132 goto out;
133 }
de9ac07b
PZ
134
135 md->prev = old;
115d2d89 136 perf_mmap__write_tail(md, old);
8d3eca20
DA
137
138out:
139 return rc;
de9ac07b
PZ
140}
141
142static volatile int done = 0;
f7b7c26e 143static volatile int signr = -1;
33e49ea7 144static volatile int child_finished = 0;
de9ac07b 145
16c8a109 146static void sig_handler(int sig)
de9ac07b 147{
33e49ea7
AK
148 if (sig == SIGCHLD)
149 child_finished = 1;
150
16c8a109 151 done = 1;
f7b7c26e
PZ
152 signr = sig;
153}
154
8c6f45a7 155static void record__sig_exit(int exit_status __maybe_unused, void *arg)
f7b7c26e 156{
8c6f45a7 157 struct record *rec = arg;
33e49ea7
AK
158 int status;
159
d20deb64 160 if (rec->evlist->workload.pid > 0) {
33e49ea7 161 if (!child_finished)
d20deb64 162 kill(rec->evlist->workload.pid, SIGTERM);
33e49ea7
AK
163
164 wait(&status);
165 if (WIFSIGNALED(status))
d20deb64 166 psignal(WTERMSIG(status), rec->progname);
33e49ea7 167 }
933da83a 168
18483b81 169 if (signr == -1 || signr == SIGUSR1)
f7b7c26e
PZ
170 return;
171
172 signal(signr, SIG_DFL);
de9ac07b
PZ
173}
174
8c6f45a7 175static int record__open(struct record *rec)
dd7927f4 176{
56e52e85 177 char msg[512];
6a4bb04c 178 struct perf_evsel *pos;
d20deb64
ACM
179 struct perf_evlist *evlist = rec->evlist;
180 struct perf_session *session = rec->session;
b4006796 181 struct record_opts *opts = &rec->opts;
8d3eca20 182 int rc = 0;
dd7927f4 183
f77a9518 184 perf_evlist__config(evlist, opts);
cac21425 185
0050f7aa 186 evlist__for_each(evlist, pos) {
dd7927f4 187try_again:
6a4bb04c 188 if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
56e52e85 189 if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
d6d901c2 190 if (verbose)
c0a54341 191 ui__warning("%s\n", msg);
d6d901c2
ZY
192 goto try_again;
193 }
ca6a4258 194
56e52e85
ACM
195 rc = -errno;
196 perf_evsel__open_strerror(pos, &opts->target,
197 errno, msg, sizeof(msg));
198 ui__error("%s\n", msg);
8d3eca20 199 goto out;
c171b552
LZ
200 }
201 }
a43d3f08 202
1491a632 203 if (perf_evlist__apply_filters(evlist)) {
0a102479
FW
204 error("failed to set filter with %d (%s)\n", errno,
205 strerror(errno));
8d3eca20
DA
206 rc = -1;
207 goto out;
0a102479
FW
208 }
209
18e60939 210 if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
8d3eca20
DA
211 if (errno == EPERM) {
212 pr_err("Permission error mapping pages.\n"
213 "Consider increasing "
214 "/proc/sys/kernel/perf_event_mlock_kb,\n"
215 "or try again with a smaller value of -m/--mmap_pages.\n"
53653d70 216 "(current value: %u)\n", opts->mmap_pages);
8d3eca20 217 rc = -errno;
8d3eca20
DA
218 } else {
219 pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
220 rc = -errno;
221 }
222 goto out;
18e60939 223 }
0a27d7f9 224
563aecb2 225 session->evlist = evlist;
7b56cce2 226 perf_session__set_id_hdr_size(session);
8d3eca20
DA
227out:
228 return rc;
16c8a109
PZ
229}
230
8c6f45a7 231static int process_buildids(struct record *rec)
6122e4e4 232{
f5fc1412
JO
233 struct perf_data_file *file = &rec->file;
234 struct perf_session *session = rec->session;
7ab75cff 235 u64 start = session->header.data_offset;
6122e4e4 236
f5fc1412 237 u64 size = lseek(file->fd, 0, SEEK_CUR);
9f591fd7
ACM
238 if (size == 0)
239 return 0;
240
7ab75cff
DA
241 return __perf_session__process_events(session, start,
242 size - start,
6122e4e4
ACM
243 size, &build_id__mark_dso_hit_ops);
244}
245
8c6f45a7 246static void record__exit(int status, void *arg)
f5970550 247{
8c6f45a7 248 struct record *rec = arg;
f5fc1412 249 struct perf_data_file *file = &rec->file;
d20deb64 250
8d3eca20
DA
251 if (status != 0)
252 return;
253
f5fc1412 254 if (!file->is_pipe) {
d20deb64
ACM
255 rec->session->header.data_size += rec->bytes_written;
256
257 if (!rec->no_buildid)
258 process_buildids(rec);
259 perf_session__write_header(rec->session, rec->evlist,
f5fc1412 260 file->fd, true);
d20deb64
ACM
261 perf_session__delete(rec->session);
262 perf_evlist__delete(rec->evlist);
d65a458b 263 symbol__exit();
c7929e47 264 }
f5970550
PZ
265}
266
8115d60c 267static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
a1645ce1
ZY
268{
269 int err;
45694aa7 270 struct perf_tool *tool = data;
a1645ce1
ZY
271 /*
272 *As for guest kernel when processing subcommand record&report,
273 *we arrange module mmap prior to guest kernel mmap and trigger
274 *a preload dso because default guest module symbols are loaded
275 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
276 *method is used to avoid symbol missing when the first addr is
277 *in module instead of in guest kernel.
278 */
45694aa7 279 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 280 machine);
a1645ce1
ZY
281 if (err < 0)
282 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 283 " relocation symbol.\n", machine->pid);
a1645ce1 284
a1645ce1
ZY
285 /*
286 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
287 * have no _text sometimes.
288 */
45694aa7 289 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
0ae617be 290 machine);
a1645ce1
ZY
291 if (err < 0)
292 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 293 " relocation symbol.\n", machine->pid);
a1645ce1
ZY
294}
295
98402807
FW
296static struct perf_event_header finished_round_event = {
297 .size = sizeof(struct perf_event_header),
298 .type = PERF_RECORD_FINISHED_ROUND,
299};
300
8c6f45a7 301static int record__mmap_read_all(struct record *rec)
98402807 302{
0e2e63dd 303 int i;
8d3eca20 304 int rc = 0;
98402807 305
d20deb64 306 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
8d3eca20 307 if (rec->evlist->mmap[i].base) {
8c6f45a7 308 if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
8d3eca20
DA
309 rc = -1;
310 goto out;
311 }
312 }
98402807
FW
313 }
314
2eeaaa09 315 if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
8c6f45a7 316 rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
8d3eca20
DA
317
318out:
319 return rc;
98402807
FW
320}
321
8c6f45a7 322static void record__init_features(struct record *rec)
57706abc 323{
57706abc
DA
324 struct perf_session *session = rec->session;
325 int feat;
326
327 for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
328 perf_header__set_feat(&session->header, feat);
329
330 if (rec->no_buildid)
331 perf_header__clear_feat(&session->header, HEADER_BUILD_ID);
332
3e2be2da 333 if (!have_tracepoints(&rec->evlist->entries))
57706abc
DA
334 perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
335
336 if (!rec->opts.branch_stack)
337 perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
338}
339
f33cbe72
ACM
340static volatile int workload_exec_errno;
341
342/*
343 * perf_evlist__prepare_workload will send a SIGUSR1
344 * if the fork fails, since we asked by setting its
345 * want_signal to true.
346 */
347static void workload_exec_failed_signal(int signo, siginfo_t *info,
348 void *ucontext __maybe_unused)
349{
350 workload_exec_errno = info->si_value.sival_int;
351 done = 1;
352 signr = signo;
353 child_finished = 1;
354}
355
8c6f45a7 356static int __cmd_record(struct record *rec, int argc, const char **argv)
16c8a109 357{
57706abc 358 int err;
8b412664 359 unsigned long waking = 0;
46be604b 360 const bool forks = argc > 0;
23346f21 361 struct machine *machine;
45694aa7 362 struct perf_tool *tool = &rec->tool;
b4006796 363 struct record_opts *opts = &rec->opts;
f5fc1412 364 struct perf_data_file *file = &rec->file;
d20deb64 365 struct perf_session *session;
2711926a 366 bool disabled = false;
de9ac07b 367
d20deb64 368 rec->progname = argv[0];
33e49ea7 369
8c6f45a7 370 on_exit(record__sig_exit, rec);
f5970550
PZ
371 signal(SIGCHLD, sig_handler);
372 signal(SIGINT, sig_handler);
804f7ac7 373 signal(SIGTERM, sig_handler);
f5970550 374
f5fc1412 375 session = perf_session__new(file, false, NULL);
94c744b6 376 if (session == NULL) {
a9a70bbc
ACM
377 pr_err("Not enough memory for reading perf file header\n");
378 return -1;
379 }
380
d20deb64
ACM
381 rec->session = session;
382
8c6f45a7 383 record__init_features(rec);
330aa675 384
d4db3f16 385 if (forks) {
3e2be2da 386 err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
f5fc1412 387 argv, file->is_pipe,
735f7e0b 388 workload_exec_failed_signal);
35b9d88e
ACM
389 if (err < 0) {
390 pr_err("Couldn't run the workload!\n");
391 goto out_delete_session;
856e9660 392 }
856e9660
PZ
393 }
394
8c6f45a7 395 if (record__open(rec) != 0) {
8d3eca20
DA
396 err = -1;
397 goto out_delete_session;
398 }
de9ac07b 399
3e2be2da 400 if (!rec->evlist->nr_groups)
a8bb559b
NK
401 perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);
402
712a4b60 403 /*
8c6f45a7 404 * perf_session__delete(session) will be called at record__exit()
712a4b60 405 */
8c6f45a7 406 on_exit(record__exit, rec);
712a4b60 407
f5fc1412
JO
408 if (file->is_pipe) {
409 err = perf_header__write_pipe(file->fd);
529870e3 410 if (err < 0)
8d3eca20 411 goto out_delete_session;
563aecb2 412 } else {
3e2be2da 413 err = perf_session__write_header(session, rec->evlist,
f5fc1412 414 file->fd, false);
d5eed904 415 if (err < 0)
8d3eca20 416 goto out_delete_session;
56b03f3c
ACM
417 }
418
d3665498 419 if (!rec->no_buildid
e20960c0 420 && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
d3665498 421 pr_err("Couldn't generate buildids. "
e20960c0 422 "Use --no-buildid to profile anyway.\n");
8d3eca20
DA
423 err = -1;
424 goto out_delete_session;
e20960c0
RR
425 }
426
34ba5122 427 machine = &session->machines.host;
743eb868 428
f5fc1412 429 if (file->is_pipe) {
45694aa7 430 err = perf_event__synthesize_attrs(tool, session,
d20deb64 431 process_synthesized_event);
2c46dbb5
TZ
432 if (err < 0) {
433 pr_err("Couldn't synthesize attrs.\n");
8d3eca20 434 goto out_delete_session;
2c46dbb5 435 }
cd19a035 436
3e2be2da 437 if (have_tracepoints(&rec->evlist->entries)) {
63e0c771
TZ
438 /*
439 * FIXME err <= 0 here actually means that
440 * there were no tracepoints so its not really
441 * an error, just that we don't need to
442 * synthesize anything. We really have to
443 * return this more properly and also
444 * propagate errors that now are calling die()
445 */
3e2be2da 446 err = perf_event__synthesize_tracing_data(tool, file->fd, rec->evlist,
743eb868 447 process_synthesized_event);
63e0c771
TZ
448 if (err <= 0) {
449 pr_err("Couldn't record tracing data.\n");
8d3eca20 450 goto out_delete_session;
63e0c771 451 }
f34b9001 452 rec->bytes_written += err;
63e0c771 453 }
2c46dbb5
TZ
454 }
455
45694aa7 456 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
0ae617be 457 machine);
c1a3a4b9
ACM
458 if (err < 0)
459 pr_err("Couldn't record kernel reference relocation symbol\n"
460 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
461 "Check /proc/kallsyms permission or run as root.\n");
b7cece76 462
45694aa7 463 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 464 machine);
c1a3a4b9
ACM
465 if (err < 0)
466 pr_err("Couldn't record kernel module information.\n"
467 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
468 "Check /proc/modules permission or run as root.\n");
469
7e383de4 470 if (perf_guest) {
876650e6
ACM
471 machines__process_guests(&session->machines,
472 perf_event__synthesize_guest_os, tool);
7e383de4 473 }
7c6a1c65 474
3e2be2da 475 err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
a33fbd56 476 process_synthesized_event, opts->sample_address);
8d3eca20
DA
477 if (err != 0)
478 goto out_delete_session;
479
d20deb64 480 if (rec->realtime_prio) {
de9ac07b
PZ
481 struct sched_param param;
482
d20deb64 483 param.sched_priority = rec->realtime_prio;
de9ac07b 484 if (sched_setscheduler(0, SCHED_FIFO, &param)) {
6beba7ad 485 pr_err("Could not set realtime priority.\n");
8d3eca20
DA
486 err = -1;
487 goto out_delete_session;
de9ac07b
PZ
488 }
489 }
490
774cb499
JO
491 /*
492 * When perf is starting the traced process, all the events
493 * (apart from group members) have enable_on_exec=1 set,
494 * so don't spoil it by prematurely enabling them.
495 */
6619a53e 496 if (!target__none(&opts->target) && !opts->initial_delay)
3e2be2da 497 perf_evlist__enable(rec->evlist);
764e16a3 498
856e9660
PZ
499 /*
500 * Let the child rip
501 */
735f7e0b 502 if (forks)
3e2be2da 503 perf_evlist__start_workload(rec->evlist);
856e9660 504
6619a53e
AK
505 if (opts->initial_delay) {
506 usleep(opts->initial_delay * 1000);
507 perf_evlist__enable(rec->evlist);
508 }
509
649c48a9 510 for (;;) {
d20deb64 511 int hits = rec->samples;
de9ac07b 512
8c6f45a7 513 if (record__mmap_read_all(rec) < 0) {
8d3eca20
DA
514 err = -1;
515 goto out_delete_session;
516 }
de9ac07b 517
d20deb64 518 if (hits == rec->samples) {
649c48a9
PZ
519 if (done)
520 break;
3e2be2da 521 err = poll(rec->evlist->pollfd, rec->evlist->nr_fds, -1);
8b412664
PZ
522 waking++;
523 }
524
774cb499
JO
525 /*
526 * When perf is starting the traced process, at the end events
527 * die with the process and we wait for that. Thus no need to
528 * disable events in this case.
529 */
602ad878 530 if (done && !disabled && !target__none(&opts->target)) {
3e2be2da 531 perf_evlist__disable(rec->evlist);
2711926a
JO
532 disabled = true;
533 }
de9ac07b
PZ
534 }
535
f33cbe72
ACM
536 if (forks && workload_exec_errno) {
537 char msg[512];
538 const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
539 pr_err("Workload failed: %s\n", emsg);
540 err = -1;
541 goto out_delete_session;
542 }
543
18483b81 544 if (quiet || signr == SIGUSR1)
b44308f5
ACM
545 return 0;
546
8b412664
PZ
547 fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);
548
021e9f47
IM
549 /*
550 * Approximate RIP event size: 24 bytes.
551 */
552 fprintf(stderr,
9486aa38 553 "[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
d20deb64 554 (double)rec->bytes_written / 1024.0 / 1024.0,
6a4d98d7 555 file->path,
d20deb64 556 rec->bytes_written / 24);
addc2785 557
de9ac07b 558 return 0;
39d17dac
ACM
559
560out_delete_session:
561 perf_session__delete(session);
562 return err;
de9ac07b 563}
0e9b20b8 564
bdfebd84
RAV
565#define BRANCH_OPT(n, m) \
566 { .name = n, .mode = (m) }
567
568#define BRANCH_END { .name = NULL }
569
570struct branch_mode {
571 const char *name;
572 int mode;
573};
574
575static const struct branch_mode branch_modes[] = {
576 BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
577 BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
578 BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
579 BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
580 BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
581 BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
582 BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
0126d493
AK
583 BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
584 BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
585 BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
bdfebd84
RAV
586 BRANCH_END
587};
588
589static int
a5aabdac 590parse_branch_stack(const struct option *opt, const char *str, int unset)
bdfebd84
RAV
591{
592#define ONLY_PLM \
593 (PERF_SAMPLE_BRANCH_USER |\
594 PERF_SAMPLE_BRANCH_KERNEL |\
595 PERF_SAMPLE_BRANCH_HV)
596
597 uint64_t *mode = (uint64_t *)opt->value;
598 const struct branch_mode *br;
a5aabdac 599 char *s, *os = NULL, *p;
bdfebd84
RAV
600 int ret = -1;
601
a5aabdac
SE
602 if (unset)
603 return 0;
bdfebd84 604
a5aabdac
SE
605 /*
606 * cannot set it twice, -b + --branch-filter for instance
607 */
608 if (*mode)
bdfebd84
RAV
609 return -1;
610
a5aabdac
SE
611 /* str may be NULL in case no arg is passed to -b */
612 if (str) {
613 /* because str is read-only */
614 s = os = strdup(str);
615 if (!s)
616 return -1;
617
618 for (;;) {
619 p = strchr(s, ',');
620 if (p)
621 *p = '\0';
622
623 for (br = branch_modes; br->name; br++) {
624 if (!strcasecmp(s, br->name))
625 break;
626 }
627 if (!br->name) {
628 ui__warning("unknown branch filter %s,"
629 " check man page\n", s);
630 goto error;
631 }
bdfebd84 632
a5aabdac 633 *mode |= br->mode;
bdfebd84 634
a5aabdac
SE
635 if (!p)
636 break;
bdfebd84 637
a5aabdac
SE
638 s = p + 1;
639 }
bdfebd84
RAV
640 }
641 ret = 0;
642
a5aabdac 643 /* default to any branch */
bdfebd84 644 if ((*mode & ~ONLY_PLM) == 0) {
a5aabdac 645 *mode = PERF_SAMPLE_BRANCH_ANY;
bdfebd84
RAV
646 }
647error:
648 free(os);
649 return ret;
650}
651
89fe808a 652#ifdef HAVE_LIBUNWIND_SUPPORT
26d33022
JO
653static int get_stack_size(char *str, unsigned long *_size)
654{
655 char *endptr;
656 unsigned long size;
657 unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
658
659 size = strtoul(str, &endptr, 0);
660
661 do {
662 if (*endptr)
663 break;
664
665 size = round_up(size, sizeof(u64));
666 if (!size || size > max_size)
667 break;
668
669 *_size = size;
670 return 0;
671
672 } while (0);
673
674 pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
675 max_size, str);
676 return -1;
677}
89fe808a 678#endif /* HAVE_LIBUNWIND_SUPPORT */
26d33022 679
b4006796 680int record_parse_callchain(const char *arg, struct record_opts *opts)
26d33022 681{
26d33022
JO
682 char *tok, *name, *saveptr = NULL;
683 char *buf;
684 int ret = -1;
685
26d33022
JO
686 /* We need buffer that we know we can write to. */
687 buf = malloc(strlen(arg) + 1);
688 if (!buf)
689 return -ENOMEM;
690
691 strcpy(buf, arg);
692
693 tok = strtok_r((char *)buf, ",", &saveptr);
694 name = tok ? : (char *)buf;
695
696 do {
697 /* Framepointer style */
698 if (!strncmp(name, "fp", sizeof("fp"))) {
699 if (!strtok_r(NULL, ",", &saveptr)) {
c5ff78c3 700 opts->call_graph = CALLCHAIN_FP;
26d33022
JO
701 ret = 0;
702 } else
703 pr_err("callchain: No more arguments "
704 "needed for -g fp\n");
705 break;
706
89fe808a 707#ifdef HAVE_LIBUNWIND_SUPPORT
26d33022
JO
708 /* Dwarf style */
709 } else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
61eaa3be
ACM
710 const unsigned long default_stack_dump_size = 8192;
711
26d33022 712 ret = 0;
c5ff78c3
ACM
713 opts->call_graph = CALLCHAIN_DWARF;
714 opts->stack_dump_size = default_stack_dump_size;
26d33022
JO
715
716 tok = strtok_r(NULL, ",", &saveptr);
717 if (tok) {
718 unsigned long size = 0;
719
720 ret = get_stack_size(tok, &size);
c5ff78c3 721 opts->stack_dump_size = size;
26d33022 722 }
89fe808a 723#endif /* HAVE_LIBUNWIND_SUPPORT */
26d33022 724 } else {
09b0fd45 725 pr_err("callchain: Unknown --call-graph option "
26d33022
JO
726 "value: %s\n", arg);
727 break;
728 }
729
730 } while (0);
731
732 free(buf);
09b0fd45
JO
733 return ret;
734}
735
b4006796 736static void callchain_debug(struct record_opts *opts)
09b0fd45
JO
737{
738 pr_debug("callchain: type %d\n", opts->call_graph);
26d33022 739
09b0fd45
JO
740 if (opts->call_graph == CALLCHAIN_DWARF)
741 pr_debug("callchain: stack dump size %d\n",
742 opts->stack_dump_size);
743}
744
745int record_parse_callchain_opt(const struct option *opt,
746 const char *arg,
747 int unset)
748{
b4006796 749 struct record_opts *opts = opt->value;
09b0fd45
JO
750 int ret;
751
752 /* --no-call-graph */
753 if (unset) {
754 opts->call_graph = CALLCHAIN_NONE;
755 pr_debug("callchain: disabled\n");
756 return 0;
757 }
758
759 ret = record_parse_callchain(arg, opts);
26d33022 760 if (!ret)
09b0fd45 761 callchain_debug(opts);
26d33022
JO
762
763 return ret;
764}
765
09b0fd45
JO
766int record_callchain_opt(const struct option *opt,
767 const char *arg __maybe_unused,
768 int unset __maybe_unused)
769{
b4006796 770 struct record_opts *opts = opt->value;
09b0fd45
JO
771
772 if (opts->call_graph == CALLCHAIN_NONE)
773 opts->call_graph = CALLCHAIN_FP;
774
775 callchain_debug(opts);
776 return 0;
777}
778
0e9b20b8 779static const char * const record_usage[] = {
9e096753
MG
780 "perf record [<options>] [<command>]",
781 "perf record [<options>] -- <command> [<options>]",
0e9b20b8
IM
782 NULL
783};
784
d20deb64 785/*
8c6f45a7
ACM
786 * XXX Ideally would be local to cmd_record() and passed to a record__new
787 * because we need to have access to it in record__exit, that is called
d20deb64
ACM
788 * after cmd_record() exits, but since record_options need to be accessible to
789 * builtin-script, leave it here.
790 *
791 * At least we don't ouch it in all the other functions here directly.
792 *
793 * Just say no to tons of global variables, sigh.
794 */
8c6f45a7 795static struct record record = {
d20deb64 796 .opts = {
d20deb64
ACM
797 .mmap_pages = UINT_MAX,
798 .user_freq = UINT_MAX,
799 .user_interval = ULLONG_MAX,
447a6013 800 .freq = 4000,
d1cb9fce
NK
801 .target = {
802 .uses_mmap = true,
3aa5939d 803 .default_per_cpu = true,
d1cb9fce 804 },
d20deb64 805 },
d20deb64 806};
7865e817 807
09b0fd45 808#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
61eaa3be 809
89fe808a 810#ifdef HAVE_LIBUNWIND_SUPPORT
09b0fd45 811const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
61eaa3be 812#else
09b0fd45 813const char record_callchain_help[] = CALLCHAIN_HELP "fp";
61eaa3be
ACM
814#endif
815
d20deb64
ACM
816/*
817 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
818 * with it and switch to use the library functions in perf_evlist that came
b4006796 819 * from builtin-record.c, i.e. use record_opts,
d20deb64
ACM
820 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
821 * using pipes, etc.
822 */
bca647aa 823const struct option record_options[] = {
d20deb64 824 OPT_CALLBACK('e', "event", &record.evlist, "event",
86847b62 825 "event selector. use 'perf list' to list available events",
f120f9d5 826 parse_events_option),
d20deb64 827 OPT_CALLBACK(0, "filter", &record.evlist, "filter",
c171b552 828 "event filter", parse_filter),
bea03405 829 OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
d6d901c2 830 "record events on existing process id"),
bea03405 831 OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
d6d901c2 832 "record events on existing thread id"),
d20deb64 833 OPT_INTEGER('r', "realtime", &record.realtime_prio,
0e9b20b8 834 "collect data with this RT SCHED_FIFO priority"),
509051ea 835 OPT_BOOLEAN(0, "no-buffering", &record.opts.no_buffering,
acac03fa 836 "collect data without buffering"),
d20deb64 837 OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
daac07b2 838 "collect raw sample records from all opened counters"),
bea03405 839 OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
0e9b20b8 840 "system-wide collection from all CPUs"),
bea03405 841 OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
c45c6ea2 842 "list of cpus to monitor"),
d20deb64 843 OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
f5fc1412 844 OPT_STRING('o', "output", &record.file.path, "file",
abaff32a 845 "output file name"),
69e7e5b0
AH
846 OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
847 &record.opts.no_inherit_set,
848 "child tasks do not inherit counters"),
d20deb64 849 OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
994a1f78
JO
850 OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
851 "number of mmap data pages",
852 perf_evlist__parse_mmap_pages),
d20deb64 853 OPT_BOOLEAN(0, "group", &record.opts.group,
43bece79 854 "put the counters into a counter group"),
09b0fd45
JO
855 OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
856 NULL, "enables call-graph recording" ,
857 &record_callchain_opt),
858 OPT_CALLBACK(0, "call-graph", &record.opts,
859 "mode[,dump_size]", record_callchain_help,
860 &record_parse_callchain_opt),
c0555642 861 OPT_INCR('v', "verbose", &verbose,
3da297a6 862 "be more verbose (show counter open errors, etc)"),
b44308f5 863 OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
d20deb64 864 OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
649c48a9 865 "per thread counts"),
d20deb64 866 OPT_BOOLEAN('d', "data", &record.opts.sample_address,
4bba828d 867 "Sample addresses"),
d20deb64 868 OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
3e76ac78 869 OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
d20deb64 870 OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
649c48a9 871 "don't sample"),
d20deb64 872 OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
a1ac1d3c 873 "do not update the buildid cache"),
d20deb64 874 OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
baa2f6ce 875 "do not collect buildids in perf.data"),
d20deb64 876 OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
023695d9
SE
877 "monitor event in cgroup name only",
878 parse_cgroups),
a6205a35 879 OPT_UINTEGER('D', "delay", &record.opts.initial_delay,
6619a53e 880 "ms to wait before starting measurement after program start"),
bea03405
NK
881 OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
882 "user to profile"),
a5aabdac
SE
883
884 OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
885 "branch any", "sample any taken branches",
886 parse_branch_stack),
887
888 OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
889 "branch filter mask", "branch stack filter modes",
bdfebd84 890 parse_branch_stack),
05484298
AK
891 OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
892 "sample by weight (on special events only)"),
475eeab9
AK
893 OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
894 "sample transaction flags (special events only)"),
3aa5939d
AH
895 OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
896 "use per-thread mmaps"),
0e9b20b8
IM
897 OPT_END()
898};
899
1d037ca1 900int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
0e9b20b8 901{
69aad6f1 902 int err = -ENOMEM;
8c6f45a7 903 struct record *rec = &record;
16ad2ffb 904 char errbuf[BUFSIZ];
0e9b20b8 905
3e2be2da
ACM
906 rec->evlist = perf_evlist__new();
907 if (rec->evlist == NULL)
361c99a6
ACM
908 return -ENOMEM;
909
bca647aa 910 argc = parse_options(argc, argv, record_options, record_usage,
655000e7 911 PARSE_OPT_STOP_AT_NON_OPTION);
602ad878 912 if (!argc && target__none(&rec->opts.target))
bca647aa 913 usage_with_options(record_usage, record_options);
0e9b20b8 914
bea03405 915 if (nr_cgroups && !rec->opts.target.system_wide) {
3780f488
NK
916 ui__error("cgroup monitoring only available in"
917 " system-wide mode\n");
023695d9
SE
918 usage_with_options(record_usage, record_options);
919 }
920
655000e7 921 symbol__init();
baa2f6ce 922
ec80fde7 923 if (symbol_conf.kptr_restrict)
646aaea6
ACM
924 pr_warning(
925"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
926"check /proc/sys/kernel/kptr_restrict.\n\n"
927"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
928"file is not found in the buildid cache or in the vmlinux path.\n\n"
929"Samples in kernel modules won't be resolved at all.\n\n"
930"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
931"even with a suitable vmlinux or kallsyms file.\n\n");
ec80fde7 932
d20deb64 933 if (rec->no_buildid_cache || rec->no_buildid)
a1ac1d3c 934 disable_buildid_cache();
655000e7 935
3e2be2da
ACM
936 if (rec->evlist->nr_entries == 0 &&
937 perf_evlist__add_default(rec->evlist) < 0) {
69aad6f1
ACM
938 pr_err("Not enough memory for event selector list\n");
939 goto out_symbol_exit;
bbd36e5e 940 }
0e9b20b8 941
69e7e5b0
AH
942 if (rec->opts.target.tid && !rec->opts.no_inherit_set)
943 rec->opts.no_inherit = true;
944
602ad878 945 err = target__validate(&rec->opts.target);
16ad2ffb 946 if (err) {
602ad878 947 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
16ad2ffb
NK
948 ui__warning("%s", errbuf);
949 }
950
602ad878 951 err = target__parse_uid(&rec->opts.target);
16ad2ffb
NK
952 if (err) {
953 int saved_errno = errno;
4bd0f2d2 954
602ad878 955 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
3780f488 956 ui__error("%s", errbuf);
16ad2ffb
NK
957
958 err = -saved_errno;
8fa60e1f 959 goto out_symbol_exit;
16ad2ffb 960 }
0d37aa34 961
16ad2ffb 962 err = -ENOMEM;
3e2be2da 963 if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
dd7927f4 964 usage_with_options(record_usage, record_options);
69aad6f1 965
b4006796 966 if (record_opts__config(&rec->opts)) {
39d17dac 967 err = -EINVAL;
03ad9747 968 goto out_symbol_exit;
7e4ff9e3
MG
969 }
970
d20deb64 971 err = __cmd_record(&record, argc, argv);
d65a458b
ACM
972out_symbol_exit:
973 symbol__exit();
39d17dac 974 return err;
0e9b20b8 975}