perf record: Use perf_data_file__write for output file
[linux-2.6-block.git] / tools / perf / builtin-record.c
CommitLineData
abaff32a 1/*
bf9e1876
IM
2 * builtin-record.c
3 *
4 * Builtin record command: Record the profile of a workload
5 * (or a CPU, or a PID) into the perf.data output file - for
6 * later analysis via perf report.
abaff32a 7 */
16f762a2 8#include "builtin.h"
bf9e1876
IM
9
10#include "perf.h"
11
6122e4e4 12#include "util/build-id.h"
6eda5838 13#include "util/util.h"
0e9b20b8 14#include "util/parse-options.h"
8ad8db37 15#include "util/parse-events.h"
6eda5838 16
7c6a1c65 17#include "util/header.h"
66e274f3 18#include "util/event.h"
361c99a6 19#include "util/evlist.h"
69aad6f1 20#include "util/evsel.h"
8f28827a 21#include "util/debug.h"
94c744b6 22#include "util/session.h"
45694aa7 23#include "util/tool.h"
8d06367f 24#include "util/symbol.h"
a12b51c4 25#include "util/cpumap.h"
fd78260b 26#include "util/thread_map.h"
f5fc1412 27#include "util/data.h"
7c6a1c65 28
97124d5e 29#include <unistd.h>
de9ac07b 30#include <sched.h>
a41794cd 31#include <sys/mman.h>
de9ac07b 32
89fe808a 33#ifndef HAVE_ON_EXIT_SUPPORT
78da39fa
BR
34#ifndef ATEXIT_MAX
35#define ATEXIT_MAX 32
36#endif
37static int __on_exit_count = 0;
38typedef void (*on_exit_func_t) (int, void *);
39static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
40static void *__on_exit_args[ATEXIT_MAX];
41static int __exitcode = 0;
42static void __handle_on_exit_funcs(void);
43static int on_exit(on_exit_func_t function, void *arg);
44#define exit(x) (exit)(__exitcode = (x))
45
46static int on_exit(on_exit_func_t function, void *arg)
47{
48 if (__on_exit_count == ATEXIT_MAX)
49 return -ENOMEM;
50 else if (__on_exit_count == 0)
51 atexit(__handle_on_exit_funcs);
52 __on_exit_funcs[__on_exit_count] = function;
53 __on_exit_args[__on_exit_count++] = arg;
54 return 0;
55}
56
57static void __handle_on_exit_funcs(void)
58{
59 int i;
60 for (i = 0; i < __on_exit_count; i++)
61 __on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
62}
63#endif
64
d20deb64 65struct perf_record {
45694aa7 66 struct perf_tool tool;
d20deb64
ACM
67 struct perf_record_opts opts;
68 u64 bytes_written;
f5fc1412 69 struct perf_data_file file;
d20deb64
ACM
70 struct perf_evlist *evlist;
71 struct perf_session *session;
72 const char *progname;
d20deb64 73 int realtime_prio;
d20deb64
ACM
74 bool no_buildid;
75 bool no_buildid_cache;
d20deb64 76 long samples;
0f82ebc4 77};
a21ca2ca 78
50a9b868
JO
79static ssize_t perf_record__write(struct perf_record *rec,
80 void *buf, size_t size)
f5970550 81{
50a9b868
JO
82 struct perf_session *session = rec->session;
83 ssize_t ret;
f5fc1412 84
50a9b868
JO
85 ret = perf_data_file__write(session->file, buf, size);
86 if (ret < 0) {
87 pr_err("failed to write perf data, error: %m\n");
88 return -1;
f5970550 89 }
8d3eca20 90
50a9b868 91 rec->bytes_written += ret;
8d3eca20 92 return 0;
f5970550
PZ
93}
94
45694aa7 95static int process_synthesized_event(struct perf_tool *tool,
d20deb64 96 union perf_event *event,
1d037ca1
IT
97 struct perf_sample *sample __maybe_unused,
98 struct machine *machine __maybe_unused)
234fbbf5 99{
45694aa7 100 struct perf_record *rec = container_of(tool, struct perf_record, tool);
6233dd5e 101 return perf_record__write(rec, event, event->header.size);
234fbbf5
ACM
102}
103
8d3eca20 104static int perf_record__mmap_read(struct perf_record *rec,
d20deb64 105 struct perf_mmap *md)
de9ac07b 106{
744bd8aa 107 unsigned int head = perf_mmap__read_head(md);
de9ac07b 108 unsigned int old = md->prev;
918512b4 109 unsigned char *data = md->base + page_size;
de9ac07b
PZ
110 unsigned long size;
111 void *buf;
8d3eca20 112 int rc = 0;
de9ac07b 113
dc82009a 114 if (old == head)
8d3eca20 115 return 0;
dc82009a 116
d20deb64 117 rec->samples++;
de9ac07b
PZ
118
119 size = head - old;
120
121 if ((old & md->mask) + size != (head & md->mask)) {
122 buf = &data[old & md->mask];
123 size = md->mask + 1 - (old & md->mask);
124 old += size;
021e9f47 125
6233dd5e 126 if (perf_record__write(rec, buf, size) < 0) {
8d3eca20
DA
127 rc = -1;
128 goto out;
129 }
de9ac07b
PZ
130 }
131
132 buf = &data[old & md->mask];
133 size = head - old;
134 old += size;
021e9f47 135
6233dd5e 136 if (perf_record__write(rec, buf, size) < 0) {
8d3eca20
DA
137 rc = -1;
138 goto out;
139 }
de9ac07b
PZ
140
141 md->prev = old;
115d2d89 142 perf_mmap__write_tail(md, old);
8d3eca20
DA
143
144out:
145 return rc;
de9ac07b
PZ
146}
147
148static volatile int done = 0;
f7b7c26e 149static volatile int signr = -1;
33e49ea7 150static volatile int child_finished = 0;
de9ac07b 151
16c8a109 152static void sig_handler(int sig)
de9ac07b 153{
33e49ea7
AK
154 if (sig == SIGCHLD)
155 child_finished = 1;
156
16c8a109 157 done = 1;
f7b7c26e
PZ
158 signr = sig;
159}
160
1d037ca1 161static void perf_record__sig_exit(int exit_status __maybe_unused, void *arg)
f7b7c26e 162{
d20deb64 163 struct perf_record *rec = arg;
33e49ea7
AK
164 int status;
165
d20deb64 166 if (rec->evlist->workload.pid > 0) {
33e49ea7 167 if (!child_finished)
d20deb64 168 kill(rec->evlist->workload.pid, SIGTERM);
33e49ea7
AK
169
170 wait(&status);
171 if (WIFSIGNALED(status))
d20deb64 172 psignal(WTERMSIG(status), rec->progname);
33e49ea7 173 }
933da83a 174
18483b81 175 if (signr == -1 || signr == SIGUSR1)
f7b7c26e
PZ
176 return;
177
178 signal(signr, SIG_DFL);
de9ac07b
PZ
179}
180
8d3eca20 181static int perf_record__open(struct perf_record *rec)
dd7927f4 182{
56e52e85 183 char msg[512];
6a4bb04c 184 struct perf_evsel *pos;
d20deb64
ACM
185 struct perf_evlist *evlist = rec->evlist;
186 struct perf_session *session = rec->session;
187 struct perf_record_opts *opts = &rec->opts;
8d3eca20 188 int rc = 0;
dd7927f4 189
f77a9518 190 perf_evlist__config(evlist, opts);
cac21425 191
dd7927f4 192 list_for_each_entry(pos, &evlist->entries, node) {
dd7927f4 193try_again:
6a4bb04c 194 if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
56e52e85 195 if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
d6d901c2 196 if (verbose)
c0a54341 197 ui__warning("%s\n", msg);
d6d901c2
ZY
198 goto try_again;
199 }
ca6a4258 200
56e52e85
ACM
201 rc = -errno;
202 perf_evsel__open_strerror(pos, &opts->target,
203 errno, msg, sizeof(msg));
204 ui__error("%s\n", msg);
8d3eca20 205 goto out;
c171b552
LZ
206 }
207 }
a43d3f08 208
1491a632 209 if (perf_evlist__apply_filters(evlist)) {
0a102479
FW
210 error("failed to set filter with %d (%s)\n", errno,
211 strerror(errno));
8d3eca20
DA
212 rc = -1;
213 goto out;
0a102479
FW
214 }
215
18e60939 216 if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
8d3eca20
DA
217 if (errno == EPERM) {
218 pr_err("Permission error mapping pages.\n"
219 "Consider increasing "
220 "/proc/sys/kernel/perf_event_mlock_kb,\n"
221 "or try again with a smaller value of -m/--mmap_pages.\n"
53653d70 222 "(current value: %u)\n", opts->mmap_pages);
8d3eca20 223 rc = -errno;
8d3eca20
DA
224 } else {
225 pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
226 rc = -errno;
227 }
228 goto out;
18e60939 229 }
0a27d7f9 230
563aecb2 231 session->evlist = evlist;
7b56cce2 232 perf_session__set_id_hdr_size(session);
8d3eca20
DA
233out:
234 return rc;
16c8a109
PZ
235}
236
d20deb64 237static int process_buildids(struct perf_record *rec)
6122e4e4 238{
f5fc1412
JO
239 struct perf_data_file *file = &rec->file;
240 struct perf_session *session = rec->session;
7ab75cff 241 u64 start = session->header.data_offset;
6122e4e4 242
f5fc1412 243 u64 size = lseek(file->fd, 0, SEEK_CUR);
9f591fd7
ACM
244 if (size == 0)
245 return 0;
246
7ab75cff
DA
247 return __perf_session__process_events(session, start,
248 size - start,
6122e4e4
ACM
249 size, &build_id__mark_dso_hit_ops);
250}
251
8d3eca20 252static void perf_record__exit(int status, void *arg)
f5970550 253{
d20deb64 254 struct perf_record *rec = arg;
f5fc1412 255 struct perf_data_file *file = &rec->file;
d20deb64 256
8d3eca20
DA
257 if (status != 0)
258 return;
259
f5fc1412 260 if (!file->is_pipe) {
d20deb64
ACM
261 rec->session->header.data_size += rec->bytes_written;
262
263 if (!rec->no_buildid)
264 process_buildids(rec);
265 perf_session__write_header(rec->session, rec->evlist,
f5fc1412 266 file->fd, true);
d20deb64
ACM
267 perf_session__delete(rec->session);
268 perf_evlist__delete(rec->evlist);
d65a458b 269 symbol__exit();
c7929e47 270 }
f5970550
PZ
271}
272
8115d60c 273static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
a1645ce1
ZY
274{
275 int err;
45694aa7 276 struct perf_tool *tool = data;
a1645ce1
ZY
277 /*
278 *As for guest kernel when processing subcommand record&report,
279 *we arrange module mmap prior to guest kernel mmap and trigger
280 *a preload dso because default guest module symbols are loaded
281 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
282 *method is used to avoid symbol missing when the first addr is
283 *in module instead of in guest kernel.
284 */
45694aa7 285 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 286 machine);
a1645ce1
ZY
287 if (err < 0)
288 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 289 " relocation symbol.\n", machine->pid);
a1645ce1 290
a1645ce1
ZY
291 /*
292 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
293 * have no _text sometimes.
294 */
45694aa7 295 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
743eb868 296 machine, "_text");
a1645ce1 297 if (err < 0)
45694aa7 298 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
743eb868 299 machine, "_stext");
a1645ce1
ZY
300 if (err < 0)
301 pr_err("Couldn't record guest kernel [%d]'s reference"
23346f21 302 " relocation symbol.\n", machine->pid);
a1645ce1
ZY
303}
304
98402807
FW
305static struct perf_event_header finished_round_event = {
306 .size = sizeof(struct perf_event_header),
307 .type = PERF_RECORD_FINISHED_ROUND,
308};
309
8d3eca20 310static int perf_record__mmap_read_all(struct perf_record *rec)
98402807 311{
0e2e63dd 312 int i;
8d3eca20 313 int rc = 0;
98402807 314
d20deb64 315 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
8d3eca20
DA
316 if (rec->evlist->mmap[i].base) {
317 if (perf_record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
318 rc = -1;
319 goto out;
320 }
321 }
98402807
FW
322 }
323
2eeaaa09 324 if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
6233dd5e
JO
325 rc = perf_record__write(rec, &finished_round_event,
326 sizeof(finished_round_event));
8d3eca20
DA
327
328out:
329 return rc;
98402807
FW
330}
331
57706abc
DA
332static void perf_record__init_features(struct perf_record *rec)
333{
334 struct perf_evlist *evsel_list = rec->evlist;
335 struct perf_session *session = rec->session;
336 int feat;
337
338 for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
339 perf_header__set_feat(&session->header, feat);
340
341 if (rec->no_buildid)
342 perf_header__clear_feat(&session->header, HEADER_BUILD_ID);
343
344 if (!have_tracepoints(&evsel_list->entries))
345 perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
346
347 if (!rec->opts.branch_stack)
348 perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
349}
350
d20deb64 351static int __cmd_record(struct perf_record *rec, int argc, const char **argv)
16c8a109 352{
57706abc 353 int err;
8b412664 354 unsigned long waking = 0;
46be604b 355 const bool forks = argc > 0;
23346f21 356 struct machine *machine;
45694aa7 357 struct perf_tool *tool = &rec->tool;
d20deb64
ACM
358 struct perf_record_opts *opts = &rec->opts;
359 struct perf_evlist *evsel_list = rec->evlist;
f5fc1412 360 struct perf_data_file *file = &rec->file;
d20deb64 361 struct perf_session *session;
2711926a 362 bool disabled = false;
de9ac07b 363
d20deb64 364 rec->progname = argv[0];
33e49ea7 365
d20deb64 366 on_exit(perf_record__sig_exit, rec);
f5970550
PZ
367 signal(SIGCHLD, sig_handler);
368 signal(SIGINT, sig_handler);
18483b81 369 signal(SIGUSR1, sig_handler);
804f7ac7 370 signal(SIGTERM, sig_handler);
f5970550 371
f5fc1412 372 session = perf_session__new(file, false, NULL);
94c744b6 373 if (session == NULL) {
a9a70bbc
ACM
374 pr_err("Not enough memory for reading perf file header\n");
375 return -1;
376 }
377
d20deb64
ACM
378 rec->session = session;
379
57706abc 380 perf_record__init_features(rec);
330aa675 381
d4db3f16 382 if (forks) {
6ef73ec4 383 err = perf_evlist__prepare_workload(evsel_list, &opts->target,
f5fc1412 384 argv, file->is_pipe,
55e162ea 385 true);
35b9d88e
ACM
386 if (err < 0) {
387 pr_err("Couldn't run the workload!\n");
388 goto out_delete_session;
856e9660 389 }
856e9660
PZ
390 }
391
8d3eca20
DA
392 if (perf_record__open(rec) != 0) {
393 err = -1;
394 goto out_delete_session;
395 }
de9ac07b 396
a8bb559b
NK
397 if (!evsel_list->nr_groups)
398 perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);
399
712a4b60 400 /*
d20deb64 401 * perf_session__delete(session) will be called at perf_record__exit()
712a4b60 402 */
d20deb64 403 on_exit(perf_record__exit, rec);
712a4b60 404
f5fc1412
JO
405 if (file->is_pipe) {
406 err = perf_header__write_pipe(file->fd);
529870e3 407 if (err < 0)
8d3eca20 408 goto out_delete_session;
563aecb2 409 } else {
a91e5431 410 err = perf_session__write_header(session, evsel_list,
f5fc1412 411 file->fd, false);
d5eed904 412 if (err < 0)
8d3eca20 413 goto out_delete_session;
56b03f3c
ACM
414 }
415
d3665498 416 if (!rec->no_buildid
e20960c0 417 && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
d3665498 418 pr_err("Couldn't generate buildids. "
e20960c0 419 "Use --no-buildid to profile anyway.\n");
8d3eca20
DA
420 err = -1;
421 goto out_delete_session;
e20960c0
RR
422 }
423
34ba5122 424 machine = &session->machines.host;
743eb868 425
f5fc1412 426 if (file->is_pipe) {
45694aa7 427 err = perf_event__synthesize_attrs(tool, session,
d20deb64 428 process_synthesized_event);
2c46dbb5
TZ
429 if (err < 0) {
430 pr_err("Couldn't synthesize attrs.\n");
8d3eca20 431 goto out_delete_session;
2c46dbb5 432 }
cd19a035 433
361c99a6 434 if (have_tracepoints(&evsel_list->entries)) {
63e0c771
TZ
435 /*
436 * FIXME err <= 0 here actually means that
437 * there were no tracepoints so its not really
438 * an error, just that we don't need to
439 * synthesize anything. We really have to
440 * return this more properly and also
441 * propagate errors that now are calling die()
442 */
f5fc1412 443 err = perf_event__synthesize_tracing_data(tool, file->fd, evsel_list,
743eb868 444 process_synthesized_event);
63e0c771
TZ
445 if (err <= 0) {
446 pr_err("Couldn't record tracing data.\n");
8d3eca20 447 goto out_delete_session;
63e0c771 448 }
f34b9001 449 rec->bytes_written += err;
63e0c771 450 }
2c46dbb5
TZ
451 }
452
45694aa7 453 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
743eb868 454 machine, "_text");
70162138 455 if (err < 0)
45694aa7 456 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
743eb868 457 machine, "_stext");
c1a3a4b9
ACM
458 if (err < 0)
459 pr_err("Couldn't record kernel reference relocation symbol\n"
460 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
461 "Check /proc/kallsyms permission or run as root.\n");
b7cece76 462
45694aa7 463 err = perf_event__synthesize_modules(tool, process_synthesized_event,
743eb868 464 machine);
c1a3a4b9
ACM
465 if (err < 0)
466 pr_err("Couldn't record kernel module information.\n"
467 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
468 "Check /proc/modules permission or run as root.\n");
469
7e383de4 470 if (perf_guest) {
876650e6
ACM
471 machines__process_guests(&session->machines,
472 perf_event__synthesize_guest_os, tool);
7e383de4 473 }
7c6a1c65 474
a33fbd56
ACM
475 err = __machine__synthesize_threads(machine, tool, &opts->target, evsel_list->threads,
476 process_synthesized_event, opts->sample_address);
8d3eca20
DA
477 if (err != 0)
478 goto out_delete_session;
479
d20deb64 480 if (rec->realtime_prio) {
de9ac07b
PZ
481 struct sched_param param;
482
d20deb64 483 param.sched_priority = rec->realtime_prio;
de9ac07b 484 if (sched_setscheduler(0, SCHED_FIFO, &param)) {
6beba7ad 485 pr_err("Could not set realtime priority.\n");
8d3eca20
DA
486 err = -1;
487 goto out_delete_session;
de9ac07b
PZ
488 }
489 }
490
774cb499
JO
491 /*
492 * When perf is starting the traced process, all the events
493 * (apart from group members) have enable_on_exec=1 set,
494 * so don't spoil it by prematurely enabling them.
495 */
602ad878 496 if (!target__none(&opts->target))
774cb499 497 perf_evlist__enable(evsel_list);
764e16a3 498
856e9660
PZ
499 /*
500 * Let the child rip
501 */
d4db3f16 502 if (forks)
35b9d88e 503 perf_evlist__start_workload(evsel_list);
856e9660 504
649c48a9 505 for (;;) {
d20deb64 506 int hits = rec->samples;
de9ac07b 507
8d3eca20
DA
508 if (perf_record__mmap_read_all(rec) < 0) {
509 err = -1;
510 goto out_delete_session;
511 }
de9ac07b 512
d20deb64 513 if (hits == rec->samples) {
649c48a9
PZ
514 if (done)
515 break;
5c581041 516 err = poll(evsel_list->pollfd, evsel_list->nr_fds, -1);
8b412664
PZ
517 waking++;
518 }
519
774cb499
JO
520 /*
521 * When perf is starting the traced process, at the end events
522 * die with the process and we wait for that. Thus no need to
523 * disable events in this case.
524 */
602ad878 525 if (done && !disabled && !target__none(&opts->target)) {
4152ab37 526 perf_evlist__disable(evsel_list);
2711926a
JO
527 disabled = true;
528 }
de9ac07b
PZ
529 }
530
18483b81 531 if (quiet || signr == SIGUSR1)
b44308f5
ACM
532 return 0;
533
8b412664
PZ
534 fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);
535
021e9f47
IM
536 /*
537 * Approximate RIP event size: 24 bytes.
538 */
539 fprintf(stderr,
9486aa38 540 "[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
d20deb64 541 (double)rec->bytes_written / 1024.0 / 1024.0,
6a4d98d7 542 file->path,
d20deb64 543 rec->bytes_written / 24);
addc2785 544
de9ac07b 545 return 0;
39d17dac
ACM
546
547out_delete_session:
548 perf_session__delete(session);
549 return err;
de9ac07b 550}
0e9b20b8 551
bdfebd84
RAV
552#define BRANCH_OPT(n, m) \
553 { .name = n, .mode = (m) }
554
555#define BRANCH_END { .name = NULL }
556
557struct branch_mode {
558 const char *name;
559 int mode;
560};
561
562static const struct branch_mode branch_modes[] = {
563 BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
564 BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
565 BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
566 BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
567 BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
568 BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
569 BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
0126d493
AK
570 BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
571 BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
572 BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
bdfebd84
RAV
573 BRANCH_END
574};
575
576static int
a5aabdac 577parse_branch_stack(const struct option *opt, const char *str, int unset)
bdfebd84
RAV
578{
579#define ONLY_PLM \
580 (PERF_SAMPLE_BRANCH_USER |\
581 PERF_SAMPLE_BRANCH_KERNEL |\
582 PERF_SAMPLE_BRANCH_HV)
583
584 uint64_t *mode = (uint64_t *)opt->value;
585 const struct branch_mode *br;
a5aabdac 586 char *s, *os = NULL, *p;
bdfebd84
RAV
587 int ret = -1;
588
a5aabdac
SE
589 if (unset)
590 return 0;
bdfebd84 591
a5aabdac
SE
592 /*
593 * cannot set it twice, -b + --branch-filter for instance
594 */
595 if (*mode)
bdfebd84
RAV
596 return -1;
597
a5aabdac
SE
598 /* str may be NULL in case no arg is passed to -b */
599 if (str) {
600 /* because str is read-only */
601 s = os = strdup(str);
602 if (!s)
603 return -1;
604
605 for (;;) {
606 p = strchr(s, ',');
607 if (p)
608 *p = '\0';
609
610 for (br = branch_modes; br->name; br++) {
611 if (!strcasecmp(s, br->name))
612 break;
613 }
614 if (!br->name) {
615 ui__warning("unknown branch filter %s,"
616 " check man page\n", s);
617 goto error;
618 }
bdfebd84 619
a5aabdac 620 *mode |= br->mode;
bdfebd84 621
a5aabdac
SE
622 if (!p)
623 break;
bdfebd84 624
a5aabdac
SE
625 s = p + 1;
626 }
bdfebd84
RAV
627 }
628 ret = 0;
629
a5aabdac 630 /* default to any branch */
bdfebd84 631 if ((*mode & ~ONLY_PLM) == 0) {
a5aabdac 632 *mode = PERF_SAMPLE_BRANCH_ANY;
bdfebd84
RAV
633 }
634error:
635 free(os);
636 return ret;
637}
638
89fe808a 639#ifdef HAVE_LIBUNWIND_SUPPORT
26d33022
JO
640static int get_stack_size(char *str, unsigned long *_size)
641{
642 char *endptr;
643 unsigned long size;
644 unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
645
646 size = strtoul(str, &endptr, 0);
647
648 do {
649 if (*endptr)
650 break;
651
652 size = round_up(size, sizeof(u64));
653 if (!size || size > max_size)
654 break;
655
656 *_size = size;
657 return 0;
658
659 } while (0);
660
661 pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
662 max_size, str);
663 return -1;
664}
89fe808a 665#endif /* HAVE_LIBUNWIND_SUPPORT */
26d33022 666
09b0fd45 667int record_parse_callchain(const char *arg, struct perf_record_opts *opts)
26d33022 668{
26d33022
JO
669 char *tok, *name, *saveptr = NULL;
670 char *buf;
671 int ret = -1;
672
26d33022
JO
673 /* We need buffer that we know we can write to. */
674 buf = malloc(strlen(arg) + 1);
675 if (!buf)
676 return -ENOMEM;
677
678 strcpy(buf, arg);
679
680 tok = strtok_r((char *)buf, ",", &saveptr);
681 name = tok ? : (char *)buf;
682
683 do {
684 /* Framepointer style */
685 if (!strncmp(name, "fp", sizeof("fp"))) {
686 if (!strtok_r(NULL, ",", &saveptr)) {
c5ff78c3 687 opts->call_graph = CALLCHAIN_FP;
26d33022
JO
688 ret = 0;
689 } else
690 pr_err("callchain: No more arguments "
691 "needed for -g fp\n");
692 break;
693
89fe808a 694#ifdef HAVE_LIBUNWIND_SUPPORT
26d33022
JO
695 /* Dwarf style */
696 } else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
61eaa3be
ACM
697 const unsigned long default_stack_dump_size = 8192;
698
26d33022 699 ret = 0;
c5ff78c3
ACM
700 opts->call_graph = CALLCHAIN_DWARF;
701 opts->stack_dump_size = default_stack_dump_size;
26d33022
JO
702
703 tok = strtok_r(NULL, ",", &saveptr);
704 if (tok) {
705 unsigned long size = 0;
706
707 ret = get_stack_size(tok, &size);
c5ff78c3 708 opts->stack_dump_size = size;
26d33022 709 }
89fe808a 710#endif /* HAVE_LIBUNWIND_SUPPORT */
26d33022 711 } else {
09b0fd45 712 pr_err("callchain: Unknown --call-graph option "
26d33022
JO
713 "value: %s\n", arg);
714 break;
715 }
716
717 } while (0);
718
719 free(buf);
09b0fd45
JO
720 return ret;
721}
722
723static void callchain_debug(struct perf_record_opts *opts)
724{
725 pr_debug("callchain: type %d\n", opts->call_graph);
26d33022 726
09b0fd45
JO
727 if (opts->call_graph == CALLCHAIN_DWARF)
728 pr_debug("callchain: stack dump size %d\n",
729 opts->stack_dump_size);
730}
731
732int record_parse_callchain_opt(const struct option *opt,
733 const char *arg,
734 int unset)
735{
736 struct perf_record_opts *opts = opt->value;
737 int ret;
738
739 /* --no-call-graph */
740 if (unset) {
741 opts->call_graph = CALLCHAIN_NONE;
742 pr_debug("callchain: disabled\n");
743 return 0;
744 }
745
746 ret = record_parse_callchain(arg, opts);
26d33022 747 if (!ret)
09b0fd45 748 callchain_debug(opts);
26d33022
JO
749
750 return ret;
751}
752
09b0fd45
JO
753int record_callchain_opt(const struct option *opt,
754 const char *arg __maybe_unused,
755 int unset __maybe_unused)
756{
757 struct perf_record_opts *opts = opt->value;
758
759 if (opts->call_graph == CALLCHAIN_NONE)
760 opts->call_graph = CALLCHAIN_FP;
761
762 callchain_debug(opts);
763 return 0;
764}
765
0e9b20b8 766static const char * const record_usage[] = {
9e096753
MG
767 "perf record [<options>] [<command>]",
768 "perf record [<options>] -- <command> [<options>]",
0e9b20b8
IM
769 NULL
770};
771
d20deb64
ACM
772/*
773 * XXX Ideally would be local to cmd_record() and passed to a perf_record__new
774 * because we need to have access to it in perf_record__exit, that is called
775 * after cmd_record() exits, but since record_options need to be accessible to
776 * builtin-script, leave it here.
777 *
778 * At least we don't ouch it in all the other functions here directly.
779 *
780 * Just say no to tons of global variables, sigh.
781 */
782static struct perf_record record = {
783 .opts = {
d20deb64
ACM
784 .mmap_pages = UINT_MAX,
785 .user_freq = UINT_MAX,
786 .user_interval = ULLONG_MAX,
447a6013 787 .freq = 4000,
d1cb9fce
NK
788 .target = {
789 .uses_mmap = true,
3aa5939d 790 .default_per_cpu = true,
d1cb9fce 791 },
d20deb64 792 },
d20deb64 793};
7865e817 794
09b0fd45 795#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
61eaa3be 796
89fe808a 797#ifdef HAVE_LIBUNWIND_SUPPORT
09b0fd45 798const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
61eaa3be 799#else
09b0fd45 800const char record_callchain_help[] = CALLCHAIN_HELP "fp";
61eaa3be
ACM
801#endif
802
d20deb64
ACM
803/*
804 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
805 * with it and switch to use the library functions in perf_evlist that came
806 * from builtin-record.c, i.e. use perf_record_opts,
807 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
808 * using pipes, etc.
809 */
bca647aa 810const struct option record_options[] = {
d20deb64 811 OPT_CALLBACK('e', "event", &record.evlist, "event",
86847b62 812 "event selector. use 'perf list' to list available events",
f120f9d5 813 parse_events_option),
d20deb64 814 OPT_CALLBACK(0, "filter", &record.evlist, "filter",
c171b552 815 "event filter", parse_filter),
bea03405 816 OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
d6d901c2 817 "record events on existing process id"),
bea03405 818 OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
d6d901c2 819 "record events on existing thread id"),
d20deb64 820 OPT_INTEGER('r', "realtime", &record.realtime_prio,
0e9b20b8 821 "collect data with this RT SCHED_FIFO priority"),
d20deb64 822 OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
acac03fa 823 "collect data without buffering"),
d20deb64 824 OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
daac07b2 825 "collect raw sample records from all opened counters"),
bea03405 826 OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
0e9b20b8 827 "system-wide collection from all CPUs"),
bea03405 828 OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
c45c6ea2 829 "list of cpus to monitor"),
d20deb64 830 OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
f5fc1412 831 OPT_STRING('o', "output", &record.file.path, "file",
abaff32a 832 "output file name"),
69e7e5b0
AH
833 OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
834 &record.opts.no_inherit_set,
835 "child tasks do not inherit counters"),
d20deb64 836 OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
994a1f78
JO
837 OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
838 "number of mmap data pages",
839 perf_evlist__parse_mmap_pages),
d20deb64 840 OPT_BOOLEAN(0, "group", &record.opts.group,
43bece79 841 "put the counters into a counter group"),
09b0fd45
JO
842 OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
843 NULL, "enables call-graph recording" ,
844 &record_callchain_opt),
845 OPT_CALLBACK(0, "call-graph", &record.opts,
846 "mode[,dump_size]", record_callchain_help,
847 &record_parse_callchain_opt),
c0555642 848 OPT_INCR('v', "verbose", &verbose,
3da297a6 849 "be more verbose (show counter open errors, etc)"),
b44308f5 850 OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
d20deb64 851 OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
649c48a9 852 "per thread counts"),
d20deb64 853 OPT_BOOLEAN('d', "data", &record.opts.sample_address,
4bba828d 854 "Sample addresses"),
d20deb64 855 OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
3e76ac78 856 OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
d20deb64 857 OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
649c48a9 858 "don't sample"),
d20deb64 859 OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
a1ac1d3c 860 "do not update the buildid cache"),
d20deb64 861 OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
baa2f6ce 862 "do not collect buildids in perf.data"),
d20deb64 863 OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
023695d9
SE
864 "monitor event in cgroup name only",
865 parse_cgroups),
bea03405
NK
866 OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
867 "user to profile"),
a5aabdac
SE
868
869 OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
870 "branch any", "sample any taken branches",
871 parse_branch_stack),
872
873 OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
874 "branch filter mask", "branch stack filter modes",
bdfebd84 875 parse_branch_stack),
05484298
AK
876 OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
877 "sample by weight (on special events only)"),
475eeab9
AK
878 OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
879 "sample transaction flags (special events only)"),
3aa5939d
AH
880 OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
881 "use per-thread mmaps"),
0e9b20b8
IM
882 OPT_END()
883};
884
1d037ca1 885int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
0e9b20b8 886{
69aad6f1 887 int err = -ENOMEM;
d20deb64
ACM
888 struct perf_evlist *evsel_list;
889 struct perf_record *rec = &record;
16ad2ffb 890 char errbuf[BUFSIZ];
0e9b20b8 891
334fe7a3 892 evsel_list = perf_evlist__new();
361c99a6
ACM
893 if (evsel_list == NULL)
894 return -ENOMEM;
895
d20deb64
ACM
896 rec->evlist = evsel_list;
897
bca647aa 898 argc = parse_options(argc, argv, record_options, record_usage,
655000e7 899 PARSE_OPT_STOP_AT_NON_OPTION);
602ad878 900 if (!argc && target__none(&rec->opts.target))
bca647aa 901 usage_with_options(record_usage, record_options);
0e9b20b8 902
bea03405 903 if (nr_cgroups && !rec->opts.target.system_wide) {
3780f488
NK
904 ui__error("cgroup monitoring only available in"
905 " system-wide mode\n");
023695d9
SE
906 usage_with_options(record_usage, record_options);
907 }
908
655000e7 909 symbol__init();
baa2f6ce 910
ec80fde7 911 if (symbol_conf.kptr_restrict)
646aaea6
ACM
912 pr_warning(
913"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
914"check /proc/sys/kernel/kptr_restrict.\n\n"
915"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
916"file is not found in the buildid cache or in the vmlinux path.\n\n"
917"Samples in kernel modules won't be resolved at all.\n\n"
918"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
919"even with a suitable vmlinux or kallsyms file.\n\n");
ec80fde7 920
d20deb64 921 if (rec->no_buildid_cache || rec->no_buildid)
a1ac1d3c 922 disable_buildid_cache();
655000e7 923
361c99a6
ACM
924 if (evsel_list->nr_entries == 0 &&
925 perf_evlist__add_default(evsel_list) < 0) {
69aad6f1
ACM
926 pr_err("Not enough memory for event selector list\n");
927 goto out_symbol_exit;
bbd36e5e 928 }
0e9b20b8 929
69e7e5b0
AH
930 if (rec->opts.target.tid && !rec->opts.no_inherit_set)
931 rec->opts.no_inherit = true;
932
602ad878 933 err = target__validate(&rec->opts.target);
16ad2ffb 934 if (err) {
602ad878 935 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
16ad2ffb
NK
936 ui__warning("%s", errbuf);
937 }
938
602ad878 939 err = target__parse_uid(&rec->opts.target);
16ad2ffb
NK
940 if (err) {
941 int saved_errno = errno;
4bd0f2d2 942
602ad878 943 target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
3780f488 944 ui__error("%s", errbuf);
16ad2ffb
NK
945
946 err = -saved_errno;
8fa60e1f 947 goto out_symbol_exit;
16ad2ffb 948 }
0d37aa34 949
16ad2ffb 950 err = -ENOMEM;
b809ac10 951 if (perf_evlist__create_maps(evsel_list, &rec->opts.target) < 0)
dd7927f4 952 usage_with_options(record_usage, record_options);
69aad6f1 953
714647bd 954 if (perf_record_opts__config(&rec->opts)) {
39d17dac 955 err = -EINVAL;
5c581041 956 goto out_free_fd;
7e4ff9e3
MG
957 }
958
d20deb64 959 err = __cmd_record(&record, argc, argv);
8fa60e1f
NK
960
961 perf_evlist__munmap(evsel_list);
962 perf_evlist__close(evsel_list);
39d17dac 963out_free_fd:
7e2ed097 964 perf_evlist__delete_maps(evsel_list);
d65a458b
ACM
965out_symbol_exit:
966 symbol__exit();
39d17dac 967 return err;
0e9b20b8 968}