Merge tag 'cgroup-for-6.11-rc4-fixes' of git://git.kernel.org/pub/scm/linux/kernel...
[linux-block.git] / tools / perf / util / kwork.h
1 #ifndef PERF_UTIL_KWORK_H
2 #define PERF_UTIL_KWORK_H
3
4 #include "util/tool.h"
5 #include "util/time-utils.h"
6
7 #include <linux/bitmap.h>
8 #include <linux/list.h>
9 #include <linux/rbtree.h>
10 #include <linux/types.h>
11
12 struct perf_sample;
13 struct perf_session;
14
15 enum kwork_class_type {
16         KWORK_CLASS_IRQ,
17         KWORK_CLASS_SOFTIRQ,
18         KWORK_CLASS_WORKQUEUE,
19         KWORK_CLASS_SCHED,
20         KWORK_CLASS_MAX,
21 };
22
23 enum kwork_report_type {
24         KWORK_REPORT_RUNTIME,
25         KWORK_REPORT_LATENCY,
26         KWORK_REPORT_TIMEHIST,
27         KWORK_REPORT_TOP,
28 };
29
30 enum kwork_trace_type {
31         KWORK_TRACE_RAISE,
32         KWORK_TRACE_ENTRY,
33         KWORK_TRACE_EXIT,
34         KWORK_TRACE_MAX,
35 };
36
37 /*
38  * data structure:
39  *
40  *                 +==================+ +============+ +======================+
41  *                 |      class       | |    work    | |         atom         |
42  *                 +==================+ +============+ +======================+
43  * +------------+  |  +-----+         | |  +------+  | |  +-------+   +-----+ |
44  * | perf_kwork | +-> | irq | --------|+-> | eth0 | --+-> | raise | - | ... | --+   +-----------+
45  * +-----+------+ ||  +-----+         |||  +------+  |||  +-------+   +-----+ | |   |           |
46  *       |        ||                  |||            |||                      | +-> | atom_page |
47  *       |        ||                  |||            |||  +-------+   +-----+ |     |           |
48  *       |  class_list                |||            |+-> | entry | - | ... | ----> |           |
49  *       |        ||                  |||            |||  +-------+   +-----+ |     |           |
50  *       |        ||                  |||            |||                      | +-> |           |
51  *       |        ||                  |||            |||  +-------+   +-----+ | |   |           |
52  *       |        ||                  |||            |+-> | exit  | - | ... | --+   +-----+-----+
53  *       |        ||                  |||            | |  +-------+   +-----+ |           |
54  *       |        ||                  |||            | |                      |           |
55  *       |        ||                  |||  +-----+   | |                      |           |
56  *       |        ||                  |+-> | ... |   | |                      |           |
57  *       |        ||                  | |  +-----+   | |                      |           |
58  *       |        ||                  | |            | |                      |           |
59  *       |        ||  +---------+     | |  +-----+   | |  +-------+   +-----+ |           |
60  *       |        +-> | softirq | -------> | RCU | ---+-> | raise | - | ... | --+   +-----+-----+
61  *       |        ||  +---------+     | |  +-----+   |||  +-------+   +-----+ | |   |           |
62  *       |        ||                  | |            |||                      | +-> | atom_page |
63  *       |        ||                  | |            |||  +-------+   +-----+ |     |           |
64  *       |        ||                  | |            |+-> | entry | - | ... | ----> |           |
65  *       |        ||                  | |            |||  +-------+   +-----+ |     |           |
66  *       |        ||                  | |            |||                      | +-> |           |
67  *       |        ||                  | |            |||  +-------+   +-----+ | |   |           |
68  *       |        ||                  | |            |+-> | exit  | - | ... | --+   +-----+-----+
69  *       |        ||                  | |            | |  +-------+   +-----+ |           |
70  *       |        ||                  | |            | |                      |           |
71  *       |        ||  +-----------+   | |  +-----+   | |                      |           |
72  *       |        +-> | workqueue | -----> | ... |   | |                      |           |
73  *       |         |  +-----------+   | |  +-----+   | |                      |           |
74  *       |         +==================+ +============+ +======================+           |
75  *       |                                                                                |
76  *       +---->  atom_page_list  ---------------------------------------------------------+
77  *
78  */
79
80 struct kwork_atom {
81         struct list_head list;
82         u64 time;
83         struct kwork_atom *prev;
84
85         void *page_addr;
86         unsigned long bit_inpage;
87 };
88
89 #define NR_ATOM_PER_PAGE 128
90 struct kwork_atom_page {
91         struct list_head list;
92         struct kwork_atom atoms[NR_ATOM_PER_PAGE];
93         DECLARE_BITMAP(bitmap, NR_ATOM_PER_PAGE);
94 };
95
96 struct perf_kwork;
97 struct kwork_class;
98 struct kwork_work {
99         /*
100          * class field
101          */
102         struct rb_node node;
103         struct kwork_class *class;
104
105         /*
106          * work field
107          */
108         u64 id;
109         int cpu;
110         char *name;
111
112         /*
113          * atom field
114          */
115         u64 nr_atoms;
116         struct list_head atom_list[KWORK_TRACE_MAX];
117
118         /*
119          * runtime report
120          */
121         u64 max_runtime;
122         u64 max_runtime_start;
123         u64 max_runtime_end;
124         u64 total_runtime;
125
126         /*
127          * latency report
128          */
129         u64 max_latency;
130         u64 max_latency_start;
131         u64 max_latency_end;
132         u64 total_latency;
133
134         /*
135          * top report
136          */
137         u32 cpu_usage;
138         u32 tgid;
139         bool is_kthread;
140 };
141
142 struct kwork_class {
143         struct list_head list;
144         const char *name;
145         enum kwork_class_type type;
146
147         unsigned int nr_tracepoints;
148         const struct evsel_str_handler *tp_handlers;
149
150         struct rb_root_cached work_root;
151
152         int (*class_init)(struct kwork_class *class,
153                           struct perf_session *session);
154
155         void (*work_init)(struct perf_kwork *kwork,
156                           struct kwork_class *class,
157                           struct kwork_work *work,
158                           enum kwork_trace_type src_type,
159                           struct evsel *evsel,
160                           struct perf_sample *sample,
161                           struct machine *machine);
162
163         void (*work_name)(struct kwork_work *work,
164                           char *buf, int len);
165 };
166
167 struct trace_kwork_handler {
168         int (*raise_event)(struct perf_kwork *kwork,
169                            struct kwork_class *class, struct evsel *evsel,
170                            struct perf_sample *sample, struct machine *machine);
171
172         int (*entry_event)(struct perf_kwork *kwork,
173                            struct kwork_class *class, struct evsel *evsel,
174                            struct perf_sample *sample, struct machine *machine);
175
176         int (*exit_event)(struct perf_kwork *kwork,
177                           struct kwork_class *class, struct evsel *evsel,
178                           struct perf_sample *sample, struct machine *machine);
179
180         int (*sched_switch_event)(struct perf_kwork *kwork,
181                                   struct kwork_class *class, struct evsel *evsel,
182                                   struct perf_sample *sample, struct machine *machine);
183 };
184
185 struct __top_cpus_runtime {
186         u64 load;
187         u64 idle;
188         u64 irq;
189         u64 softirq;
190         u64 total;
191 };
192
193 struct kwork_top_stat {
194         DECLARE_BITMAP(all_cpus_bitmap, MAX_NR_CPUS);
195         struct __top_cpus_runtime *cpus_runtime;
196 };
197
198 struct perf_kwork {
199         /*
200          * metadata
201          */
202         struct perf_tool tool;
203         struct list_head class_list;
204         struct list_head atom_page_list;
205         struct list_head sort_list, cmp_id;
206         struct rb_root_cached sorted_work_root;
207         const struct trace_kwork_handler *tp_handler;
208
209         /*
210          * profile filters
211          */
212         const char *profile_name;
213
214         const char *cpu_list;
215         DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
216
217         const char *time_str;
218         struct perf_time_interval ptime;
219
220         /*
221          * options for command
222          */
223         bool force;
224         const char *event_list_str;
225         enum kwork_report_type report;
226
227         /*
228          * options for subcommand
229          */
230         bool summary;
231         const char *sort_order;
232         bool show_callchain;
233         unsigned int max_stack;
234         bool use_bpf;
235
236         /*
237          * statistics
238          */
239         u64 timestart;
240         u64 timeend;
241
242         unsigned long nr_events;
243         unsigned long nr_lost_chunks;
244         unsigned long nr_lost_events;
245
246         u64 all_runtime;
247         u64 all_count;
248         u64 nr_skipped_events[KWORK_TRACE_MAX + 1];
249
250         /*
251          * perf kwork top data
252          */
253         struct kwork_top_stat top_stat;
254 };
255
256 struct kwork_work *perf_kwork_add_work(struct perf_kwork *kwork,
257                                        struct kwork_class *class,
258                                        struct kwork_work *key);
259
260 #ifdef HAVE_BPF_SKEL
261
262 int perf_kwork__trace_prepare_bpf(struct perf_kwork *kwork);
263 int perf_kwork__report_read_bpf(struct perf_kwork *kwork);
264 void perf_kwork__report_cleanup_bpf(void);
265
266 void perf_kwork__trace_start(void);
267 void perf_kwork__trace_finish(void);
268
269 int perf_kwork__top_prepare_bpf(struct perf_kwork *kwork);
270 int perf_kwork__top_read_bpf(struct perf_kwork *kwork);
271 void perf_kwork__top_cleanup_bpf(void);
272
273 void perf_kwork__top_start(void);
274 void perf_kwork__top_finish(void);
275
276 #else  /* !HAVE_BPF_SKEL */
277
278 static inline int
279 perf_kwork__trace_prepare_bpf(struct perf_kwork *kwork __maybe_unused)
280 {
281         return -1;
282 }
283
284 static inline int
285 perf_kwork__report_read_bpf(struct perf_kwork *kwork __maybe_unused)
286 {
287         return -1;
288 }
289
290 static inline void perf_kwork__report_cleanup_bpf(void) {}
291
292 static inline void perf_kwork__trace_start(void) {}
293 static inline void perf_kwork__trace_finish(void) {}
294
295 static inline int
296 perf_kwork__top_prepare_bpf(struct perf_kwork *kwork __maybe_unused)
297 {
298         return -1;
299 }
300
301 static inline int
302 perf_kwork__top_read_bpf(struct perf_kwork *kwork __maybe_unused)
303 {
304         return -1;
305 }
306
307 static inline void perf_kwork__top_cleanup_bpf(void) {}
308
309 static inline void perf_kwork__top_start(void) {}
310 static inline void perf_kwork__top_finish(void) {}
311
312 #endif  /* HAVE_BPF_SKEL */
313
314 #endif  /* PERF_UTIL_KWORK_H */