Merge branch 'smp-hotplug-for-linus' of git://git.kernel.org/pub/scm/linux/kernel...
[linux-2.6-block.git] / include / trace / events / kmem.h
CommitLineData
d0b6e04a
LZ
1#undef TRACE_SYSTEM
2#define TRACE_SYSTEM kmem
3
ea20d929 4#if !defined(_TRACE_KMEM_H) || defined(TRACE_HEADER_MULTI_READ)
02af61bb 5#define _TRACE_KMEM_H
b9ce08c0
EGM
6
7#include <linux/types.h>
fc182a43 8#include <linux/tracepoint.h>
a1ce3928 9#include <trace/events/gfpflags.h>
62ba180e 10
53d0422c 11DECLARE_EVENT_CLASS(kmem_alloc,
ea20d929
SR
12
13 TP_PROTO(unsigned long call_site,
14 const void *ptr,
15 size_t bytes_req,
16 size_t bytes_alloc,
17 gfp_t gfp_flags),
18
19 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags),
20
21 TP_STRUCT__entry(
22 __field( unsigned long, call_site )
23 __field( const void *, ptr )
24 __field( size_t, bytes_req )
25 __field( size_t, bytes_alloc )
26 __field( gfp_t, gfp_flags )
27 ),
28
29 TP_fast_assign(
30 __entry->call_site = call_site;
31 __entry->ptr = ptr;
32 __entry->bytes_req = bytes_req;
33 __entry->bytes_alloc = bytes_alloc;
34 __entry->gfp_flags = gfp_flags;
35 ),
36
62ba180e 37 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s",
ea20d929
SR
38 __entry->call_site,
39 __entry->ptr,
40 __entry->bytes_req,
41 __entry->bytes_alloc,
62ba180e 42 show_gfp_flags(__entry->gfp_flags))
ea20d929
SR
43);
44
53d0422c 45DEFINE_EVENT(kmem_alloc, kmalloc,
ea20d929 46
53d0422c
LZ
47 TP_PROTO(unsigned long call_site, const void *ptr,
48 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
ea20d929 49
53d0422c
LZ
50 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
51);
ea20d929 52
53d0422c 53DEFINE_EVENT(kmem_alloc, kmem_cache_alloc,
ea20d929 54
53d0422c
LZ
55 TP_PROTO(unsigned long call_site, const void *ptr,
56 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags),
ea20d929 57
53d0422c 58 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags)
ea20d929
SR
59);
60
53d0422c 61DECLARE_EVENT_CLASS(kmem_alloc_node,
ea20d929
SR
62
63 TP_PROTO(unsigned long call_site,
64 const void *ptr,
65 size_t bytes_req,
66 size_t bytes_alloc,
67 gfp_t gfp_flags,
68 int node),
69
70 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node),
71
72 TP_STRUCT__entry(
73 __field( unsigned long, call_site )
74 __field( const void *, ptr )
75 __field( size_t, bytes_req )
76 __field( size_t, bytes_alloc )
77 __field( gfp_t, gfp_flags )
78 __field( int, node )
79 ),
80
81 TP_fast_assign(
82 __entry->call_site = call_site;
83 __entry->ptr = ptr;
84 __entry->bytes_req = bytes_req;
85 __entry->bytes_alloc = bytes_alloc;
86 __entry->gfp_flags = gfp_flags;
87 __entry->node = node;
88 ),
89
62ba180e 90 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d",
ea20d929
SR
91 __entry->call_site,
92 __entry->ptr,
93 __entry->bytes_req,
94 __entry->bytes_alloc,
62ba180e 95 show_gfp_flags(__entry->gfp_flags),
ea20d929
SR
96 __entry->node)
97);
98
53d0422c 99DEFINE_EVENT(kmem_alloc_node, kmalloc_node,
ea20d929 100
53d0422c
LZ
101 TP_PROTO(unsigned long call_site, const void *ptr,
102 size_t bytes_req, size_t bytes_alloc,
103 gfp_t gfp_flags, int node),
ea20d929 104
53d0422c
LZ
105 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
106);
ea20d929 107
53d0422c 108DEFINE_EVENT(kmem_alloc_node, kmem_cache_alloc_node,
ea20d929 109
53d0422c
LZ
110 TP_PROTO(unsigned long call_site, const void *ptr,
111 size_t bytes_req, size_t bytes_alloc,
112 gfp_t gfp_flags, int node),
ea20d929 113
53d0422c 114 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node)
ea20d929
SR
115);
116
53d0422c 117DECLARE_EVENT_CLASS(kmem_free,
ea20d929
SR
118
119 TP_PROTO(unsigned long call_site, const void *ptr),
120
121 TP_ARGS(call_site, ptr),
122
123 TP_STRUCT__entry(
124 __field( unsigned long, call_site )
125 __field( const void *, ptr )
126 ),
127
128 TP_fast_assign(
129 __entry->call_site = call_site;
130 __entry->ptr = ptr;
131 ),
132
133 TP_printk("call_site=%lx ptr=%p", __entry->call_site, __entry->ptr)
134);
135
53d0422c 136DEFINE_EVENT(kmem_free, kfree,
ea20d929
SR
137
138 TP_PROTO(unsigned long call_site, const void *ptr),
139
53d0422c
LZ
140 TP_ARGS(call_site, ptr)
141);
ea20d929 142
e5feb1eb 143DEFINE_EVENT_CONDITION(kmem_free, kmem_cache_free,
ea20d929 144
53d0422c 145 TP_PROTO(unsigned long call_site, const void *ptr),
ea20d929 146
e5feb1eb
SP
147 TP_ARGS(call_site, ptr),
148
149 /*
150 * This trace can be potentially called from an offlined cpu.
151 * Since trace points use RCU and RCU should not be used from
152 * offline cpus, filter such calls out.
153 * While this trace can be called from a preemptable section,
154 * it has no impact on the condition since tasks can migrate
155 * only from online cpus to other online cpus. Thus its safe
156 * to use raw_smp_processor_id.
157 */
158 TP_CONDITION(cpu_online(raw_smp_processor_id()))
ea20d929 159);
4b4f278c 160
1f0c27b5 161TRACE_EVENT_CONDITION(mm_page_free,
4b4f278c
MG
162
163 TP_PROTO(struct page *page, unsigned int order),
164
165 TP_ARGS(page, order),
166
1f0c27b5
SP
167
168 /*
169 * This trace can be potentially called from an offlined cpu.
170 * Since trace points use RCU and RCU should not be used from
171 * offline cpus, filter such calls out.
172 * While this trace can be called from a preemptable section,
173 * it has no impact on the condition since tasks can migrate
174 * only from online cpus to other online cpus. Thus its safe
175 * to use raw_smp_processor_id.
176 */
177 TP_CONDITION(cpu_online(raw_smp_processor_id())),
178
4b4f278c 179 TP_STRUCT__entry(
9fdd8a87 180 __field( unsigned long, pfn )
4b4f278c
MG
181 __field( unsigned int, order )
182 ),
183
184 TP_fast_assign(
9fdd8a87 185 __entry->pfn = page_to_pfn(page);
4b4f278c
MG
186 __entry->order = order;
187 ),
188
189 TP_printk("page=%p pfn=%lu order=%d",
9fdd8a87
NK
190 pfn_to_page(__entry->pfn),
191 __entry->pfn,
4b4f278c
MG
192 __entry->order)
193);
194
b413d48a 195TRACE_EVENT(mm_page_free_batched,
4b4f278c
MG
196
197 TP_PROTO(struct page *page, int cold),
198
199 TP_ARGS(page, cold),
200
201 TP_STRUCT__entry(
9fdd8a87 202 __field( unsigned long, pfn )
4b4f278c
MG
203 __field( int, cold )
204 ),
205
206 TP_fast_assign(
9fdd8a87 207 __entry->pfn = page_to_pfn(page);
4b4f278c
MG
208 __entry->cold = cold;
209 ),
210
211 TP_printk("page=%p pfn=%lu order=0 cold=%d",
9fdd8a87
NK
212 pfn_to_page(__entry->pfn),
213 __entry->pfn,
4b4f278c
MG
214 __entry->cold)
215);
216
217TRACE_EVENT(mm_page_alloc,
218
219 TP_PROTO(struct page *page, unsigned int order,
220 gfp_t gfp_flags, int migratetype),
221
222 TP_ARGS(page, order, gfp_flags, migratetype),
223
224 TP_STRUCT__entry(
9fdd8a87 225 __field( unsigned long, pfn )
4b4f278c
MG
226 __field( unsigned int, order )
227 __field( gfp_t, gfp_flags )
228 __field( int, migratetype )
229 ),
230
231 TP_fast_assign(
9fdd8a87 232 __entry->pfn = page ? page_to_pfn(page) : -1UL;
4b4f278c
MG
233 __entry->order = order;
234 __entry->gfp_flags = gfp_flags;
235 __entry->migratetype = migratetype;
236 ),
237
238 TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s",
9fdd8a87
NK
239 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
240 __entry->pfn != -1UL ? __entry->pfn : 0,
4b4f278c
MG
241 __entry->order,
242 __entry->migratetype,
243 show_gfp_flags(__entry->gfp_flags))
244);
245
53d0422c 246DECLARE_EVENT_CLASS(mm_page,
0d3d062a
MG
247
248 TP_PROTO(struct page *page, unsigned int order, int migratetype),
249
250 TP_ARGS(page, order, migratetype),
251
252 TP_STRUCT__entry(
9fdd8a87 253 __field( unsigned long, pfn )
0d3d062a
MG
254 __field( unsigned int, order )
255 __field( int, migratetype )
256 ),
257
258 TP_fast_assign(
9fdd8a87 259 __entry->pfn = page ? page_to_pfn(page) : -1UL;
0d3d062a
MG
260 __entry->order = order;
261 __entry->migratetype = migratetype;
262 ),
263
264 TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d",
9fdd8a87
NK
265 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
266 __entry->pfn != -1UL ? __entry->pfn : 0,
0d3d062a
MG
267 __entry->order,
268 __entry->migratetype,
269 __entry->order == 0)
270);
271
53d0422c 272DEFINE_EVENT(mm_page, mm_page_alloc_zone_locked,
0d3d062a 273
53d0422c 274 TP_PROTO(struct page *page, unsigned int order, int migratetype),
0d3d062a 275
53d0422c
LZ
276 TP_ARGS(page, order, migratetype)
277);
0d3d062a 278
649b8de2 279TRACE_EVENT_CONDITION(mm_page_pcpu_drain,
0d3d062a 280
53d0422c
LZ
281 TP_PROTO(struct page *page, unsigned int order, int migratetype),
282
283 TP_ARGS(page, order, migratetype),
0d3d062a 284
649b8de2
SP
285 /*
286 * This trace can be potentially called from an offlined cpu.
287 * Since trace points use RCU and RCU should not be used from
288 * offline cpus, filter such calls out.
289 * While this trace can be called from a preemptable section,
290 * it has no impact on the condition since tasks can migrate
291 * only from online cpus to other online cpus. Thus its safe
292 * to use raw_smp_processor_id.
293 */
294 TP_CONDITION(cpu_online(raw_smp_processor_id())),
295
296 TP_STRUCT__entry(
297 __field( unsigned long, pfn )
298 __field( unsigned int, order )
299 __field( int, migratetype )
300 ),
301
302 TP_fast_assign(
303 __entry->pfn = page ? page_to_pfn(page) : -1UL;
304 __entry->order = order;
305 __entry->migratetype = migratetype;
306 ),
307
0d3d062a 308 TP_printk("page=%p pfn=%lu order=%d migratetype=%d",
9fdd8a87 309 pfn_to_page(__entry->pfn), __entry->pfn,
53d0422c 310 __entry->order, __entry->migratetype)
0d3d062a
MG
311);
312
e0fff1bd
MG
313TRACE_EVENT(mm_page_alloc_extfrag,
314
315 TP_PROTO(struct page *page,
52c8f6a5 316 int alloc_order, int fallback_order,
99592d59 317 int alloc_migratetype, int fallback_migratetype),
e0fff1bd
MG
318
319 TP_ARGS(page,
320 alloc_order, fallback_order,
99592d59 321 alloc_migratetype, fallback_migratetype),
e0fff1bd
MG
322
323 TP_STRUCT__entry(
9fdd8a87 324 __field( unsigned long, pfn )
e0fff1bd
MG
325 __field( int, alloc_order )
326 __field( int, fallback_order )
327 __field( int, alloc_migratetype )
328 __field( int, fallback_migratetype )
f92310c1 329 __field( int, change_ownership )
e0fff1bd
MG
330 ),
331
332 TP_fast_assign(
9fdd8a87 333 __entry->pfn = page_to_pfn(page);
e0fff1bd
MG
334 __entry->alloc_order = alloc_order;
335 __entry->fallback_order = fallback_order;
336 __entry->alloc_migratetype = alloc_migratetype;
337 __entry->fallback_migratetype = fallback_migratetype;
99592d59
VB
338 __entry->change_ownership = (alloc_migratetype ==
339 get_pageblock_migratetype(page));
e0fff1bd
MG
340 ),
341
342 TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d",
9fdd8a87
NK
343 pfn_to_page(__entry->pfn),
344 __entry->pfn,
e0fff1bd
MG
345 __entry->alloc_order,
346 __entry->fallback_order,
347 pageblock_order,
348 __entry->alloc_migratetype,
349 __entry->fallback_migratetype,
350 __entry->fallback_order < pageblock_order,
f92310c1 351 __entry->change_ownership)
e0fff1bd
MG
352);
353
a8d154b0 354#endif /* _TRACE_KMEM_H */
ea20d929 355
a8d154b0
SR
356/* This part must be outside protection */
357#include <trace/define_trace.h>