Enable lguest drivers in Kconfig
[linux-block.git] / kernel / sched_debug.c
CommitLineData
43ae34cb
IM
1/*
2 * kernel/time/sched_debug.c
3 *
4 * Print the CFS rbtree
5 *
6 * Copyright(C) 2007, Red Hat, Inc., Ingo Molnar
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
11 */
12
13#include <linux/proc_fs.h>
14#include <linux/sched.h>
15#include <linux/seq_file.h>
16#include <linux/kallsyms.h>
17#include <linux/utsname.h>
18
19/*
20 * This allows printing both to /proc/sched_debug and
21 * to the console
22 */
23#define SEQ_printf(m, x...) \
24 do { \
25 if (m) \
26 seq_printf(m, x); \
27 else \
28 printk(x); \
29 } while (0)
30
31static void
32print_task(struct seq_file *m, struct rq *rq, struct task_struct *p, u64 now)
33{
34 if (rq->curr == p)
35 SEQ_printf(m, "R");
36 else
37 SEQ_printf(m, " ");
38
39 SEQ_printf(m, "%15s %5d %15Ld %13Ld %13Ld %9Ld %5d "
40 "%15Ld %15Ld %15Ld %15Ld %15Ld\n",
41 p->comm, p->pid,
42 (long long)p->se.fair_key,
43 (long long)(p->se.fair_key - rq->cfs.fair_clock),
44 (long long)p->se.wait_runtime,
45 (long long)(p->nvcsw + p->nivcsw),
46 p->prio,
6cfb0d5d 47#ifdef CONFIG_SCHEDSTATS
43ae34cb
IM
48 (long long)p->se.sum_exec_runtime,
49 (long long)p->se.sum_wait_runtime,
50 (long long)p->se.sum_sleep_runtime,
51 (long long)p->se.wait_runtime_overruns,
6cfb0d5d
IM
52 (long long)p->se.wait_runtime_underruns
53#else
54 0LL, 0LL, 0LL, 0LL, 0LL
55#endif
56 );
43ae34cb
IM
57}
58
59static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu, u64 now)
60{
61 struct task_struct *g, *p;
62
63 SEQ_printf(m,
64 "\nrunnable tasks:\n"
65 " task PID tree-key delta waiting"
66 " switches prio"
67 " sum-exec sum-wait sum-sleep"
68 " wait-overrun wait-underrun\n"
69 "------------------------------------------------------------------"
70 "----------------"
71 "------------------------------------------------"
72 "--------------------------------\n");
73
74 read_lock_irq(&tasklist_lock);
75
76 do_each_thread(g, p) {
77 if (!p->se.on_rq || task_cpu(p) != rq_cpu)
78 continue;
79
80 print_task(m, rq, p, now);
81 } while_each_thread(g, p);
82
83 read_unlock_irq(&tasklist_lock);
84}
85
86static void
87print_cfs_rq_runtime_sum(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
88{
89 s64 wait_runtime_rq_sum = 0;
90 struct task_struct *p;
91 struct rb_node *curr;
92 unsigned long flags;
93 struct rq *rq = &per_cpu(runqueues, cpu);
94
95 spin_lock_irqsave(&rq->lock, flags);
96 curr = first_fair(cfs_rq);
97 while (curr) {
98 p = rb_entry(curr, struct task_struct, se.run_node);
99 wait_runtime_rq_sum += p->se.wait_runtime;
100
101 curr = rb_next(curr);
102 }
103 spin_unlock_irqrestore(&rq->lock, flags);
104
105 SEQ_printf(m, " .%-30s: %Ld\n", "wait_runtime_rq_sum",
106 (long long)wait_runtime_rq_sum);
107}
108
109void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq, u64 now)
110{
111 SEQ_printf(m, "\ncfs_rq %p\n", cfs_rq);
112
113#define P(x) \
114 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(cfs_rq->x))
115
116 P(fair_clock);
117 P(exec_clock);
118 P(wait_runtime);
119 P(wait_runtime_overruns);
120 P(wait_runtime_underruns);
121 P(sleeper_bonus);
122#undef P
123
124 print_cfs_rq_runtime_sum(m, cpu, cfs_rq);
125}
126
127static void print_cpu(struct seq_file *m, int cpu, u64 now)
128{
129 struct rq *rq = &per_cpu(runqueues, cpu);
130
131#ifdef CONFIG_X86
132 {
133 unsigned int freq = cpu_khz ? : 1;
134
135 SEQ_printf(m, "\ncpu#%d, %u.%03u MHz\n",
136 cpu, freq / 1000, (freq % 1000));
137 }
138#else
139 SEQ_printf(m, "\ncpu#%d\n", cpu);
140#endif
141
142#define P(x) \
143 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x))
144
145 P(nr_running);
146 SEQ_printf(m, " .%-30s: %lu\n", "load",
147 rq->ls.load.weight);
148 P(ls.delta_fair);
149 P(ls.delta_exec);
150 P(nr_switches);
151 P(nr_load_updates);
152 P(nr_uninterruptible);
153 SEQ_printf(m, " .%-30s: %lu\n", "jiffies", jiffies);
154 P(next_balance);
155 P(curr->pid);
156 P(clock);
157 P(prev_clock_raw);
158 P(clock_warps);
159 P(clock_overflows);
160 P(clock_unstable_events);
161 P(clock_max_delta);
162 P(cpu_load[0]);
163 P(cpu_load[1]);
164 P(cpu_load[2]);
165 P(cpu_load[3]);
166 P(cpu_load[4]);
167#undef P
168
169 print_cfs_stats(m, cpu, now);
170
171 print_rq(m, rq, cpu, now);
172}
173
174static int sched_debug_show(struct seq_file *m, void *v)
175{
176 u64 now = ktime_to_ns(ktime_get());
177 int cpu;
178
6cfb0d5d 179 SEQ_printf(m, "Sched Debug Version: v0.05-v20, %s %.*s\n",
43ae34cb
IM
180 init_utsname()->release,
181 (int)strcspn(init_utsname()->version, " "),
182 init_utsname()->version);
183
184 SEQ_printf(m, "now at %Lu nsecs\n", (unsigned long long)now);
185
186 for_each_online_cpu(cpu)
187 print_cpu(m, cpu, now);
188
189 SEQ_printf(m, "\n");
190
191 return 0;
192}
193
f3373461 194static void sysrq_sched_debug_show(void)
43ae34cb
IM
195{
196 sched_debug_show(NULL, NULL);
197}
198
199static int sched_debug_open(struct inode *inode, struct file *filp)
200{
201 return single_open(filp, sched_debug_show, NULL);
202}
203
204static struct file_operations sched_debug_fops = {
205 .open = sched_debug_open,
206 .read = seq_read,
207 .llseek = seq_lseek,
5ea473a1 208 .release = single_release,
43ae34cb
IM
209};
210
211static int __init init_sched_debug_procfs(void)
212{
213 struct proc_dir_entry *pe;
214
215 pe = create_proc_entry("sched_debug", 0644, NULL);
216 if (!pe)
217 return -ENOMEM;
218
219 pe->proc_fops = &sched_debug_fops;
220
221 return 0;
222}
223
224__initcall(init_sched_debug_procfs);
225
226void proc_sched_show_task(struct task_struct *p, struct seq_file *m)
227{
228 unsigned long flags;
229 int num_threads = 1;
230
231 rcu_read_lock();
232 if (lock_task_sighand(p, &flags)) {
233 num_threads = atomic_read(&p->signal->count);
234 unlock_task_sighand(p, &flags);
235 }
236 rcu_read_unlock();
237
238 SEQ_printf(m, "%s (%d, #threads: %d)\n", p->comm, p->pid, num_threads);
239 SEQ_printf(m, "----------------------------------------------\n");
240#define P(F) \
241 SEQ_printf(m, "%-25s:%20Ld\n", #F, (long long)p->F)
242
6cfb0d5d 243 P(se.wait_runtime);
43ae34cb
IM
244 P(se.wait_start_fair);
245 P(se.exec_start);
43ae34cb 246 P(se.sleep_start_fair);
6cfb0d5d
IM
247 P(se.sum_exec_runtime);
248
249#ifdef CONFIG_SCHEDSTATS
250 P(se.wait_start);
251 P(se.sleep_start);
43ae34cb
IM
252 P(se.block_start);
253 P(se.sleep_max);
254 P(se.block_max);
255 P(se.exec_max);
256 P(se.wait_max);
43ae34cb
IM
257 P(se.wait_runtime_overruns);
258 P(se.wait_runtime_underruns);
259 P(se.sum_wait_runtime);
6cfb0d5d 260#endif
43ae34cb
IM
261 SEQ_printf(m, "%-25s:%20Ld\n",
262 "nr_switches", (long long)(p->nvcsw + p->nivcsw));
263 P(se.load.weight);
264 P(policy);
265 P(prio);
266#undef P
267
268 {
269 u64 t0, t1;
270
271 t0 = sched_clock();
272 t1 = sched_clock();
273 SEQ_printf(m, "%-25s:%20Ld\n",
274 "clock-delta", (long long)(t1-t0));
275 }
276}
277
278void proc_sched_set_task(struct task_struct *p)
279{
6cfb0d5d 280#ifdef CONFIG_SCHEDSTATS
43ae34cb
IM
281 p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0;
282 p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0;
6cfb0d5d 283#endif
43ae34cb
IM
284 p->se.sum_exec_runtime = 0;
285}