blob: 59cb603081a32b1ebe8f33b9ad11a6d3fd25837d [file] [log] [blame]
Ingo Molnar43ae34c2007-07-09 18:52:00 +02001/*
Peter Zijlstra391e43d2011-11-15 17:14:39 +01002 * kernel/sched/debug.c
Ingo Molnar43ae34c2007-07-09 18:52:00 +02003 *
4 * Print the CFS rbtree
5 *
6 * Copyright(C) 2007, Red Hat, Inc., Ingo Molnar
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
11 */
12
13#include <linux/proc_fs.h>
14#include <linux/sched.h>
15#include <linux/seq_file.h>
16#include <linux/kallsyms.h>
17#include <linux/utsname.h>
Ingo Molnarb32e86b2013-10-07 11:29:30 +010018#include <linux/mempolicy.h>
Ingo Molnar43ae34c2007-07-09 18:52:00 +020019
Peter Zijlstra029632f2011-10-25 10:00:11 +020020#include "sched.h"
21
Bharata B Raoefe25c22011-01-11 15:41:54 +053022static DEFINE_SPINLOCK(sched_debug_lock);
23
Ingo Molnar43ae34c2007-07-09 18:52:00 +020024/*
25 * This allows printing both to /proc/sched_debug and
26 * to the console
27 */
28#define SEQ_printf(m, x...) \
29 do { \
30 if (m) \
31 seq_printf(m, x); \
32 else \
33 printk(x); \
34 } while (0)
35
Ingo Molnaref83a572007-10-15 17:00:08 +020036/*
37 * Ease the printing of nsec fields:
38 */
Ingo Molnar90b26282007-12-30 17:24:35 +010039static long long nsec_high(unsigned long long nsec)
Ingo Molnaref83a572007-10-15 17:00:08 +020040{
Ingo Molnar90b26282007-12-30 17:24:35 +010041 if ((long long)nsec < 0) {
Ingo Molnaref83a572007-10-15 17:00:08 +020042 nsec = -nsec;
43 do_div(nsec, 1000000);
44 return -nsec;
45 }
46 do_div(nsec, 1000000);
47
48 return nsec;
49}
50
Ingo Molnar90b26282007-12-30 17:24:35 +010051static unsigned long nsec_low(unsigned long long nsec)
Ingo Molnaref83a572007-10-15 17:00:08 +020052{
Ingo Molnar90b26282007-12-30 17:24:35 +010053 if ((long long)nsec < 0)
Ingo Molnaref83a572007-10-15 17:00:08 +020054 nsec = -nsec;
55
56 return do_div(nsec, 1000000);
57}
58
59#define SPLIT_NS(x) nsec_high(x), nsec_low(x)
60
Bharata B Raoff9b48c2008-11-10 21:34:09 +053061#ifdef CONFIG_FAIR_GROUP_SCHED
Mike Galbraith5091faa2010-11-30 14:18:03 +010062static void print_cfs_group_stats(struct seq_file *m, int cpu, struct task_group *tg)
Bharata B Raoff9b48c2008-11-10 21:34:09 +053063{
64 struct sched_entity *se = tg->se[cpu];
Bharata B Raoff9b48c2008-11-10 21:34:09 +053065
66#define P(F) \
67 SEQ_printf(m, " .%-30s: %lld\n", #F, (long long)F)
68#define PN(F) \
69 SEQ_printf(m, " .%-30s: %lld.%06ld\n", #F, SPLIT_NS((long long)F))
70
Ben Segall18bf2802012-10-04 12:51:20 +020071 if (!se) {
72 struct sched_avg *avg = &cpu_rq(cpu)->avg;
73 P(avg->runnable_avg_sum);
Vincent Guittot36ee28e2015-02-27 16:54:04 +010074 P(avg->avg_period);
Ben Segall18bf2802012-10-04 12:51:20 +020075 return;
76 }
77
78
Bharata B Raoff9b48c2008-11-10 21:34:09 +053079 PN(se->exec_start);
80 PN(se->vruntime);
81 PN(se->sum_exec_runtime);
82#ifdef CONFIG_SCHEDSTATS
Lucas De Marchi41acab82010-03-10 23:37:45 -030083 PN(se->statistics.wait_start);
84 PN(se->statistics.sleep_start);
85 PN(se->statistics.block_start);
86 PN(se->statistics.sleep_max);
87 PN(se->statistics.block_max);
88 PN(se->statistics.exec_max);
89 PN(se->statistics.slice_max);
90 PN(se->statistics.wait_max);
91 PN(se->statistics.wait_sum);
92 P(se->statistics.wait_count);
Bharata B Raoff9b48c2008-11-10 21:34:09 +053093#endif
94 P(se->load.weight);
Paul Turner9d85f212012-10-04 13:18:29 +020095#ifdef CONFIG_SMP
96 P(se->avg.runnable_avg_sum);
Morten Rasmussen21f44862015-02-27 16:54:05 +010097 P(se->avg.running_avg_sum);
Vincent Guittot36ee28e2015-02-27 16:54:04 +010098 P(se->avg.avg_period);
Paul Turner2dac7542012-10-04 13:18:30 +020099 P(se->avg.load_avg_contrib);
Morten Rasmussen21f44862015-02-27 16:54:05 +0100100 P(se->avg.utilization_avg_contrib);
Paul Turner9ee474f2012-10-04 13:18:30 +0200101 P(se->avg.decay_count);
Paul Turner9d85f212012-10-04 13:18:29 +0200102#endif
Bharata B Raoff9b48c2008-11-10 21:34:09 +0530103#undef PN
104#undef P
105}
106#endif
107
Bharata B Raoefe25c22011-01-11 15:41:54 +0530108#ifdef CONFIG_CGROUP_SCHED
109static char group_path[PATH_MAX];
110
111static char *task_group_path(struct task_group *tg)
112{
Bharata B Rao8ecedd72011-01-11 15:42:57 +0530113 if (autogroup_path(tg, group_path, PATH_MAX))
114 return group_path;
115
Tejun Heoe61734c2014-02-12 09:29:50 -0500116 return cgroup_path(tg->css.cgroup, group_path, PATH_MAX);
Bharata B Raoefe25c22011-01-11 15:41:54 +0530117}
118#endif
119
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200120static void
Ingo Molnara48da482007-08-09 11:16:51 +0200121print_task(struct seq_file *m, struct rq *rq, struct task_struct *p)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200122{
123 if (rq->curr == p)
124 SEQ_printf(m, "R");
125 else
126 SEQ_printf(m, " ");
127
Ingo Molnaref83a572007-10-15 17:00:08 +0200128 SEQ_printf(m, "%15s %5d %9Ld.%06ld %9Ld %5d ",
Peter Zijlstrafc840912013-09-09 13:01:41 +0200129 p->comm, task_pid_nr(p),
Ingo Molnaref83a572007-10-15 17:00:08 +0200130 SPLIT_NS(p->se.vruntime),
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200131 (long long)(p->nvcsw + p->nivcsw),
Al Viro6f605d82007-08-06 04:26:59 +0100132 p->prio);
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200133#ifdef CONFIG_SCHEDSTATS
Peter Zijlstrad19ca302008-04-19 19:45:00 +0200134 SEQ_printf(m, "%9Ld.%06ld %9Ld.%06ld %9Ld.%06ld",
Ingo Molnaref83a572007-10-15 17:00:08 +0200135 SPLIT_NS(p->se.vruntime),
136 SPLIT_NS(p->se.sum_exec_runtime),
Lucas De Marchi41acab82010-03-10 23:37:45 -0300137 SPLIT_NS(p->se.statistics.sum_sleep_runtime));
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200138#else
Srikar Dronamraju33d61762015-06-08 13:40:39 +0530139 SEQ_printf(m, "%9Ld.%06ld %9Ld.%06ld %9Ld.%06ld",
140 0LL, 0L,
141 SPLIT_NS(p->se.sum_exec_runtime),
142 0LL, 0L);
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200143#endif
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100144#ifdef CONFIG_NUMA_BALANCING
Wanpeng Lide1b3012013-12-12 15:23:24 +0800145 SEQ_printf(m, " %d", task_node(p));
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100146#endif
Bharata B Raoefe25c22011-01-11 15:41:54 +0530147#ifdef CONFIG_CGROUP_SCHED
148 SEQ_printf(m, " %s", task_group_path(task_group(p)));
149#endif
Peter Zijlstrad19ca302008-04-19 19:45:00 +0200150
Peter Zijlstrad19ca302008-04-19 19:45:00 +0200151 SEQ_printf(m, "\n");
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200152}
153
Ingo Molnara48da482007-08-09 11:16:51 +0200154static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200155{
156 struct task_struct *g, *p;
157
158 SEQ_printf(m,
159 "\nrunnable tasks:\n"
Mike Galbraithc86da3a2007-10-15 17:00:08 +0200160 " task PID tree-key switches prio"
161 " exec-runtime sum-exec sum-sleep\n"
Ingo Molnar1a75b942007-10-15 17:00:08 +0200162 "------------------------------------------------------"
Mike Galbraithc86da3a2007-10-15 17:00:08 +0200163 "----------------------------------------------------\n");
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200164
Oleg Nesterov5bd96ab2014-09-21 21:33:41 +0200165 rcu_read_lock();
Oleg Nesterovd38e83c2014-08-13 21:19:56 +0200166 for_each_process_thread(g, p) {
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100167 if (task_cpu(p) != rq_cpu)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200168 continue;
169
Ingo Molnara48da482007-08-09 11:16:51 +0200170 print_task(m, rq, p);
Oleg Nesterovd38e83c2014-08-13 21:19:56 +0200171 }
Oleg Nesterov5bd96ab2014-09-21 21:33:41 +0200172 rcu_read_unlock();
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200173}
174
Ingo Molnar5cef9ec2007-08-09 11:16:47 +0200175void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200176{
Ingo Molnar86d95602007-10-15 17:00:06 +0200177 s64 MIN_vruntime = -1, min_vruntime, max_vruntime = -1,
178 spread, rq0_min_vruntime, spread0;
Hitoshi Mitake348ec612009-06-17 22:20:55 +0900179 struct rq *rq = cpu_rq(cpu);
Ingo Molnar67e12ea2007-10-15 17:00:05 +0200180 struct sched_entity *last;
181 unsigned long flags;
182
Bharata B Raoefe25c22011-01-11 15:41:54 +0530183#ifdef CONFIG_FAIR_GROUP_SCHED
184 SEQ_printf(m, "\ncfs_rq[%d]:%s\n", cpu, task_group_path(cfs_rq->tg));
185#else
Peter Zijlstraada18de2008-06-19 14:22:24 +0200186 SEQ_printf(m, "\ncfs_rq[%d]:\n", cpu);
Bharata B Raoefe25c22011-01-11 15:41:54 +0530187#endif
Ingo Molnaref83a572007-10-15 17:00:08 +0200188 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "exec_clock",
189 SPLIT_NS(cfs_rq->exec_clock));
Ingo Molnar67e12ea2007-10-15 17:00:05 +0200190
Thomas Gleixner05fa7852009-11-17 14:28:38 +0100191 raw_spin_lock_irqsave(&rq->lock, flags);
Ingo Molnar67e12ea2007-10-15 17:00:05 +0200192 if (cfs_rq->rb_leftmost)
Rik van Rielac53db52011-02-01 09:51:03 -0500193 MIN_vruntime = (__pick_first_entity(cfs_rq))->vruntime;
Ingo Molnar67e12ea2007-10-15 17:00:05 +0200194 last = __pick_last_entity(cfs_rq);
195 if (last)
196 max_vruntime = last->vruntime;
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100197 min_vruntime = cfs_rq->min_vruntime;
Hitoshi Mitake348ec612009-06-17 22:20:55 +0900198 rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime;
Thomas Gleixner05fa7852009-11-17 14:28:38 +0100199 raw_spin_unlock_irqrestore(&rq->lock, flags);
Ingo Molnaref83a572007-10-15 17:00:08 +0200200 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "MIN_vruntime",
201 SPLIT_NS(MIN_vruntime));
202 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "min_vruntime",
203 SPLIT_NS(min_vruntime));
204 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "max_vruntime",
205 SPLIT_NS(max_vruntime));
Ingo Molnar67e12ea2007-10-15 17:00:05 +0200206 spread = max_vruntime - MIN_vruntime;
Ingo Molnaref83a572007-10-15 17:00:08 +0200207 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "spread",
208 SPLIT_NS(spread));
Ingo Molnar86d95602007-10-15 17:00:06 +0200209 spread0 = min_vruntime - rq0_min_vruntime;
Ingo Molnaref83a572007-10-15 17:00:08 +0200210 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "spread0",
211 SPLIT_NS(spread0));
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100212 SEQ_printf(m, " .%-30s: %d\n", "nr_spread_over",
Peter Zijlstraddc97292007-10-15 17:00:10 +0200213 cfs_rq->nr_spread_over);
Peter Zijlstrac82513e2012-04-26 13:12:27 +0200214 SEQ_printf(m, " .%-30s: %d\n", "nr_running", cfs_rq->nr_running);
Peter Zijlstra2069dd72010-11-15 15:47:00 -0800215 SEQ_printf(m, " .%-30s: %ld\n", "load", cfs_rq->load.weight);
Peter Zijlstrac09595f2008-06-27 13:41:14 +0200216#ifdef CONFIG_SMP
Alex Shi72a4cf22013-06-20 10:18:53 +0800217 SEQ_printf(m, " .%-30s: %ld\n", "runnable_load_avg",
Paul Turner2dac7542012-10-04 13:18:30 +0200218 cfs_rq->runnable_load_avg);
Alex Shi72a4cf22013-06-20 10:18:53 +0800219 SEQ_printf(m, " .%-30s: %ld\n", "blocked_load_avg",
Paul Turner9ee474f2012-10-04 13:18:30 +0200220 cfs_rq->blocked_load_avg);
Vincent Guittot36ee28e2015-02-27 16:54:04 +0100221 SEQ_printf(m, " .%-30s: %ld\n", "utilization_load_avg",
222 cfs_rq->utilization_load_avg);
Alex Shi333bb862013-06-28 19:10:35 +0800223#ifdef CONFIG_FAIR_GROUP_SCHED
Alex Shibf5b9862013-06-20 10:18:54 +0800224 SEQ_printf(m, " .%-30s: %ld\n", "tg_load_contrib",
Paul Turnerc566e8e2012-10-04 13:18:30 +0200225 cfs_rq->tg_load_contrib);
Paul Turnerbb17f652012-10-04 13:18:31 +0200226 SEQ_printf(m, " .%-30s: %d\n", "tg_runnable_contrib",
227 cfs_rq->tg_runnable_contrib);
Alex Shi333bb862013-06-28 19:10:35 +0800228 SEQ_printf(m, " .%-30s: %ld\n", "tg_load_avg",
229 atomic_long_read(&cfs_rq->tg->load_avg));
Paul Turnerbb17f652012-10-04 13:18:31 +0200230 SEQ_printf(m, " .%-30s: %d\n", "tg->runnable_avg",
231 atomic_read(&cfs_rq->tg->runnable_avg));
Peter Zijlstrac09595f2008-06-27 13:41:14 +0200232#endif
Alex Shi333bb862013-06-28 19:10:35 +0800233#endif
Ben Segallf9f9ffc2013-10-16 11:16:32 -0700234#ifdef CONFIG_CFS_BANDWIDTH
235 SEQ_printf(m, " .%-30s: %d\n", "tg->cfs_bandwidth.timer_active",
236 cfs_rq->tg->cfs_bandwidth.timer_active);
237 SEQ_printf(m, " .%-30s: %d\n", "throttled",
238 cfs_rq->throttled);
239 SEQ_printf(m, " .%-30s: %d\n", "throttle_count",
240 cfs_rq->throttle_count);
241#endif
Peter Zijlstra2069dd72010-11-15 15:47:00 -0800242
Alex Shi333bb862013-06-28 19:10:35 +0800243#ifdef CONFIG_FAIR_GROUP_SCHED
Bharata B Raoff9b48c2008-11-10 21:34:09 +0530244 print_cfs_group_stats(m, cpu, cfs_rq->tg);
Peter Zijlstrac09595f2008-06-27 13:41:14 +0200245#endif
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200246}
247
Peter Zijlstraada18de2008-06-19 14:22:24 +0200248void print_rt_rq(struct seq_file *m, int cpu, struct rt_rq *rt_rq)
249{
Bharata B Raoefe25c22011-01-11 15:41:54 +0530250#ifdef CONFIG_RT_GROUP_SCHED
251 SEQ_printf(m, "\nrt_rq[%d]:%s\n", cpu, task_group_path(rt_rq->tg));
252#else
Peter Zijlstraada18de2008-06-19 14:22:24 +0200253 SEQ_printf(m, "\nrt_rq[%d]:\n", cpu);
Bharata B Raoefe25c22011-01-11 15:41:54 +0530254#endif
Peter Zijlstraada18de2008-06-19 14:22:24 +0200255
256#define P(x) \
257 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rt_rq->x))
258#define PN(x) \
259 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rt_rq->x))
260
261 P(rt_nr_running);
262 P(rt_throttled);
263 PN(rt_time);
264 PN(rt_runtime);
265
266#undef PN
267#undef P
268}
269
Wanpeng Liacb32132014-10-31 06:39:33 +0800270void print_dl_rq(struct seq_file *m, int cpu, struct dl_rq *dl_rq)
271{
272 SEQ_printf(m, "\ndl_rq[%d]:\n", cpu);
273 SEQ_printf(m, " .%-30s: %ld\n", "dl_nr_running", dl_rq->dl_nr_running);
274}
275
Peter Zijlstra5bb6b1e2010-11-19 21:11:09 +0100276extern __read_mostly int sched_clock_running;
277
Ingo Molnara48da482007-08-09 11:16:51 +0200278static void print_cpu(struct seq_file *m, int cpu)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200279{
Hitoshi Mitake348ec612009-06-17 22:20:55 +0900280 struct rq *rq = cpu_rq(cpu);
Bharata B Raoefe25c22011-01-11 15:41:54 +0530281 unsigned long flags;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200282
283#ifdef CONFIG_X86
284 {
285 unsigned int freq = cpu_khz ? : 1;
286
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800287 SEQ_printf(m, "cpu#%d, %u.%03u MHz\n",
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200288 cpu, freq / 1000, (freq % 1000));
289 }
290#else
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800291 SEQ_printf(m, "cpu#%d\n", cpu);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200292#endif
293
Peter Zijlstra13e099d2012-05-14 14:34:00 +0200294#define P(x) \
295do { \
296 if (sizeof(rq->x) == 4) \
297 SEQ_printf(m, " .%-30s: %ld\n", #x, (long)(rq->x)); \
298 else \
299 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x));\
300} while (0)
301
Ingo Molnaref83a572007-10-15 17:00:08 +0200302#define PN(x) \
303 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rq->x))
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200304
305 P(nr_running);
306 SEQ_printf(m, " .%-30s: %lu\n", "load",
Dmitry Adamushko495eca42007-10-15 17:00:06 +0200307 rq->load.weight);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200308 P(nr_switches);
309 P(nr_load_updates);
310 P(nr_uninterruptible);
Ingo Molnaref83a572007-10-15 17:00:08 +0200311 PN(next_balance);
Peter Zijlstrafc840912013-09-09 13:01:41 +0200312 SEQ_printf(m, " .%-30s: %ld\n", "curr->pid", (long)(task_pid_nr(rq->curr)));
Ingo Molnaref83a572007-10-15 17:00:08 +0200313 PN(clock);
Peter Zijlstra5a537592015-01-05 11:18:12 +0100314 PN(clock_task);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200315 P(cpu_load[0]);
316 P(cpu_load[1]);
317 P(cpu_load[2]);
318 P(cpu_load[3]);
319 P(cpu_load[4]);
320#undef P
Ingo Molnaref83a572007-10-15 17:00:08 +0200321#undef PN
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200322
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100323#ifdef CONFIG_SCHEDSTATS
324#define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, rq->n);
Mike Galbraith1b9508f2009-11-04 17:53:50 +0100325#define P64(n) SEQ_printf(m, " .%-30s: %Ld\n", #n, rq->n);
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100326
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100327 P(yld_count);
328
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100329 P(sched_count);
330 P(sched_goidle);
Mike Galbraith1b9508f2009-11-04 17:53:50 +0100331#ifdef CONFIG_SMP
332 P64(avg_idle);
Alex Shi37e6bae2014-01-23 18:39:54 +0800333 P64(max_idle_balance_cost);
Mike Galbraith1b9508f2009-11-04 17:53:50 +0100334#endif
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100335
336 P(ttwu_count);
337 P(ttwu_local);
338
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100339#undef P
Yong Zhangfce20972011-01-14 15:57:39 +0800340#undef P64
Peter Zijlstra5ac5c4d2008-11-10 10:46:32 +0100341#endif
Bharata B Raoefe25c22011-01-11 15:41:54 +0530342 spin_lock_irqsave(&sched_debug_lock, flags);
Ingo Molnar5cef9ec2007-08-09 11:16:47 +0200343 print_cfs_stats(m, cpu);
Peter Zijlstraada18de2008-06-19 14:22:24 +0200344 print_rt_stats(m, cpu);
Wanpeng Liacb32132014-10-31 06:39:33 +0800345 print_dl_stats(m, cpu);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200346
Ingo Molnara48da482007-08-09 11:16:51 +0200347 print_rq(m, rq, cpu);
Bharata B Raoefe25c22011-01-11 15:41:54 +0530348 spin_unlock_irqrestore(&sched_debug_lock, flags);
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800349 SEQ_printf(m, "\n");
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200350}
351
Christian Ehrhardt1983a922009-11-30 12:16:47 +0100352static const char *sched_tunable_scaling_names[] = {
353 "none",
354 "logaritmic",
355 "linear"
356};
357
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800358static void sched_debug_header(struct seq_file *m)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200359{
Peter Zijlstra5bb6b1e2010-11-19 21:11:09 +0100360 u64 ktime, sched_clk, cpu_clk;
361 unsigned long flags;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200362
Peter Zijlstra5bb6b1e2010-11-19 21:11:09 +0100363 local_irq_save(flags);
364 ktime = ktime_to_ns(ktime_get());
365 sched_clk = sched_clock();
366 cpu_clk = local_clock();
367 local_irq_restore(flags);
368
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100369 SEQ_printf(m, "Sched Debug Version: v0.11, %s %.*s\n",
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200370 init_utsname()->release,
371 (int)strcspn(init_utsname()->version, " "),
372 init_utsname()->version);
373
Peter Zijlstra5bb6b1e2010-11-19 21:11:09 +0100374#define P(x) \
375 SEQ_printf(m, "%-40s: %Ld\n", #x, (long long)(x))
376#define PN(x) \
377 SEQ_printf(m, "%-40s: %Ld.%06ld\n", #x, SPLIT_NS(x))
378 PN(ktime);
379 PN(sched_clk);
380 PN(cpu_clk);
381 P(jiffies);
382#ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
Peter Zijlstra35af99e2013-11-28 19:38:42 +0100383 P(sched_clock_stable());
Peter Zijlstra5bb6b1e2010-11-19 21:11:09 +0100384#endif
385#undef PN
386#undef P
387
388 SEQ_printf(m, "\n");
389 SEQ_printf(m, "sysctl_sched\n");
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200390
Ingo Molnar1aa47312007-10-15 17:00:10 +0200391#define P(x) \
Ingo Molnard822cec2007-10-15 17:00:10 +0200392 SEQ_printf(m, " .%-40s: %Ld\n", #x, (long long)(x))
Ingo Molnar1aa47312007-10-15 17:00:10 +0200393#define PN(x) \
Ingo Molnard822cec2007-10-15 17:00:10 +0200394 SEQ_printf(m, " .%-40s: %Ld.%06ld\n", #x, SPLIT_NS(x))
Ingo Molnar1aa47312007-10-15 17:00:10 +0200395 PN(sysctl_sched_latency);
Peter Zijlstrab2be5e92007-11-09 22:39:37 +0100396 PN(sysctl_sched_min_granularity);
Ingo Molnar1aa47312007-10-15 17:00:10 +0200397 PN(sysctl_sched_wakeup_granularity);
Josh Hunteebef742010-07-19 12:31:16 -0700398 P(sysctl_sched_child_runs_first);
Ingo Molnar1aa47312007-10-15 17:00:10 +0200399 P(sysctl_sched_features);
400#undef PN
401#undef P
402
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800403 SEQ_printf(m, " .%-40s: %d (%s)\n",
404 "sysctl_sched_tunable_scaling",
Christian Ehrhardt1983a922009-11-30 12:16:47 +0100405 sysctl_sched_tunable_scaling,
406 sched_tunable_scaling_names[sysctl_sched_tunable_scaling]);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200407 SEQ_printf(m, "\n");
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800408}
409
410static int sched_debug_show(struct seq_file *m, void *v)
411{
412 int cpu = (unsigned long)(v - 2);
413
414 if (cpu != -1)
415 print_cpu(m, cpu);
416 else
417 sched_debug_header(m);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200418
419 return 0;
420}
421
Peter Zijlstra029632f2011-10-25 10:00:11 +0200422void sysrq_sched_debug_show(void)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200423{
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800424 int cpu;
425
426 sched_debug_header(NULL);
427 for_each_online_cpu(cpu)
428 print_cpu(NULL, cpu);
429
430}
431
432/*
433 * This itererator needs some explanation.
434 * It returns 1 for the header position.
435 * This means 2 is cpu 0.
436 * In a hotplugged system some cpus, including cpu 0, may be missing so we have
437 * to use cpumask_* to iterate over the cpus.
438 */
439static void *sched_debug_start(struct seq_file *file, loff_t *offset)
440{
441 unsigned long n = *offset;
442
443 if (n == 0)
444 return (void *) 1;
445
446 n--;
447
448 if (n > 0)
449 n = cpumask_next(n - 1, cpu_online_mask);
450 else
451 n = cpumask_first(cpu_online_mask);
452
453 *offset = n + 1;
454
455 if (n < nr_cpu_ids)
456 return (void *)(unsigned long)(n + 2);
457 return NULL;
458}
459
460static void *sched_debug_next(struct seq_file *file, void *data, loff_t *offset)
461{
462 (*offset)++;
463 return sched_debug_start(file, offset);
464}
465
466static void sched_debug_stop(struct seq_file *file, void *data)
467{
468}
469
470static const struct seq_operations sched_debug_sops = {
471 .start = sched_debug_start,
472 .next = sched_debug_next,
473 .stop = sched_debug_stop,
474 .show = sched_debug_show,
475};
476
477static int sched_debug_release(struct inode *inode, struct file *file)
478{
479 seq_release(inode, file);
480
481 return 0;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200482}
483
484static int sched_debug_open(struct inode *inode, struct file *filp)
485{
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800486 int ret = 0;
487
488 ret = seq_open(filp, &sched_debug_sops);
489
490 return ret;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200491}
492
Arjan van de Ven0dbee3a2007-10-15 17:00:19 +0200493static const struct file_operations sched_debug_fops = {
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200494 .open = sched_debug_open,
495 .read = seq_read,
496 .llseek = seq_lseek,
Nathan Zimmerbbbfeac2013-02-21 15:15:09 -0800497 .release = sched_debug_release,
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200498};
499
500static int __init init_sched_debug_procfs(void)
501{
502 struct proc_dir_entry *pe;
503
Li Zefana9cf4dd2008-10-30 15:23:34 +0800504 pe = proc_create("sched_debug", 0444, NULL, &sched_debug_fops);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200505 if (!pe)
506 return -ENOMEM;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200507 return 0;
508}
509
510__initcall(init_sched_debug_procfs);
511
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100512#define __P(F) \
513 SEQ_printf(m, "%-45s:%21Ld\n", #F, (long long)F)
514#define P(F) \
515 SEQ_printf(m, "%-45s:%21Ld\n", #F, (long long)p->F)
516#define __PN(F) \
517 SEQ_printf(m, "%-45s:%14Ld.%06ld\n", #F, SPLIT_NS((long long)F))
518#define PN(F) \
519 SEQ_printf(m, "%-45s:%14Ld.%06ld\n", #F, SPLIT_NS((long long)p->F))
520
521
522static void sched_show_numa(struct task_struct *p, struct seq_file *m)
523{
524#ifdef CONFIG_NUMA_BALANCING
525 struct mempolicy *pol;
526 int node, i;
527
528 if (p->mm)
529 P(mm->numa_scan_seq);
530
531 task_lock(p);
532 pol = p->mempolicy;
533 if (pol && !(pol->flags & MPOL_F_MORON))
534 pol = NULL;
535 mpol_get(pol);
536 task_unlock(p);
537
538 SEQ_printf(m, "numa_migrations, %ld\n", xchg(&p->numa_pages_migrated, 0));
539
540 for_each_online_node(node) {
541 for (i = 0; i < 2; i++) {
542 unsigned long nr_faults = -1;
543 int cpu_current, home_node;
544
Iulia Manda44dba3d2014-10-31 02:13:31 +0200545 if (p->numa_faults)
546 nr_faults = p->numa_faults[2*node + i];
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100547
548 cpu_current = !i ? (task_node(p) == node) :
549 (pol && node_isset(node, pol->v.nodes));
550
551 home_node = (p->numa_preferred_nid == node);
552
Rik van Rielff1df892014-01-27 17:03:41 -0500553 SEQ_printf(m, "numa_faults_memory, %d, %d, %d, %d, %ld\n",
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100554 i, node, cpu_current, home_node, nr_faults);
555 }
556 }
557
558 mpol_put(pol);
559#endif
560}
561
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200562void proc_sched_show_task(struct task_struct *p, struct seq_file *m)
563{
Ingo Molnarcc367732007-10-15 17:00:18 +0200564 unsigned long nr_switches;
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200565
Peter Zijlstrafc840912013-09-09 13:01:41 +0200566 SEQ_printf(m, "%s (%d, #threads: %d)\n", p->comm, task_pid_nr(p),
Oleg Nesterov5089a972010-05-26 14:43:22 -0700567 get_nr_threads(p));
Ingo Molnar2d92f222007-10-15 17:00:18 +0200568 SEQ_printf(m,
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530569 "---------------------------------------------------------"
570 "----------\n");
Ingo Molnarcc367732007-10-15 17:00:18 +0200571#define __P(F) \
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530572 SEQ_printf(m, "%-45s:%21Ld\n", #F, (long long)F)
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200573#define P(F) \
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530574 SEQ_printf(m, "%-45s:%21Ld\n", #F, (long long)p->F)
Ingo Molnarcc367732007-10-15 17:00:18 +0200575#define __PN(F) \
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530576 SEQ_printf(m, "%-45s:%14Ld.%06ld\n", #F, SPLIT_NS((long long)F))
Ingo Molnaref83a572007-10-15 17:00:08 +0200577#define PN(F) \
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530578 SEQ_printf(m, "%-45s:%14Ld.%06ld\n", #F, SPLIT_NS((long long)p->F))
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200579
Ingo Molnaref83a572007-10-15 17:00:08 +0200580 PN(se.exec_start);
581 PN(se.vruntime);
582 PN(se.sum_exec_runtime);
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200583
Ingo Molnarcc367732007-10-15 17:00:18 +0200584 nr_switches = p->nvcsw + p->nivcsw;
585
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200586#ifdef CONFIG_SCHEDSTATS
Lucas De Marchi41acab82010-03-10 23:37:45 -0300587 PN(se.statistics.wait_start);
588 PN(se.statistics.sleep_start);
589 PN(se.statistics.block_start);
590 PN(se.statistics.sleep_max);
591 PN(se.statistics.block_max);
592 PN(se.statistics.exec_max);
593 PN(se.statistics.slice_max);
594 PN(se.statistics.wait_max);
595 PN(se.statistics.wait_sum);
596 P(se.statistics.wait_count);
597 PN(se.statistics.iowait_sum);
598 P(se.statistics.iowait_count);
Ingo Molnarcc367732007-10-15 17:00:18 +0200599 P(se.nr_migrations);
Lucas De Marchi41acab82010-03-10 23:37:45 -0300600 P(se.statistics.nr_migrations_cold);
601 P(se.statistics.nr_failed_migrations_affine);
602 P(se.statistics.nr_failed_migrations_running);
603 P(se.statistics.nr_failed_migrations_hot);
604 P(se.statistics.nr_forced_migrations);
605 P(se.statistics.nr_wakeups);
606 P(se.statistics.nr_wakeups_sync);
607 P(se.statistics.nr_wakeups_migrate);
608 P(se.statistics.nr_wakeups_local);
609 P(se.statistics.nr_wakeups_remote);
610 P(se.statistics.nr_wakeups_affine);
611 P(se.statistics.nr_wakeups_affine_attempts);
612 P(se.statistics.nr_wakeups_passive);
613 P(se.statistics.nr_wakeups_idle);
Ingo Molnarcc367732007-10-15 17:00:18 +0200614
615 {
616 u64 avg_atom, avg_per_cpu;
617
618 avg_atom = p->se.sum_exec_runtime;
619 if (nr_switches)
Mateusz Guzikb0ab99e2014-06-14 15:00:09 +0200620 avg_atom = div64_ul(avg_atom, nr_switches);
Ingo Molnarcc367732007-10-15 17:00:18 +0200621 else
622 avg_atom = -1LL;
623
624 avg_per_cpu = p->se.sum_exec_runtime;
Ingo Molnarc1a897402007-11-28 15:52:56 +0100625 if (p->se.nr_migrations) {
Roman Zippel6f6d6a12008-05-01 04:34:28 -0700626 avg_per_cpu = div64_u64(avg_per_cpu,
627 p->se.nr_migrations);
Ingo Molnarc1a897402007-11-28 15:52:56 +0100628 } else {
Ingo Molnarcc367732007-10-15 17:00:18 +0200629 avg_per_cpu = -1LL;
Ingo Molnarc1a897402007-11-28 15:52:56 +0100630 }
Ingo Molnarcc367732007-10-15 17:00:18 +0200631
632 __PN(avg_atom);
633 __PN(avg_per_cpu);
634 }
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200635#endif
Ingo Molnarcc367732007-10-15 17:00:18 +0200636 __P(nr_switches);
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530637 SEQ_printf(m, "%-45s:%21Ld\n",
Ingo Molnarcc367732007-10-15 17:00:18 +0200638 "nr_voluntary_switches", (long long)p->nvcsw);
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530639 SEQ_printf(m, "%-45s:%21Ld\n",
Ingo Molnarcc367732007-10-15 17:00:18 +0200640 "nr_involuntary_switches", (long long)p->nivcsw);
641
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200642 P(se.load.weight);
Alex Shi333bb862013-06-28 19:10:35 +0800643#ifdef CONFIG_SMP
Kamalesh Babulal939fd732013-06-25 13:33:36 +0530644 P(se.avg.runnable_avg_sum);
Vincent Guittot36ee28e2015-02-27 16:54:04 +0100645 P(se.avg.running_avg_sum);
646 P(se.avg.avg_period);
Kamalesh Babulal939fd732013-06-25 13:33:36 +0530647 P(se.avg.load_avg_contrib);
Vincent Guittot36ee28e2015-02-27 16:54:04 +0100648 P(se.avg.utilization_avg_contrib);
Kamalesh Babulal939fd732013-06-25 13:33:36 +0530649 P(se.avg.decay_count);
650#endif
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200651 P(policy);
652 P(prio);
Ingo Molnaref83a572007-10-15 17:00:08 +0200653#undef PN
Ingo Molnarcc367732007-10-15 17:00:18 +0200654#undef __PN
655#undef P
656#undef __P
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200657
658 {
Ingo Molnar29d7b902008-11-16 08:07:15 +0100659 unsigned int this_cpu = raw_smp_processor_id();
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200660 u64 t0, t1;
661
Ingo Molnar29d7b902008-11-16 08:07:15 +0100662 t0 = cpu_clock(this_cpu);
663 t1 = cpu_clock(this_cpu);
Kamalesh Babulaladd332a2013-06-27 22:20:05 +0530664 SEQ_printf(m, "%-45s:%21Ld\n",
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200665 "clock-delta", (long long)(t1-t0));
666 }
Ingo Molnarb32e86b2013-10-07 11:29:30 +0100667
668 sched_show_numa(p, m);
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200669}
670
671void proc_sched_set_task(struct task_struct *p)
672{
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200673#ifdef CONFIG_SCHEDSTATS
Lucas De Marchi41acab82010-03-10 23:37:45 -0300674 memset(&p->se.statistics, 0, sizeof(p->se.statistics));
Ingo Molnar6cfb0d52007-08-02 17:41:40 +0200675#endif
Ingo Molnar43ae34c2007-07-09 18:52:00 +0200676}