diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/sched/core.c | 22 | ||||
| -rw-r--r-- | kernel/sched/cputime.c | 2 | ||||
| -rw-r--r-- | kernel/sched/debug.c | 90 | ||||
| -rw-r--r-- | kernel/sched/stats.c | 79 | 
4 files changed, 141 insertions, 52 deletions
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index fc9103e9ff0..2b5243176ab 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -1979,11 +1979,10 @@ context_switch(struct rq *rq, struct task_struct *prev,  }  /* - * nr_running, nr_uninterruptible and nr_context_switches: + * nr_running and nr_context_switches:   *   * externally visible scheduler statistics: current number of runnable - * threads, current number of uninterruptible-sleeping threads, total - * number of context switches performed since bootup. + * threads, total number of context switches performed since bootup.   */  unsigned long nr_running(void)  { @@ -1995,23 +1994,6 @@ unsigned long nr_running(void)  	return sum;  } -unsigned long nr_uninterruptible(void) -{ -	unsigned long i, sum = 0; - -	for_each_possible_cpu(i) -		sum += cpu_rq(i)->nr_uninterruptible; - -	/* -	 * Since we read the counters lockless, it might be slightly -	 * inaccurate. Do not allow it to go below zero though: -	 */ -	if (unlikely((long)sum < 0)) -		sum = 0; - -	return sum; -} -  unsigned long long nr_context_switches(void)  {  	int i; diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c index 9857329ed28..ed12cbb135f 100644 --- a/kernel/sched/cputime.c +++ b/kernel/sched/cputime.c @@ -604,7 +604,7 @@ static unsigned long long vtime_delta(struct task_struct *tsk)  {  	unsigned long long clock; -	clock = sched_clock(); +	clock = local_clock();  	if (clock < tsk->vtime_snap)  		return 0; diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c index 557e7b53b32..75024a67352 100644 --- a/kernel/sched/debug.c +++ b/kernel/sched/debug.c @@ -262,11 +262,11 @@ static void print_cpu(struct seq_file *m, int cpu)  	{  		unsigned int freq = cpu_khz ? : 1; -		SEQ_printf(m, "\ncpu#%d, %u.%03u MHz\n", +		SEQ_printf(m, "cpu#%d, %u.%03u MHz\n",  			   cpu, freq / 1000, (freq % 1000));  	}  #else -	SEQ_printf(m, "\ncpu#%d\n", cpu); +	SEQ_printf(m, "cpu#%d\n", cpu);  #endif  #define P(x)								\ @@ -323,6 +323,7 @@ do {									\  	print_rq(m, rq, cpu);  	rcu_read_unlock();  	spin_unlock_irqrestore(&sched_debug_lock, flags); +	SEQ_printf(m, "\n");  }  static const char *sched_tunable_scaling_names[] = { @@ -331,11 +332,10 @@ static const char *sched_tunable_scaling_names[] = {  	"linear"  }; -static int sched_debug_show(struct seq_file *m, void *v) +static void sched_debug_header(struct seq_file *m)  {  	u64 ktime, sched_clk, cpu_clk;  	unsigned long flags; -	int cpu;  	local_irq_save(flags);  	ktime = ktime_to_ns(ktime_get()); @@ -377,33 +377,101 @@ static int sched_debug_show(struct seq_file *m, void *v)  #undef PN  #undef P -	SEQ_printf(m, "  .%-40s: %d (%s)\n", "sysctl_sched_tunable_scaling", +	SEQ_printf(m, "  .%-40s: %d (%s)\n", +		"sysctl_sched_tunable_scaling",  		sysctl_sched_tunable_scaling,  		sched_tunable_scaling_names[sysctl_sched_tunable_scaling]); +	SEQ_printf(m, "\n"); +} -	for_each_online_cpu(cpu) -		print_cpu(m, cpu); +static int sched_debug_show(struct seq_file *m, void *v) +{ +	int cpu = (unsigned long)(v - 2); -	SEQ_printf(m, "\n"); +	if (cpu != -1) +		print_cpu(m, cpu); +	else +		sched_debug_header(m);  	return 0;  }  void sysrq_sched_debug_show(void)  { -	sched_debug_show(NULL, NULL); +	int cpu; + +	sched_debug_header(NULL); +	for_each_online_cpu(cpu) +		print_cpu(NULL, cpu); + +} + +/* + * This itererator needs some explanation. + * It returns 1 for the header position. + * This means 2 is cpu 0. + * In a hotplugged system some cpus, including cpu 0, may be missing so we have + * to use cpumask_* to iterate over the cpus. + */ +static void *sched_debug_start(struct seq_file *file, loff_t *offset) +{ +	unsigned long n = *offset; + +	if (n == 0) +		return (void *) 1; + +	n--; + +	if (n > 0) +		n = cpumask_next(n - 1, cpu_online_mask); +	else +		n = cpumask_first(cpu_online_mask); + +	*offset = n + 1; + +	if (n < nr_cpu_ids) +		return (void *)(unsigned long)(n + 2); +	return NULL; +} + +static void *sched_debug_next(struct seq_file *file, void *data, loff_t *offset) +{ +	(*offset)++; +	return sched_debug_start(file, offset); +} + +static void sched_debug_stop(struct seq_file *file, void *data) +{ +} + +static const struct seq_operations sched_debug_sops = { +	.start = sched_debug_start, +	.next = sched_debug_next, +	.stop = sched_debug_stop, +	.show = sched_debug_show, +}; + +static int sched_debug_release(struct inode *inode, struct file *file) +{ +	seq_release(inode, file); + +	return 0;  }  static int sched_debug_open(struct inode *inode, struct file *filp)  { -	return single_open(filp, sched_debug_show, NULL); +	int ret = 0; + +	ret = seq_open(filp, &sched_debug_sops); + +	return ret;  }  static const struct file_operations sched_debug_fops = {  	.open		= sched_debug_open,  	.read		= seq_read,  	.llseek		= seq_lseek, -	.release	= single_release, +	.release	= sched_debug_release,  };  static int __init init_sched_debug_procfs(void) diff --git a/kernel/sched/stats.c b/kernel/sched/stats.c index 903ffa9e887..e036eda1a9c 100644 --- a/kernel/sched/stats.c +++ b/kernel/sched/stats.c @@ -21,14 +21,17 @@ static int show_schedstat(struct seq_file *seq, void *v)  	if (mask_str == NULL)  		return -ENOMEM; -	seq_printf(seq, "version %d\n", SCHEDSTAT_VERSION); -	seq_printf(seq, "timestamp %lu\n", jiffies); -	for_each_online_cpu(cpu) { -		struct rq *rq = cpu_rq(cpu); +	if (v == (void *)1) { +		seq_printf(seq, "version %d\n", SCHEDSTAT_VERSION); +		seq_printf(seq, "timestamp %lu\n", jiffies); +	} else { +		struct rq *rq;  #ifdef CONFIG_SMP  		struct sched_domain *sd;  		int dcount = 0;  #endif +		cpu = (unsigned long)(v - 2); +		rq = cpu_rq(cpu);  		/* runqueue-specific stats */  		seq_printf(seq, @@ -77,30 +80,66 @@ static int show_schedstat(struct seq_file *seq, void *v)  	return 0;  } -static int schedstat_open(struct inode *inode, struct file *file) +/* + * This itererator needs some explanation. + * It returns 1 for the header position. + * This means 2 is cpu 0. + * In a hotplugged system some cpus, including cpu 0, may be missing so we have + * to use cpumask_* to iterate over the cpus. + */ +static void *schedstat_start(struct seq_file *file, loff_t *offset)  { -	unsigned int size = PAGE_SIZE * (1 + num_online_cpus() / 32); -	char *buf = kmalloc(size, GFP_KERNEL); -	struct seq_file *m; -	int res; +	unsigned long n = *offset; -	if (!buf) -		return -ENOMEM; -	res = single_open(file, show_schedstat, NULL); -	if (!res) { -		m = file->private_data; -		m->buf = buf; -		m->size = size; -	} else -		kfree(buf); -	return res; +	if (n == 0) +		return (void *) 1; + +	n--; + +	if (n > 0) +		n = cpumask_next(n - 1, cpu_online_mask); +	else +		n = cpumask_first(cpu_online_mask); + +	*offset = n + 1; + +	if (n < nr_cpu_ids) +		return (void *)(unsigned long)(n + 2); +	return NULL; +} + +static void *schedstat_next(struct seq_file *file, void *data, loff_t *offset) +{ +	(*offset)++; +	return schedstat_start(file, offset); +} + +static void schedstat_stop(struct seq_file *file, void *data) +{ +} + +static const struct seq_operations schedstat_sops = { +	.start = schedstat_start, +	.next  = schedstat_next, +	.stop  = schedstat_stop, +	.show  = show_schedstat, +}; + +static int schedstat_open(struct inode *inode, struct file *file) +{ +	return seq_open(file, &schedstat_sops);  } +static int schedstat_release(struct inode *inode, struct file *file) +{ +	return 0; +}; +  static const struct file_operations proc_schedstat_operations = {  	.open    = schedstat_open,  	.read    = seq_read,  	.llseek  = seq_lseek, -	.release = single_release, +	.release = schedstat_release,  };  static int __init proc_schedstat_init(void)  |