summaryrefslogtreecommitdiff
path: root/kernel/sched_debug.c
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2008-05-03 18:29:28 +0200
committerIngo Molnar <mingo@elte.hu>2008-05-05 23:56:18 +0200
commit3e51f33fcc7f55e6df25d15b55ed10c8b4da84cd (patch)
tree3752f9ea8e014ec40e95a1b197b0a3d18e1056a8 /kernel/sched_debug.c
parenta5574cf65b5f03ce9ade3918764fe22e5e2371e3 (diff)
downloadkernel-common-3e51f33fcc7f55e6df25d15b55ed10c8b4da84cd.tar.gz
kernel-common-3e51f33fcc7f55e6df25d15b55ed10c8b4da84cd.tar.bz2
kernel-common-3e51f33fcc7f55e6df25d15b55ed10c8b4da84cd.zip
sched: add optional support for CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
this replaces the rq->clock stuff (and possibly cpu_clock()). - architectures that have an 'imperfect' hardware clock can set CONFIG_HAVE_UNSTABLE_SCHED_CLOCK - the 'jiffie' window might be superfulous when we update tick_gtod before the __update_sched_clock() call in sched_clock_tick() - cpu_clock() might be implemented as: sched_clock_cpu(smp_processor_id()) if the accuracy proves good enough - how far can TSC drift in a single jiffie when considering the filtering and idle hooks? [ mingo@elte.hu: various fixes and cleanups ] Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_debug.c')
-rw-r--r--kernel/sched_debug.c7
1 files changed, 0 insertions, 7 deletions
diff --git a/kernel/sched_debug.c b/kernel/sched_debug.c
index 6b4a12558e88..5f06118fbc31 100644
--- a/kernel/sched_debug.c
+++ b/kernel/sched_debug.c
@@ -204,13 +204,6 @@ static void print_cpu(struct seq_file *m, int cpu)
PN(next_balance);
P(curr->pid);
PN(clock);
- PN(idle_clock);
- PN(prev_clock_raw);
- P(clock_warps);
- P(clock_overflows);
- P(clock_underflows);
- P(clock_deep_idle_events);
- PN(clock_max_delta);
P(cpu_load[0]);
P(cpu_load[1]);
P(cpu_load[2]);