From 43b82306d4f97e803271a001a90f550a721e7ddb Mon Sep 17 00:00:00 2001 From: Ingo Molnar Date: Tue, 28 Aug 2007 12:53:24 +0200 Subject: [PATCH] --- yaml --- r: 64656 b: refs/heads/master c: 213c8af67f21c1dc0d50940b159d9521c95f3c89 h: refs/heads/master v: v3 --- [refs] | 2 +- trunk/kernel/sched_fair.c | 4 +++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/[refs] b/[refs] index dc1bbbd56341..1b4e09911798 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: b77d69db9f4ba03b2ed17e383c2d73ca89f5ab14 +refs/heads/master: 213c8af67f21c1dc0d50940b159d9521c95f3c89 diff --git a/trunk/kernel/sched_fair.c b/trunk/kernel/sched_fair.c index 0c718857176f..75f025da6f7c 100644 --- a/trunk/kernel/sched_fair.c +++ b/trunk/kernel/sched_fair.c @@ -1121,8 +1121,10 @@ static void task_new_fair(struct rq *rq, struct task_struct *p) * The statistical average of wait_runtime is about * -granularity/2, so initialize the task with that: */ - if (sysctl_sched_features & SCHED_FEAT_START_DEBIT) + if (sysctl_sched_features & SCHED_FEAT_START_DEBIT) { p->se.wait_runtime = -(sched_granularity(cfs_rq) / 2); + schedstat_add(cfs_rq, wait_runtime, se->wait_runtime); + } __enqueue_entity(cfs_rq, se); }