sched: call update_curr() in task_tick_fair()
authorTing Yang <tingy@cs.umass.edu>
Tue, 28 Aug 2007 10:53:24 +0000 (12:53 +0200)
committerIngo Molnar <mingo@elte.hu>
Tue, 28 Aug 2007 10:53:24 +0000 (12:53 +0200)
update the fair-clock before using it for the key value.

[ mingo@elte.hu: small cleanups. ]

Signed-off-by: Ting Yang <tingy@cs.umass.edu>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
kernel/sched_fair.c

index 721fe77448740b922050e9e207f5e9bdffed2878..9f06094e527576658ba338ee1906092d07029d49 100644 (file)
@@ -1094,10 +1094,11 @@ static void task_tick_fair(struct rq *rq, struct task_struct *curr)
 static void task_new_fair(struct rq *rq, struct task_struct *p)
 {
        struct cfs_rq *cfs_rq = task_cfs_rq(p);
-       struct sched_entity *se = &p->se;
+       struct sched_entity *se = &p->se, *curr = cfs_rq_curr(cfs_rq);
 
        sched_info_queued(p);
 
+       update_curr(cfs_rq);
        update_stats_enqueue(cfs_rq, se);
        /*
         * Child runs first: we let it run before the parent
@@ -1105,7 +1106,7 @@ static void task_new_fair(struct rq *rq, struct task_struct *p)
         * it will preempt the parent:
         */
        p->se.fair_key = current->se.fair_key -
-               niced_granularity(&rq->curr->se, sched_granularity(cfs_rq)) - 1;
+               niced_granularity(curr, sched_granularity(cfs_rq)) - 1;
        /*
         * The first wait is dominated by the child-runs-first logic,
         * so do not credit it with that waiting time yet: