summaryrefslogtreecommitdiffstats
path: root/kernel/sched_fair.c
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2009-09-11 12:45:38 +0200
committerIngo Molnar <mingo@elte.hu>2009-09-15 16:01:07 +0200
commitd7c33c4930f569caf6b2ece597432853c4151a45 (patch)
tree96a0eca96b0109a0dbc2c7a92065cafcf5660718 /kernel/sched_fair.c
parentsched: for_each_domain() vs RCU (diff)
downloadlinux-d7c33c4930f569caf6b2ece597432853c4151a45.tar.xz
linux-d7c33c4930f569caf6b2ece597432853c4151a45.zip
sched: Fix task affinity for select_task_rq_fair
While merging select_task_rq_fair() and sched_balance_self() I made a mistake that leads to testing the wrong task affinty. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> LKML-Reference: <new-submission> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_fair.c')
-rw-r--r--kernel/sched_fair.c5
1 files changed, 2 insertions, 3 deletions
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index 43dc6d1d9e88..8b3eddbcf9a4 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -1318,7 +1318,6 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
*/
static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
{
- struct task_struct *t = current;
struct sched_domain *tmp, *sd = NULL;
int cpu = smp_processor_id();
int prev_cpu = task_cpu(p);
@@ -1393,13 +1392,13 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
continue;
}
- group = find_idlest_group(sd, t, cpu);
+ group = find_idlest_group(sd, p, cpu);
if (!group) {
sd = sd->child;
continue;
}
- new_cpu = find_idlest_cpu(group, t, cpu);
+ new_cpu = find_idlest_cpu(group, p, cpu);
if (new_cpu == -1 || new_cpu == cpu) {
/* Now try balancing at a lower domain level of cpu */
sd = sd->child;