diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2008-06-27 13:41:19 +0200 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-06-27 14:31:33 +0200 |
commit | 4d8d595dfa69e1c807bf928f364668a7f30da5dc (patch) | |
tree | af61c1d6d53aea66fac272e7dad67ae93a832a66 /kernel/sched_fair.c | |
parent | b6a86c746f5b708012809958462234d19e9c8177 (diff) | |
download | kernel_samsung_smdk4412-4d8d595dfa69e1c807bf928f364668a7f30da5dc.zip kernel_samsung_smdk4412-4d8d595dfa69e1c807bf928f364668a7f30da5dc.tar.gz kernel_samsung_smdk4412-4d8d595dfa69e1c807bf928f364668a7f30da5dc.tar.bz2 |
sched: update aggregate when holding the RQs
It was observed that in __update_group_shares_cpu()
rq_weight > aggregate()->rq_weight
This is caused by forks/wakeups in between the initial aggregate pass and
locking of the RQs for load balance. To avoid this situation partially re-do
the aggregation once we have the RQs locked (which avoids new tasks from
appearing).
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Cc: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_fair.c')
0 files changed, 0 insertions, 0 deletions