summaryrefslogtreecommitdiff
path: root/include/linux/raid/raid1.h
diff options
context:
space:
mode:
authorAnton Blanchard <anton@samba.org>2006-03-23 13:59:20 +0300
committerLinus Torvalds <torvalds@g5.osdl.org>2006-03-23 18:38:02 +0300
commite9028b0ff2bad1954568604dc17725692c8524d6 (patch)
treedbff742d39520574e5985930333f1d421282e080 /include/linux/raid/raid1.h
parent5be0e9511990dc307670dc66a42073db96b20f26 (diff)
downloadlinux-e9028b0ff2bad1954568604dc17725692c8524d6.tar.xz
[PATCH] fix scheduler deadlock
We have noticed lockups during boot when stress testing kexec on ppc64. Two cpus would deadlock in scheduler code trying to grab already taken spinlocks. The double_rq_lock code uses the address of the runqueue to order the taking of multiple locks. This address is a per cpu variable: if (rq1 < rq2) { spin_lock(&rq1->lock); spin_lock(&rq2->lock); } else { spin_lock(&rq2->lock); spin_lock(&rq1->lock); } On the other hand, the code in wake_sleeping_dependent uses the cpu id order to grab locks: for_each_cpu_mask(i, sibling_map) spin_lock(&cpu_rq(i)->lock); This means we rely on the address of per cpu data increasing as cpu ids increase. While this will be true for the generic percpu implementation it may not be true for arch specific implementations. One way to solve this is to always take runqueues in cpu id order. To do this we add a cpu variable to the runqueue and check it in the double runqueue locking functions. Signed-off-by: Anton Blanchard <anton@samba.org> Acked-by: Ingo Molnar <mingo@elte.hu> Cc: <stable@kernel.org> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'include/linux/raid/raid1.h')
0 files changed, 0 insertions, 0 deletions