From: Ingo Molnar Date: Mon, 18 Dec 2006 19:05:09 +0000 (+0100) Subject: [PATCH] workqueue: fix schedule_on_each_cpu() X-Git-Tag: MMI-PSA29.97-13-9~45805 X-Git-Url: https://git.stricted.de/?a=commitdiff_plain;h=9bfb18392ef586467277fa25d8f3a7a93611f6df;p=GitHub%2FMotorolaMobilityLLC%2Fkernel-slsi.git [PATCH] workqueue: fix schedule_on_each_cpu() fix the schedule_on_each_cpu() implementation: __queue_work() is now stricter, hence set the work-pending bit before passing in the new work. (found in the -rt tree, using Peter Zijlstra's files-lock scalability patchset) Signed-off-by: Ingo Molnar Signed-off-by: Linus Torvalds --- diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 742cbbe49bdc..180a8ce11535 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -637,9 +637,11 @@ int schedule_on_each_cpu(work_func_t func) mutex_lock(&workqueue_mutex); for_each_online_cpu(cpu) { - INIT_WORK(per_cpu_ptr(works, cpu), func); - __queue_work(per_cpu_ptr(keventd_wq->cpu_wq, cpu), - per_cpu_ptr(works, cpu)); + struct work_struct *work = per_cpu_ptr(works, cpu); + + INIT_WORK(work, func); + set_bit(WORK_STRUCT_PENDING, work_data_bits(work)); + __queue_work(per_cpu_ptr(keventd_wq->cpu_wq, cpu), work); } mutex_unlock(&workqueue_mutex); flush_workqueue(keventd_wq);