Hello, Lai.

On Tue, Aug 28, 2012 at 07:34:37PM +0800, Lai Jiangshan wrote:
> So this implement adds an "all_done", thus rebind_workers() can't leave until
> idle_worker_rebind() successful wait something until all other idle also done,
> so this "wait something" will not cause deadlock as the old code.
> 
> The code of "idle_worker_rebind() wait something until all other idle also 
> done"
> is also changed. It is changed to wait on "worker_rebind.idle_done". With
> the help of "all_done" this "worker_rebind" is valid when they wait on
> "worker_rebind.idle_done".
> 
> The busy_worker_rebind_fn() also explicitly wait on all idle done. It adds
> a small overhead on cold path, but it makes the rebind_workers() as single 
> pass.
> Clean Code/Readability wins.
> 
> "all_cnt" 's decreasing is done without any lock, because this decreasing
> only happens on the bound CPU, no lock needed. (the bound CPU can't go
> until we notify on "all_done")

I really hope the fix and reimplementation are done in separate steps.
All we need is an additional completion wait before leaving
rebind_workers()(), so let's please just add that to fix the immediate
bug.  If you think it can be further simplified by reimplmenting
rebind_workers(), that's great but let's please do that as a separate
step.

Also, I asked this previously but have you encounted this problem or
is it only from code review?

>  static void idle_worker_rebind(struct worker *worker)
>  {
> -     struct global_cwq *gcwq = worker->pool->gcwq;
> -
>       /* CPU must be online at this point */
>       WARN_ON(!worker_maybe_bind_and_lock(worker));
> -     if (!--worker->idle_rebind->cnt)
> -             complete(&worker->idle_rebind->done);
> +     worker_clr_flags(worker, WORKER_REBIND);
> +     if (!--worker->worker_rebind->idle_cnt)
> +             complete_all(&worker->worker_rebind->idle_done);
>       spin_unlock_irq(&worker->pool->gcwq->lock);
>  
> -     /* we did our part, wait for rebind_workers() to finish up */
> -     wait_event(gcwq->rebind_hold, !(worker->flags & WORKER_REBIND));
> +     /* It did its part, wait for all other idle to finish up */
> +     wait_for_completion(&worker->worker_rebind->idle_done);
> +
> +     /* all_cnt is only accessed by the bound CPU, don't need any lock */
> +     if (!--worker->worker_rebind->all_cnt)

What if this worker gets preempted by another worker?  There's no
point in this type of optimization in this path.  Let's please keep
things straight-forward and robust.

>  static void busy_worker_rebind_fn(struct work_struct *work)
>  {
>       struct worker *worker = container_of(work, struct worker, rebind_work);
>       struct global_cwq *gcwq = worker->pool->gcwq;
>  
> -     if (worker_maybe_bind_and_lock(worker))
> -             worker_clr_flags(worker, WORKER_REBIND);
> +     /* Wait for all idle to finish up */
> +     wait_for_completion(&worker->worker_rebind->idle_done);
>  
> +     /* CPU must be online at this point */
> +     WARN_ON(!worker_maybe_bind_and_lock(worker));
> +     worker_clr_flags(worker, WORKER_REBIND);
>       spin_unlock_irq(&gcwq->lock);
> +
> +     /* all_cnt is only accessed by the bound CPU, don't need any lock */
> +     if (!--worker->worker_rebind->all_cnt)
> +             complete(&worker->worker_rebind->all_done);

You can't do this.  There is no guarantee that busy worker will
execute busy_worker_rebind_fn() in definite amount of time.  A given
work item may run indefinitely.

Let's please fix the discovered issue first.

Thanks.

-- 
tejun
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to