[PATCH 09/10] workqueue: separate pool-binding code out from create_worker()
The code of pool-binding is unfolded in create_worker(). Separating this code out will make the codes more clear. Signed-off-by: Lai Jiangshan --- kernel/workqueue.c | 53 --- 1 files changed, 33 insertions(+), 20 deletions(-) diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 9e0e606..0c575b1 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -66,7 +66,7 @@ enum { * * Note that DISASSOCIATED should be flipped only while holding * bind_mutex to avoid changing binding state while -* create_worker() is in progress. +* worker_bind_pool() is in progress. */ POOL_DISASSOCIATED = 1 << 2, /* cpu can't serve workers */ POOL_FREEZING = 1 << 3, /* freeze in progress */ @@ -1668,6 +1668,37 @@ static struct worker *alloc_worker(void) } /** + * worker_bind_pool() - bind the worker to the pool + * @worker: worker to be bound + * @pool: the target pool + * + * bind the worker to the pool, thus the concurrency and cpu-binding of + * the worker are kept coordination with the pool across cpu-[un]hotplug. + */ +static void worker_bind_pool(struct worker *worker, struct worker_pool *pool) +{ + mutex_lock(>bind_mutex); + + /* +* set_cpus_allowed_ptr() will fail if the cpumask doesn't have any +* online CPUs. It'll be re-applied when any of the CPUs come up. +*/ + set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); + + /* +* The pool->bind_mutex ensures %POOL_DISASSOCIATED remains +* stable across this function. See the comments above the +* flag definition for details. +*/ + if (pool->flags & POOL_DISASSOCIATED) + worker->flags |= WORKER_UNBOUND; + + list_add_tail(>bind_entry, >bind_list); + + mutex_unlock(>bind_mutex); +} + +/** * worker_unbind_pool() - unbind the worker from the pool * @worker: worker which is bound to its pool * @@ -1731,26 +1762,8 @@ static struct worker *create_worker(struct worker_pool *pool) /* prevent userland from meddling with cpumask of workqueue workers */ worker->task->flags |= PF_NO_SETAFFINITY; - mutex_lock(>bind_mutex); - - /* -* set_cpus_allowed_ptr() will fail if the cpumask doesn't have any -* online CPUs. It'll be re-applied when any of the CPUs come up. -*/ - set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); - - /* -* The pool->bind_mutex ensures %POOL_DISASSOCIATED -* remains stable across this function. See the comments above the -* flag definition for details. -*/ - if (pool->flags & POOL_DISASSOCIATED) - worker->flags |= WORKER_UNBOUND; - /* successful, bind the worker to the pool */ - list_add_tail(>bind_entry, >bind_list); - - mutex_unlock(>bind_mutex); + worker_bind_pool(worker, pool); return worker; -- 1.7.4.4 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
[PATCH 09/10] workqueue: separate pool-binding code out from create_worker()
The code of pool-binding is unfolded in create_worker(). Separating this code out will make the codes more clear. Signed-off-by: Lai Jiangshan la...@cn.fujitsu.com --- kernel/workqueue.c | 53 --- 1 files changed, 33 insertions(+), 20 deletions(-) diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 9e0e606..0c575b1 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -66,7 +66,7 @@ enum { * * Note that DISASSOCIATED should be flipped only while holding * bind_mutex to avoid changing binding state while -* create_worker() is in progress. +* worker_bind_pool() is in progress. */ POOL_DISASSOCIATED = 1 2, /* cpu can't serve workers */ POOL_FREEZING = 1 3, /* freeze in progress */ @@ -1668,6 +1668,37 @@ static struct worker *alloc_worker(void) } /** + * worker_bind_pool() - bind the worker to the pool + * @worker: worker to be bound + * @pool: the target pool + * + * bind the worker to the pool, thus the concurrency and cpu-binding of + * the worker are kept coordination with the pool across cpu-[un]hotplug. + */ +static void worker_bind_pool(struct worker *worker, struct worker_pool *pool) +{ + mutex_lock(pool-bind_mutex); + + /* +* set_cpus_allowed_ptr() will fail if the cpumask doesn't have any +* online CPUs. It'll be re-applied when any of the CPUs come up. +*/ + set_cpus_allowed_ptr(worker-task, pool-attrs-cpumask); + + /* +* The pool-bind_mutex ensures %POOL_DISASSOCIATED remains +* stable across this function. See the comments above the +* flag definition for details. +*/ + if (pool-flags POOL_DISASSOCIATED) + worker-flags |= WORKER_UNBOUND; + + list_add_tail(worker-bind_entry, pool-bind_list); + + mutex_unlock(pool-bind_mutex); +} + +/** * worker_unbind_pool() - unbind the worker from the pool * @worker: worker which is bound to its pool * @@ -1731,26 +1762,8 @@ static struct worker *create_worker(struct worker_pool *pool) /* prevent userland from meddling with cpumask of workqueue workers */ worker-task-flags |= PF_NO_SETAFFINITY; - mutex_lock(pool-bind_mutex); - - /* -* set_cpus_allowed_ptr() will fail if the cpumask doesn't have any -* online CPUs. It'll be re-applied when any of the CPUs come up. -*/ - set_cpus_allowed_ptr(worker-task, pool-attrs-cpumask); - - /* -* The pool-bind_mutex ensures %POOL_DISASSOCIATED -* remains stable across this function. See the comments above the -* flag definition for details. -*/ - if (pool-flags POOL_DISASSOCIATED) - worker-flags |= WORKER_UNBOUND; - /* successful, bind the worker to the pool */ - list_add_tail(worker-bind_entry, pool-bind_list); - - mutex_unlock(pool-bind_mutex); + worker_bind_pool(worker, pool); return worker; -- 1.7.4.4 -- To unsubscribe from this list: send the line unsubscribe linux-kernel in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/