diff options
author | Tejun Heo <tj@kernel.org> | 2013-03-12 14:30:03 -0400 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2013-03-12 14:30:03 -0400 |
commit | f02ae73aaa4f285199683862ac59972877a11c5d (patch) | |
tree | ca517312e1ed5bfaff9b18ebefcd8faf064ad040 /kernel/workqueue.c | |
parent | 7a62c2c87e3bc174fe4b9e9720e148427510fcfb (diff) |
workqueue: drop "std" from cpu_std_worker_pools and for_each_std_worker_pool()
All per-cpu pools are standard, so there's no need to use both "cpu"
and "std" and for_each_std_worker_pool() is confusing in that it can
be used only for per-cpu pools.
* s/cpu_std_worker_pools/cpu_worker_pools/
* s/for_each_std_worker_pool()/for_each_cpu_worker_pool()/
Signed-off-by: Tejun Heo <tj@kernel.org>
Reviewed-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Diffstat (limited to 'kernel/workqueue.c')
-rw-r--r-- | kernel/workqueue.c | 22 |
1 files changed, 11 insertions, 11 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 7642bb7b70ee..2c5073214774 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
@@ -252,9 +252,9 @@ EXPORT_SYMBOL_GPL(system_freezable_wq); | |||
252 | lockdep_is_held(&workqueue_lock), \ | 252 | lockdep_is_held(&workqueue_lock), \ |
253 | "sched RCU or workqueue lock should be held") | 253 | "sched RCU or workqueue lock should be held") |
254 | 254 | ||
255 | #define for_each_std_worker_pool(pool, cpu) \ | 255 | #define for_each_cpu_worker_pool(pool, cpu) \ |
256 | for ((pool) = &per_cpu(cpu_std_worker_pools, cpu)[0]; \ | 256 | for ((pool) = &per_cpu(cpu_worker_pools, cpu)[0]; \ |
257 | (pool) < &per_cpu(cpu_std_worker_pools, cpu)[NR_STD_WORKER_POOLS]; \ | 257 | (pool) < &per_cpu(cpu_worker_pools, cpu)[NR_STD_WORKER_POOLS]; \ |
258 | (pool)++) | 258 | (pool)++) |
259 | 259 | ||
260 | #define for_each_busy_worker(worker, i, pool) \ | 260 | #define for_each_busy_worker(worker, i, pool) \ |
@@ -420,7 +420,7 @@ static bool workqueue_freezing; /* W: have wqs started freezing? */ | |||
420 | * POOL_DISASSOCIATED set, and their workers have WORKER_UNBOUND set. | 420 | * POOL_DISASSOCIATED set, and their workers have WORKER_UNBOUND set. |
421 | */ | 421 | */ |
422 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct worker_pool [NR_STD_WORKER_POOLS], | 422 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct worker_pool [NR_STD_WORKER_POOLS], |
423 | cpu_std_worker_pools); | 423 | cpu_worker_pools); |
424 | 424 | ||
425 | /* | 425 | /* |
426 | * idr of all pools. Modifications are protected by workqueue_lock. Read | 426 | * idr of all pools. Modifications are protected by workqueue_lock. Read |
@@ -3342,7 +3342,7 @@ static int alloc_and_link_pwqs(struct workqueue_struct *wq) | |||
3342 | struct pool_workqueue *pwq = | 3342 | struct pool_workqueue *pwq = |
3343 | per_cpu_ptr(wq->cpu_pwqs, cpu); | 3343 | per_cpu_ptr(wq->cpu_pwqs, cpu); |
3344 | struct worker_pool *cpu_pools = | 3344 | struct worker_pool *cpu_pools = |
3345 | per_cpu(cpu_std_worker_pools, cpu); | 3345 | per_cpu(cpu_worker_pools, cpu); |
3346 | 3346 | ||
3347 | pwq->pool = &cpu_pools[highpri]; | 3347 | pwq->pool = &cpu_pools[highpri]; |
3348 | list_add_tail_rcu(&pwq->pwqs_node, &wq->pwqs); | 3348 | list_add_tail_rcu(&pwq->pwqs_node, &wq->pwqs); |
@@ -3694,7 +3694,7 @@ static void wq_unbind_fn(struct work_struct *work) | |||
3694 | struct worker *worker; | 3694 | struct worker *worker; |
3695 | int i; | 3695 | int i; |
3696 | 3696 | ||
3697 | for_each_std_worker_pool(pool, cpu) { | 3697 | for_each_cpu_worker_pool(pool, cpu) { |
3698 | WARN_ON_ONCE(cpu != smp_processor_id()); | 3698 | WARN_ON_ONCE(cpu != smp_processor_id()); |
3699 | 3699 | ||
3700 | mutex_lock(&pool->assoc_mutex); | 3700 | mutex_lock(&pool->assoc_mutex); |
@@ -3737,7 +3737,7 @@ static void wq_unbind_fn(struct work_struct *work) | |||
3737 | * unbound chain execution of pending work items if other workers | 3737 | * unbound chain execution of pending work items if other workers |
3738 | * didn't already. | 3738 | * didn't already. |
3739 | */ | 3739 | */ |
3740 | for_each_std_worker_pool(pool, cpu) | 3740 | for_each_cpu_worker_pool(pool, cpu) |
3741 | atomic_set(&pool->nr_running, 0); | 3741 | atomic_set(&pool->nr_running, 0); |
3742 | } | 3742 | } |
3743 | 3743 | ||
@@ -3754,7 +3754,7 @@ static int __cpuinit workqueue_cpu_up_callback(struct notifier_block *nfb, | |||
3754 | 3754 | ||
3755 | switch (action & ~CPU_TASKS_FROZEN) { | 3755 | switch (action & ~CPU_TASKS_FROZEN) { |
3756 | case CPU_UP_PREPARE: | 3756 | case CPU_UP_PREPARE: |
3757 | for_each_std_worker_pool(pool, cpu) { | 3757 | for_each_cpu_worker_pool(pool, cpu) { |
3758 | struct worker *worker; | 3758 | struct worker *worker; |
3759 | 3759 | ||
3760 | if (pool->nr_workers) | 3760 | if (pool->nr_workers) |
@@ -3772,7 +3772,7 @@ static int __cpuinit workqueue_cpu_up_callback(struct notifier_block *nfb, | |||
3772 | 3772 | ||
3773 | case CPU_DOWN_FAILED: | 3773 | case CPU_DOWN_FAILED: |
3774 | case CPU_ONLINE: | 3774 | case CPU_ONLINE: |
3775 | for_each_std_worker_pool(pool, cpu) { | 3775 | for_each_cpu_worker_pool(pool, cpu) { |
3776 | mutex_lock(&pool->assoc_mutex); | 3776 | mutex_lock(&pool->assoc_mutex); |
3777 | spin_lock_irq(&pool->lock); | 3777 | spin_lock_irq(&pool->lock); |
3778 | 3778 | ||
@@ -4012,7 +4012,7 @@ static int __init init_workqueues(void) | |||
4012 | struct worker_pool *pool; | 4012 | struct worker_pool *pool; |
4013 | 4013 | ||
4014 | i = 0; | 4014 | i = 0; |
4015 | for_each_std_worker_pool(pool, cpu) { | 4015 | for_each_cpu_worker_pool(pool, cpu) { |
4016 | BUG_ON(init_worker_pool(pool)); | 4016 | BUG_ON(init_worker_pool(pool)); |
4017 | pool->cpu = cpu; | 4017 | pool->cpu = cpu; |
4018 | cpumask_copy(pool->attrs->cpumask, cpumask_of(cpu)); | 4018 | cpumask_copy(pool->attrs->cpumask, cpumask_of(cpu)); |
@@ -4027,7 +4027,7 @@ static int __init init_workqueues(void) | |||
4027 | for_each_online_cpu(cpu) { | 4027 | for_each_online_cpu(cpu) { |
4028 | struct worker_pool *pool; | 4028 | struct worker_pool *pool; |
4029 | 4029 | ||
4030 | for_each_std_worker_pool(pool, cpu) { | 4030 | for_each_cpu_worker_pool(pool, cpu) { |
4031 | struct worker *worker; | 4031 | struct worker *worker; |
4032 | 4032 | ||
4033 | pool->flags &= ~POOL_DISASSOCIATED; | 4033 | pool->flags &= ~POOL_DISASSOCIATED; |