diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/async.c | 43 |
1 files changed, 41 insertions, 2 deletions
diff --git a/kernel/async.c b/kernel/async.c index ba5491dfa991..9d3118384858 100644 --- a/kernel/async.c +++ b/kernel/async.c | |||
| @@ -63,7 +63,9 @@ static async_cookie_t next_cookie = 1; | |||
| 63 | 63 | ||
| 64 | static LIST_HEAD(async_pending); | 64 | static LIST_HEAD(async_pending); |
| 65 | static ASYNC_DOMAIN(async_running); | 65 | static ASYNC_DOMAIN(async_running); |
| 66 | static LIST_HEAD(async_domains); | ||
| 66 | static DEFINE_SPINLOCK(async_lock); | 67 | static DEFINE_SPINLOCK(async_lock); |
| 68 | static DEFINE_MUTEX(async_register_mutex); | ||
| 67 | 69 | ||
| 68 | struct async_entry { | 70 | struct async_entry { |
| 69 | struct list_head list; | 71 | struct list_head list; |
| @@ -145,6 +147,8 @@ static void async_run_entry_fn(struct work_struct *work) | |||
| 145 | /* 3) remove self from the running queue */ | 147 | /* 3) remove self from the running queue */ |
| 146 | spin_lock_irqsave(&async_lock, flags); | 148 | spin_lock_irqsave(&async_lock, flags); |
| 147 | list_del(&entry->list); | 149 | list_del(&entry->list); |
| 150 | if (running->registered && --running->count == 0) | ||
| 151 | list_del_init(&running->node); | ||
| 148 | 152 | ||
| 149 | /* 4) free the entry */ | 153 | /* 4) free the entry */ |
| 150 | kfree(entry); | 154 | kfree(entry); |
| @@ -187,6 +191,8 @@ static async_cookie_t __async_schedule(async_func_ptr *ptr, void *data, struct a | |||
| 187 | spin_lock_irqsave(&async_lock, flags); | 191 | spin_lock_irqsave(&async_lock, flags); |
| 188 | newcookie = entry->cookie = next_cookie++; | 192 | newcookie = entry->cookie = next_cookie++; |
| 189 | list_add_tail(&entry->list, &async_pending); | 193 | list_add_tail(&entry->list, &async_pending); |
| 194 | if (running->registered && running->count++ == 0) | ||
| 195 | list_add_tail(&running->node, &async_domains); | ||
| 190 | atomic_inc(&entry_count); | 196 | atomic_inc(&entry_count); |
| 191 | spin_unlock_irqrestore(&async_lock, flags); | 197 | spin_unlock_irqrestore(&async_lock, flags); |
| 192 | 198 | ||
| @@ -236,13 +242,43 @@ EXPORT_SYMBOL_GPL(async_schedule_domain); | |||
| 236 | */ | 242 | */ |
| 237 | void async_synchronize_full(void) | 243 | void async_synchronize_full(void) |
| 238 | { | 244 | { |
| 245 | mutex_lock(&async_register_mutex); | ||
| 239 | do { | 246 | do { |
| 240 | async_synchronize_cookie(next_cookie); | 247 | struct async_domain *domain = NULL; |
| 241 | } while (!list_empty(&async_running.domain) || !list_empty(&async_pending)); | 248 | |
| 249 | spin_lock_irq(&async_lock); | ||
| 250 | if (!list_empty(&async_domains)) | ||
| 251 | domain = list_first_entry(&async_domains, typeof(*domain), node); | ||
| 252 | spin_unlock_irq(&async_lock); | ||
| 253 | |||
| 254 | async_synchronize_cookie_domain(next_cookie, domain); | ||
| 255 | } while (!list_empty(&async_domains)); | ||
| 256 | mutex_unlock(&async_register_mutex); | ||
| 242 | } | 257 | } |
| 243 | EXPORT_SYMBOL_GPL(async_synchronize_full); | 258 | EXPORT_SYMBOL_GPL(async_synchronize_full); |
| 244 | 259 | ||
| 245 | /** | 260 | /** |
| 261 | * async_unregister_domain - ensure no more anonymous waiters on this domain | ||
| 262 | * @domain: idle domain to flush out of any async_synchronize_full instances | ||
| 263 | * | ||
| 264 | * async_synchronize_{cookie|full}_domain() are not flushed since callers | ||
| 265 | * of these routines should know the lifetime of @domain | ||
| 266 | * | ||
| 267 | * Prefer ASYNC_DOMAIN_EXCLUSIVE() declarations over flushing | ||
| 268 | */ | ||
| 269 | void async_unregister_domain(struct async_domain *domain) | ||
| 270 | { | ||
| 271 | mutex_lock(&async_register_mutex); | ||
| 272 | spin_lock_irq(&async_lock); | ||
| 273 | WARN_ON(!domain->registered || !list_empty(&domain->node) || | ||
| 274 | !list_empty(&domain->domain)); | ||
| 275 | domain->registered = 0; | ||
| 276 | spin_unlock_irq(&async_lock); | ||
| 277 | mutex_unlock(&async_register_mutex); | ||
| 278 | } | ||
| 279 | EXPORT_SYMBOL_GPL(async_unregister_domain); | ||
| 280 | |||
| 281 | /** | ||
| 246 | * async_synchronize_full_domain - synchronize all asynchronous function within a certain domain | 282 | * async_synchronize_full_domain - synchronize all asynchronous function within a certain domain |
| 247 | * @domain: running list to synchronize on | 283 | * @domain: running list to synchronize on |
| 248 | * | 284 | * |
| @@ -268,6 +304,9 @@ void async_synchronize_cookie_domain(async_cookie_t cookie, struct async_domain | |||
| 268 | { | 304 | { |
| 269 | ktime_t uninitialized_var(starttime), delta, endtime; | 305 | ktime_t uninitialized_var(starttime), delta, endtime; |
| 270 | 306 | ||
| 307 | if (!running) | ||
| 308 | return; | ||
| 309 | |||
| 271 | if (initcall_debug && system_state == SYSTEM_BOOTING) { | 310 | if (initcall_debug && system_state == SYSTEM_BOOTING) { |
| 272 | printk(KERN_DEBUG "async_waiting @ %i\n", task_pid_nr(current)); | 311 | printk(KERN_DEBUG "async_waiting @ %i\n", task_pid_nr(current)); |
| 273 | starttime = ktime_get(); | 312 | starttime = ktime_get(); |
