diff options
Diffstat (limited to 'kernel/kthread.c')
-rw-r--r-- | kernel/kthread.c | 149 |
1 files changed, 149 insertions, 0 deletions
diff --git a/kernel/kthread.c b/kernel/kthread.c index 83911c780175..8b63c7fee73b 100644 --- a/kernel/kthread.c +++ b/kernel/kthread.c | |||
@@ -14,6 +14,8 @@ | |||
14 | #include <linux/file.h> | 14 | #include <linux/file.h> |
15 | #include <linux/module.h> | 15 | #include <linux/module.h> |
16 | #include <linux/mutex.h> | 16 | #include <linux/mutex.h> |
17 | #include <linux/slab.h> | ||
18 | #include <linux/freezer.h> | ||
17 | #include <trace/events/sched.h> | 19 | #include <trace/events/sched.h> |
18 | 20 | ||
19 | static DEFINE_SPINLOCK(kthread_create_lock); | 21 | static DEFINE_SPINLOCK(kthread_create_lock); |
@@ -247,3 +249,150 @@ int kthreadd(void *unused) | |||
247 | 249 | ||
248 | return 0; | 250 | return 0; |
249 | } | 251 | } |
252 | |||
253 | /** | ||
254 | * kthread_worker_fn - kthread function to process kthread_worker | ||
255 | * @worker_ptr: pointer to initialized kthread_worker | ||
256 | * | ||
257 | * This function can be used as @threadfn to kthread_create() or | ||
258 | * kthread_run() with @worker_ptr argument pointing to an initialized | ||
259 | * kthread_worker. The started kthread will process work_list until | ||
260 | * the it is stopped with kthread_stop(). A kthread can also call | ||
261 | * this function directly after extra initialization. | ||
262 | * | ||
263 | * Different kthreads can be used for the same kthread_worker as long | ||
264 | * as there's only one kthread attached to it at any given time. A | ||
265 | * kthread_worker without an attached kthread simply collects queued | ||
266 | * kthread_works. | ||
267 | */ | ||
268 | int kthread_worker_fn(void *worker_ptr) | ||
269 | { | ||
270 | struct kthread_worker *worker = worker_ptr; | ||
271 | struct kthread_work *work; | ||
272 | |||
273 | WARN_ON(worker->task); | ||
274 | worker->task = current; | ||
275 | repeat: | ||
276 | set_current_state(TASK_INTERRUPTIBLE); /* mb paired w/ kthread_stop */ | ||
277 | |||
278 | if (kthread_should_stop()) { | ||
279 | __set_current_state(TASK_RUNNING); | ||
280 | spin_lock_irq(&worker->lock); | ||
281 | worker->task = NULL; | ||
282 | spin_unlock_irq(&worker->lock); | ||
283 | return 0; | ||
284 | } | ||
285 | |||
286 | work = NULL; | ||
287 | spin_lock_irq(&worker->lock); | ||
288 | if (!list_empty(&worker->work_list)) { | ||
289 | work = list_first_entry(&worker->work_list, | ||
290 | struct kthread_work, node); | ||
291 | list_del_init(&work->node); | ||
292 | } | ||
293 | spin_unlock_irq(&worker->lock); | ||
294 | |||
295 | if (work) { | ||
296 | __set_current_state(TASK_RUNNING); | ||
297 | work->func(work); | ||
298 | smp_wmb(); /* wmb worker-b0 paired with flush-b1 */ | ||
299 | work->done_seq = work->queue_seq; | ||
300 | smp_mb(); /* mb worker-b1 paired with flush-b0 */ | ||
301 | if (atomic_read(&work->flushing)) | ||
302 | wake_up_all(&work->done); | ||
303 | } else if (!freezing(current)) | ||
304 | schedule(); | ||
305 | |||
306 | try_to_freeze(); | ||
307 | goto repeat; | ||
308 | } | ||
309 | EXPORT_SYMBOL_GPL(kthread_worker_fn); | ||
310 | |||
311 | /** | ||
312 | * queue_kthread_work - queue a kthread_work | ||
313 | * @worker: target kthread_worker | ||
314 | * @work: kthread_work to queue | ||
315 | * | ||
316 | * Queue @work to work processor @task for async execution. @task | ||
317 | * must have been created with kthread_worker_create(). Returns %true | ||
318 | * if @work was successfully queued, %false if it was already pending. | ||
319 | */ | ||
320 | bool queue_kthread_work(struct kthread_worker *worker, | ||
321 | struct kthread_work *work) | ||
322 | { | ||
323 | bool ret = false; | ||
324 | unsigned long flags; | ||
325 | |||
326 | spin_lock_irqsave(&worker->lock, flags); | ||
327 | if (list_empty(&work->node)) { | ||
328 | list_add_tail(&work->node, &worker->work_list); | ||
329 | work->queue_seq++; | ||
330 | if (likely(worker->task)) | ||
331 | wake_up_process(worker->task); | ||
332 | ret = true; | ||
333 | } | ||
334 | spin_unlock_irqrestore(&worker->lock, flags); | ||
335 | return ret; | ||
336 | } | ||
337 | EXPORT_SYMBOL_GPL(queue_kthread_work); | ||
338 | |||
339 | /** | ||
340 | * flush_kthread_work - flush a kthread_work | ||
341 | * @work: work to flush | ||
342 | * | ||
343 | * If @work is queued or executing, wait for it to finish execution. | ||
344 | */ | ||
345 | void flush_kthread_work(struct kthread_work *work) | ||
346 | { | ||
347 | int seq = work->queue_seq; | ||
348 | |||
349 | atomic_inc(&work->flushing); | ||
350 | |||
351 | /* | ||
352 | * mb flush-b0 paired with worker-b1, to make sure either | ||
353 | * worker sees the above increment or we see done_seq update. | ||
354 | */ | ||
355 | smp_mb__after_atomic_inc(); | ||
356 | |||
357 | /* A - B <= 0 tests whether B is in front of A regardless of overflow */ | ||
358 | wait_event(work->done, seq - work->done_seq <= 0); | ||
359 | atomic_dec(&work->flushing); | ||
360 | |||
361 | /* | ||
362 | * rmb flush-b1 paired with worker-b0, to make sure our caller | ||
363 | * sees every change made by work->func(). | ||
364 | */ | ||
365 | smp_mb__after_atomic_dec(); | ||
366 | } | ||
367 | EXPORT_SYMBOL_GPL(flush_kthread_work); | ||
368 | |||
369 | struct kthread_flush_work { | ||
370 | struct kthread_work work; | ||
371 | struct completion done; | ||
372 | }; | ||
373 | |||
374 | static void kthread_flush_work_fn(struct kthread_work *work) | ||
375 | { | ||
376 | struct kthread_flush_work *fwork = | ||
377 | container_of(work, struct kthread_flush_work, work); | ||
378 | complete(&fwork->done); | ||
379 | } | ||
380 | |||
381 | /** | ||
382 | * flush_kthread_worker - flush all current works on a kthread_worker | ||
383 | * @worker: worker to flush | ||
384 | * | ||
385 | * Wait until all currently executing or pending works on @worker are | ||
386 | * finished. | ||
387 | */ | ||
388 | void flush_kthread_worker(struct kthread_worker *worker) | ||
389 | { | ||
390 | struct kthread_flush_work fwork = { | ||
391 | KTHREAD_WORK_INIT(fwork.work, kthread_flush_work_fn), | ||
392 | COMPLETION_INITIALIZER_ONSTACK(fwork.done), | ||
393 | }; | ||
394 | |||
395 | queue_kthread_work(worker, &fwork.work); | ||
396 | wait_for_completion(&fwork.done); | ||
397 | } | ||
398 | EXPORT_SYMBOL_GPL(flush_kthread_worker); | ||