diff options
-rw-r--r-- | fs/aio.c | 67 | ||||
-rw-r--r-- | include/linux/aio.h | 19 |
2 files changed, 39 insertions, 47 deletions
@@ -191,6 +191,43 @@ static int aio_setup_ring(struct kioctx *ctx) | |||
191 | kunmap_atomic((void *)((unsigned long)__event & PAGE_MASK), km); \ | 191 | kunmap_atomic((void *)((unsigned long)__event & PAGE_MASK), km); \ |
192 | } while(0) | 192 | } while(0) |
193 | 193 | ||
194 | |||
195 | /* __put_ioctx | ||
196 | * Called when the last user of an aio context has gone away, | ||
197 | * and the struct needs to be freed. | ||
198 | */ | ||
199 | static void __put_ioctx(struct kioctx *ctx) | ||
200 | { | ||
201 | unsigned nr_events = ctx->max_reqs; | ||
202 | |||
203 | BUG_ON(ctx->reqs_active); | ||
204 | |||
205 | cancel_delayed_work(&ctx->wq); | ||
206 | cancel_work_sync(&ctx->wq.work); | ||
207 | aio_free_ring(ctx); | ||
208 | mmdrop(ctx->mm); | ||
209 | ctx->mm = NULL; | ||
210 | pr_debug("__put_ioctx: freeing %p\n", ctx); | ||
211 | kmem_cache_free(kioctx_cachep, ctx); | ||
212 | |||
213 | if (nr_events) { | ||
214 | spin_lock(&aio_nr_lock); | ||
215 | BUG_ON(aio_nr - nr_events > aio_nr); | ||
216 | aio_nr -= nr_events; | ||
217 | spin_unlock(&aio_nr_lock); | ||
218 | } | ||
219 | } | ||
220 | |||
221 | #define get_ioctx(kioctx) do { \ | ||
222 | BUG_ON(atomic_read(&(kioctx)->users) <= 0); \ | ||
223 | atomic_inc(&(kioctx)->users); \ | ||
224 | } while (0) | ||
225 | #define put_ioctx(kioctx) do { \ | ||
226 | BUG_ON(atomic_read(&(kioctx)->users) <= 0); \ | ||
227 | if (unlikely(atomic_dec_and_test(&(kioctx)->users))) \ | ||
228 | __put_ioctx(kioctx); \ | ||
229 | } while (0) | ||
230 | |||
194 | /* ioctx_alloc | 231 | /* ioctx_alloc |
195 | * Allocates and initializes an ioctx. Returns an ERR_PTR if it failed. | 232 | * Allocates and initializes an ioctx. Returns an ERR_PTR if it failed. |
196 | */ | 233 | */ |
@@ -361,32 +398,6 @@ void exit_aio(struct mm_struct *mm) | |||
361 | } | 398 | } |
362 | } | 399 | } |
363 | 400 | ||
364 | /* __put_ioctx | ||
365 | * Called when the last user of an aio context has gone away, | ||
366 | * and the struct needs to be freed. | ||
367 | */ | ||
368 | void __put_ioctx(struct kioctx *ctx) | ||
369 | { | ||
370 | unsigned nr_events = ctx->max_reqs; | ||
371 | |||
372 | BUG_ON(ctx->reqs_active); | ||
373 | |||
374 | cancel_delayed_work(&ctx->wq); | ||
375 | cancel_work_sync(&ctx->wq.work); | ||
376 | aio_free_ring(ctx); | ||
377 | mmdrop(ctx->mm); | ||
378 | ctx->mm = NULL; | ||
379 | pr_debug("__put_ioctx: freeing %p\n", ctx); | ||
380 | kmem_cache_free(kioctx_cachep, ctx); | ||
381 | |||
382 | if (nr_events) { | ||
383 | spin_lock(&aio_nr_lock); | ||
384 | BUG_ON(aio_nr - nr_events > aio_nr); | ||
385 | aio_nr -= nr_events; | ||
386 | spin_unlock(&aio_nr_lock); | ||
387 | } | ||
388 | } | ||
389 | |||
390 | /* aio_get_req | 401 | /* aio_get_req |
391 | * Allocate a slot for an aio request. Increments the users count | 402 | * Allocate a slot for an aio request. Increments the users count |
392 | * of the kioctx so that the kioctx stays around until all requests are | 403 | * of the kioctx so that the kioctx stays around until all requests are |
@@ -545,7 +556,7 @@ int aio_put_req(struct kiocb *req) | |||
545 | /* Lookup an ioctx id. ioctx_list is lockless for reads. | 556 | /* Lookup an ioctx id. ioctx_list is lockless for reads. |
546 | * FIXME: this is O(n) and is only suitable for development. | 557 | * FIXME: this is O(n) and is only suitable for development. |
547 | */ | 558 | */ |
548 | struct kioctx *lookup_ioctx(unsigned long ctx_id) | 559 | static struct kioctx *lookup_ioctx(unsigned long ctx_id) |
549 | { | 560 | { |
550 | struct kioctx *ioctx; | 561 | struct kioctx *ioctx; |
551 | struct mm_struct *mm; | 562 | struct mm_struct *mm; |
@@ -1552,7 +1563,7 @@ static int aio_wake_function(wait_queue_t *wait, unsigned mode, | |||
1552 | return 1; | 1563 | return 1; |
1553 | } | 1564 | } |
1554 | 1565 | ||
1555 | int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, | 1566 | static int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, |
1556 | struct iocb *iocb) | 1567 | struct iocb *iocb) |
1557 | { | 1568 | { |
1558 | struct kiocb *req; | 1569 | struct kiocb *req; |
diff --git a/include/linux/aio.h b/include/linux/aio.h index 0d0b7f629bd3..b51ddd28444e 100644 --- a/include/linux/aio.h +++ b/include/linux/aio.h | |||
@@ -209,27 +209,8 @@ extern ssize_t wait_on_sync_kiocb(struct kiocb *iocb); | |||
209 | extern int aio_put_req(struct kiocb *iocb); | 209 | extern int aio_put_req(struct kiocb *iocb); |
210 | extern void kick_iocb(struct kiocb *iocb); | 210 | extern void kick_iocb(struct kiocb *iocb); |
211 | extern int aio_complete(struct kiocb *iocb, long res, long res2); | 211 | extern int aio_complete(struct kiocb *iocb, long res, long res2); |
212 | extern void __put_ioctx(struct kioctx *ctx); | ||
213 | struct mm_struct; | 212 | struct mm_struct; |
214 | extern void exit_aio(struct mm_struct *mm); | 213 | extern void exit_aio(struct mm_struct *mm); |
215 | extern struct kioctx *lookup_ioctx(unsigned long ctx_id); | ||
216 | extern int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, | ||
217 | struct iocb *iocb); | ||
218 | |||
219 | /* semi private, but used by the 32bit emulations: */ | ||
220 | struct kioctx *lookup_ioctx(unsigned long ctx_id); | ||
221 | int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb, | ||
222 | struct iocb *iocb); | ||
223 | |||
224 | #define get_ioctx(kioctx) do { \ | ||
225 | BUG_ON(atomic_read(&(kioctx)->users) <= 0); \ | ||
226 | atomic_inc(&(kioctx)->users); \ | ||
227 | } while (0) | ||
228 | #define put_ioctx(kioctx) do { \ | ||
229 | BUG_ON(atomic_read(&(kioctx)->users) <= 0); \ | ||
230 | if (unlikely(atomic_dec_and_test(&(kioctx)->users))) \ | ||
231 | __put_ioctx(kioctx); \ | ||
232 | } while (0) | ||
233 | 214 | ||
234 | #define io_wait_to_kiocb(wait) container_of(wait, struct kiocb, ki_wait) | 215 | #define io_wait_to_kiocb(wait) container_of(wait, struct kiocb, ki_wait) |
235 | 216 | ||