aboutsummaryrefslogtreecommitdiffstats
path: root/mm/slab.c
diff options
context:
space:
mode:
authorChristoph Lameter <clameter@sgi.com>2006-09-26 02:31:50 -0400
committerLinus Torvalds <torvalds@g5.osdl.org>2006-09-26 11:48:51 -0400
commitd00bcc98d7ec2c87391c9d9e1cca519ef64d33ef (patch)
tree08b7d0fafba03d7b1d4d1d861897f78658aba173 /mm/slab.c
parent39bbcb8f88154c4ac9853baf3f1134af4c987517 (diff)
[PATCH] Extract the allocpercpu functions from the slab allocator
The allocpercpu functions __alloc_percpu and __free_percpu() are heavily using the slab allocator. However, they are conceptually slab. This also simplifies SLOB (at this point slob may be broken in mm. This should fix it). Signed-off-by: Christoph Lameter <clameter@sgi.com> Cc: Matt Mackall <mpm@selenic.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'mm/slab.c')
-rw-r--r--mm/slab.c124
1 files changed, 0 insertions, 124 deletions
diff --git a/mm/slab.c b/mm/slab.c
index 619337a5cb2b..13b5050f84cc 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -3440,130 +3440,6 @@ void *__kmalloc_track_caller(size_t size, gfp_t flags, void *caller)
3440EXPORT_SYMBOL(__kmalloc_track_caller); 3440EXPORT_SYMBOL(__kmalloc_track_caller);
3441#endif 3441#endif
3442 3442
3443#ifdef CONFIG_SMP
3444/**
3445 * percpu_depopulate - depopulate per-cpu data for given cpu
3446 * @__pdata: per-cpu data to depopulate
3447 * @cpu: depopulate per-cpu data for this cpu
3448 *
3449 * Depopulating per-cpu data for a cpu going offline would be a typical
3450 * use case. You need to register a cpu hotplug handler for that purpose.
3451 */
3452void percpu_depopulate(void *__pdata, int cpu)
3453{
3454 struct percpu_data *pdata = __percpu_disguise(__pdata);
3455 if (pdata->ptrs[cpu]) {
3456 kfree(pdata->ptrs[cpu]);
3457 pdata->ptrs[cpu] = NULL;
3458 }
3459}
3460EXPORT_SYMBOL_GPL(percpu_depopulate);
3461
3462/**
3463 * percpu_depopulate_mask - depopulate per-cpu data for some cpu's
3464 * @__pdata: per-cpu data to depopulate
3465 * @mask: depopulate per-cpu data for cpu's selected through mask bits
3466 */
3467void __percpu_depopulate_mask(void *__pdata, cpumask_t *mask)
3468{
3469 int cpu;
3470 for_each_cpu_mask(cpu, *mask)
3471 percpu_depopulate(__pdata, cpu);
3472}
3473EXPORT_SYMBOL_GPL(__percpu_depopulate_mask);
3474
3475/**
3476 * percpu_populate - populate per-cpu data for given cpu
3477 * @__pdata: per-cpu data to populate further
3478 * @size: size of per-cpu object
3479 * @gfp: may sleep or not etc.
3480 * @cpu: populate per-data for this cpu
3481 *
3482 * Populating per-cpu data for a cpu coming online would be a typical
3483 * use case. You need to register a cpu hotplug handler for that purpose.
3484 * Per-cpu object is populated with zeroed buffer.
3485 */
3486void *percpu_populate(void *__pdata, size_t size, gfp_t gfp, int cpu)
3487{
3488 struct percpu_data *pdata = __percpu_disguise(__pdata);
3489 int node = cpu_to_node(cpu);
3490
3491 BUG_ON(pdata->ptrs[cpu]);
3492 if (node_online(node)) {
3493 /* FIXME: kzalloc_node(size, gfp, node) */
3494 pdata->ptrs[cpu] = kmalloc_node(size, gfp, node);
3495 if (pdata->ptrs[cpu])
3496 memset(pdata->ptrs[cpu], 0, size);
3497 } else
3498 pdata->ptrs[cpu] = kzalloc(size, gfp);
3499 return pdata->ptrs[cpu];
3500}
3501EXPORT_SYMBOL_GPL(percpu_populate);
3502
3503/**
3504 * percpu_populate_mask - populate per-cpu data for more cpu's
3505 * @__pdata: per-cpu data to populate further
3506 * @size: size of per-cpu object
3507 * @gfp: may sleep or not etc.
3508 * @mask: populate per-cpu data for cpu's selected through mask bits
3509 *
3510 * Per-cpu objects are populated with zeroed buffers.
3511 */
3512int __percpu_populate_mask(void *__pdata, size_t size, gfp_t gfp,
3513 cpumask_t *mask)
3514{
3515 cpumask_t populated = CPU_MASK_NONE;
3516 int cpu;
3517
3518 for_each_cpu_mask(cpu, *mask)
3519 if (unlikely(!percpu_populate(__pdata, size, gfp, cpu))) {
3520 __percpu_depopulate_mask(__pdata, &populated);
3521 return -ENOMEM;
3522 } else
3523 cpu_set(cpu, populated);
3524 return 0;
3525}
3526EXPORT_SYMBOL_GPL(__percpu_populate_mask);
3527
3528/**
3529 * percpu_alloc_mask - initial setup of per-cpu data
3530 * @size: size of per-cpu object
3531 * @gfp: may sleep or not etc.
3532 * @mask: populate per-data for cpu's selected through mask bits
3533 *
3534 * Populating per-cpu data for all online cpu's would be a typical use case,
3535 * which is simplified by the percpu_alloc() wrapper.
3536 * Per-cpu objects are populated with zeroed buffers.
3537 */
3538void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask)
3539{
3540 void *pdata = kzalloc(sizeof(struct percpu_data), gfp);
3541 void *__pdata = __percpu_disguise(pdata);
3542
3543 if (unlikely(!pdata))
3544 return NULL;
3545 if (likely(!__percpu_populate_mask(__pdata, size, gfp, mask)))
3546 return __pdata;
3547 kfree(pdata);
3548 return NULL;
3549}
3550EXPORT_SYMBOL_GPL(__percpu_alloc_mask);
3551
3552/**
3553 * percpu_free - final cleanup of per-cpu data
3554 * @__pdata: object to clean up
3555 *
3556 * We simply clean up any per-cpu object left. No need for the client to
3557 * track and specify through a bis mask which per-cpu objects are to free.
3558 */
3559void percpu_free(void *__pdata)
3560{
3561 __percpu_depopulate_mask(__pdata, &cpu_possible_map);
3562 kfree(__percpu_disguise(__pdata));
3563}
3564EXPORT_SYMBOL_GPL(percpu_free);
3565#endif /* CONFIG_SMP */
3566
3567/** 3443/**
3568 * kmem_cache_free - Deallocate an object 3444 * kmem_cache_free - Deallocate an object
3569 * @cachep: The cache the allocation was from. 3445 * @cachep: The cache the allocation was from.