aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/slab.c29
1 files changed, 27 insertions, 2 deletions
diff --git a/mm/slab.c b/mm/slab.c
index d47d0e186973..3ad2f64998fd 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -2206,8 +2206,17 @@ kmem_cache_create (const char *name, size_t size, size_t align,
2206 cachep->gfpflags |= GFP_DMA; 2206 cachep->gfpflags |= GFP_DMA;
2207 cachep->buffer_size = size; 2207 cachep->buffer_size = size;
2208 2208
2209 if (flags & CFLGS_OFF_SLAB) 2209 if (flags & CFLGS_OFF_SLAB) {
2210 cachep->slabp_cache = kmem_find_general_cachep(slab_size, 0u); 2210 cachep->slabp_cache = kmem_find_general_cachep(slab_size, 0u);
2211 /*
2212 * This is a possibility for one of the malloc_sizes caches.
2213 * But since we go off slab only for object size greater than
2214 * PAGE_SIZE/8, and malloc_sizes gets created in ascending order,
2215 * this should not happen at all.
2216 * But leave a BUG_ON for some lucky dude.
2217 */
2218 BUG_ON(!cachep->slabp_cache);
2219 }
2211 cachep->ctor = ctor; 2220 cachep->ctor = ctor;
2212 cachep->dtor = dtor; 2221 cachep->dtor = dtor;
2213 cachep->name = name; 2222 cachep->name = name;
@@ -2441,7 +2450,17 @@ int kmem_cache_destroy(struct kmem_cache *cachep)
2441} 2450}
2442EXPORT_SYMBOL(kmem_cache_destroy); 2451EXPORT_SYMBOL(kmem_cache_destroy);
2443 2452
2444/* Get the memory for a slab management obj. */ 2453/*
2454 * Get the memory for a slab management obj.
2455 * For a slab cache when the slab descriptor is off-slab, slab descriptors
2456 * always come from malloc_sizes caches. The slab descriptor cannot
2457 * come from the same cache which is getting created because,
2458 * when we are searching for an appropriate cache for these
2459 * descriptors in kmem_cache_create, we search through the malloc_sizes array.
2460 * If we are creating a malloc_sizes cache here it would not be visible to
2461 * kmem_find_general_cachep till the initialization is complete.
2462 * Hence we cannot have slabp_cache same as the original cache.
2463 */
2445static struct slab *alloc_slabmgmt(struct kmem_cache *cachep, void *objp, 2464static struct slab *alloc_slabmgmt(struct kmem_cache *cachep, void *objp,
2446 int colour_off, gfp_t local_flags, 2465 int colour_off, gfp_t local_flags,
2447 int nodeid) 2466 int nodeid)
@@ -3125,6 +3144,12 @@ static void free_block(struct kmem_cache *cachep, void **objpp, int nr_objects,
3125 if (slabp->inuse == 0) { 3144 if (slabp->inuse == 0) {
3126 if (l3->free_objects > l3->free_limit) { 3145 if (l3->free_objects > l3->free_limit) {
3127 l3->free_objects -= cachep->num; 3146 l3->free_objects -= cachep->num;
3147 /* No need to drop any previously held
3148 * lock here, even if we have a off-slab slab
3149 * descriptor it is guaranteed to come from
3150 * a different cache, refer to comments before
3151 * alloc_slabmgmt.
3152 */
3128 slab_destroy(cachep, slabp); 3153 slab_destroy(cachep, slabp);
3129 } else { 3154 } else {
3130 list_add(&slabp->list, &l3->slabs_free); 3155 list_add(&slabp->list, &l3->slabs_free);