aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorChristoph Lameter <cl@linux.com>2010-08-20 13:37:13 -0400
committerPekka Enberg <penberg@kernel.org>2010-10-02 03:24:26 -0400
commit55136592fedf5cc58c03d9b5eb66103a9013d4e5 (patch)
tree891d60718ad8bf9d7da35709a1f90fa86566506e /mm
parent1537066c69bee9ebc4b998626c532ce1724d4e67 (diff)
slub: Remove dynamic dma slab allocation
Remove the dynamic dma slab allocation since this causes too many issues with nested locks etc etc. The change avoids passing gfpflags into many functions. V3->V4: - Create dma caches in kmem_cache_init() instead of kmem_cache_init_late(). Acked-by: David Rientjes <rientjes@google.com> Signed-off-by: Christoph Lameter <cl@linux-foundation.org> Signed-off-by: Pekka Enberg <penberg@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/slub.c150
1 files changed, 39 insertions, 111 deletions
diff --git a/mm/slub.c b/mm/slub.c
index d28ee64bcdb5..38c73a3364c6 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -2064,7 +2064,7 @@ init_kmem_cache_node(struct kmem_cache_node *n, struct kmem_cache *s)
2064 2064
2065static DEFINE_PER_CPU(struct kmem_cache_cpu, kmalloc_percpu[KMALLOC_CACHES]); 2065static DEFINE_PER_CPU(struct kmem_cache_cpu, kmalloc_percpu[KMALLOC_CACHES]);
2066 2066
2067static inline int alloc_kmem_cache_cpus(struct kmem_cache *s, gfp_t flags) 2067static inline int alloc_kmem_cache_cpus(struct kmem_cache *s)
2068{ 2068{
2069 if (s < kmalloc_caches + KMALLOC_CACHES && s >= kmalloc_caches) 2069 if (s < kmalloc_caches + KMALLOC_CACHES && s >= kmalloc_caches)
2070 /* 2070 /*
@@ -2091,7 +2091,7 @@ static inline int alloc_kmem_cache_cpus(struct kmem_cache *s, gfp_t flags)
2091 * when allocating for the kmalloc_node_cache. This is used for bootstrapping 2091 * when allocating for the kmalloc_node_cache. This is used for bootstrapping
2092 * memory on a fresh node that has no slab structures yet. 2092 * memory on a fresh node that has no slab structures yet.
2093 */ 2093 */
2094static void early_kmem_cache_node_alloc(gfp_t gfpflags, int node) 2094static void early_kmem_cache_node_alloc(int node)
2095{ 2095{
2096 struct page *page; 2096 struct page *page;
2097 struct kmem_cache_node *n; 2097 struct kmem_cache_node *n;
@@ -2099,7 +2099,7 @@ static void early_kmem_cache_node_alloc(gfp_t gfpflags, int node)
2099 2099
2100 BUG_ON(kmalloc_caches->size < sizeof(struct kmem_cache_node)); 2100 BUG_ON(kmalloc_caches->size < sizeof(struct kmem_cache_node));
2101 2101
2102 page = new_slab(kmalloc_caches, gfpflags, node); 2102 page = new_slab(kmalloc_caches, GFP_NOWAIT, node);
2103 2103
2104 BUG_ON(!page); 2104 BUG_ON(!page);
2105 if (page_to_nid(page) != node) { 2105 if (page_to_nid(page) != node) {
@@ -2143,7 +2143,7 @@ static void free_kmem_cache_nodes(struct kmem_cache *s)
2143 } 2143 }
2144} 2144}
2145 2145
2146static int init_kmem_cache_nodes(struct kmem_cache *s, gfp_t gfpflags) 2146static int init_kmem_cache_nodes(struct kmem_cache *s)
2147{ 2147{
2148 int node; 2148 int node;
2149 2149
@@ -2151,11 +2151,11 @@ static int init_kmem_cache_nodes(struct kmem_cache *s, gfp_t gfpflags)
2151 struct kmem_cache_node *n; 2151 struct kmem_cache_node *n;
2152 2152
2153 if (slab_state == DOWN) { 2153 if (slab_state == DOWN) {
2154 early_kmem_cache_node_alloc(gfpflags, node); 2154 early_kmem_cache_node_alloc(node);
2155 continue; 2155 continue;
2156 } 2156 }
2157 n = kmem_cache_alloc_node(kmalloc_caches, 2157 n = kmem_cache_alloc_node(kmalloc_caches,
2158 gfpflags, node); 2158 GFP_KERNEL, node);
2159 2159
2160 if (!n) { 2160 if (!n) {
2161 free_kmem_cache_nodes(s); 2161 free_kmem_cache_nodes(s);
@@ -2172,7 +2172,7 @@ static void free_kmem_cache_nodes(struct kmem_cache *s)
2172{ 2172{
2173} 2173}
2174 2174
2175static int init_kmem_cache_nodes(struct kmem_cache *s, gfp_t gfpflags) 2175static int init_kmem_cache_nodes(struct kmem_cache *s)
2176{ 2176{
2177 init_kmem_cache_node(&s->local_node, s); 2177 init_kmem_cache_node(&s->local_node, s);
2178 return 1; 2178 return 1;
@@ -2312,7 +2312,7 @@ static int calculate_sizes(struct kmem_cache *s, int forced_order)
2312 2312
2313} 2313}
2314 2314
2315static int kmem_cache_open(struct kmem_cache *s, gfp_t gfpflags, 2315static int kmem_cache_open(struct kmem_cache *s,
2316 const char *name, size_t size, 2316 const char *name, size_t size,
2317 size_t align, unsigned long flags, 2317 size_t align, unsigned long flags,
2318 void (*ctor)(void *)) 2318 void (*ctor)(void *))
@@ -2348,10 +2348,10 @@ static int kmem_cache_open(struct kmem_cache *s, gfp_t gfpflags,
2348#ifdef CONFIG_NUMA 2348#ifdef CONFIG_NUMA
2349 s->remote_node_defrag_ratio = 1000; 2349 s->remote_node_defrag_ratio = 1000;
2350#endif 2350#endif
2351 if (!init_kmem_cache_nodes(s, gfpflags & ~SLUB_DMA)) 2351 if (!init_kmem_cache_nodes(s))
2352 goto error; 2352 goto error;
2353 2353
2354 if (alloc_kmem_cache_cpus(s, gfpflags & ~SLUB_DMA)) 2354 if (alloc_kmem_cache_cpus(s))
2355 return 1; 2355 return 1;
2356 2356
2357 free_kmem_cache_nodes(s); 2357 free_kmem_cache_nodes(s);
@@ -2510,6 +2510,10 @@ EXPORT_SYMBOL(kmem_cache_destroy);
2510struct kmem_cache kmalloc_caches[KMALLOC_CACHES] __cacheline_aligned; 2510struct kmem_cache kmalloc_caches[KMALLOC_CACHES] __cacheline_aligned;
2511EXPORT_SYMBOL(kmalloc_caches); 2511EXPORT_SYMBOL(kmalloc_caches);
2512 2512
2513#ifdef CONFIG_ZONE_DMA
2514static struct kmem_cache kmalloc_dma_caches[SLUB_PAGE_SHIFT];
2515#endif
2516
2513static int __init setup_slub_min_order(char *str) 2517static int __init setup_slub_min_order(char *str)
2514{ 2518{
2515 get_option(&str, &slub_min_order); 2519 get_option(&str, &slub_min_order);
@@ -2546,116 +2550,26 @@ static int __init setup_slub_nomerge(char *str)
2546 2550
2547__setup("slub_nomerge", setup_slub_nomerge); 2551__setup("slub_nomerge", setup_slub_nomerge);
2548 2552
2549static struct kmem_cache *create_kmalloc_cache(struct kmem_cache *s, 2553static void create_kmalloc_cache(struct kmem_cache *s,
2550 const char *name, int size, gfp_t gfp_flags) 2554 const char *name, int size, unsigned int flags)
2551{ 2555{
2552 unsigned int flags = 0;
2553
2554 if (gfp_flags & SLUB_DMA)
2555 flags = SLAB_CACHE_DMA;
2556
2557 /* 2556 /*
2558 * This function is called with IRQs disabled during early-boot on 2557 * This function is called with IRQs disabled during early-boot on
2559 * single CPU so there's no need to take slub_lock here. 2558 * single CPU so there's no need to take slub_lock here.
2560 */ 2559 */
2561 if (!kmem_cache_open(s, gfp_flags, name, size, ARCH_KMALLOC_MINALIGN, 2560 if (!kmem_cache_open(s, name, size, ARCH_KMALLOC_MINALIGN,
2562 flags, NULL)) 2561 flags, NULL))
2563 goto panic; 2562 goto panic;
2564 2563
2565 list_add(&s->list, &slab_caches); 2564 list_add(&s->list, &slab_caches);
2566 2565
2567 if (sysfs_slab_add(s)) 2566 if (!sysfs_slab_add(s))
2568 goto panic; 2567 return;
2569 return s;
2570 2568
2571panic: 2569panic:
2572 panic("Creation of kmalloc slab %s size=%d failed.\n", name, size); 2570 panic("Creation of kmalloc slab %s size=%d failed.\n", name, size);
2573} 2571}
2574 2572
2575#ifdef CONFIG_ZONE_DMA
2576static struct kmem_cache *kmalloc_caches_dma[SLUB_PAGE_SHIFT];
2577
2578static void sysfs_add_func(struct work_struct *w)
2579{
2580 struct kmem_cache *s;
2581
2582 down_write(&slub_lock);
2583 list_for_each_entry(s, &slab_caches, list) {
2584 if (s->flags & __SYSFS_ADD_DEFERRED) {
2585 s->flags &= ~__SYSFS_ADD_DEFERRED;
2586 sysfs_slab_add(s);
2587 }
2588 }
2589 up_write(&slub_lock);
2590}
2591
2592static DECLARE_WORK(sysfs_add_work, sysfs_add_func);
2593
2594static noinline struct kmem_cache *dma_kmalloc_cache(int index, gfp_t flags)
2595{
2596 struct kmem_cache *s;
2597 char *text;
2598 size_t realsize;
2599 unsigned long slabflags;
2600 int i;
2601
2602 s = kmalloc_caches_dma[index];
2603 if (s)
2604 return s;
2605
2606 /* Dynamically create dma cache */
2607 if (flags & __GFP_WAIT)
2608 down_write(&slub_lock);
2609 else {
2610 if (!down_write_trylock(&slub_lock))
2611 goto out;
2612 }
2613
2614 if (kmalloc_caches_dma[index])
2615 goto unlock_out;
2616
2617 realsize = kmalloc_caches[index].objsize;
2618 text = kasprintf(flags & ~SLUB_DMA, "kmalloc_dma-%d",
2619 (unsigned int)realsize);
2620
2621 s = NULL;
2622 for (i = 0; i < KMALLOC_CACHES; i++)
2623 if (!kmalloc_caches[i].size)
2624 break;
2625
2626 BUG_ON(i >= KMALLOC_CACHES);
2627 s = kmalloc_caches + i;
2628
2629 /*
2630 * Must defer sysfs creation to a workqueue because we don't know
2631 * what context we are called from. Before sysfs comes up, we don't
2632 * need to do anything because our sysfs initcall will start by
2633 * adding all existing slabs to sysfs.
2634 */
2635 slabflags = SLAB_CACHE_DMA|SLAB_NOTRACK;
2636 if (slab_state >= SYSFS)
2637 slabflags |= __SYSFS_ADD_DEFERRED;
2638
2639 if (!text || !kmem_cache_open(s, flags, text,
2640 realsize, ARCH_KMALLOC_MINALIGN, slabflags, NULL)) {
2641 s->size = 0;
2642 kfree(text);
2643 goto unlock_out;
2644 }
2645
2646 list_add(&s->list, &slab_caches);
2647 kmalloc_caches_dma[index] = s;
2648
2649 if (slab_state >= SYSFS)
2650 schedule_work(&sysfs_add_work);
2651
2652unlock_out:
2653 up_write(&slub_lock);
2654out:
2655 return kmalloc_caches_dma[index];
2656}
2657#endif
2658
2659/* 2573/*
2660 * Conversion table for small slabs sizes / 8 to the index in the 2574 * Conversion table for small slabs sizes / 8 to the index in the
2661 * kmalloc array. This is necessary for slabs < 192 since we have non power 2575 * kmalloc array. This is necessary for slabs < 192 since we have non power
@@ -2708,7 +2622,7 @@ static struct kmem_cache *get_slab(size_t size, gfp_t flags)
2708 2622
2709#ifdef CONFIG_ZONE_DMA 2623#ifdef CONFIG_ZONE_DMA
2710 if (unlikely((flags & SLUB_DMA))) 2624 if (unlikely((flags & SLUB_DMA)))
2711 return dma_kmalloc_cache(index, flags); 2625 return &kmalloc_dma_caches[index];
2712 2626
2713#endif 2627#endif
2714 return &kmalloc_caches[index]; 2628 return &kmalloc_caches[index];
@@ -3047,7 +2961,7 @@ void __init kmem_cache_init(void)
3047 * kmem_cache_open for slab_state == DOWN. 2961 * kmem_cache_open for slab_state == DOWN.
3048 */ 2962 */
3049 create_kmalloc_cache(&kmalloc_caches[0], "kmem_cache_node", 2963 create_kmalloc_cache(&kmalloc_caches[0], "kmem_cache_node",
3050 sizeof(struct kmem_cache_node), GFP_NOWAIT); 2964 sizeof(struct kmem_cache_node), 0);
3051 kmalloc_caches[0].refcount = -1; 2965 kmalloc_caches[0].refcount = -1;
3052 caches++; 2966 caches++;
3053 2967
@@ -3060,18 +2974,18 @@ void __init kmem_cache_init(void)
3060 /* Caches that are not of the two-to-the-power-of size */ 2974 /* Caches that are not of the two-to-the-power-of size */
3061 if (KMALLOC_MIN_SIZE <= 32) { 2975 if (KMALLOC_MIN_SIZE <= 32) {
3062 create_kmalloc_cache(&kmalloc_caches[1], 2976 create_kmalloc_cache(&kmalloc_caches[1],
3063 "kmalloc-96", 96, GFP_NOWAIT); 2977 "kmalloc-96", 96, 0);
3064 caches++; 2978 caches++;
3065 } 2979 }
3066 if (KMALLOC_MIN_SIZE <= 64) { 2980 if (KMALLOC_MIN_SIZE <= 64) {
3067 create_kmalloc_cache(&kmalloc_caches[2], 2981 create_kmalloc_cache(&kmalloc_caches[2],
3068 "kmalloc-192", 192, GFP_NOWAIT); 2982 "kmalloc-192", 192, 0);
3069 caches++; 2983 caches++;
3070 } 2984 }
3071 2985
3072 for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++) { 2986 for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++) {
3073 create_kmalloc_cache(&kmalloc_caches[i], 2987 create_kmalloc_cache(&kmalloc_caches[i],
3074 "kmalloc", 1 << i, GFP_NOWAIT); 2988 "kmalloc", 1 << i, 0);
3075 caches++; 2989 caches++;
3076 } 2990 }
3077 2991
@@ -3134,6 +3048,20 @@ void __init kmem_cache_init(void)
3134 kmem_size = sizeof(struct kmem_cache); 3048 kmem_size = sizeof(struct kmem_cache);
3135#endif 3049#endif
3136 3050
3051#ifdef CONFIG_ZONE_DMA
3052 for (i = 1; i < SLUB_PAGE_SHIFT; i++) {
3053 struct kmem_cache *s = &kmalloc_caches[i];
3054
3055 if (s->size) {
3056 char *name = kasprintf(GFP_NOWAIT,
3057 "dma-kmalloc-%d", s->objsize);
3058
3059 BUG_ON(!name);
3060 create_kmalloc_cache(&kmalloc_dma_caches[i],
3061 name, s->objsize, SLAB_CACHE_DMA);
3062 }
3063 }
3064#endif
3137 printk(KERN_INFO 3065 printk(KERN_INFO
3138 "SLUB: Genslabs=%d, HWalign=%d, Order=%d-%d, MinObjects=%d," 3066 "SLUB: Genslabs=%d, HWalign=%d, Order=%d-%d, MinObjects=%d,"
3139 " CPUs=%d, Nodes=%d\n", 3067 " CPUs=%d, Nodes=%d\n",
@@ -3236,7 +3164,7 @@ struct kmem_cache *kmem_cache_create(const char *name, size_t size,
3236 3164
3237 s = kmalloc(kmem_size, GFP_KERNEL); 3165 s = kmalloc(kmem_size, GFP_KERNEL);
3238 if (s) { 3166 if (s) {
3239 if (kmem_cache_open(s, GFP_KERNEL, name, 3167 if (kmem_cache_open(s, name,
3240 size, align, flags, ctor)) { 3168 size, align, flags, ctor)) {
3241 list_add(&s->list, &slab_caches); 3169 list_add(&s->list, &slab_caches);
3242 if (sysfs_slab_add(s)) { 3170 if (sysfs_slab_add(s)) {