diff options
| -rw-r--r-- | mm/slab.c | 4 | ||||
| -rw-r--r-- | mm/slub.c | 2 |
2 files changed, 4 insertions, 2 deletions
| @@ -1481,7 +1481,7 @@ void __init kmem_cache_init(void) | |||
| 1481 | list_add(&cache_cache.next, &cache_chain); | 1481 | list_add(&cache_cache.next, &cache_chain); |
| 1482 | cache_cache.colour_off = cache_line_size(); | 1482 | cache_cache.colour_off = cache_line_size(); |
| 1483 | cache_cache.array[smp_processor_id()] = &initarray_cache.cache; | 1483 | cache_cache.array[smp_processor_id()] = &initarray_cache.cache; |
| 1484 | cache_cache.nodelists[node] = &initkmem_list3[CACHE_CACHE]; | 1484 | cache_cache.nodelists[node] = &initkmem_list3[CACHE_CACHE + node]; |
| 1485 | 1485 | ||
| 1486 | /* | 1486 | /* |
| 1487 | * struct kmem_cache size depends on nr_node_ids, which | 1487 | * struct kmem_cache size depends on nr_node_ids, which |
| @@ -1602,7 +1602,7 @@ void __init kmem_cache_init(void) | |||
| 1602 | int nid; | 1602 | int nid; |
| 1603 | 1603 | ||
| 1604 | for_each_online_node(nid) { | 1604 | for_each_online_node(nid) { |
| 1605 | init_list(&cache_cache, &initkmem_list3[CACHE_CACHE], nid); | 1605 | init_list(&cache_cache, &initkmem_list3[CACHE_CACHE + nid], nid); |
| 1606 | 1606 | ||
| 1607 | init_list(malloc_sizes[INDEX_AC].cs_cachep, | 1607 | init_list(malloc_sizes[INDEX_AC].cs_cachep, |
| 1608 | &initkmem_list3[SIZE_AC + nid], nid); | 1608 | &initkmem_list3[SIZE_AC + nid], nid); |
| @@ -2685,6 +2685,7 @@ void kfree(const void *x) | |||
| 2685 | } | 2685 | } |
| 2686 | EXPORT_SYMBOL(kfree); | 2686 | EXPORT_SYMBOL(kfree); |
| 2687 | 2687 | ||
| 2688 | #if defined(SLUB_DEBUG) || defined(CONFIG_SLABINFO) | ||
| 2688 | static unsigned long count_partial(struct kmem_cache_node *n) | 2689 | static unsigned long count_partial(struct kmem_cache_node *n) |
| 2689 | { | 2690 | { |
| 2690 | unsigned long flags; | 2691 | unsigned long flags; |
| @@ -2697,6 +2698,7 @@ static unsigned long count_partial(struct kmem_cache_node *n) | |||
| 2697 | spin_unlock_irqrestore(&n->list_lock, flags); | 2698 | spin_unlock_irqrestore(&n->list_lock, flags); |
| 2698 | return x; | 2699 | return x; |
| 2699 | } | 2700 | } |
| 2701 | #endif | ||
| 2700 | 2702 | ||
| 2701 | /* | 2703 | /* |
| 2702 | * kmem_cache_shrink removes empty slabs from the partial lists and sorts | 2704 | * kmem_cache_shrink removes empty slabs from the partial lists and sorts |
