aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorChristoph Lameter <cl@linux.com>2011-08-09 17:12:25 -0400
committerPekka Enberg <penberg@kernel.org>2011-08-19 12:34:26 -0400
commitacd19fd1a7b5152cf29f67aaab23aa61078aaa74 (patch)
tree8695f48ccaa1baa3bc26df2d7b1b0648dd82078b /mm
parente6e82ea1127d899af7eee876f1c1103716d13772 (diff)
slub: pass kmem_cache_cpu pointer to get_partial()
Pass the kmem_cache_cpu pointer to get_partial(). That way we can avoid the this_cpu_write() statements. Signed-off-by: Christoph Lameter <cl@linux.com> Signed-off-by: Pekka Enberg <penberg@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/slub.c30
1 files changed, 15 insertions, 15 deletions
diff --git a/mm/slub.c b/mm/slub.c
index 83bb81de4468..cb53affecca7 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -1557,7 +1557,8 @@ static inline void remove_partial(struct kmem_cache_node *n,
1557 * Must hold list_lock. 1557 * Must hold list_lock.
1558 */ 1558 */
1559static inline int acquire_slab(struct kmem_cache *s, 1559static inline int acquire_slab(struct kmem_cache *s,
1560 struct kmem_cache_node *n, struct page *page) 1560 struct kmem_cache_node *n, struct page *page,
1561 struct kmem_cache_cpu *c)
1561{ 1562{
1562 void *freelist; 1563 void *freelist;
1563 unsigned long counters; 1564 unsigned long counters;
@@ -1586,9 +1587,9 @@ static inline int acquire_slab(struct kmem_cache *s,
1586 1587
1587 if (freelist) { 1588 if (freelist) {
1588 /* Populate the per cpu freelist */ 1589 /* Populate the per cpu freelist */
1589 this_cpu_write(s->cpu_slab->freelist, freelist); 1590 c->freelist = freelist;
1590 this_cpu_write(s->cpu_slab->page, page); 1591 c->page = page;
1591 this_cpu_write(s->cpu_slab->node, page_to_nid(page)); 1592 c->node = page_to_nid(page);
1592 return 1; 1593 return 1;
1593 } else { 1594 } else {
1594 /* 1595 /*
@@ -1606,7 +1607,7 @@ static inline int acquire_slab(struct kmem_cache *s,
1606 * Try to allocate a partial slab from a specific node. 1607 * Try to allocate a partial slab from a specific node.
1607 */ 1608 */
1608static struct page *get_partial_node(struct kmem_cache *s, 1609static struct page *get_partial_node(struct kmem_cache *s,
1609 struct kmem_cache_node *n) 1610 struct kmem_cache_node *n, struct kmem_cache_cpu *c)
1610{ 1611{
1611 struct page *page; 1612 struct page *page;
1612 1613
@@ -1621,7 +1622,7 @@ static struct page *get_partial_node(struct kmem_cache *s,
1621 1622
1622 spin_lock(&n->list_lock); 1623 spin_lock(&n->list_lock);
1623 list_for_each_entry(page, &n->partial, lru) 1624 list_for_each_entry(page, &n->partial, lru)
1624 if (acquire_slab(s, n, page)) 1625 if (acquire_slab(s, n, page, c))
1625 goto out; 1626 goto out;
1626 page = NULL; 1627 page = NULL;
1627out: 1628out:
@@ -1632,7 +1633,8 @@ out:
1632/* 1633/*
1633 * Get a page from somewhere. Search in increasing NUMA distances. 1634 * Get a page from somewhere. Search in increasing NUMA distances.
1634 */ 1635 */
1635static struct page *get_any_partial(struct kmem_cache *s, gfp_t flags) 1636static struct page *get_any_partial(struct kmem_cache *s, gfp_t flags,
1637 struct kmem_cache_cpu *c)
1636{ 1638{
1637#ifdef CONFIG_NUMA 1639#ifdef CONFIG_NUMA
1638 struct zonelist *zonelist; 1640 struct zonelist *zonelist;
@@ -1672,7 +1674,7 @@ static struct page *get_any_partial(struct kmem_cache *s, gfp_t flags)
1672 1674
1673 if (n && cpuset_zone_allowed_hardwall(zone, flags) && 1675 if (n && cpuset_zone_allowed_hardwall(zone, flags) &&
1674 n->nr_partial > s->min_partial) { 1676 n->nr_partial > s->min_partial) {
1675 page = get_partial_node(s, n); 1677 page = get_partial_node(s, n, c);
1676 if (page) { 1678 if (page) {
1677 put_mems_allowed(); 1679 put_mems_allowed();
1678 return page; 1680 return page;
@@ -1687,16 +1689,17 @@ static struct page *get_any_partial(struct kmem_cache *s, gfp_t flags)
1687/* 1689/*
1688 * Get a partial page, lock it and return it. 1690 * Get a partial page, lock it and return it.
1689 */ 1691 */
1690static struct page *get_partial(struct kmem_cache *s, gfp_t flags, int node) 1692static struct page *get_partial(struct kmem_cache *s, gfp_t flags, int node,
1693 struct kmem_cache_cpu *c)
1691{ 1694{
1692 struct page *page; 1695 struct page *page;
1693 int searchnode = (node == NUMA_NO_NODE) ? numa_node_id() : node; 1696 int searchnode = (node == NUMA_NO_NODE) ? numa_node_id() : node;
1694 1697
1695 page = get_partial_node(s, get_node(s, searchnode)); 1698 page = get_partial_node(s, get_node(s, searchnode), c);
1696 if (page || node != NUMA_NO_NODE) 1699 if (page || node != NUMA_NO_NODE)
1697 return page; 1700 return page;
1698 1701
1699 return get_any_partial(s, flags); 1702 return get_any_partial(s, flags, c);
1700} 1703}
1701 1704
1702#ifdef CONFIG_PREEMPT 1705#ifdef CONFIG_PREEMPT
@@ -1765,9 +1768,6 @@ void init_kmem_cache_cpus(struct kmem_cache *s)
1765 for_each_possible_cpu(cpu) 1768 for_each_possible_cpu(cpu)
1766 per_cpu_ptr(s->cpu_slab, cpu)->tid = init_tid(cpu); 1769 per_cpu_ptr(s->cpu_slab, cpu)->tid = init_tid(cpu);
1767} 1770}
1768/*
1769 * Remove the cpu slab
1770 */
1771 1771
1772/* 1772/*
1773 * Remove the cpu slab 1773 * Remove the cpu slab
@@ -2116,7 +2116,7 @@ load_freelist:
2116 return object; 2116 return object;
2117 2117
2118new_slab: 2118new_slab:
2119 page = get_partial(s, gfpflags, node); 2119 page = get_partial(s, gfpflags, node, c);
2120 if (page) { 2120 if (page) {
2121 stat(s, ALLOC_FROM_PARTIAL); 2121 stat(s, ALLOC_FROM_PARTIAL);
2122 object = c->freelist; 2122 object = c->freelist;