diff options
author | Lee Schermerhorn <lee.schermerhorn@hp.com> | 2008-04-28 05:13:21 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-04-28 11:58:24 -0400 |
commit | fc36b8d3d819047eb4d23ca079fb4d3af20ff076 (patch) | |
tree | 65ee215a6bdca1e8d4ac4b57525445d7d1829c1d /mm/mempolicy.c | |
parent | 53f2556b6792ed99fde965f5e061749edd455623 (diff) |
mempolicy: use MPOL_F_LOCAL to Indicate Preferred Local Policy
Now that we're using "preferred local" policy for system default, we need to
make this as fast as possible. Because of the variable size of the mempolicy
structure [based on size of nodemasks], the preferred_node may be in a
different cacheline from the mode. This can result in accessing an extra
cacheline in the normal case of system default policy. Suspect this is the
cause of an observed 2-3% slowdown in page fault testing relative to kernel
without this patch series.
To alleviate this, use an internal mode flag, MPOL_F_LOCAL in the mempolicy
flags member which is guaranteed [?] to be in the same cacheline as the mode
itself.
Verified that reworked mempolicy now performs slightly better on 25-rc8-mm1
for both anon and shmem segments with system default and vma [preferred local]
policy.
Signed-off-by: Lee Schermerhorn <lee.schermerhorn@hp.com>
Cc: Christoph Lameter <clameter@sgi.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Mel Gorman <mel@csn.ul.ie>
Cc: Andi Kleen <ak@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/mempolicy.c')
-rw-r--r-- | mm/mempolicy.c | 47 |
1 files changed, 22 insertions, 25 deletions
diff --git a/mm/mempolicy.c b/mm/mempolicy.c index 7b3ae977b158..143b019e9834 100644 --- a/mm/mempolicy.c +++ b/mm/mempolicy.c | |||
@@ -110,7 +110,7 @@ enum zone_type policy_zone = 0; | |||
110 | struct mempolicy default_policy = { | 110 | struct mempolicy default_policy = { |
111 | .refcnt = ATOMIC_INIT(1), /* never free it */ | 111 | .refcnt = ATOMIC_INIT(1), /* never free it */ |
112 | .mode = MPOL_PREFERRED, | 112 | .mode = MPOL_PREFERRED, |
113 | .v = { .preferred_node = -1 }, | 113 | .flags = MPOL_F_LOCAL, |
114 | }; | 114 | }; |
115 | 115 | ||
116 | static const struct mempolicy_operations { | 116 | static const struct mempolicy_operations { |
@@ -163,7 +163,7 @@ static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) | |||
163 | static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) | 163 | static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) |
164 | { | 164 | { |
165 | if (!nodes) | 165 | if (!nodes) |
166 | pol->v.preferred_node = -1; /* local allocation */ | 166 | pol->flags |= MPOL_F_LOCAL; /* local allocation */ |
167 | else if (nodes_empty(*nodes)) | 167 | else if (nodes_empty(*nodes)) |
168 | return -EINVAL; /* no allowed nodes */ | 168 | return -EINVAL; /* no allowed nodes */ |
169 | else | 169 | else |
@@ -290,14 +290,15 @@ static void mpol_rebind_preferred(struct mempolicy *pol, | |||
290 | if (pol->flags & MPOL_F_STATIC_NODES) { | 290 | if (pol->flags & MPOL_F_STATIC_NODES) { |
291 | int node = first_node(pol->w.user_nodemask); | 291 | int node = first_node(pol->w.user_nodemask); |
292 | 292 | ||
293 | if (node_isset(node, *nodes)) | 293 | if (node_isset(node, *nodes)) { |
294 | pol->v.preferred_node = node; | 294 | pol->v.preferred_node = node; |
295 | else | 295 | pol->flags &= ~MPOL_F_LOCAL; |
296 | pol->v.preferred_node = -1; | 296 | } else |
297 | pol->flags |= MPOL_F_LOCAL; | ||
297 | } else if (pol->flags & MPOL_F_RELATIVE_NODES) { | 298 | } else if (pol->flags & MPOL_F_RELATIVE_NODES) { |
298 | mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); | 299 | mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); |
299 | pol->v.preferred_node = first_node(tmp); | 300 | pol->v.preferred_node = first_node(tmp); |
300 | } else if (pol->v.preferred_node != -1) { | 301 | } else if (!(pol->flags & MPOL_F_LOCAL)) { |
301 | pol->v.preferred_node = node_remap(pol->v.preferred_node, | 302 | pol->v.preferred_node = node_remap(pol->v.preferred_node, |
302 | pol->w.cpuset_mems_allowed, | 303 | pol->w.cpuset_mems_allowed, |
303 | *nodes); | 304 | *nodes); |
@@ -645,7 +646,7 @@ static void get_policy_nodemask(struct mempolicy *p, nodemask_t *nodes) | |||
645 | *nodes = p->v.nodes; | 646 | *nodes = p->v.nodes; |
646 | break; | 647 | break; |
647 | case MPOL_PREFERRED: | 648 | case MPOL_PREFERRED: |
648 | if (p->v.preferred_node >= 0) | 649 | if (!(p->flags & MPOL_F_LOCAL)) |
649 | node_set(p->v.preferred_node, *nodes); | 650 | node_set(p->v.preferred_node, *nodes); |
650 | /* else return empty node mask for local allocation */ | 651 | /* else return empty node mask for local allocation */ |
651 | break; | 652 | break; |
@@ -1324,13 +1325,12 @@ static nodemask_t *policy_nodemask(gfp_t gfp, struct mempolicy *policy) | |||
1324 | /* Return a zonelist indicated by gfp for node representing a mempolicy */ | 1325 | /* Return a zonelist indicated by gfp for node representing a mempolicy */ |
1325 | static struct zonelist *policy_zonelist(gfp_t gfp, struct mempolicy *policy) | 1326 | static struct zonelist *policy_zonelist(gfp_t gfp, struct mempolicy *policy) |
1326 | { | 1327 | { |
1327 | int nd; | 1328 | int nd = numa_node_id(); |
1328 | 1329 | ||
1329 | switch (policy->mode) { | 1330 | switch (policy->mode) { |
1330 | case MPOL_PREFERRED: | 1331 | case MPOL_PREFERRED: |
1331 | nd = policy->v.preferred_node; | 1332 | if (!(policy->flags & MPOL_F_LOCAL)) |
1332 | if (nd < 0) | 1333 | nd = policy->v.preferred_node; |
1333 | nd = numa_node_id(); | ||
1334 | break; | 1334 | break; |
1335 | case MPOL_BIND: | 1335 | case MPOL_BIND: |
1336 | /* | 1336 | /* |
@@ -1339,16 +1339,13 @@ static struct zonelist *policy_zonelist(gfp_t gfp, struct mempolicy *policy) | |||
1339 | * current node is part of the mask, we use the zonelist for | 1339 | * current node is part of the mask, we use the zonelist for |
1340 | * the first node in the mask instead. | 1340 | * the first node in the mask instead. |
1341 | */ | 1341 | */ |
1342 | nd = numa_node_id(); | ||
1343 | if (unlikely(gfp & __GFP_THISNODE) && | 1342 | if (unlikely(gfp & __GFP_THISNODE) && |
1344 | unlikely(!node_isset(nd, policy->v.nodes))) | 1343 | unlikely(!node_isset(nd, policy->v.nodes))) |
1345 | nd = first_node(policy->v.nodes); | 1344 | nd = first_node(policy->v.nodes); |
1346 | break; | 1345 | break; |
1347 | case MPOL_INTERLEAVE: /* should not happen */ | 1346 | case MPOL_INTERLEAVE: /* should not happen */ |
1348 | nd = numa_node_id(); | ||
1349 | break; | 1347 | break; |
1350 | default: | 1348 | default: |
1351 | nd = 0; | ||
1352 | BUG(); | 1349 | BUG(); |
1353 | } | 1350 | } |
1354 | return node_zonelist(nd, gfp); | 1351 | return node_zonelist(nd, gfp); |
@@ -1379,14 +1376,15 @@ static unsigned interleave_nodes(struct mempolicy *policy) | |||
1379 | */ | 1376 | */ |
1380 | unsigned slab_node(struct mempolicy *policy) | 1377 | unsigned slab_node(struct mempolicy *policy) |
1381 | { | 1378 | { |
1382 | if (!policy) | 1379 | if (!policy || policy->flags & MPOL_F_LOCAL) |
1383 | return numa_node_id(); | 1380 | return numa_node_id(); |
1384 | 1381 | ||
1385 | switch (policy->mode) { | 1382 | switch (policy->mode) { |
1386 | case MPOL_PREFERRED: | 1383 | case MPOL_PREFERRED: |
1387 | if (unlikely(policy->v.preferred_node >= 0)) | 1384 | /* |
1388 | return policy->v.preferred_node; | 1385 | * handled MPOL_F_LOCAL above |
1389 | return numa_node_id(); | 1386 | */ |
1387 | return policy->v.preferred_node; | ||
1390 | 1388 | ||
1391 | case MPOL_INTERLEAVE: | 1389 | case MPOL_INTERLEAVE: |
1392 | return interleave_nodes(policy); | 1390 | return interleave_nodes(policy); |
@@ -1666,7 +1664,8 @@ int __mpol_equal(struct mempolicy *a, struct mempolicy *b) | |||
1666 | case MPOL_INTERLEAVE: | 1664 | case MPOL_INTERLEAVE: |
1667 | return nodes_equal(a->v.nodes, b->v.nodes); | 1665 | return nodes_equal(a->v.nodes, b->v.nodes); |
1668 | case MPOL_PREFERRED: | 1666 | case MPOL_PREFERRED: |
1669 | return a->v.preferred_node == b->v.preferred_node; | 1667 | return a->v.preferred_node == b->v.preferred_node && |
1668 | a->flags == b->flags; | ||
1670 | default: | 1669 | default: |
1671 | BUG(); | 1670 | BUG(); |
1672 | return 0; | 1671 | return 0; |
@@ -1946,7 +1945,7 @@ void numa_default_policy(void) | |||
1946 | } | 1945 | } |
1947 | 1946 | ||
1948 | /* | 1947 | /* |
1949 | * "local" is pseudo-policy: MPOL_PREFERRED with preferred_node == -1 | 1948 | * "local" is pseudo-policy: MPOL_PREFERRED with MPOL_F_LOCAL flag |
1950 | * Used only for mpol_to_str() | 1949 | * Used only for mpol_to_str() |
1951 | */ | 1950 | */ |
1952 | #define MPOL_LOCAL (MPOL_INTERLEAVE + 1) | 1951 | #define MPOL_LOCAL (MPOL_INTERLEAVE + 1) |
@@ -1962,7 +1961,6 @@ static inline int mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol) | |||
1962 | { | 1961 | { |
1963 | char *p = buffer; | 1962 | char *p = buffer; |
1964 | int l; | 1963 | int l; |
1965 | int nid; | ||
1966 | nodemask_t nodes; | 1964 | nodemask_t nodes; |
1967 | unsigned short mode; | 1965 | unsigned short mode; |
1968 | unsigned short flags = pol ? pol->flags : 0; | 1966 | unsigned short flags = pol ? pol->flags : 0; |
@@ -1979,11 +1977,10 @@ static inline int mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol) | |||
1979 | 1977 | ||
1980 | case MPOL_PREFERRED: | 1978 | case MPOL_PREFERRED: |
1981 | nodes_clear(nodes); | 1979 | nodes_clear(nodes); |
1982 | nid = pol->v.preferred_node; | 1980 | if (flags & MPOL_F_LOCAL) |
1983 | if (nid < 0) | ||
1984 | mode = MPOL_LOCAL; /* pseudo-policy */ | 1981 | mode = MPOL_LOCAL; /* pseudo-policy */ |
1985 | else | 1982 | else |
1986 | node_set(nid, nodes); | 1983 | node_set(pol->v.preferred_node, nodes); |
1987 | break; | 1984 | break; |
1988 | 1985 | ||
1989 | case MPOL_BIND: | 1986 | case MPOL_BIND: |
@@ -2004,7 +2001,7 @@ static inline int mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol) | |||
2004 | strcpy(p, policy_types[mode]); | 2001 | strcpy(p, policy_types[mode]); |
2005 | p += l; | 2002 | p += l; |
2006 | 2003 | ||
2007 | if (flags) { | 2004 | if (flags & MPOL_MODE_FLAGS) { |
2008 | int need_bar = 0; | 2005 | int need_bar = 0; |
2009 | 2006 | ||
2010 | if (buffer + maxlen < p + 2) | 2007 | if (buffer + maxlen < p + 2) |