aboutsummaryrefslogtreecommitdiffstats
path: root/mm/mempolicy.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/mempolicy.c')
-rw-r--r--mm/mempolicy.c46
1 files changed, 23 insertions, 23 deletions
diff --git a/mm/mempolicy.c b/mm/mempolicy.c
index 8924aaf4665c..5e7eea2dc8b4 100644
--- a/mm/mempolicy.c
+++ b/mm/mempolicy.c
@@ -106,7 +106,7 @@ enum zone_type policy_zone = 0;
106 106
107struct mempolicy default_policy = { 107struct mempolicy default_policy = {
108 .refcnt = ATOMIC_INIT(1), /* never free it */ 108 .refcnt = ATOMIC_INIT(1), /* never free it */
109 .policy = MPOL_DEFAULT, 109 .mode = MPOL_DEFAULT,
110}; 110};
111 111
112static const struct mempolicy_operations { 112static const struct mempolicy_operations {
@@ -211,7 +211,7 @@ static struct mempolicy *mpol_new(unsigned short mode, unsigned short flags,
211 if (!policy) 211 if (!policy)
212 return ERR_PTR(-ENOMEM); 212 return ERR_PTR(-ENOMEM);
213 atomic_set(&policy->refcnt, 1); 213 atomic_set(&policy->refcnt, 1);
214 policy->policy = mode; 214 policy->mode = mode;
215 policy->flags = flags; 215 policy->flags = flags;
216 216
217 if (nodes) { 217 if (nodes) {
@@ -302,7 +302,7 @@ static void mpol_rebind_policy(struct mempolicy *pol,
302 if (!mpol_store_user_nodemask(pol) && 302 if (!mpol_store_user_nodemask(pol) &&
303 nodes_equal(pol->w.cpuset_mems_allowed, *newmask)) 303 nodes_equal(pol->w.cpuset_mems_allowed, *newmask))
304 return; 304 return;
305 mpol_ops[pol->policy].rebind(pol, newmask); 305 mpol_ops[pol->mode].rebind(pol, newmask);
306} 306}
307 307
308/* 308/*
@@ -608,7 +608,7 @@ static long do_set_mempolicy(unsigned short mode, unsigned short flags,
608 mpol_put(current->mempolicy); 608 mpol_put(current->mempolicy);
609 current->mempolicy = new; 609 current->mempolicy = new;
610 mpol_set_task_struct_flag(); 610 mpol_set_task_struct_flag();
611 if (new && new->policy == MPOL_INTERLEAVE && 611 if (new && new->mode == MPOL_INTERLEAVE &&
612 nodes_weight(new->v.nodes)) 612 nodes_weight(new->v.nodes))
613 current->il_next = first_node(new->v.nodes); 613 current->il_next = first_node(new->v.nodes);
614 if (mm) 614 if (mm)
@@ -621,7 +621,7 @@ static long do_set_mempolicy(unsigned short mode, unsigned short flags,
621static void get_zonemask(struct mempolicy *p, nodemask_t *nodes) 621static void get_zonemask(struct mempolicy *p, nodemask_t *nodes)
622{ 622{
623 nodes_clear(*nodes); 623 nodes_clear(*nodes);
624 switch (p->policy) { 624 switch (p->mode) {
625 case MPOL_DEFAULT: 625 case MPOL_DEFAULT:
626 break; 626 break;
627 case MPOL_BIND: 627 case MPOL_BIND:
@@ -700,14 +700,14 @@ static long do_get_mempolicy(int *policy, nodemask_t *nmask,
700 goto out; 700 goto out;
701 *policy = err; 701 *policy = err;
702 } else if (pol == current->mempolicy && 702 } else if (pol == current->mempolicy &&
703 pol->policy == MPOL_INTERLEAVE) { 703 pol->mode == MPOL_INTERLEAVE) {
704 *policy = current->il_next; 704 *policy = current->il_next;
705 } else { 705 } else {
706 err = -EINVAL; 706 err = -EINVAL;
707 goto out; 707 goto out;
708 } 708 }
709 } else 709 } else
710 *policy = pol->policy | pol->flags; 710 *policy = pol->mode | pol->flags;
711 711
712 if (vma) { 712 if (vma) {
713 up_read(&current->mm->mmap_sem); 713 up_read(&current->mm->mmap_sem);
@@ -1276,7 +1276,7 @@ static struct mempolicy *get_vma_policy(struct task_struct *task,
1276 pol = vpol; 1276 pol = vpol;
1277 shared_pol = 1; /* if pol non-NULL, add ref below */ 1277 shared_pol = 1; /* if pol non-NULL, add ref below */
1278 } else if (vma->vm_policy && 1278 } else if (vma->vm_policy &&
1279 vma->vm_policy->policy != MPOL_DEFAULT) 1279 vma->vm_policy->mode != MPOL_DEFAULT)
1280 pol = vma->vm_policy; 1280 pol = vma->vm_policy;
1281 } 1281 }
1282 if (!pol) 1282 if (!pol)
@@ -1290,7 +1290,7 @@ static struct mempolicy *get_vma_policy(struct task_struct *task,
1290static nodemask_t *nodemask_policy(gfp_t gfp, struct mempolicy *policy) 1290static nodemask_t *nodemask_policy(gfp_t gfp, struct mempolicy *policy)
1291{ 1291{
1292 /* Lower zones don't get a nodemask applied for MPOL_BIND */ 1292 /* Lower zones don't get a nodemask applied for MPOL_BIND */
1293 if (unlikely(policy->policy == MPOL_BIND) && 1293 if (unlikely(policy->mode == MPOL_BIND) &&
1294 gfp_zone(gfp) >= policy_zone && 1294 gfp_zone(gfp) >= policy_zone &&
1295 cpuset_nodemask_valid_mems_allowed(&policy->v.nodes)) 1295 cpuset_nodemask_valid_mems_allowed(&policy->v.nodes))
1296 return &policy->v.nodes; 1296 return &policy->v.nodes;
@@ -1303,7 +1303,7 @@ static struct zonelist *zonelist_policy(gfp_t gfp, struct mempolicy *policy)
1303{ 1303{
1304 int nd; 1304 int nd;
1305 1305
1306 switch (policy->policy) { 1306 switch (policy->mode) {
1307 case MPOL_PREFERRED: 1307 case MPOL_PREFERRED:
1308 nd = policy->v.preferred_node; 1308 nd = policy->v.preferred_node;
1309 if (nd < 0) 1309 if (nd < 0)
@@ -1353,7 +1353,7 @@ static unsigned interleave_nodes(struct mempolicy *policy)
1353 */ 1353 */
1354unsigned slab_node(struct mempolicy *policy) 1354unsigned slab_node(struct mempolicy *policy)
1355{ 1355{
1356 unsigned short pol = policy ? policy->policy : MPOL_DEFAULT; 1356 unsigned short pol = policy ? policy->mode : MPOL_DEFAULT;
1357 1357
1358 switch (pol) { 1358 switch (pol) {
1359 case MPOL_INTERLEAVE: 1359 case MPOL_INTERLEAVE:
@@ -1454,9 +1454,9 @@ struct zonelist *huge_zonelist(struct vm_area_struct *vma, unsigned long addr,
1454 1454
1455 *mpol = NULL; /* probably no unref needed */ 1455 *mpol = NULL; /* probably no unref needed */
1456 *nodemask = NULL; /* assume !MPOL_BIND */ 1456 *nodemask = NULL; /* assume !MPOL_BIND */
1457 if (pol->policy == MPOL_BIND) { 1457 if (pol->mode == MPOL_BIND) {
1458 *nodemask = &pol->v.nodes; 1458 *nodemask = &pol->v.nodes;
1459 } else if (pol->policy == MPOL_INTERLEAVE) { 1459 } else if (pol->mode == MPOL_INTERLEAVE) {
1460 unsigned nid; 1460 unsigned nid;
1461 1461
1462 nid = interleave_nid(pol, vma, addr, HPAGE_SHIFT); 1462 nid = interleave_nid(pol, vma, addr, HPAGE_SHIFT);
@@ -1468,7 +1468,7 @@ struct zonelist *huge_zonelist(struct vm_area_struct *vma, unsigned long addr,
1468 1468
1469 zl = zonelist_policy(GFP_HIGHUSER, pol); 1469 zl = zonelist_policy(GFP_HIGHUSER, pol);
1470 if (unlikely(pol != &default_policy && pol != current->mempolicy)) { 1470 if (unlikely(pol != &default_policy && pol != current->mempolicy)) {
1471 if (pol->policy != MPOL_BIND) 1471 if (pol->mode != MPOL_BIND)
1472 __mpol_put(pol); /* finished with pol */ 1472 __mpol_put(pol); /* finished with pol */
1473 else 1473 else
1474 *mpol = pol; /* unref needed after allocation */ 1474 *mpol = pol; /* unref needed after allocation */
@@ -1522,7 +1522,7 @@ alloc_page_vma(gfp_t gfp, struct vm_area_struct *vma, unsigned long addr)
1522 1522
1523 cpuset_update_task_memory_state(); 1523 cpuset_update_task_memory_state();
1524 1524
1525 if (unlikely(pol->policy == MPOL_INTERLEAVE)) { 1525 if (unlikely(pol->mode == MPOL_INTERLEAVE)) {
1526 unsigned nid; 1526 unsigned nid;
1527 1527
1528 nid = interleave_nid(pol, vma, addr, PAGE_SHIFT); 1528 nid = interleave_nid(pol, vma, addr, PAGE_SHIFT);
@@ -1574,7 +1574,7 @@ struct page *alloc_pages_current(gfp_t gfp, unsigned order)
1574 cpuset_update_task_memory_state(); 1574 cpuset_update_task_memory_state();
1575 if (!pol || in_interrupt() || (gfp & __GFP_THISNODE)) 1575 if (!pol || in_interrupt() || (gfp & __GFP_THISNODE))
1576 pol = &default_policy; 1576 pol = &default_policy;
1577 if (pol->policy == MPOL_INTERLEAVE) 1577 if (pol->mode == MPOL_INTERLEAVE)
1578 return alloc_page_interleave(gfp, order, interleave_nodes(pol)); 1578 return alloc_page_interleave(gfp, order, interleave_nodes(pol));
1579 return __alloc_pages_nodemask(gfp, order, 1579 return __alloc_pages_nodemask(gfp, order,
1580 zonelist_policy(gfp, pol), nodemask_policy(gfp, pol)); 1580 zonelist_policy(gfp, pol), nodemask_policy(gfp, pol));
@@ -1620,11 +1620,11 @@ int __mpol_equal(struct mempolicy *a, struct mempolicy *b)
1620{ 1620{
1621 if (!a || !b) 1621 if (!a || !b)
1622 return 0; 1622 return 0;
1623 if (a->policy != b->policy) 1623 if (a->mode != b->mode)
1624 return 0; 1624 return 0;
1625 if (a->policy != MPOL_DEFAULT && !mpol_match_intent(a, b)) 1625 if (a->mode != MPOL_DEFAULT && !mpol_match_intent(a, b))
1626 return 0; 1626 return 0;
1627 switch (a->policy) { 1627 switch (a->mode) {
1628 case MPOL_DEFAULT: 1628 case MPOL_DEFAULT:
1629 return 1; 1629 return 1;
1630 case MPOL_BIND: 1630 case MPOL_BIND:
@@ -1644,7 +1644,7 @@ void __mpol_put(struct mempolicy *p)
1644{ 1644{
1645 if (!atomic_dec_and_test(&p->refcnt)) 1645 if (!atomic_dec_and_test(&p->refcnt))
1646 return; 1646 return;
1647 p->policy = MPOL_DEFAULT; 1647 p->mode = MPOL_DEFAULT;
1648 kmem_cache_free(policy_cache, p); 1648 kmem_cache_free(policy_cache, p);
1649} 1649}
1650 1650
@@ -1710,7 +1710,7 @@ static void sp_insert(struct shared_policy *sp, struct sp_node *new)
1710 rb_link_node(&new->nd, parent, p); 1710 rb_link_node(&new->nd, parent, p);
1711 rb_insert_color(&new->nd, &sp->root); 1711 rb_insert_color(&new->nd, &sp->root);
1712 pr_debug("inserting %lx-%lx: %d\n", new->start, new->end, 1712 pr_debug("inserting %lx-%lx: %d\n", new->start, new->end,
1713 new->policy ? new->policy->policy : 0); 1713 new->policy ? new->policy->mode : 0);
1714} 1714}
1715 1715
1716/* Find shared policy intersecting idx */ 1716/* Find shared policy intersecting idx */
@@ -1835,7 +1835,7 @@ int mpol_set_shared_policy(struct shared_policy *info,
1835 1835
1836 pr_debug("set_shared_policy %lx sz %lu %d %d %lx\n", 1836 pr_debug("set_shared_policy %lx sz %lu %d %d %lx\n",
1837 vma->vm_pgoff, 1837 vma->vm_pgoff,
1838 sz, npol ? npol->policy : -1, 1838 sz, npol ? npol->mode : -1,
1839 npol ? npol->flags : -1, 1839 npol ? npol->flags : -1,
1840 npol ? nodes_addr(npol->v.nodes)[0] : -1); 1840 npol ? nodes_addr(npol->v.nodes)[0] : -1);
1841 1841
@@ -1935,7 +1935,7 @@ static inline int mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol)
1935 char *p = buffer; 1935 char *p = buffer;
1936 int l; 1936 int l;
1937 nodemask_t nodes; 1937 nodemask_t nodes;
1938 unsigned short mode = pol ? pol->policy : MPOL_DEFAULT; 1938 unsigned short mode = pol ? pol->mode : MPOL_DEFAULT;
1939 unsigned short flags = pol ? pol->flags : 0; 1939 unsigned short flags = pol ? pol->flags : 0;
1940 1940
1941 switch (mode) { 1941 switch (mode) {