aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMel Gorman <mgorman@suse.de>2013-10-07 06:29:40 -0400
committerIngo Molnar <mingo@kernel.org>2013-10-09 08:48:23 -0400
commit989348b5fc2367d6880d23a1c779a90bbb6f9baf (patch)
treecc5b883e5e2534b6a29c01bee153efd8daf445d5
parentde1c9ce6f07fec0381a39a9d0b379ea35aa1167f (diff)
sched/numa: Use unsigned longs for numa group fault stats
As Peter says "If you're going to hold locks you can also do away with all that atomic_long_*() nonsense". Lock aquisition moved slightly to protect the updates. Signed-off-by: Mel Gorman <mgorman@suse.de> Reviewed-by: Rik van Riel <riel@redhat.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Srikar Dronamraju <srikar@linux.vnet.ibm.com> Signed-off-by: Peter Zijlstra <peterz@infradead.org> Link: http://lkml.kernel.org/r/1381141781-10992-63-git-send-email-mgorman@suse.de Signed-off-by: Ingo Molnar <mingo@kernel.org>
-rw-r--r--kernel/sched/fair.c49
1 files changed, 20 insertions, 29 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index e7884dc3416d..5b2208e504a4 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -916,8 +916,8 @@ struct numa_group {
916 struct list_head task_list; 916 struct list_head task_list;
917 917
918 struct rcu_head rcu; 918 struct rcu_head rcu;
919 atomic_long_t total_faults; 919 unsigned long total_faults;
920 atomic_long_t faults[0]; 920 unsigned long faults[0];
921}; 921};
922 922
923pid_t task_numa_group_id(struct task_struct *p) 923pid_t task_numa_group_id(struct task_struct *p)
@@ -944,8 +944,7 @@ static inline unsigned long group_faults(struct task_struct *p, int nid)
944 if (!p->numa_group) 944 if (!p->numa_group)
945 return 0; 945 return 0;
946 946
947 return atomic_long_read(&p->numa_group->faults[2*nid]) + 947 return p->numa_group->faults[2*nid] + p->numa_group->faults[2*nid+1];
948 atomic_long_read(&p->numa_group->faults[2*nid+1]);
949} 948}
950 949
951/* 950/*
@@ -971,17 +970,10 @@ static inline unsigned long task_weight(struct task_struct *p, int nid)
971 970
972static inline unsigned long group_weight(struct task_struct *p, int nid) 971static inline unsigned long group_weight(struct task_struct *p, int nid)
973{ 972{
974 unsigned long total_faults; 973 if (!p->numa_group || !p->numa_group->total_faults)
975
976 if (!p->numa_group)
977 return 0;
978
979 total_faults = atomic_long_read(&p->numa_group->total_faults);
980
981 if (!total_faults)
982 return 0; 974 return 0;
983 975
984 return 1000 * group_faults(p, nid) / total_faults; 976 return 1000 * group_faults(p, nid) / p->numa_group->total_faults;
985} 977}
986 978
987static unsigned long weighted_cpuload(const int cpu); 979static unsigned long weighted_cpuload(const int cpu);
@@ -1397,9 +1389,9 @@ static void task_numa_placement(struct task_struct *p)
1397 p->total_numa_faults += diff; 1389 p->total_numa_faults += diff;
1398 if (p->numa_group) { 1390 if (p->numa_group) {
1399 /* safe because we can only change our own group */ 1391 /* safe because we can only change our own group */
1400 atomic_long_add(diff, &p->numa_group->faults[i]); 1392 p->numa_group->faults[i] += diff;
1401 atomic_long_add(diff, &p->numa_group->total_faults); 1393 p->numa_group->total_faults += diff;
1402 group_faults += atomic_long_read(&p->numa_group->faults[i]); 1394 group_faults += p->numa_group->faults[i];
1403 } 1395 }
1404 } 1396 }
1405 1397
@@ -1475,7 +1467,7 @@ static void task_numa_group(struct task_struct *p, int cpupid, int flags,
1475 1467
1476 if (unlikely(!p->numa_group)) { 1468 if (unlikely(!p->numa_group)) {
1477 unsigned int size = sizeof(struct numa_group) + 1469 unsigned int size = sizeof(struct numa_group) +
1478 2*nr_node_ids*sizeof(atomic_long_t); 1470 2*nr_node_ids*sizeof(unsigned long);
1479 1471
1480 grp = kzalloc(size, GFP_KERNEL | __GFP_NOWARN); 1472 grp = kzalloc(size, GFP_KERNEL | __GFP_NOWARN);
1481 if (!grp) 1473 if (!grp)
@@ -1487,9 +1479,9 @@ static void task_numa_group(struct task_struct *p, int cpupid, int flags,
1487 grp->gid = p->pid; 1479 grp->gid = p->pid;
1488 1480
1489 for (i = 0; i < 2*nr_node_ids; i++) 1481 for (i = 0; i < 2*nr_node_ids; i++)
1490 atomic_long_set(&grp->faults[i], p->numa_faults[i]); 1482 grp->faults[i] = p->numa_faults[i];
1491 1483
1492 atomic_long_set(&grp->total_faults, p->total_numa_faults); 1484 grp->total_faults = p->total_numa_faults;
1493 1485
1494 list_add(&p->numa_entry, &grp->task_list); 1486 list_add(&p->numa_entry, &grp->task_list);
1495 grp->nr_tasks++; 1487 grp->nr_tasks++;
@@ -1543,14 +1535,14 @@ unlock:
1543 if (!join) 1535 if (!join)
1544 return; 1536 return;
1545 1537
1538 double_lock(&my_grp->lock, &grp->lock);
1539
1546 for (i = 0; i < 2*nr_node_ids; i++) { 1540 for (i = 0; i < 2*nr_node_ids; i++) {
1547 atomic_long_sub(p->numa_faults[i], &my_grp->faults[i]); 1541 my_grp->faults[i] -= p->numa_faults[i];
1548 atomic_long_add(p->numa_faults[i], &grp->faults[i]); 1542 grp->faults[i] += p->numa_faults[i];
1549 } 1543 }
1550 atomic_long_sub(p->total_numa_faults, &my_grp->total_faults); 1544 my_grp->total_faults -= p->total_numa_faults;
1551 atomic_long_add(p->total_numa_faults, &grp->total_faults); 1545 grp->total_faults += p->total_numa_faults;
1552
1553 double_lock(&my_grp->lock, &grp->lock);
1554 1546
1555 list_move(&p->numa_entry, &grp->task_list); 1547 list_move(&p->numa_entry, &grp->task_list);
1556 my_grp->nr_tasks--; 1548 my_grp->nr_tasks--;
@@ -1571,12 +1563,11 @@ void task_numa_free(struct task_struct *p)
1571 void *numa_faults = p->numa_faults; 1563 void *numa_faults = p->numa_faults;
1572 1564
1573 if (grp) { 1565 if (grp) {
1566 spin_lock(&grp->lock);
1574 for (i = 0; i < 2*nr_node_ids; i++) 1567 for (i = 0; i < 2*nr_node_ids; i++)
1575 atomic_long_sub(p->numa_faults[i], &grp->faults[i]); 1568 grp->faults[i] -= p->numa_faults[i];
1576 1569 grp->total_faults -= p->total_numa_faults;
1577 atomic_long_sub(p->total_numa_faults, &grp->total_faults);
1578 1570
1579 spin_lock(&grp->lock);
1580 list_del(&p->numa_entry); 1571 list_del(&p->numa_entry);
1581 grp->nr_tasks--; 1572 grp->nr_tasks--;
1582 spin_unlock(&grp->lock); 1573 spin_unlock(&grp->lock);