diff options
author | Mel Gorman <mgorman@suse.de> | 2013-10-07 06:29:40 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2013-10-09 08:48:23 -0400 |
commit | 989348b5fc2367d6880d23a1c779a90bbb6f9baf (patch) | |
tree | cc5b883e5e2534b6a29c01bee153efd8daf445d5 | |
parent | de1c9ce6f07fec0381a39a9d0b379ea35aa1167f (diff) |
sched/numa: Use unsigned longs for numa group fault stats
As Peter says "If you're going to hold locks you can also do away with all
that atomic_long_*() nonsense". Lock aquisition moved slightly to protect
the updates.
Signed-off-by: Mel Gorman <mgorman@suse.de>
Reviewed-by: Rik van Riel <riel@redhat.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Srikar Dronamraju <srikar@linux.vnet.ibm.com>
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Link: http://lkml.kernel.org/r/1381141781-10992-63-git-send-email-mgorman@suse.de
Signed-off-by: Ingo Molnar <mingo@kernel.org>
-rw-r--r-- | kernel/sched/fair.c | 49 |
1 files changed, 20 insertions, 29 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index e7884dc3416d..5b2208e504a4 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -916,8 +916,8 @@ struct numa_group { | |||
916 | struct list_head task_list; | 916 | struct list_head task_list; |
917 | 917 | ||
918 | struct rcu_head rcu; | 918 | struct rcu_head rcu; |
919 | atomic_long_t total_faults; | 919 | unsigned long total_faults; |
920 | atomic_long_t faults[0]; | 920 | unsigned long faults[0]; |
921 | }; | 921 | }; |
922 | 922 | ||
923 | pid_t task_numa_group_id(struct task_struct *p) | 923 | pid_t task_numa_group_id(struct task_struct *p) |
@@ -944,8 +944,7 @@ static inline unsigned long group_faults(struct task_struct *p, int nid) | |||
944 | if (!p->numa_group) | 944 | if (!p->numa_group) |
945 | return 0; | 945 | return 0; |
946 | 946 | ||
947 | return atomic_long_read(&p->numa_group->faults[2*nid]) + | 947 | return p->numa_group->faults[2*nid] + p->numa_group->faults[2*nid+1]; |
948 | atomic_long_read(&p->numa_group->faults[2*nid+1]); | ||
949 | } | 948 | } |
950 | 949 | ||
951 | /* | 950 | /* |
@@ -971,17 +970,10 @@ static inline unsigned long task_weight(struct task_struct *p, int nid) | |||
971 | 970 | ||
972 | static inline unsigned long group_weight(struct task_struct *p, int nid) | 971 | static inline unsigned long group_weight(struct task_struct *p, int nid) |
973 | { | 972 | { |
974 | unsigned long total_faults; | 973 | if (!p->numa_group || !p->numa_group->total_faults) |
975 | |||
976 | if (!p->numa_group) | ||
977 | return 0; | ||
978 | |||
979 | total_faults = atomic_long_read(&p->numa_group->total_faults); | ||
980 | |||
981 | if (!total_faults) | ||
982 | return 0; | 974 | return 0; |
983 | 975 | ||
984 | return 1000 * group_faults(p, nid) / total_faults; | 976 | return 1000 * group_faults(p, nid) / p->numa_group->total_faults; |
985 | } | 977 | } |
986 | 978 | ||
987 | static unsigned long weighted_cpuload(const int cpu); | 979 | static unsigned long weighted_cpuload(const int cpu); |
@@ -1397,9 +1389,9 @@ static void task_numa_placement(struct task_struct *p) | |||
1397 | p->total_numa_faults += diff; | 1389 | p->total_numa_faults += diff; |
1398 | if (p->numa_group) { | 1390 | if (p->numa_group) { |
1399 | /* safe because we can only change our own group */ | 1391 | /* safe because we can only change our own group */ |
1400 | atomic_long_add(diff, &p->numa_group->faults[i]); | 1392 | p->numa_group->faults[i] += diff; |
1401 | atomic_long_add(diff, &p->numa_group->total_faults); | 1393 | p->numa_group->total_faults += diff; |
1402 | group_faults += atomic_long_read(&p->numa_group->faults[i]); | 1394 | group_faults += p->numa_group->faults[i]; |
1403 | } | 1395 | } |
1404 | } | 1396 | } |
1405 | 1397 | ||
@@ -1475,7 +1467,7 @@ static void task_numa_group(struct task_struct *p, int cpupid, int flags, | |||
1475 | 1467 | ||
1476 | if (unlikely(!p->numa_group)) { | 1468 | if (unlikely(!p->numa_group)) { |
1477 | unsigned int size = sizeof(struct numa_group) + | 1469 | unsigned int size = sizeof(struct numa_group) + |
1478 | 2*nr_node_ids*sizeof(atomic_long_t); | 1470 | 2*nr_node_ids*sizeof(unsigned long); |
1479 | 1471 | ||
1480 | grp = kzalloc(size, GFP_KERNEL | __GFP_NOWARN); | 1472 | grp = kzalloc(size, GFP_KERNEL | __GFP_NOWARN); |
1481 | if (!grp) | 1473 | if (!grp) |
@@ -1487,9 +1479,9 @@ static void task_numa_group(struct task_struct *p, int cpupid, int flags, | |||
1487 | grp->gid = p->pid; | 1479 | grp->gid = p->pid; |
1488 | 1480 | ||
1489 | for (i = 0; i < 2*nr_node_ids; i++) | 1481 | for (i = 0; i < 2*nr_node_ids; i++) |
1490 | atomic_long_set(&grp->faults[i], p->numa_faults[i]); | 1482 | grp->faults[i] = p->numa_faults[i]; |
1491 | 1483 | ||
1492 | atomic_long_set(&grp->total_faults, p->total_numa_faults); | 1484 | grp->total_faults = p->total_numa_faults; |
1493 | 1485 | ||
1494 | list_add(&p->numa_entry, &grp->task_list); | 1486 | list_add(&p->numa_entry, &grp->task_list); |
1495 | grp->nr_tasks++; | 1487 | grp->nr_tasks++; |
@@ -1543,14 +1535,14 @@ unlock: | |||
1543 | if (!join) | 1535 | if (!join) |
1544 | return; | 1536 | return; |
1545 | 1537 | ||
1538 | double_lock(&my_grp->lock, &grp->lock); | ||
1539 | |||
1546 | for (i = 0; i < 2*nr_node_ids; i++) { | 1540 | for (i = 0; i < 2*nr_node_ids; i++) { |
1547 | atomic_long_sub(p->numa_faults[i], &my_grp->faults[i]); | 1541 | my_grp->faults[i] -= p->numa_faults[i]; |
1548 | atomic_long_add(p->numa_faults[i], &grp->faults[i]); | 1542 | grp->faults[i] += p->numa_faults[i]; |
1549 | } | 1543 | } |
1550 | atomic_long_sub(p->total_numa_faults, &my_grp->total_faults); | 1544 | my_grp->total_faults -= p->total_numa_faults; |
1551 | atomic_long_add(p->total_numa_faults, &grp->total_faults); | 1545 | grp->total_faults += p->total_numa_faults; |
1552 | |||
1553 | double_lock(&my_grp->lock, &grp->lock); | ||
1554 | 1546 | ||
1555 | list_move(&p->numa_entry, &grp->task_list); | 1547 | list_move(&p->numa_entry, &grp->task_list); |
1556 | my_grp->nr_tasks--; | 1548 | my_grp->nr_tasks--; |
@@ -1571,12 +1563,11 @@ void task_numa_free(struct task_struct *p) | |||
1571 | void *numa_faults = p->numa_faults; | 1563 | void *numa_faults = p->numa_faults; |
1572 | 1564 | ||
1573 | if (grp) { | 1565 | if (grp) { |
1566 | spin_lock(&grp->lock); | ||
1574 | for (i = 0; i < 2*nr_node_ids; i++) | 1567 | for (i = 0; i < 2*nr_node_ids; i++) |
1575 | atomic_long_sub(p->numa_faults[i], &grp->faults[i]); | 1568 | grp->faults[i] -= p->numa_faults[i]; |
1576 | 1569 | grp->total_faults -= p->total_numa_faults; | |
1577 | atomic_long_sub(p->total_numa_faults, &grp->total_faults); | ||
1578 | 1570 | ||
1579 | spin_lock(&grp->lock); | ||
1580 | list_del(&p->numa_entry); | 1571 | list_del(&p->numa_entry); |
1581 | grp->nr_tasks--; | 1572 | grp->nr_tasks--; |
1582 | spin_unlock(&grp->lock); | 1573 | spin_unlock(&grp->lock); |