diff options
-rw-r--r-- | include/linux/percpu_counter.h | 6 | ||||
-rw-r--r-- | mm/swap.c | 25 |
2 files changed, 29 insertions, 2 deletions
diff --git a/include/linux/percpu_counter.h b/include/linux/percpu_counter.h index bd6708e2c027..682525511c9e 100644 --- a/include/linux/percpu_counter.h +++ b/include/linux/percpu_counter.h | |||
@@ -39,6 +39,7 @@ static inline void percpu_counter_destroy(struct percpu_counter *fbc) | |||
39 | } | 39 | } |
40 | 40 | ||
41 | void percpu_counter_mod(struct percpu_counter *fbc, long amount); | 41 | void percpu_counter_mod(struct percpu_counter *fbc, long amount); |
42 | long percpu_counter_sum(struct percpu_counter *fbc); | ||
42 | 43 | ||
43 | static inline long percpu_counter_read(struct percpu_counter *fbc) | 44 | static inline long percpu_counter_read(struct percpu_counter *fbc) |
44 | { | 45 | { |
@@ -92,6 +93,11 @@ static inline long percpu_counter_read_positive(struct percpu_counter *fbc) | |||
92 | return fbc->count; | 93 | return fbc->count; |
93 | } | 94 | } |
94 | 95 | ||
96 | static inline long percpu_counter_sum(struct percpu_counter *fbc) | ||
97 | { | ||
98 | return percpu_counter_read_positive(fbc); | ||
99 | } | ||
100 | |||
95 | #endif /* CONFIG_SMP */ | 101 | #endif /* CONFIG_SMP */ |
96 | 102 | ||
97 | static inline void percpu_counter_inc(struct percpu_counter *fbc) | 103 | static inline void percpu_counter_inc(struct percpu_counter *fbc) |
@@ -489,13 +489,34 @@ void percpu_counter_mod(struct percpu_counter *fbc, long amount) | |||
489 | if (count >= FBC_BATCH || count <= -FBC_BATCH) { | 489 | if (count >= FBC_BATCH || count <= -FBC_BATCH) { |
490 | spin_lock(&fbc->lock); | 490 | spin_lock(&fbc->lock); |
491 | fbc->count += count; | 491 | fbc->count += count; |
492 | *pcount = 0; | ||
492 | spin_unlock(&fbc->lock); | 493 | spin_unlock(&fbc->lock); |
493 | count = 0; | 494 | } else { |
495 | *pcount = count; | ||
494 | } | 496 | } |
495 | *pcount = count; | ||
496 | put_cpu(); | 497 | put_cpu(); |
497 | } | 498 | } |
498 | EXPORT_SYMBOL(percpu_counter_mod); | 499 | EXPORT_SYMBOL(percpu_counter_mod); |
500 | |||
501 | /* | ||
502 | * Add up all the per-cpu counts, return the result. This is a more accurate | ||
503 | * but much slower version of percpu_counter_read_positive() | ||
504 | */ | ||
505 | long percpu_counter_sum(struct percpu_counter *fbc) | ||
506 | { | ||
507 | long ret; | ||
508 | int cpu; | ||
509 | |||
510 | spin_lock(&fbc->lock); | ||
511 | ret = fbc->count; | ||
512 | for_each_cpu(cpu) { | ||
513 | long *pcount = per_cpu_ptr(fbc->counters, cpu); | ||
514 | ret += *pcount; | ||
515 | } | ||
516 | spin_unlock(&fbc->lock); | ||
517 | return ret < 0 ? 0 : ret; | ||
518 | } | ||
519 | EXPORT_SYMBOL(percpu_counter_sum); | ||
499 | #endif | 520 | #endif |
500 | 521 | ||
501 | /* | 522 | /* |