diff options
Diffstat (limited to 'kernel/rcupreempt.c')
-rw-r--r-- | kernel/rcupreempt.c | 10 |
1 files changed, 5 insertions, 5 deletions
diff --git a/kernel/rcupreempt.c b/kernel/rcupreempt.c index 5e02b7740702..5cbd69edf5d9 100644 --- a/kernel/rcupreempt.c +++ b/kernel/rcupreempt.c | |||
@@ -655,7 +655,7 @@ rcu_try_flip_idle(void) | |||
655 | 655 | ||
656 | /* Now ask each CPU for acknowledgement of the flip. */ | 656 | /* Now ask each CPU for acknowledgement of the flip. */ |
657 | 657 | ||
658 | for_each_cpu_mask(cpu, rcu_cpu_online_map) { | 658 | for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) { |
659 | per_cpu(rcu_flip_flag, cpu) = rcu_flipped; | 659 | per_cpu(rcu_flip_flag, cpu) = rcu_flipped; |
660 | dyntick_save_progress_counter(cpu); | 660 | dyntick_save_progress_counter(cpu); |
661 | } | 661 | } |
@@ -673,7 +673,7 @@ rcu_try_flip_waitack(void) | |||
673 | int cpu; | 673 | int cpu; |
674 | 674 | ||
675 | RCU_TRACE_ME(rcupreempt_trace_try_flip_a1); | 675 | RCU_TRACE_ME(rcupreempt_trace_try_flip_a1); |
676 | for_each_cpu_mask(cpu, rcu_cpu_online_map) | 676 | for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) |
677 | if (rcu_try_flip_waitack_needed(cpu) && | 677 | if (rcu_try_flip_waitack_needed(cpu) && |
678 | per_cpu(rcu_flip_flag, cpu) != rcu_flip_seen) { | 678 | per_cpu(rcu_flip_flag, cpu) != rcu_flip_seen) { |
679 | RCU_TRACE_ME(rcupreempt_trace_try_flip_ae1); | 679 | RCU_TRACE_ME(rcupreempt_trace_try_flip_ae1); |
@@ -705,7 +705,7 @@ rcu_try_flip_waitzero(void) | |||
705 | /* Check to see if the sum of the "last" counters is zero. */ | 705 | /* Check to see if the sum of the "last" counters is zero. */ |
706 | 706 | ||
707 | RCU_TRACE_ME(rcupreempt_trace_try_flip_z1); | 707 | RCU_TRACE_ME(rcupreempt_trace_try_flip_z1); |
708 | for_each_cpu_mask(cpu, rcu_cpu_online_map) | 708 | for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) |
709 | sum += RCU_DATA_CPU(cpu)->rcu_flipctr[lastidx]; | 709 | sum += RCU_DATA_CPU(cpu)->rcu_flipctr[lastidx]; |
710 | if (sum != 0) { | 710 | if (sum != 0) { |
711 | RCU_TRACE_ME(rcupreempt_trace_try_flip_ze1); | 711 | RCU_TRACE_ME(rcupreempt_trace_try_flip_ze1); |
@@ -720,7 +720,7 @@ rcu_try_flip_waitzero(void) | |||
720 | smp_mb(); /* ^^^^^^^^^^^^ */ | 720 | smp_mb(); /* ^^^^^^^^^^^^ */ |
721 | 721 | ||
722 | /* Call for a memory barrier from each CPU. */ | 722 | /* Call for a memory barrier from each CPU. */ |
723 | for_each_cpu_mask(cpu, rcu_cpu_online_map) { | 723 | for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) { |
724 | per_cpu(rcu_mb_flag, cpu) = rcu_mb_needed; | 724 | per_cpu(rcu_mb_flag, cpu) = rcu_mb_needed; |
725 | dyntick_save_progress_counter(cpu); | 725 | dyntick_save_progress_counter(cpu); |
726 | } | 726 | } |
@@ -740,7 +740,7 @@ rcu_try_flip_waitmb(void) | |||
740 | int cpu; | 740 | int cpu; |
741 | 741 | ||
742 | RCU_TRACE_ME(rcupreempt_trace_try_flip_m1); | 742 | RCU_TRACE_ME(rcupreempt_trace_try_flip_m1); |
743 | for_each_cpu_mask(cpu, rcu_cpu_online_map) | 743 | for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) |
744 | if (rcu_try_flip_waitmb_needed(cpu) && | 744 | if (rcu_try_flip_waitmb_needed(cpu) && |
745 | per_cpu(rcu_mb_flag, cpu) != rcu_mb_done) { | 745 | per_cpu(rcu_mb_flag, cpu) != rcu_mb_done) { |
746 | RCU_TRACE_ME(rcupreempt_trace_try_flip_me1); | 746 | RCU_TRACE_ME(rcupreempt_trace_try_flip_me1); |