diff options
Diffstat (limited to 'kernel/rcutree.c')
-rw-r--r-- | kernel/rcutree.c | 115 |
1 files changed, 77 insertions, 38 deletions
diff --git a/kernel/rcutree.c b/kernel/rcutree.c index 18e33313873e..28fd92a9e0d0 100644 --- a/kernel/rcutree.c +++ b/kernel/rcutree.c | |||
@@ -81,6 +81,8 @@ DEFINE_PER_CPU(struct rcu_data, rcu_sched_data); | |||
81 | struct rcu_state rcu_bh_state = RCU_STATE_INITIALIZER(rcu_bh_state); | 81 | struct rcu_state rcu_bh_state = RCU_STATE_INITIALIZER(rcu_bh_state); |
82 | DEFINE_PER_CPU(struct rcu_data, rcu_bh_data); | 82 | DEFINE_PER_CPU(struct rcu_data, rcu_bh_data); |
83 | 83 | ||
84 | static struct rcu_state *rcu_state; | ||
85 | |||
84 | int rcu_scheduler_active __read_mostly; | 86 | int rcu_scheduler_active __read_mostly; |
85 | EXPORT_SYMBOL_GPL(rcu_scheduler_active); | 87 | EXPORT_SYMBOL_GPL(rcu_scheduler_active); |
86 | 88 | ||
@@ -94,7 +96,7 @@ static DEFINE_PER_CPU(char, rcu_cpu_has_work); | |||
94 | static char rcu_kthreads_spawnable; | 96 | static char rcu_kthreads_spawnable; |
95 | 97 | ||
96 | static void rcu_node_kthread_setaffinity(struct rcu_node *rnp); | 98 | static void rcu_node_kthread_setaffinity(struct rcu_node *rnp); |
97 | static void invoke_rcu_kthread(void); | 99 | static void invoke_rcu_cpu_kthread(void); |
98 | 100 | ||
99 | #define RCU_KTHREAD_PRIO 1 /* RT priority for per-CPU kthreads. */ | 101 | #define RCU_KTHREAD_PRIO 1 /* RT priority for per-CPU kthreads. */ |
100 | 102 | ||
@@ -791,6 +793,7 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags) | |||
791 | rnp->completed = rsp->completed; | 793 | rnp->completed = rsp->completed; |
792 | rsp->signaled = RCU_SIGNAL_INIT; /* force_quiescent_state OK. */ | 794 | rsp->signaled = RCU_SIGNAL_INIT; /* force_quiescent_state OK. */ |
793 | rcu_start_gp_per_cpu(rsp, rnp, rdp); | 795 | rcu_start_gp_per_cpu(rsp, rnp, rdp); |
796 | rcu_preempt_boost_start_gp(rnp); | ||
794 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 797 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
795 | return; | 798 | return; |
796 | } | 799 | } |
@@ -826,6 +829,7 @@ rcu_start_gp(struct rcu_state *rsp, unsigned long flags) | |||
826 | rnp->completed = rsp->completed; | 829 | rnp->completed = rsp->completed; |
827 | if (rnp == rdp->mynode) | 830 | if (rnp == rdp->mynode) |
828 | rcu_start_gp_per_cpu(rsp, rnp, rdp); | 831 | rcu_start_gp_per_cpu(rsp, rnp, rdp); |
832 | rcu_preempt_boost_start_gp(rnp); | ||
829 | raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */ | 833 | raw_spin_unlock(&rnp->lock); /* irqs remain disabled. */ |
830 | } | 834 | } |
831 | 835 | ||
@@ -882,7 +886,7 @@ rcu_report_qs_rnp(unsigned long mask, struct rcu_state *rsp, | |||
882 | return; | 886 | return; |
883 | } | 887 | } |
884 | rnp->qsmask &= ~mask; | 888 | rnp->qsmask &= ~mask; |
885 | if (rnp->qsmask != 0 || rcu_preempted_readers(rnp)) { | 889 | if (rnp->qsmask != 0 || rcu_preempt_blocked_readers_cgp(rnp)) { |
886 | 890 | ||
887 | /* Other bits still set at this level, so done. */ | 891 | /* Other bits still set at this level, so done. */ |
888 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 892 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
@@ -1089,8 +1093,11 @@ static void __rcu_offline_cpu(int cpu, struct rcu_state *rsp) | |||
1089 | t = rnp->node_kthread_task; | 1093 | t = rnp->node_kthread_task; |
1090 | if (t != NULL && | 1094 | if (t != NULL && |
1091 | rnp->qsmaskinit == 0) { | 1095 | rnp->qsmaskinit == 0) { |
1092 | kthread_stop(t); | 1096 | raw_spin_lock_irqsave(&rnp->lock, flags); |
1093 | rnp->node_kthread_task = NULL; | 1097 | rnp->node_kthread_task = NULL; |
1098 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | ||
1099 | kthread_stop(t); | ||
1100 | rcu_stop_boost_kthread(rnp); | ||
1094 | } else | 1101 | } else |
1095 | rcu_node_kthread_setaffinity(rnp); | 1102 | rcu_node_kthread_setaffinity(rnp); |
1096 | } | 1103 | } |
@@ -1190,7 +1197,7 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp) | |||
1190 | 1197 | ||
1191 | /* Re-raise the RCU softirq if there are callbacks remaining. */ | 1198 | /* Re-raise the RCU softirq if there are callbacks remaining. */ |
1192 | if (cpu_has_callbacks_ready_to_invoke(rdp)) | 1199 | if (cpu_has_callbacks_ready_to_invoke(rdp)) |
1193 | invoke_rcu_kthread(); | 1200 | invoke_rcu_cpu_kthread(); |
1194 | } | 1201 | } |
1195 | 1202 | ||
1196 | /* | 1203 | /* |
@@ -1236,7 +1243,7 @@ void rcu_check_callbacks(int cpu, int user) | |||
1236 | } | 1243 | } |
1237 | rcu_preempt_check_callbacks(cpu); | 1244 | rcu_preempt_check_callbacks(cpu); |
1238 | if (rcu_pending(cpu)) | 1245 | if (rcu_pending(cpu)) |
1239 | invoke_rcu_kthread(); | 1246 | invoke_rcu_cpu_kthread(); |
1240 | } | 1247 | } |
1241 | 1248 | ||
1242 | #ifdef CONFIG_SMP | 1249 | #ifdef CONFIG_SMP |
@@ -1244,6 +1251,8 @@ void rcu_check_callbacks(int cpu, int user) | |||
1244 | /* | 1251 | /* |
1245 | * Scan the leaf rcu_node structures, processing dyntick state for any that | 1252 | * Scan the leaf rcu_node structures, processing dyntick state for any that |
1246 | * have not yet encountered a quiescent state, using the function specified. | 1253 | * have not yet encountered a quiescent state, using the function specified. |
1254 | * Also initiate boosting for any threads blocked on the root rcu_node. | ||
1255 | * | ||
1247 | * The caller must have suppressed start of new grace periods. | 1256 | * The caller must have suppressed start of new grace periods. |
1248 | */ | 1257 | */ |
1249 | static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *)) | 1258 | static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *)) |
@@ -1262,6 +1271,7 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *)) | |||
1262 | return; | 1271 | return; |
1263 | } | 1272 | } |
1264 | if (rnp->qsmask == 0) { | 1273 | if (rnp->qsmask == 0) { |
1274 | rcu_initiate_boost(rnp); | ||
1265 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 1275 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
1266 | continue; | 1276 | continue; |
1267 | } | 1277 | } |
@@ -1280,6 +1290,11 @@ static void force_qs_rnp(struct rcu_state *rsp, int (*f)(struct rcu_data *)) | |||
1280 | } | 1290 | } |
1281 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 1291 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
1282 | } | 1292 | } |
1293 | rnp = rcu_get_root(rsp); | ||
1294 | raw_spin_lock_irqsave(&rnp->lock, flags); | ||
1295 | if (rnp->qsmask == 0) | ||
1296 | rcu_initiate_boost(rnp); | ||
1297 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | ||
1283 | } | 1298 | } |
1284 | 1299 | ||
1285 | /* | 1300 | /* |
@@ -1417,7 +1432,7 @@ static void rcu_process_callbacks(void) | |||
1417 | * the current CPU with interrupts disabled, the rcu_cpu_kthread_task | 1432 | * the current CPU with interrupts disabled, the rcu_cpu_kthread_task |
1418 | * cannot disappear out from under us. | 1433 | * cannot disappear out from under us. |
1419 | */ | 1434 | */ |
1420 | static void invoke_rcu_kthread(void) | 1435 | static void invoke_rcu_cpu_kthread(void) |
1421 | { | 1436 | { |
1422 | unsigned long flags; | 1437 | unsigned long flags; |
1423 | wait_queue_head_t *q; | 1438 | wait_queue_head_t *q; |
@@ -1436,24 +1451,33 @@ static void invoke_rcu_kthread(void) | |||
1436 | } | 1451 | } |
1437 | 1452 | ||
1438 | /* | 1453 | /* |
1454 | * Wake up the specified per-rcu_node-structure kthread. | ||
1455 | * The caller must hold ->lock. | ||
1456 | */ | ||
1457 | static void invoke_rcu_node_kthread(struct rcu_node *rnp) | ||
1458 | { | ||
1459 | struct task_struct *t; | ||
1460 | |||
1461 | t = rnp->node_kthread_task; | ||
1462 | if (t != NULL) | ||
1463 | wake_up_process(t); | ||
1464 | } | ||
1465 | |||
1466 | /* | ||
1439 | * Timer handler to initiate the waking up of per-CPU kthreads that | 1467 | * Timer handler to initiate the waking up of per-CPU kthreads that |
1440 | * have yielded the CPU due to excess numbers of RCU callbacks. | 1468 | * have yielded the CPU due to excess numbers of RCU callbacks. |
1469 | * We wake up the per-rcu_node kthread, which in turn will wake up | ||
1470 | * the booster kthread. | ||
1441 | */ | 1471 | */ |
1442 | static void rcu_cpu_kthread_timer(unsigned long arg) | 1472 | static void rcu_cpu_kthread_timer(unsigned long arg) |
1443 | { | 1473 | { |
1444 | unsigned long flags; | 1474 | unsigned long flags; |
1445 | struct rcu_data *rdp = (struct rcu_data *)arg; | 1475 | struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, arg); |
1446 | struct rcu_node *rnp = rdp->mynode; | 1476 | struct rcu_node *rnp = rdp->mynode; |
1447 | struct task_struct *t; | ||
1448 | 1477 | ||
1449 | raw_spin_lock_irqsave(&rnp->lock, flags); | 1478 | raw_spin_lock_irqsave(&rnp->lock, flags); |
1450 | rnp->wakemask |= rdp->grpmask; | 1479 | rnp->wakemask |= rdp->grpmask; |
1451 | t = rnp->node_kthread_task; | 1480 | invoke_rcu_node_kthread(rnp); |
1452 | if (t == NULL) { | ||
1453 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | ||
1454 | return; | ||
1455 | } | ||
1456 | wake_up_process(t); | ||
1457 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 1481 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
1458 | } | 1482 | } |
1459 | 1483 | ||
@@ -1463,13 +1487,12 @@ static void rcu_cpu_kthread_timer(unsigned long arg) | |||
1463 | * remain preempted. Either way, we restore our real-time priority | 1487 | * remain preempted. Either way, we restore our real-time priority |
1464 | * before returning. | 1488 | * before returning. |
1465 | */ | 1489 | */ |
1466 | static void rcu_yield(int cpu) | 1490 | static void rcu_yield(void (*f)(unsigned long), unsigned long arg) |
1467 | { | 1491 | { |
1468 | struct rcu_data *rdp = per_cpu_ptr(rcu_sched_state.rda, cpu); | ||
1469 | struct sched_param sp; | 1492 | struct sched_param sp; |
1470 | struct timer_list yield_timer; | 1493 | struct timer_list yield_timer; |
1471 | 1494 | ||
1472 | setup_timer_on_stack(&yield_timer, rcu_cpu_kthread_timer, (unsigned long)rdp); | 1495 | setup_timer_on_stack(&yield_timer, f, arg); |
1473 | mod_timer(&yield_timer, jiffies + 2); | 1496 | mod_timer(&yield_timer, jiffies + 2); |
1474 | sp.sched_priority = 0; | 1497 | sp.sched_priority = 0; |
1475 | sched_setscheduler_nocheck(current, SCHED_NORMAL, &sp); | 1498 | sched_setscheduler_nocheck(current, SCHED_NORMAL, &sp); |
@@ -1540,7 +1563,7 @@ static int rcu_cpu_kthread(void *arg) | |||
1540 | else | 1563 | else |
1541 | spincnt = 0; | 1564 | spincnt = 0; |
1542 | if (spincnt > 10) { | 1565 | if (spincnt > 10) { |
1543 | rcu_yield(cpu); | 1566 | rcu_yield(rcu_cpu_kthread_timer, (unsigned long)cpu); |
1544 | spincnt = 0; | 1567 | spincnt = 0; |
1545 | } | 1568 | } |
1546 | } | 1569 | } |
@@ -1597,6 +1620,7 @@ static int rcu_node_kthread(void *arg) | |||
1597 | raw_spin_lock_irqsave(&rnp->lock, flags); | 1620 | raw_spin_lock_irqsave(&rnp->lock, flags); |
1598 | mask = rnp->wakemask; | 1621 | mask = rnp->wakemask; |
1599 | rnp->wakemask = 0; | 1622 | rnp->wakemask = 0; |
1623 | rcu_initiate_boost(rnp); | ||
1600 | raw_spin_unlock_irqrestore(&rnp->lock, flags); | 1624 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
1601 | for (cpu = rnp->grplo; cpu <= rnp->grphi; cpu++, mask >>= 1) { | 1625 | for (cpu = rnp->grplo; cpu <= rnp->grphi; cpu++, mask >>= 1) { |
1602 | if ((mask & 0x1) == 0) | 1626 | if ((mask & 0x1) == 0) |
@@ -1618,7 +1642,8 @@ static int rcu_node_kthread(void *arg) | |||
1618 | 1642 | ||
1619 | /* | 1643 | /* |
1620 | * Set the per-rcu_node kthread's affinity to cover all CPUs that are | 1644 | * Set the per-rcu_node kthread's affinity to cover all CPUs that are |
1621 | * served by the rcu_node in question. | 1645 | * served by the rcu_node in question. The CPU hotplug lock is still |
1646 | * held, so the value of rnp->qsmaskinit will be stable. | ||
1622 | */ | 1647 | */ |
1623 | static void rcu_node_kthread_setaffinity(struct rcu_node *rnp) | 1648 | static void rcu_node_kthread_setaffinity(struct rcu_node *rnp) |
1624 | { | 1649 | { |
@@ -1626,8 +1651,7 @@ static void rcu_node_kthread_setaffinity(struct rcu_node *rnp) | |||
1626 | int cpu; | 1651 | int cpu; |
1627 | unsigned long mask = rnp->qsmaskinit; | 1652 | unsigned long mask = rnp->qsmaskinit; |
1628 | 1653 | ||
1629 | if (rnp->node_kthread_task == NULL || | 1654 | if (rnp->node_kthread_task == NULL || mask == 0) |
1630 | rnp->qsmaskinit == 0) | ||
1631 | return; | 1655 | return; |
1632 | if (!alloc_cpumask_var(&cm, GFP_KERNEL)) | 1656 | if (!alloc_cpumask_var(&cm, GFP_KERNEL)) |
1633 | return; | 1657 | return; |
@@ -1636,31 +1660,40 @@ static void rcu_node_kthread_setaffinity(struct rcu_node *rnp) | |||
1636 | if (mask & 0x1) | 1660 | if (mask & 0x1) |
1637 | cpumask_set_cpu(cpu, cm); | 1661 | cpumask_set_cpu(cpu, cm); |
1638 | set_cpus_allowed_ptr(rnp->node_kthread_task, cm); | 1662 | set_cpus_allowed_ptr(rnp->node_kthread_task, cm); |
1663 | rcu_boost_kthread_setaffinity(rnp, cm); | ||
1639 | free_cpumask_var(cm); | 1664 | free_cpumask_var(cm); |
1640 | } | 1665 | } |
1641 | 1666 | ||
1642 | /* | 1667 | /* |
1643 | * Spawn a per-rcu_node kthread, setting priority and affinity. | 1668 | * Spawn a per-rcu_node kthread, setting priority and affinity. |
1669 | * Called during boot before online/offline can happen, or, if | ||
1670 | * during runtime, with the main CPU-hotplug locks held. So only | ||
1671 | * one of these can be executing at a time. | ||
1644 | */ | 1672 | */ |
1645 | static int __cpuinit rcu_spawn_one_node_kthread(struct rcu_state *rsp, | 1673 | static int __cpuinit rcu_spawn_one_node_kthread(struct rcu_state *rsp, |
1646 | struct rcu_node *rnp) | 1674 | struct rcu_node *rnp) |
1647 | { | 1675 | { |
1676 | unsigned long flags; | ||
1648 | int rnp_index = rnp - &rsp->node[0]; | 1677 | int rnp_index = rnp - &rsp->node[0]; |
1649 | struct sched_param sp; | 1678 | struct sched_param sp; |
1650 | struct task_struct *t; | 1679 | struct task_struct *t; |
1651 | 1680 | ||
1652 | if (!rcu_kthreads_spawnable || | 1681 | if (!rcu_kthreads_spawnable || |
1653 | rnp->qsmaskinit == 0 || | 1682 | rnp->qsmaskinit == 0) |
1654 | rnp->node_kthread_task != NULL) | ||
1655 | return 0; | 1683 | return 0; |
1656 | t = kthread_create(rcu_node_kthread, (void *)rnp, "rcun%d", rnp_index); | 1684 | if (rnp->node_kthread_task == NULL) { |
1657 | if (IS_ERR(t)) | 1685 | t = kthread_create(rcu_node_kthread, (void *)rnp, |
1658 | return PTR_ERR(t); | 1686 | "rcun%d", rnp_index); |
1659 | rnp->node_kthread_task = t; | 1687 | if (IS_ERR(t)) |
1660 | wake_up_process(t); | 1688 | return PTR_ERR(t); |
1661 | sp.sched_priority = 99; | 1689 | raw_spin_lock_irqsave(&rnp->lock, flags); |
1662 | sched_setscheduler_nocheck(t, SCHED_FIFO, &sp); | 1690 | rnp->node_kthread_task = t; |
1663 | return 0; | 1691 | raw_spin_unlock_irqrestore(&rnp->lock, flags); |
1692 | wake_up_process(t); | ||
1693 | sp.sched_priority = 99; | ||
1694 | sched_setscheduler_nocheck(t, SCHED_FIFO, &sp); | ||
1695 | } | ||
1696 | return rcu_spawn_one_boost_kthread(rsp, rnp, rnp_index); | ||
1664 | } | 1697 | } |
1665 | 1698 | ||
1666 | /* | 1699 | /* |
@@ -1678,10 +1711,16 @@ static int __init rcu_spawn_kthreads(void) | |||
1678 | if (cpu_online(cpu)) | 1711 | if (cpu_online(cpu)) |
1679 | (void)rcu_spawn_one_cpu_kthread(cpu); | 1712 | (void)rcu_spawn_one_cpu_kthread(cpu); |
1680 | } | 1713 | } |
1681 | rcu_for_each_leaf_node(&rcu_sched_state, rnp) { | 1714 | rnp = rcu_get_root(rcu_state); |
1682 | init_waitqueue_head(&rnp->node_wq); | 1715 | init_waitqueue_head(&rnp->node_wq); |
1683 | (void)rcu_spawn_one_node_kthread(&rcu_sched_state, rnp); | 1716 | rcu_init_boost_waitqueue(rnp); |
1684 | } | 1717 | (void)rcu_spawn_one_node_kthread(rcu_state, rnp); |
1718 | if (NUM_RCU_NODES > 1) | ||
1719 | rcu_for_each_leaf_node(rcu_state, rnp) { | ||
1720 | init_waitqueue_head(&rnp->node_wq); | ||
1721 | rcu_init_boost_waitqueue(rnp); | ||
1722 | (void)rcu_spawn_one_node_kthread(rcu_state, rnp); | ||
1723 | } | ||
1685 | return 0; | 1724 | return 0; |
1686 | } | 1725 | } |
1687 | early_initcall(rcu_spawn_kthreads); | 1726 | early_initcall(rcu_spawn_kthreads); |
@@ -2087,14 +2126,14 @@ static void __cpuinit rcu_online_cpu(int cpu) | |||
2087 | 2126 | ||
2088 | static void __cpuinit rcu_online_kthreads(int cpu) | 2127 | static void __cpuinit rcu_online_kthreads(int cpu) |
2089 | { | 2128 | { |
2090 | struct rcu_data *rdp = per_cpu_ptr(rcu_sched_state.rda, cpu); | 2129 | struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, cpu); |
2091 | struct rcu_node *rnp = rdp->mynode; | 2130 | struct rcu_node *rnp = rdp->mynode; |
2092 | 2131 | ||
2093 | /* Fire up the incoming CPU's kthread and leaf rcu_node kthread. */ | 2132 | /* Fire up the incoming CPU's kthread and leaf rcu_node kthread. */ |
2094 | if (rcu_kthreads_spawnable) { | 2133 | if (rcu_kthreads_spawnable) { |
2095 | (void)rcu_spawn_one_cpu_kthread(cpu); | 2134 | (void)rcu_spawn_one_cpu_kthread(cpu); |
2096 | if (rnp->node_kthread_task == NULL) | 2135 | if (rnp->node_kthread_task == NULL) |
2097 | (void)rcu_spawn_one_node_kthread(&rcu_sched_state, rnp); | 2136 | (void)rcu_spawn_one_node_kthread(rcu_state, rnp); |
2098 | } | 2137 | } |
2099 | } | 2138 | } |
2100 | 2139 | ||
@@ -2105,7 +2144,7 @@ static int __cpuinit rcu_cpu_notify(struct notifier_block *self, | |||
2105 | unsigned long action, void *hcpu) | 2144 | unsigned long action, void *hcpu) |
2106 | { | 2145 | { |
2107 | long cpu = (long)hcpu; | 2146 | long cpu = (long)hcpu; |
2108 | struct rcu_data *rdp = per_cpu_ptr(rcu_sched_state.rda, cpu); | 2147 | struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, cpu); |
2109 | struct rcu_node *rnp = rdp->mynode; | 2148 | struct rcu_node *rnp = rdp->mynode; |
2110 | 2149 | ||
2111 | switch (action) { | 2150 | switch (action) { |