diff options
Diffstat (limited to 'kernel/hrtimer.c')
-rw-r--r-- | kernel/hrtimer.c | 95 |
1 files changed, 85 insertions, 10 deletions
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 2bd230be1cb5..51ee90bca2de 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c | |||
@@ -672,13 +672,14 @@ static inline int hrtimer_enqueue_reprogram(struct hrtimer *timer, | |||
672 | */ | 672 | */ |
673 | BUG_ON(timer->function(timer) != HRTIMER_NORESTART); | 673 | BUG_ON(timer->function(timer) != HRTIMER_NORESTART); |
674 | return 1; | 674 | return 1; |
675 | case HRTIMER_CB_IRQSAFE_NO_SOFTIRQ: | 675 | case HRTIMER_CB_IRQSAFE_PERCPU: |
676 | case HRTIMER_CB_IRQSAFE_UNLOCKED: | ||
676 | /* | 677 | /* |
677 | * This is solely for the sched tick emulation with | 678 | * This is solely for the sched tick emulation with |
678 | * dynamic tick support to ensure that we do not | 679 | * dynamic tick support to ensure that we do not |
679 | * restart the tick right on the edge and end up with | 680 | * restart the tick right on the edge and end up with |
680 | * the tick timer in the softirq ! The calling site | 681 | * the tick timer in the softirq ! The calling site |
681 | * takes care of this. | 682 | * takes care of this. Also used for hrtimer sleeper ! |
682 | */ | 683 | */ |
683 | debug_hrtimer_deactivate(timer); | 684 | debug_hrtimer_deactivate(timer); |
684 | return 1; | 685 | return 1; |
@@ -1266,7 +1267,8 @@ static void __run_hrtimer(struct hrtimer *timer) | |||
1266 | timer_stats_account_hrtimer(timer); | 1267 | timer_stats_account_hrtimer(timer); |
1267 | 1268 | ||
1268 | fn = timer->function; | 1269 | fn = timer->function; |
1269 | if (timer->cb_mode == HRTIMER_CB_IRQSAFE_NO_SOFTIRQ) { | 1270 | if (timer->cb_mode == HRTIMER_CB_IRQSAFE_PERCPU || |
1271 | timer->cb_mode == HRTIMER_CB_IRQSAFE_UNLOCKED) { | ||
1270 | /* | 1272 | /* |
1271 | * Used for scheduler timers, avoid lock inversion with | 1273 | * Used for scheduler timers, avoid lock inversion with |
1272 | * rq->lock and tasklist_lock. | 1274 | * rq->lock and tasklist_lock. |
@@ -1517,7 +1519,7 @@ void hrtimer_init_sleeper(struct hrtimer_sleeper *sl, struct task_struct *task) | |||
1517 | sl->timer.function = hrtimer_wakeup; | 1519 | sl->timer.function = hrtimer_wakeup; |
1518 | sl->task = task; | 1520 | sl->task = task; |
1519 | #ifdef CONFIG_HIGH_RES_TIMERS | 1521 | #ifdef CONFIG_HIGH_RES_TIMERS |
1520 | sl->timer.cb_mode = HRTIMER_CB_IRQSAFE_NO_SOFTIRQ; | 1522 | sl->timer.cb_mode = HRTIMER_CB_IRQSAFE_UNLOCKED; |
1521 | #endif | 1523 | #endif |
1522 | } | 1524 | } |
1523 | 1525 | ||
@@ -1661,29 +1663,95 @@ static void __cpuinit init_hrtimers_cpu(int cpu) | |||
1661 | 1663 | ||
1662 | #ifdef CONFIG_HOTPLUG_CPU | 1664 | #ifdef CONFIG_HOTPLUG_CPU |
1663 | 1665 | ||
1664 | static void migrate_hrtimer_list(struct hrtimer_clock_base *old_base, | 1666 | static int migrate_hrtimer_list(struct hrtimer_clock_base *old_base, |
1665 | struct hrtimer_clock_base *new_base) | 1667 | struct hrtimer_clock_base *new_base, int dcpu) |
1666 | { | 1668 | { |
1667 | struct hrtimer *timer; | 1669 | struct hrtimer *timer; |
1668 | struct rb_node *node; | 1670 | struct rb_node *node; |
1671 | int raise = 0; | ||
1669 | 1672 | ||
1670 | while ((node = rb_first(&old_base->active))) { | 1673 | while ((node = rb_first(&old_base->active))) { |
1671 | timer = rb_entry(node, struct hrtimer, node); | 1674 | timer = rb_entry(node, struct hrtimer, node); |
1672 | BUG_ON(hrtimer_callback_running(timer)); | 1675 | BUG_ON(hrtimer_callback_running(timer)); |
1673 | debug_hrtimer_deactivate(timer); | 1676 | debug_hrtimer_deactivate(timer); |
1674 | __remove_hrtimer(timer, old_base, HRTIMER_STATE_INACTIVE, 0); | 1677 | |
1678 | /* | ||
1679 | * Should not happen. Per CPU timers should be | ||
1680 | * canceled _before_ the migration code is called | ||
1681 | */ | ||
1682 | if (timer->cb_mode == HRTIMER_CB_IRQSAFE_PERCPU) { | ||
1683 | __remove_hrtimer(timer, old_base, | ||
1684 | HRTIMER_STATE_INACTIVE, 0); | ||
1685 | WARN(1, "hrtimer (%p %p)active but cpu %d dead\n", | ||
1686 | timer, timer->function, dcpu); | ||
1687 | continue; | ||
1688 | } | ||
1689 | |||
1690 | /* | ||
1691 | * Mark it as STATE_MIGRATE not INACTIVE otherwise the | ||
1692 | * timer could be seen as !active and just vanish away | ||
1693 | * under us on another CPU | ||
1694 | */ | ||
1695 | __remove_hrtimer(timer, old_base, HRTIMER_STATE_MIGRATE, 0); | ||
1675 | timer->base = new_base; | 1696 | timer->base = new_base; |
1676 | /* | 1697 | /* |
1677 | * Enqueue the timer. Allow reprogramming of the event device | 1698 | * Enqueue the timer. Allow reprogramming of the event device |
1678 | */ | 1699 | */ |
1679 | enqueue_hrtimer(timer, new_base, 1); | 1700 | enqueue_hrtimer(timer, new_base, 1); |
1701 | |||
1702 | #ifdef CONFIG_HIGH_RES_TIMERS | ||
1703 | /* | ||
1704 | * Happens with high res enabled when the timer was | ||
1705 | * already expired and the callback mode is | ||
1706 | * HRTIMER_CB_IRQSAFE_UNLOCKED (hrtimer_sleeper). The | ||
1707 | * enqueue code does not move them to the soft irq | ||
1708 | * pending list for performance/latency reasons, but | ||
1709 | * in the migration state, we need to do that | ||
1710 | * otherwise we end up with a stale timer. | ||
1711 | */ | ||
1712 | if (timer->state == HRTIMER_STATE_MIGRATE) { | ||
1713 | timer->state = HRTIMER_STATE_PENDING; | ||
1714 | list_add_tail(&timer->cb_entry, | ||
1715 | &new_base->cpu_base->cb_pending); | ||
1716 | raise = 1; | ||
1717 | } | ||
1718 | #endif | ||
1719 | /* Clear the migration state bit */ | ||
1720 | timer->state &= ~HRTIMER_STATE_MIGRATE; | ||
1721 | } | ||
1722 | return raise; | ||
1723 | } | ||
1724 | |||
1725 | #ifdef CONFIG_HIGH_RES_TIMERS | ||
1726 | static int migrate_hrtimer_pending(struct hrtimer_cpu_base *old_base, | ||
1727 | struct hrtimer_cpu_base *new_base) | ||
1728 | { | ||
1729 | struct hrtimer *timer; | ||
1730 | int raise = 0; | ||
1731 | |||
1732 | while (!list_empty(&old_base->cb_pending)) { | ||
1733 | timer = list_entry(old_base->cb_pending.next, | ||
1734 | struct hrtimer, cb_entry); | ||
1735 | |||
1736 | __remove_hrtimer(timer, timer->base, HRTIMER_STATE_PENDING, 0); | ||
1737 | timer->base = &new_base->clock_base[timer->base->index]; | ||
1738 | list_add_tail(&timer->cb_entry, &new_base->cb_pending); | ||
1739 | raise = 1; | ||
1680 | } | 1740 | } |
1741 | return raise; | ||
1742 | } | ||
1743 | #else | ||
1744 | static int migrate_hrtimer_pending(struct hrtimer_cpu_base *old_base, | ||
1745 | struct hrtimer_cpu_base *new_base) | ||
1746 | { | ||
1747 | return 0; | ||
1681 | } | 1748 | } |
1749 | #endif | ||
1682 | 1750 | ||
1683 | static void migrate_hrtimers(int cpu) | 1751 | static void migrate_hrtimers(int cpu) |
1684 | { | 1752 | { |
1685 | struct hrtimer_cpu_base *old_base, *new_base; | 1753 | struct hrtimer_cpu_base *old_base, *new_base; |
1686 | int i; | 1754 | int i, raise = 0; |
1687 | 1755 | ||
1688 | BUG_ON(cpu_online(cpu)); | 1756 | BUG_ON(cpu_online(cpu)); |
1689 | old_base = &per_cpu(hrtimer_bases, cpu); | 1757 | old_base = &per_cpu(hrtimer_bases, cpu); |
@@ -1696,14 +1764,21 @@ static void migrate_hrtimers(int cpu) | |||
1696 | spin_lock_nested(&old_base->lock, SINGLE_DEPTH_NESTING); | 1764 | spin_lock_nested(&old_base->lock, SINGLE_DEPTH_NESTING); |
1697 | 1765 | ||
1698 | for (i = 0; i < HRTIMER_MAX_CLOCK_BASES; i++) { | 1766 | for (i = 0; i < HRTIMER_MAX_CLOCK_BASES; i++) { |
1699 | migrate_hrtimer_list(&old_base->clock_base[i], | 1767 | if (migrate_hrtimer_list(&old_base->clock_base[i], |
1700 | &new_base->clock_base[i]); | 1768 | &new_base->clock_base[i], cpu)) |
1769 | raise = 1; | ||
1701 | } | 1770 | } |
1702 | 1771 | ||
1772 | if (migrate_hrtimer_pending(old_base, new_base)) | ||
1773 | raise = 1; | ||
1774 | |||
1703 | spin_unlock(&old_base->lock); | 1775 | spin_unlock(&old_base->lock); |
1704 | spin_unlock(&new_base->lock); | 1776 | spin_unlock(&new_base->lock); |
1705 | local_irq_enable(); | 1777 | local_irq_enable(); |
1706 | put_cpu_var(hrtimer_bases); | 1778 | put_cpu_var(hrtimer_bases); |
1779 | |||
1780 | if (raise) | ||
1781 | hrtimer_raise_softirq(); | ||
1707 | } | 1782 | } |
1708 | #endif /* CONFIG_HOTPLUG_CPU */ | 1783 | #endif /* CONFIG_HOTPLUG_CPU */ |
1709 | 1784 | ||