aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/ia64/kernel/time.c6
-rw-r--r--arch/powerpc/kernel/time.c10
-rw-r--r--arch/s390/kernel/vtime.c6
-rw-r--r--include/linux/hardirq.h8
-rw-r--r--include/linux/kernel_stat.h4
-rw-r--r--include/linux/kvm_host.h4
-rw-r--r--kernel/sched/core.c2
-rw-r--r--kernel/sched/cputime.c8
-rw-r--r--kernel/softirq.c6
9 files changed, 27 insertions, 27 deletions
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c
index 6247197b9877..16bb6eda879d 100644
--- a/arch/ia64/kernel/time.c
+++ b/arch/ia64/kernel/time.c
@@ -88,7 +88,7 @@ extern cputime_t cycle_to_cputime(u64 cyc);
88 * accumulated times to the current process, and to prepare accounting on 88 * accumulated times to the current process, and to prepare accounting on
89 * the next process. 89 * the next process.
90 */ 90 */
91void account_switch_vtime(struct task_struct *prev) 91void vtime_task_switch(struct task_struct *prev)
92{ 92{
93 struct thread_info *pi = task_thread_info(prev); 93 struct thread_info *pi = task_thread_info(prev);
94 struct thread_info *ni = task_thread_info(current); 94 struct thread_info *ni = task_thread_info(current);
@@ -116,7 +116,7 @@ void account_switch_vtime(struct task_struct *prev)
116 * Account time for a transition between system, hard irq or soft irq state. 116 * Account time for a transition between system, hard irq or soft irq state.
117 * Note that this function is called with interrupts enabled. 117 * Note that this function is called with interrupts enabled.
118 */ 118 */
119void account_system_vtime(struct task_struct *tsk) 119void vtime_account(struct task_struct *tsk)
120{ 120{
121 struct thread_info *ti = task_thread_info(tsk); 121 struct thread_info *ti = task_thread_info(tsk);
122 unsigned long flags; 122 unsigned long flags;
@@ -138,7 +138,7 @@ void account_system_vtime(struct task_struct *tsk)
138 138
139 local_irq_restore(flags); 139 local_irq_restore(flags);
140} 140}
141EXPORT_SYMBOL_GPL(account_system_vtime); 141EXPORT_SYMBOL_GPL(vtime_account);
142 142
143/* 143/*
144 * Called from the timer interrupt handler to charge accumulated user time 144 * Called from the timer interrupt handler to charge accumulated user time
diff --git a/arch/powerpc/kernel/time.c b/arch/powerpc/kernel/time.c
index 49da7f06e643..39899d7ebda0 100644
--- a/arch/powerpc/kernel/time.c
+++ b/arch/powerpc/kernel/time.c
@@ -291,7 +291,7 @@ static inline u64 calculate_stolen_time(u64 stop_tb)
291 * Account time for a transition between system, hard irq 291 * Account time for a transition between system, hard irq
292 * or soft irq state. 292 * or soft irq state.
293 */ 293 */
294void account_system_vtime(struct task_struct *tsk) 294void vtime_account(struct task_struct *tsk)
295{ 295{
296 u64 now, nowscaled, delta, deltascaled; 296 u64 now, nowscaled, delta, deltascaled;
297 unsigned long flags; 297 unsigned long flags;
@@ -343,14 +343,14 @@ void account_system_vtime(struct task_struct *tsk)
343 } 343 }
344 local_irq_restore(flags); 344 local_irq_restore(flags);
345} 345}
346EXPORT_SYMBOL_GPL(account_system_vtime); 346EXPORT_SYMBOL_GPL(vtime_account);
347 347
348/* 348/*
349 * Transfer the user and system times accumulated in the paca 349 * Transfer the user and system times accumulated in the paca
350 * by the exception entry and exit code to the generic process 350 * by the exception entry and exit code to the generic process
351 * user and system time records. 351 * user and system time records.
352 * Must be called with interrupts disabled. 352 * Must be called with interrupts disabled.
353 * Assumes that account_system_vtime() has been called recently 353 * Assumes that vtime_account() has been called recently
354 * (i.e. since the last entry from usermode) so that 354 * (i.e. since the last entry from usermode) so that
355 * get_paca()->user_time_scaled is up to date. 355 * get_paca()->user_time_scaled is up to date.
356 */ 356 */
@@ -366,9 +366,9 @@ void account_process_tick(struct task_struct *tsk, int user_tick)
366 account_user_time(tsk, utime, utimescaled); 366 account_user_time(tsk, utime, utimescaled);
367} 367}
368 368
369void account_switch_vtime(struct task_struct *prev) 369void vtime_task_switch(struct task_struct *prev)
370{ 370{
371 account_system_vtime(prev); 371 vtime_account(prev);
372 account_process_tick(prev, 0); 372 account_process_tick(prev, 0);
373} 373}
374 374
diff --git a/arch/s390/kernel/vtime.c b/arch/s390/kernel/vtime.c
index 449ac22cc71b..cb5093c26d16 100644
--- a/arch/s390/kernel/vtime.c
+++ b/arch/s390/kernel/vtime.c
@@ -99,7 +99,7 @@ static int do_account_vtime(struct task_struct *tsk, int hardirq_offset)
99 return virt_timer_forward(user + system); 99 return virt_timer_forward(user + system);
100} 100}
101 101
102void account_switch_vtime(struct task_struct *prev) 102void vtime_task_switch(struct task_struct *prev)
103{ 103{
104 struct thread_info *ti; 104 struct thread_info *ti;
105 105
@@ -122,7 +122,7 @@ void account_process_tick(struct task_struct *tsk, int user_tick)
122 * Update process times based on virtual cpu times stored by entry.S 122 * Update process times based on virtual cpu times stored by entry.S
123 * to the lowcore fields user_timer, system_timer & steal_clock. 123 * to the lowcore fields user_timer, system_timer & steal_clock.
124 */ 124 */
125void account_system_vtime(struct task_struct *tsk) 125void vtime_account(struct task_struct *tsk)
126{ 126{
127 struct thread_info *ti = task_thread_info(tsk); 127 struct thread_info *ti = task_thread_info(tsk);
128 u64 timer, system; 128 u64 timer, system;
@@ -138,7 +138,7 @@ void account_system_vtime(struct task_struct *tsk)
138 138
139 virt_timer_forward(system); 139 virt_timer_forward(system);
140} 140}
141EXPORT_SYMBOL_GPL(account_system_vtime); 141EXPORT_SYMBOL_GPL(vtime_account);
142 142
143void __kprobes vtime_stop_cpu(void) 143void __kprobes vtime_stop_cpu(void)
144{ 144{
diff --git a/include/linux/hardirq.h b/include/linux/hardirq.h
index 305f23cd7cff..cab3da3d0949 100644
--- a/include/linux/hardirq.h
+++ b/include/linux/hardirq.h
@@ -132,11 +132,11 @@ extern void synchronize_irq(unsigned int irq);
132struct task_struct; 132struct task_struct;
133 133
134#if !defined(CONFIG_VIRT_CPU_ACCOUNTING) && !defined(CONFIG_IRQ_TIME_ACCOUNTING) 134#if !defined(CONFIG_VIRT_CPU_ACCOUNTING) && !defined(CONFIG_IRQ_TIME_ACCOUNTING)
135static inline void account_system_vtime(struct task_struct *tsk) 135static inline void vtime_account(struct task_struct *tsk)
136{ 136{
137} 137}
138#else 138#else
139extern void account_system_vtime(struct task_struct *tsk); 139extern void vtime_account(struct task_struct *tsk);
140#endif 140#endif
141 141
142#if defined(CONFIG_TINY_RCU) || defined(CONFIG_TINY_PREEMPT_RCU) 142#if defined(CONFIG_TINY_RCU) || defined(CONFIG_TINY_PREEMPT_RCU)
@@ -162,7 +162,7 @@ extern void rcu_nmi_exit(void);
162 */ 162 */
163#define __irq_enter() \ 163#define __irq_enter() \
164 do { \ 164 do { \
165 account_system_vtime(current); \ 165 vtime_account(current); \
166 add_preempt_count(HARDIRQ_OFFSET); \ 166 add_preempt_count(HARDIRQ_OFFSET); \
167 trace_hardirq_enter(); \ 167 trace_hardirq_enter(); \
168 } while (0) 168 } while (0)
@@ -178,7 +178,7 @@ extern void irq_enter(void);
178#define __irq_exit() \ 178#define __irq_exit() \
179 do { \ 179 do { \
180 trace_hardirq_exit(); \ 180 trace_hardirq_exit(); \
181 account_system_vtime(current); \ 181 vtime_account(current); \
182 sub_preempt_count(HARDIRQ_OFFSET); \ 182 sub_preempt_count(HARDIRQ_OFFSET); \
183 } while (0) 183 } while (0)
184 184
diff --git a/include/linux/kernel_stat.h b/include/linux/kernel_stat.h
index bbe5d15d6597..ca0944b92f4a 100644
--- a/include/linux/kernel_stat.h
+++ b/include/linux/kernel_stat.h
@@ -131,9 +131,9 @@ extern void account_steal_ticks(unsigned long ticks);
131extern void account_idle_ticks(unsigned long ticks); 131extern void account_idle_ticks(unsigned long ticks);
132 132
133#ifdef CONFIG_VIRT_CPU_ACCOUNTING 133#ifdef CONFIG_VIRT_CPU_ACCOUNTING
134extern void account_switch_vtime(struct task_struct *prev); 134extern void vtime_task_switch(struct task_struct *prev);
135#else 135#else
136static inline void account_switch_vtime(struct task_struct *prev) { } 136static inline void vtime_task_switch(struct task_struct *prev) { }
137#endif 137#endif
138 138
139#endif /* _LINUX_KERNEL_STAT_H */ 139#endif /* _LINUX_KERNEL_STAT_H */
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index b70b48b01098..8a59e0abe5fa 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -685,7 +685,7 @@ static inline int kvm_deassign_device(struct kvm *kvm,
685static inline void kvm_guest_enter(void) 685static inline void kvm_guest_enter(void)
686{ 686{
687 BUG_ON(preemptible()); 687 BUG_ON(preemptible());
688 account_system_vtime(current); 688 vtime_account(current);
689 current->flags |= PF_VCPU; 689 current->flags |= PF_VCPU;
690 /* KVM does not hold any references to rcu protected data when it 690 /* KVM does not hold any references to rcu protected data when it
691 * switches CPU into a guest mode. In fact switching to a guest mode 691 * switches CPU into a guest mode. In fact switching to a guest mode
@@ -699,7 +699,7 @@ static inline void kvm_guest_enter(void)
699 699
700static inline void kvm_guest_exit(void) 700static inline void kvm_guest_exit(void)
701{ 701{
702 account_system_vtime(current); 702 vtime_account(current);
703 current->flags &= ~PF_VCPU; 703 current->flags &= ~PF_VCPU;
704} 704}
705 705
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index ba144b121f3d..21e4dcff18f3 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -1764,7 +1764,7 @@ static void finish_task_switch(struct rq *rq, struct task_struct *prev)
1764 * Manfred Spraul <manfred@colorfullife.com> 1764 * Manfred Spraul <manfred@colorfullife.com>
1765 */ 1765 */
1766 prev_state = prev->state; 1766 prev_state = prev->state;
1767 account_switch_vtime(prev); 1767 vtime_task_switch(prev);
1768 finish_arch_switch(prev); 1768 finish_arch_switch(prev);
1769 perf_event_task_sched_in(prev, current); 1769 perf_event_task_sched_in(prev, current);
1770 finish_lock_switch(rq, prev); 1770 finish_lock_switch(rq, prev);
diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c
index 372692bd5376..53f5b12f2821 100644
--- a/kernel/sched/cputime.c
+++ b/kernel/sched/cputime.c
@@ -10,11 +10,11 @@
10 10
11/* 11/*
12 * There are no locks covering percpu hardirq/softirq time. 12 * There are no locks covering percpu hardirq/softirq time.
13 * They are only modified in account_system_vtime, on corresponding CPU 13 * They are only modified in vtime_account, on corresponding CPU
14 * with interrupts disabled. So, writes are safe. 14 * with interrupts disabled. So, writes are safe.
15 * They are read and saved off onto struct rq in update_rq_clock(). 15 * They are read and saved off onto struct rq in update_rq_clock().
16 * This may result in other CPU reading this CPU's irq time and can 16 * This may result in other CPU reading this CPU's irq time and can
17 * race with irq/account_system_vtime on this CPU. We would either get old 17 * race with irq/vtime_account on this CPU. We would either get old
18 * or new value with a side effect of accounting a slice of irq time to wrong 18 * or new value with a side effect of accounting a slice of irq time to wrong
19 * task when irq is in progress while we read rq->clock. That is a worthy 19 * task when irq is in progress while we read rq->clock. That is a worthy
20 * compromise in place of having locks on each irq in account_system_time. 20 * compromise in place of having locks on each irq in account_system_time.
@@ -43,7 +43,7 @@ DEFINE_PER_CPU(seqcount_t, irq_time_seq);
43 * Called before incrementing preempt_count on {soft,}irq_enter 43 * Called before incrementing preempt_count on {soft,}irq_enter
44 * and before decrementing preempt_count on {soft,}irq_exit. 44 * and before decrementing preempt_count on {soft,}irq_exit.
45 */ 45 */
46void account_system_vtime(struct task_struct *curr) 46void vtime_account(struct task_struct *curr)
47{ 47{
48 unsigned long flags; 48 unsigned long flags;
49 s64 delta; 49 s64 delta;
@@ -73,7 +73,7 @@ void account_system_vtime(struct task_struct *curr)
73 irq_time_write_end(); 73 irq_time_write_end();
74 local_irq_restore(flags); 74 local_irq_restore(flags);
75} 75}
76EXPORT_SYMBOL_GPL(account_system_vtime); 76EXPORT_SYMBOL_GPL(vtime_account);
77 77
78static int irqtime_account_hi_update(void) 78static int irqtime_account_hi_update(void)
79{ 79{
diff --git a/kernel/softirq.c b/kernel/softirq.c
index b73e681df09e..d55e3159f928 100644
--- a/kernel/softirq.c
+++ b/kernel/softirq.c
@@ -220,7 +220,7 @@ asmlinkage void __do_softirq(void)
220 current->flags &= ~PF_MEMALLOC; 220 current->flags &= ~PF_MEMALLOC;
221 221
222 pending = local_softirq_pending(); 222 pending = local_softirq_pending();
223 account_system_vtime(current); 223 vtime_account(current);
224 224
225 __local_bh_disable((unsigned long)__builtin_return_address(0), 225 __local_bh_disable((unsigned long)__builtin_return_address(0),
226 SOFTIRQ_OFFSET); 226 SOFTIRQ_OFFSET);
@@ -271,7 +271,7 @@ restart:
271 271
272 lockdep_softirq_exit(); 272 lockdep_softirq_exit();
273 273
274 account_system_vtime(current); 274 vtime_account(current);
275 __local_bh_enable(SOFTIRQ_OFFSET); 275 __local_bh_enable(SOFTIRQ_OFFSET);
276 tsk_restore_flags(current, old_flags, PF_MEMALLOC); 276 tsk_restore_flags(current, old_flags, PF_MEMALLOC);
277} 277}
@@ -340,7 +340,7 @@ static inline void invoke_softirq(void)
340 */ 340 */
341void irq_exit(void) 341void irq_exit(void)
342{ 342{
343 account_system_vtime(current); 343 vtime_account(current);
344 trace_hardirq_exit(); 344 trace_hardirq_exit();
345 sub_preempt_count(IRQ_EXIT_OFFSET); 345 sub_preempt_count(IRQ_EXIT_OFFSET);
346 if (!in_interrupt() && local_softirq_pending()) 346 if (!in_interrupt() && local_softirq_pending())