diff options
| author | Linus Torvalds <torvalds@linux-foundation.org> | 2013-02-19 21:19:48 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@linux-foundation.org> | 2013-02-19 21:19:48 -0500 |
| commit | d652e1eb8e7b739fccbfb503a3da3e9f640fbf3d (patch) | |
| tree | 55ab77bad0cbb045eac0b84b80d63f88f1ae09e6 | |
| parent | 8f55cea410dbc56114bb71a3742032070c8108d0 (diff) | |
| parent | 77852fea6e2442a0e654a9292060489895de18c7 (diff) | |
Merge branch 'sched-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull scheduler changes from Ingo Molnar:
"Main changes:
- scheduler side full-dynticks (user-space execution is undisturbed
and receives no timer IRQs) preparation changes that convert the
cputime accounting code to be full-dynticks ready, from Frederic
Weisbecker.
- Initial sched.h split-up changes, by Clark Williams
- select_idle_sibling() performance improvement by Mike Galbraith:
" 1 tbench pair (worst case) in a 10 core + SMT package:
pre 15.22 MB/sec 1 procs
post 252.01 MB/sec 1 procs "
- sched_rr_get_interval() ABI fix/change. We think this detail is not
used by apps (so it's not an ABI in practice), but lets keep it
under observation.
- misc RT scheduling cleanups, optimizations"
* 'sched-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: (24 commits)
sched/rt: Add <linux/sched/rt.h> header to <linux/init_task.h>
cputime: Remove irqsave from seqlock readers
sched, powerpc: Fix sched.h split-up build failure
cputime: Restore CPU_ACCOUNTING config defaults for PPC64
sched/rt: Move rt specific bits into new header file
sched/rt: Add a tuning knob to allow changing SCHED_RR timeslice
sched: Move sched.h sysctl bits into separate header
sched: Fix signedness bug in yield_to()
sched: Fix select_idle_sibling() bouncing cow syndrome
sched/rt: Further simplify pick_rt_task()
sched/rt: Do not account zero delta_exec in update_curr_rt()
cputime: Safely read cputime of full dynticks CPUs
kvm: Prepare to add generic guest entry/exit callbacks
cputime: Use accessors to read task cputime stats
cputime: Allow dynamic switch between tick/virtual based cputime accounting
cputime: Generic on-demand virtual cputime accounting
cputime: Move default nsecs_to_cputime() to jiffies based cputime file
cputime: Librarize per nsecs resolution cputime definitions
cputime: Avoid multiplication overflow on utime scaling
context_tracking: Export context state for generic vtime
...
Fix up conflict in kernel/context_tracking.c due to comment additions.
83 files changed, 1105 insertions, 507 deletions
diff --git a/arch/alpha/kernel/osf_sys.c b/arch/alpha/kernel/osf_sys.c index 14db93e4c8a8..dbc1760f418b 100644 --- a/arch/alpha/kernel/osf_sys.c +++ b/arch/alpha/kernel/osf_sys.c | |||
| @@ -1139,6 +1139,7 @@ struct rusage32 { | |||
| 1139 | SYSCALL_DEFINE2(osf_getrusage, int, who, struct rusage32 __user *, ru) | 1139 | SYSCALL_DEFINE2(osf_getrusage, int, who, struct rusage32 __user *, ru) |
| 1140 | { | 1140 | { |
| 1141 | struct rusage32 r; | 1141 | struct rusage32 r; |
| 1142 | cputime_t utime, stime; | ||
| 1142 | 1143 | ||
| 1143 | if (who != RUSAGE_SELF && who != RUSAGE_CHILDREN) | 1144 | if (who != RUSAGE_SELF && who != RUSAGE_CHILDREN) |
| 1144 | return -EINVAL; | 1145 | return -EINVAL; |
| @@ -1146,8 +1147,9 @@ SYSCALL_DEFINE2(osf_getrusage, int, who, struct rusage32 __user *, ru) | |||
| 1146 | memset(&r, 0, sizeof(r)); | 1147 | memset(&r, 0, sizeof(r)); |
| 1147 | switch (who) { | 1148 | switch (who) { |
| 1148 | case RUSAGE_SELF: | 1149 | case RUSAGE_SELF: |
| 1149 | jiffies_to_timeval32(current->utime, &r.ru_utime); | 1150 | task_cputime(current, &utime, &stime); |
| 1150 | jiffies_to_timeval32(current->stime, &r.ru_stime); | 1151 | jiffies_to_timeval32(utime, &r.ru_utime); |
| 1152 | jiffies_to_timeval32(stime, &r.ru_stime); | ||
| 1151 | r.ru_minflt = current->min_flt; | 1153 | r.ru_minflt = current->min_flt; |
| 1152 | r.ru_majflt = current->maj_flt; | 1154 | r.ru_majflt = current->maj_flt; |
| 1153 | break; | 1155 | break; |
diff --git a/arch/ia64/include/asm/cputime.h b/arch/ia64/include/asm/cputime.h index 7fcf7f08ab06..e2d3f5baf265 100644 --- a/arch/ia64/include/asm/cputime.h +++ b/arch/ia64/include/asm/cputime.h | |||
| @@ -11,99 +11,19 @@ | |||
| 11 | * as published by the Free Software Foundation; either version | 11 | * as published by the Free Software Foundation; either version |
| 12 | * 2 of the License, or (at your option) any later version. | 12 | * 2 of the License, or (at your option) any later version. |
| 13 | * | 13 | * |
| 14 | * If we have CONFIG_VIRT_CPU_ACCOUNTING, we measure cpu time in nsec. | 14 | * If we have CONFIG_VIRT_CPU_ACCOUNTING_NATIVE, we measure cpu time in nsec. |
| 15 | * Otherwise we measure cpu time in jiffies using the generic definitions. | 15 | * Otherwise we measure cpu time in jiffies using the generic definitions. |
| 16 | */ | 16 | */ |
| 17 | 17 | ||
| 18 | #ifndef __IA64_CPUTIME_H | 18 | #ifndef __IA64_CPUTIME_H |
| 19 | #define __IA64_CPUTIME_H | 19 | #define __IA64_CPUTIME_H |
| 20 | 20 | ||
| 21 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 21 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 22 | #include <asm-generic/cputime.h> | 22 | # include <asm-generic/cputime.h> |
| 23 | #else | 23 | #else |
| 24 | 24 | # include <asm/processor.h> | |
| 25 | #include <linux/time.h> | 25 | # include <asm-generic/cputime_nsecs.h> |
| 26 | #include <linux/jiffies.h> | ||
| 27 | #include <asm/processor.h> | ||
| 28 | |||
| 29 | typedef u64 __nocast cputime_t; | ||
| 30 | typedef u64 __nocast cputime64_t; | ||
| 31 | |||
| 32 | #define cputime_one_jiffy jiffies_to_cputime(1) | ||
| 33 | |||
| 34 | /* | ||
| 35 | * Convert cputime <-> jiffies (HZ) | ||
| 36 | */ | ||
| 37 | #define cputime_to_jiffies(__ct) \ | ||
| 38 | ((__force u64)(__ct) / (NSEC_PER_SEC / HZ)) | ||
| 39 | #define jiffies_to_cputime(__jif) \ | ||
| 40 | (__force cputime_t)((__jif) * (NSEC_PER_SEC / HZ)) | ||
| 41 | #define cputime64_to_jiffies64(__ct) \ | ||
| 42 | ((__force u64)(__ct) / (NSEC_PER_SEC / HZ)) | ||
| 43 | #define jiffies64_to_cputime64(__jif) \ | ||
| 44 | (__force cputime64_t)((__jif) * (NSEC_PER_SEC / HZ)) | ||
| 45 | |||
| 46 | /* | ||
| 47 | * Convert cputime <-> microseconds | ||
| 48 | */ | ||
| 49 | #define cputime_to_usecs(__ct) \ | ||
| 50 | ((__force u64)(__ct) / NSEC_PER_USEC) | ||
| 51 | #define usecs_to_cputime(__usecs) \ | ||
| 52 | (__force cputime_t)((__usecs) * NSEC_PER_USEC) | ||
| 53 | #define usecs_to_cputime64(__usecs) \ | ||
| 54 | (__force cputime64_t)((__usecs) * NSEC_PER_USEC) | ||
| 55 | |||
| 56 | /* | ||
| 57 | * Convert cputime <-> seconds | ||
| 58 | */ | ||
| 59 | #define cputime_to_secs(__ct) \ | ||
| 60 | ((__force u64)(__ct) / NSEC_PER_SEC) | ||
| 61 | #define secs_to_cputime(__secs) \ | ||
| 62 | (__force cputime_t)((__secs) * NSEC_PER_SEC) | ||
| 63 | |||
| 64 | /* | ||
| 65 | * Convert cputime <-> timespec (nsec) | ||
| 66 | */ | ||
| 67 | static inline cputime_t timespec_to_cputime(const struct timespec *val) | ||
| 68 | { | ||
| 69 | u64 ret = val->tv_sec * NSEC_PER_SEC + val->tv_nsec; | ||
| 70 | return (__force cputime_t) ret; | ||
| 71 | } | ||
| 72 | static inline void cputime_to_timespec(const cputime_t ct, struct timespec *val) | ||
| 73 | { | ||
| 74 | val->tv_sec = (__force u64) ct / NSEC_PER_SEC; | ||
| 75 | val->tv_nsec = (__force u64) ct % NSEC_PER_SEC; | ||
| 76 | } | ||
| 77 | |||
| 78 | /* | ||
| 79 | * Convert cputime <-> timeval (msec) | ||
| 80 | */ | ||
| 81 | static inline cputime_t timeval_to_cputime(struct timeval *val) | ||
| 82 | { | ||
| 83 | u64 ret = val->tv_sec * NSEC_PER_SEC + val->tv_usec * NSEC_PER_USEC; | ||
| 84 | return (__force cputime_t) ret; | ||
| 85 | } | ||
| 86 | static inline void cputime_to_timeval(const cputime_t ct, struct timeval *val) | ||
| 87 | { | ||
| 88 | val->tv_sec = (__force u64) ct / NSEC_PER_SEC; | ||
| 89 | val->tv_usec = ((__force u64) ct % NSEC_PER_SEC) / NSEC_PER_USEC; | ||
| 90 | } | ||
| 91 | |||
| 92 | /* | ||
| 93 | * Convert cputime <-> clock (USER_HZ) | ||
| 94 | */ | ||
| 95 | #define cputime_to_clock_t(__ct) \ | ||
| 96 | ((__force u64)(__ct) / (NSEC_PER_SEC / USER_HZ)) | ||
| 97 | #define clock_t_to_cputime(__x) \ | ||
| 98 | (__force cputime_t)((__x) * (NSEC_PER_SEC / USER_HZ)) | ||
| 99 | |||
| 100 | /* | ||
| 101 | * Convert cputime64 to clock. | ||
| 102 | */ | ||
| 103 | #define cputime64_to_clock_t(__ct) \ | ||
| 104 | cputime_to_clock_t((__force cputime_t)__ct) | ||
| 105 | |||
| 106 | extern void arch_vtime_task_switch(struct task_struct *tsk); | 26 | extern void arch_vtime_task_switch(struct task_struct *tsk); |
| 27 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ | ||
| 107 | 28 | ||
| 108 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | ||
| 109 | #endif /* __IA64_CPUTIME_H */ | 29 | #endif /* __IA64_CPUTIME_H */ |
diff --git a/arch/ia64/include/asm/thread_info.h b/arch/ia64/include/asm/thread_info.h index ff2ae4136584..020d655ed082 100644 --- a/arch/ia64/include/asm/thread_info.h +++ b/arch/ia64/include/asm/thread_info.h | |||
| @@ -31,7 +31,7 @@ struct thread_info { | |||
| 31 | mm_segment_t addr_limit; /* user-level address space limit */ | 31 | mm_segment_t addr_limit; /* user-level address space limit */ |
| 32 | int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ | 32 | int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */ |
| 33 | struct restart_block restart_block; | 33 | struct restart_block restart_block; |
| 34 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 34 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 35 | __u64 ac_stamp; | 35 | __u64 ac_stamp; |
| 36 | __u64 ac_leave; | 36 | __u64 ac_leave; |
| 37 | __u64 ac_stime; | 37 | __u64 ac_stime; |
| @@ -69,7 +69,7 @@ struct thread_info { | |||
| 69 | #define task_stack_page(tsk) ((void *)(tsk)) | 69 | #define task_stack_page(tsk) ((void *)(tsk)) |
| 70 | 70 | ||
| 71 | #define __HAVE_THREAD_FUNCTIONS | 71 | #define __HAVE_THREAD_FUNCTIONS |
| 72 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 72 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 73 | #define setup_thread_stack(p, org) \ | 73 | #define setup_thread_stack(p, org) \ |
| 74 | *task_thread_info(p) = *task_thread_info(org); \ | 74 | *task_thread_info(p) = *task_thread_info(org); \ |
| 75 | task_thread_info(p)->ac_stime = 0; \ | 75 | task_thread_info(p)->ac_stime = 0; \ |
diff --git a/arch/ia64/include/asm/xen/minstate.h b/arch/ia64/include/asm/xen/minstate.h index c57fa910f2c9..00cf03e0cb82 100644 --- a/arch/ia64/include/asm/xen/minstate.h +++ b/arch/ia64/include/asm/xen/minstate.h | |||
| @@ -1,5 +1,5 @@ | |||
| 1 | 1 | ||
| 2 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 2 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 3 | /* read ar.itc in advance, and use it before leaving bank 0 */ | 3 | /* read ar.itc in advance, and use it before leaving bank 0 */ |
| 4 | #define XEN_ACCOUNT_GET_STAMP \ | 4 | #define XEN_ACCOUNT_GET_STAMP \ |
| 5 | MOV_FROM_ITC(pUStk, p6, r20, r2); | 5 | MOV_FROM_ITC(pUStk, p6, r20, r2); |
diff --git a/arch/ia64/kernel/asm-offsets.c b/arch/ia64/kernel/asm-offsets.c index a48bd9a9927b..46c9e3007315 100644 --- a/arch/ia64/kernel/asm-offsets.c +++ b/arch/ia64/kernel/asm-offsets.c | |||
| @@ -41,7 +41,7 @@ void foo(void) | |||
| 41 | DEFINE(TI_FLAGS, offsetof(struct thread_info, flags)); | 41 | DEFINE(TI_FLAGS, offsetof(struct thread_info, flags)); |
| 42 | DEFINE(TI_CPU, offsetof(struct thread_info, cpu)); | 42 | DEFINE(TI_CPU, offsetof(struct thread_info, cpu)); |
| 43 | DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); | 43 | DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count)); |
| 44 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 44 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 45 | DEFINE(TI_AC_STAMP, offsetof(struct thread_info, ac_stamp)); | 45 | DEFINE(TI_AC_STAMP, offsetof(struct thread_info, ac_stamp)); |
| 46 | DEFINE(TI_AC_LEAVE, offsetof(struct thread_info, ac_leave)); | 46 | DEFINE(TI_AC_LEAVE, offsetof(struct thread_info, ac_leave)); |
| 47 | DEFINE(TI_AC_STIME, offsetof(struct thread_info, ac_stime)); | 47 | DEFINE(TI_AC_STIME, offsetof(struct thread_info, ac_stime)); |
diff --git a/arch/ia64/kernel/entry.S b/arch/ia64/kernel/entry.S index 6bfd8429ee0f..7a53530f22c2 100644 --- a/arch/ia64/kernel/entry.S +++ b/arch/ia64/kernel/entry.S | |||
| @@ -724,7 +724,7 @@ GLOBAL_ENTRY(__paravirt_leave_syscall) | |||
| 724 | #endif | 724 | #endif |
| 725 | .global __paravirt_work_processed_syscall; | 725 | .global __paravirt_work_processed_syscall; |
| 726 | __paravirt_work_processed_syscall: | 726 | __paravirt_work_processed_syscall: |
| 727 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 727 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 728 | adds r2=PT(LOADRS)+16,r12 | 728 | adds r2=PT(LOADRS)+16,r12 |
| 729 | MOV_FROM_ITC(pUStk, p9, r22, r19) // fetch time at leave | 729 | MOV_FROM_ITC(pUStk, p9, r22, r19) // fetch time at leave |
| 730 | adds r18=TI_FLAGS+IA64_TASK_SIZE,r13 | 730 | adds r18=TI_FLAGS+IA64_TASK_SIZE,r13 |
| @@ -762,7 +762,7 @@ __paravirt_work_processed_syscall: | |||
| 762 | 762 | ||
| 763 | ld8 r29=[r2],16 // M0|1 load cr.ipsr | 763 | ld8 r29=[r2],16 // M0|1 load cr.ipsr |
| 764 | ld8 r28=[r3],16 // M0|1 load cr.iip | 764 | ld8 r28=[r3],16 // M0|1 load cr.iip |
| 765 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 765 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 766 | (pUStk) add r14=TI_AC_LEAVE+IA64_TASK_SIZE,r13 | 766 | (pUStk) add r14=TI_AC_LEAVE+IA64_TASK_SIZE,r13 |
| 767 | ;; | 767 | ;; |
| 768 | ld8 r30=[r2],16 // M0|1 load cr.ifs | 768 | ld8 r30=[r2],16 // M0|1 load cr.ifs |
| @@ -793,7 +793,7 @@ __paravirt_work_processed_syscall: | |||
| 793 | ld8.fill r1=[r3],16 // M0|1 load r1 | 793 | ld8.fill r1=[r3],16 // M0|1 load r1 |
| 794 | (pUStk) mov r17=1 // A | 794 | (pUStk) mov r17=1 // A |
| 795 | ;; | 795 | ;; |
| 796 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 796 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 797 | (pUStk) st1 [r15]=r17 // M2|3 | 797 | (pUStk) st1 [r15]=r17 // M2|3 |
| 798 | #else | 798 | #else |
| 799 | (pUStk) st1 [r14]=r17 // M2|3 | 799 | (pUStk) st1 [r14]=r17 // M2|3 |
| @@ -813,7 +813,7 @@ __paravirt_work_processed_syscall: | |||
| 813 | shr.u r18=r19,16 // I0|1 get byte size of existing "dirty" partition | 813 | shr.u r18=r19,16 // I0|1 get byte size of existing "dirty" partition |
| 814 | COVER // B add current frame into dirty partition & set cr.ifs | 814 | COVER // B add current frame into dirty partition & set cr.ifs |
| 815 | ;; | 815 | ;; |
| 816 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 816 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 817 | mov r19=ar.bsp // M2 get new backing store pointer | 817 | mov r19=ar.bsp // M2 get new backing store pointer |
| 818 | st8 [r14]=r22 // M save time at leave | 818 | st8 [r14]=r22 // M save time at leave |
| 819 | mov f10=f0 // F clear f10 | 819 | mov f10=f0 // F clear f10 |
| @@ -948,7 +948,7 @@ GLOBAL_ENTRY(__paravirt_leave_kernel) | |||
| 948 | adds r16=PT(CR_IPSR)+16,r12 | 948 | adds r16=PT(CR_IPSR)+16,r12 |
| 949 | adds r17=PT(CR_IIP)+16,r12 | 949 | adds r17=PT(CR_IIP)+16,r12 |
| 950 | 950 | ||
| 951 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 951 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 952 | .pred.rel.mutex pUStk,pKStk | 952 | .pred.rel.mutex pUStk,pKStk |
| 953 | MOV_FROM_PSR(pKStk, r22, r29) // M2 read PSR now that interrupts are disabled | 953 | MOV_FROM_PSR(pKStk, r22, r29) // M2 read PSR now that interrupts are disabled |
| 954 | MOV_FROM_ITC(pUStk, p9, r22, r29) // M fetch time at leave | 954 | MOV_FROM_ITC(pUStk, p9, r22, r29) // M fetch time at leave |
| @@ -981,7 +981,7 @@ GLOBAL_ENTRY(__paravirt_leave_kernel) | |||
| 981 | ;; | 981 | ;; |
| 982 | ld8.fill r12=[r16],16 | 982 | ld8.fill r12=[r16],16 |
| 983 | ld8.fill r13=[r17],16 | 983 | ld8.fill r13=[r17],16 |
| 984 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 984 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 985 | (pUStk) adds r3=TI_AC_LEAVE+IA64_TASK_SIZE,r18 | 985 | (pUStk) adds r3=TI_AC_LEAVE+IA64_TASK_SIZE,r18 |
| 986 | #else | 986 | #else |
| 987 | (pUStk) adds r18=IA64_TASK_THREAD_ON_USTACK_OFFSET,r18 | 987 | (pUStk) adds r18=IA64_TASK_THREAD_ON_USTACK_OFFSET,r18 |
| @@ -989,7 +989,7 @@ GLOBAL_ENTRY(__paravirt_leave_kernel) | |||
| 989 | ;; | 989 | ;; |
| 990 | ld8 r20=[r16],16 // ar.fpsr | 990 | ld8 r20=[r16],16 // ar.fpsr |
| 991 | ld8.fill r15=[r17],16 | 991 | ld8.fill r15=[r17],16 |
| 992 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 992 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 993 | (pUStk) adds r18=IA64_TASK_THREAD_ON_USTACK_OFFSET,r18 // deferred | 993 | (pUStk) adds r18=IA64_TASK_THREAD_ON_USTACK_OFFSET,r18 // deferred |
| 994 | #endif | 994 | #endif |
| 995 | ;; | 995 | ;; |
| @@ -997,7 +997,7 @@ GLOBAL_ENTRY(__paravirt_leave_kernel) | |||
| 997 | ld8.fill r2=[r17] | 997 | ld8.fill r2=[r17] |
| 998 | (pUStk) mov r17=1 | 998 | (pUStk) mov r17=1 |
| 999 | ;; | 999 | ;; |
| 1000 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 1000 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 1001 | // mmi_ : ld8 st1 shr;; mmi_ : st8 st1 shr;; | 1001 | // mmi_ : ld8 st1 shr;; mmi_ : st8 st1 shr;; |
| 1002 | // mib : mov add br -> mib : ld8 add br | 1002 | // mib : mov add br -> mib : ld8 add br |
| 1003 | // bbb_ : br nop cover;; mbb_ : mov br cover;; | 1003 | // bbb_ : br nop cover;; mbb_ : mov br cover;; |
diff --git a/arch/ia64/kernel/fsys.S b/arch/ia64/kernel/fsys.S index e662f178b990..c4cd45d97749 100644 --- a/arch/ia64/kernel/fsys.S +++ b/arch/ia64/kernel/fsys.S | |||
| @@ -529,7 +529,7 @@ GLOBAL_ENTRY(paravirt_fsys_bubble_down) | |||
| 529 | nop.i 0 | 529 | nop.i 0 |
| 530 | ;; | 530 | ;; |
| 531 | mov ar.rsc=0 // M2 set enforced lazy mode, pl 0, LE, loadrs=0 | 531 | mov ar.rsc=0 // M2 set enforced lazy mode, pl 0, LE, loadrs=0 |
| 532 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 532 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 533 | MOV_FROM_ITC(p0, p6, r30, r23) // M get cycle for accounting | 533 | MOV_FROM_ITC(p0, p6, r30, r23) // M get cycle for accounting |
| 534 | #else | 534 | #else |
| 535 | nop.m 0 | 535 | nop.m 0 |
| @@ -555,7 +555,7 @@ GLOBAL_ENTRY(paravirt_fsys_bubble_down) | |||
| 555 | cmp.ne pKStk,pUStk=r0,r0 // A set pKStk <- 0, pUStk <- 1 | 555 | cmp.ne pKStk,pUStk=r0,r0 // A set pKStk <- 0, pUStk <- 1 |
| 556 | br.call.sptk.many b7=ia64_syscall_setup // B | 556 | br.call.sptk.many b7=ia64_syscall_setup // B |
| 557 | ;; | 557 | ;; |
| 558 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 558 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 559 | // mov.m r30=ar.itc is called in advance | 559 | // mov.m r30=ar.itc is called in advance |
| 560 | add r16=TI_AC_STAMP+IA64_TASK_SIZE,r2 | 560 | add r16=TI_AC_STAMP+IA64_TASK_SIZE,r2 |
| 561 | add r17=TI_AC_LEAVE+IA64_TASK_SIZE,r2 | 561 | add r17=TI_AC_LEAVE+IA64_TASK_SIZE,r2 |
diff --git a/arch/ia64/kernel/head.S b/arch/ia64/kernel/head.S index 4738ff7bd66a..9be4e497f3d3 100644 --- a/arch/ia64/kernel/head.S +++ b/arch/ia64/kernel/head.S | |||
| @@ -1073,7 +1073,7 @@ END(ia64_native_sched_clock) | |||
| 1073 | sched_clock = ia64_native_sched_clock | 1073 | sched_clock = ia64_native_sched_clock |
| 1074 | #endif | 1074 | #endif |
| 1075 | 1075 | ||
| 1076 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 1076 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 1077 | GLOBAL_ENTRY(cycle_to_cputime) | 1077 | GLOBAL_ENTRY(cycle_to_cputime) |
| 1078 | alloc r16=ar.pfs,1,0,0,0 | 1078 | alloc r16=ar.pfs,1,0,0,0 |
| 1079 | addl r8=THIS_CPU(ia64_cpu_info) + IA64_CPUINFO_NSEC_PER_CYC_OFFSET,r0 | 1079 | addl r8=THIS_CPU(ia64_cpu_info) + IA64_CPUINFO_NSEC_PER_CYC_OFFSET,r0 |
| @@ -1091,7 +1091,7 @@ GLOBAL_ENTRY(cycle_to_cputime) | |||
| 1091 | shrp r8=r9,r8,IA64_NSEC_PER_CYC_SHIFT | 1091 | shrp r8=r9,r8,IA64_NSEC_PER_CYC_SHIFT |
| 1092 | br.ret.sptk.many rp | 1092 | br.ret.sptk.many rp |
| 1093 | END(cycle_to_cputime) | 1093 | END(cycle_to_cputime) |
| 1094 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 1094 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 1095 | 1095 | ||
| 1096 | #ifdef CONFIG_IA64_BRL_EMU | 1096 | #ifdef CONFIG_IA64_BRL_EMU |
| 1097 | 1097 | ||
diff --git a/arch/ia64/kernel/ivt.S b/arch/ia64/kernel/ivt.S index fa25689fc453..689ffcaa284e 100644 --- a/arch/ia64/kernel/ivt.S +++ b/arch/ia64/kernel/ivt.S | |||
| @@ -784,7 +784,7 @@ ENTRY(break_fault) | |||
| 784 | 784 | ||
| 785 | (p8) adds r28=16,r28 // A switch cr.iip to next bundle | 785 | (p8) adds r28=16,r28 // A switch cr.iip to next bundle |
| 786 | (p9) adds r8=1,r8 // A increment ei to next slot | 786 | (p9) adds r8=1,r8 // A increment ei to next slot |
| 787 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 787 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 788 | ;; | 788 | ;; |
| 789 | mov b6=r30 // I0 setup syscall handler branch reg early | 789 | mov b6=r30 // I0 setup syscall handler branch reg early |
| 790 | #else | 790 | #else |
| @@ -801,7 +801,7 @@ ENTRY(break_fault) | |||
| 801 | // | 801 | // |
| 802 | /////////////////////////////////////////////////////////////////////// | 802 | /////////////////////////////////////////////////////////////////////// |
| 803 | st1 [r16]=r0 // M2|3 clear current->thread.on_ustack flag | 803 | st1 [r16]=r0 // M2|3 clear current->thread.on_ustack flag |
| 804 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 804 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 805 | MOV_FROM_ITC(p0, p14, r30, r18) // M get cycle for accounting | 805 | MOV_FROM_ITC(p0, p14, r30, r18) // M get cycle for accounting |
| 806 | #else | 806 | #else |
| 807 | mov b6=r30 // I0 setup syscall handler branch reg early | 807 | mov b6=r30 // I0 setup syscall handler branch reg early |
| @@ -817,7 +817,7 @@ ENTRY(break_fault) | |||
| 817 | cmp.eq p14,p0=r9,r0 // A are syscalls being traced/audited? | 817 | cmp.eq p14,p0=r9,r0 // A are syscalls being traced/audited? |
| 818 | br.call.sptk.many b7=ia64_syscall_setup // B | 818 | br.call.sptk.many b7=ia64_syscall_setup // B |
| 819 | 1: | 819 | 1: |
| 820 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 820 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 821 | // mov.m r30=ar.itc is called in advance, and r13 is current | 821 | // mov.m r30=ar.itc is called in advance, and r13 is current |
| 822 | add r16=TI_AC_STAMP+IA64_TASK_SIZE,r13 // A | 822 | add r16=TI_AC_STAMP+IA64_TASK_SIZE,r13 // A |
| 823 | add r17=TI_AC_LEAVE+IA64_TASK_SIZE,r13 // A | 823 | add r17=TI_AC_LEAVE+IA64_TASK_SIZE,r13 // A |
| @@ -1043,7 +1043,7 @@ END(ia64_syscall_setup) | |||
| 1043 | DBG_FAULT(16) | 1043 | DBG_FAULT(16) |
| 1044 | FAULT(16) | 1044 | FAULT(16) |
| 1045 | 1045 | ||
| 1046 | #if defined(CONFIG_VIRT_CPU_ACCOUNTING) && defined(__IA64_ASM_PARAVIRTUALIZED_NATIVE) | 1046 | #if defined(CONFIG_VIRT_CPU_ACCOUNTING_NATIVE) && defined(__IA64_ASM_PARAVIRTUALIZED_NATIVE) |
| 1047 | /* | 1047 | /* |
| 1048 | * There is no particular reason for this code to be here, other than | 1048 | * There is no particular reason for this code to be here, other than |
| 1049 | * that there happens to be space here that would go unused otherwise. | 1049 | * that there happens to be space here that would go unused otherwise. |
diff --git a/arch/ia64/kernel/minstate.h b/arch/ia64/kernel/minstate.h index d56753a11636..cc82a7d744c9 100644 --- a/arch/ia64/kernel/minstate.h +++ b/arch/ia64/kernel/minstate.h | |||
| @@ -4,7 +4,7 @@ | |||
| 4 | #include "entry.h" | 4 | #include "entry.h" |
| 5 | #include "paravirt_inst.h" | 5 | #include "paravirt_inst.h" |
| 6 | 6 | ||
| 7 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 7 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 8 | /* read ar.itc in advance, and use it before leaving bank 0 */ | 8 | /* read ar.itc in advance, and use it before leaving bank 0 */ |
| 9 | #define ACCOUNT_GET_STAMP \ | 9 | #define ACCOUNT_GET_STAMP \ |
| 10 | (pUStk) mov.m r20=ar.itc; | 10 | (pUStk) mov.m r20=ar.itc; |
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c index 88a794536bc0..fbaac1afb844 100644 --- a/arch/ia64/kernel/time.c +++ b/arch/ia64/kernel/time.c | |||
| @@ -77,7 +77,7 @@ static struct clocksource clocksource_itc = { | |||
| 77 | }; | 77 | }; |
| 78 | static struct clocksource *itc_clocksource; | 78 | static struct clocksource *itc_clocksource; |
| 79 | 79 | ||
| 80 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 80 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 81 | 81 | ||
| 82 | #include <linux/kernel_stat.h> | 82 | #include <linux/kernel_stat.h> |
| 83 | 83 | ||
| @@ -136,13 +136,14 @@ void vtime_account_system(struct task_struct *tsk) | |||
| 136 | 136 | ||
| 137 | account_system_time(tsk, 0, delta, delta); | 137 | account_system_time(tsk, 0, delta, delta); |
| 138 | } | 138 | } |
| 139 | EXPORT_SYMBOL_GPL(vtime_account_system); | ||
| 139 | 140 | ||
| 140 | void vtime_account_idle(struct task_struct *tsk) | 141 | void vtime_account_idle(struct task_struct *tsk) |
| 141 | { | 142 | { |
| 142 | account_idle_time(vtime_delta(tsk)); | 143 | account_idle_time(vtime_delta(tsk)); |
| 143 | } | 144 | } |
| 144 | 145 | ||
| 145 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 146 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 146 | 147 | ||
| 147 | static irqreturn_t | 148 | static irqreturn_t |
| 148 | timer_interrupt (int irq, void *dev_id) | 149 | timer_interrupt (int irq, void *dev_id) |
diff --git a/arch/powerpc/configs/chroma_defconfig b/arch/powerpc/configs/chroma_defconfig index 29bb11ec6c64..4f35fc462385 100644 --- a/arch/powerpc/configs/chroma_defconfig +++ b/arch/powerpc/configs/chroma_defconfig | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | CONFIG_PPC64=y | 1 | CONFIG_PPC64=y |
| 2 | CONFIG_PPC_BOOK3E_64=y | 2 | CONFIG_PPC_BOOK3E_64=y |
| 3 | # CONFIG_VIRT_CPU_ACCOUNTING is not set | 3 | # CONFIG_VIRT_CPU_ACCOUNTING_NATIVE is not set |
| 4 | CONFIG_SMP=y | 4 | CONFIG_SMP=y |
| 5 | CONFIG_NR_CPUS=256 | 5 | CONFIG_NR_CPUS=256 |
| 6 | CONFIG_EXPERIMENTAL=y | 6 | CONFIG_EXPERIMENTAL=y |
diff --git a/arch/powerpc/configs/corenet64_smp_defconfig b/arch/powerpc/configs/corenet64_smp_defconfig index 88fa5c46f66f..f7df8362911f 100644 --- a/arch/powerpc/configs/corenet64_smp_defconfig +++ b/arch/powerpc/configs/corenet64_smp_defconfig | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | CONFIG_PPC64=y | 1 | CONFIG_PPC64=y |
| 2 | CONFIG_PPC_BOOK3E_64=y | 2 | CONFIG_PPC_BOOK3E_64=y |
| 3 | # CONFIG_VIRT_CPU_ACCOUNTING is not set | 3 | # CONFIG_VIRT_CPU_ACCOUNTING_NATIVE is not set |
| 4 | CONFIG_SMP=y | 4 | CONFIG_SMP=y |
| 5 | CONFIG_NR_CPUS=2 | 5 | CONFIG_NR_CPUS=2 |
| 6 | CONFIG_EXPERIMENTAL=y | 6 | CONFIG_EXPERIMENTAL=y |
diff --git a/arch/powerpc/configs/pasemi_defconfig b/arch/powerpc/configs/pasemi_defconfig index 840a2c2d0430..bcedeea0df89 100644 --- a/arch/powerpc/configs/pasemi_defconfig +++ b/arch/powerpc/configs/pasemi_defconfig | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | CONFIG_PPC64=y | 1 | CONFIG_PPC64=y |
| 2 | CONFIG_ALTIVEC=y | 2 | CONFIG_ALTIVEC=y |
| 3 | # CONFIG_VIRT_CPU_ACCOUNTING is not set | 3 | # CONFIG_VIRT_CPU_ACCOUNTING_NATIVE is not set |
| 4 | CONFIG_SMP=y | 4 | CONFIG_SMP=y |
| 5 | CONFIG_NR_CPUS=2 | 5 | CONFIG_NR_CPUS=2 |
| 6 | CONFIG_EXPERIMENTAL=y | 6 | CONFIG_EXPERIMENTAL=y |
diff --git a/arch/powerpc/include/asm/cputime.h b/arch/powerpc/include/asm/cputime.h index 483733bd06d4..607559ab271f 100644 --- a/arch/powerpc/include/asm/cputime.h +++ b/arch/powerpc/include/asm/cputime.h | |||
| @@ -8,7 +8,7 @@ | |||
| 8 | * as published by the Free Software Foundation; either version | 8 | * as published by the Free Software Foundation; either version |
| 9 | * 2 of the License, or (at your option) any later version. | 9 | * 2 of the License, or (at your option) any later version. |
| 10 | * | 10 | * |
| 11 | * If we have CONFIG_VIRT_CPU_ACCOUNTING, we measure cpu time in | 11 | * If we have CONFIG_VIRT_CPU_ACCOUNTING_NATIVE, we measure cpu time in |
| 12 | * the same units as the timebase. Otherwise we measure cpu time | 12 | * the same units as the timebase. Otherwise we measure cpu time |
| 13 | * in jiffies using the generic definitions. | 13 | * in jiffies using the generic definitions. |
| 14 | */ | 14 | */ |
| @@ -16,7 +16,7 @@ | |||
| 16 | #ifndef __POWERPC_CPUTIME_H | 16 | #ifndef __POWERPC_CPUTIME_H |
| 17 | #define __POWERPC_CPUTIME_H | 17 | #define __POWERPC_CPUTIME_H |
| 18 | 18 | ||
| 19 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 19 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 20 | #include <asm-generic/cputime.h> | 20 | #include <asm-generic/cputime.h> |
| 21 | #ifdef __KERNEL__ | 21 | #ifdef __KERNEL__ |
| 22 | static inline void setup_cputime_one_jiffy(void) { } | 22 | static inline void setup_cputime_one_jiffy(void) { } |
| @@ -231,5 +231,5 @@ static inline cputime_t clock_t_to_cputime(const unsigned long clk) | |||
| 231 | static inline void arch_vtime_task_switch(struct task_struct *tsk) { } | 231 | static inline void arch_vtime_task_switch(struct task_struct *tsk) { } |
| 232 | 232 | ||
| 233 | #endif /* __KERNEL__ */ | 233 | #endif /* __KERNEL__ */ |
| 234 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 234 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 235 | #endif /* __POWERPC_CPUTIME_H */ | 235 | #endif /* __POWERPC_CPUTIME_H */ |
diff --git a/arch/powerpc/include/asm/lppaca.h b/arch/powerpc/include/asm/lppaca.h index 531fe0c3108f..b1e7f2af1016 100644 --- a/arch/powerpc/include/asm/lppaca.h +++ b/arch/powerpc/include/asm/lppaca.h | |||
| @@ -145,7 +145,7 @@ struct dtl_entry { | |||
| 145 | extern struct kmem_cache *dtl_cache; | 145 | extern struct kmem_cache *dtl_cache; |
| 146 | 146 | ||
| 147 | /* | 147 | /* |
| 148 | * When CONFIG_VIRT_CPU_ACCOUNTING = y, the cpu accounting code controls | 148 | * When CONFIG_VIRT_CPU_ACCOUNTING_NATIVE = y, the cpu accounting code controls |
| 149 | * reading from the dispatch trace log. If other code wants to consume | 149 | * reading from the dispatch trace log. If other code wants to consume |
| 150 | * DTL entries, it can set this pointer to a function that will get | 150 | * DTL entries, it can set this pointer to a function that will get |
| 151 | * called once for each DTL entry that gets processed. | 151 | * called once for each DTL entry that gets processed. |
diff --git a/arch/powerpc/include/asm/ppc_asm.h b/arch/powerpc/include/asm/ppc_asm.h index ea2a86e8ff95..2d0e1f5d8339 100644 --- a/arch/powerpc/include/asm/ppc_asm.h +++ b/arch/powerpc/include/asm/ppc_asm.h | |||
| @@ -24,7 +24,7 @@ | |||
| 24 | * user_time and system_time fields in the paca. | 24 | * user_time and system_time fields in the paca. |
| 25 | */ | 25 | */ |
| 26 | 26 | ||
| 27 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 27 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 28 | #define ACCOUNT_CPU_USER_ENTRY(ra, rb) | 28 | #define ACCOUNT_CPU_USER_ENTRY(ra, rb) |
| 29 | #define ACCOUNT_CPU_USER_EXIT(ra, rb) | 29 | #define ACCOUNT_CPU_USER_EXIT(ra, rb) |
| 30 | #define ACCOUNT_STOLEN_TIME | 30 | #define ACCOUNT_STOLEN_TIME |
| @@ -70,7 +70,7 @@ END_FW_FTR_SECTION_IFSET(FW_FEATURE_SPLPAR) | |||
| 70 | 70 | ||
| 71 | #endif /* CONFIG_PPC_SPLPAR */ | 71 | #endif /* CONFIG_PPC_SPLPAR */ |
| 72 | 72 | ||
| 73 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 73 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 74 | 74 | ||
| 75 | /* | 75 | /* |
| 76 | * Macros for storing registers into and loading registers from | 76 | * Macros for storing registers into and loading registers from |
diff --git a/arch/powerpc/kernel/entry_64.S b/arch/powerpc/kernel/entry_64.S index 3d990d3bd8ba..ac057013f9fd 100644 --- a/arch/powerpc/kernel/entry_64.S +++ b/arch/powerpc/kernel/entry_64.S | |||
| @@ -94,7 +94,7 @@ system_call_common: | |||
| 94 | addi r9,r1,STACK_FRAME_OVERHEAD | 94 | addi r9,r1,STACK_FRAME_OVERHEAD |
| 95 | ld r11,exception_marker@toc(r2) | 95 | ld r11,exception_marker@toc(r2) |
| 96 | std r11,-16(r9) /* "regshere" marker */ | 96 | std r11,-16(r9) /* "regshere" marker */ |
| 97 | #if defined(CONFIG_VIRT_CPU_ACCOUNTING) && defined(CONFIG_PPC_SPLPAR) | 97 | #if defined(CONFIG_VIRT_CPU_ACCOUNTING_NATIVE) && defined(CONFIG_PPC_SPLPAR) |
| 98 | BEGIN_FW_FTR_SECTION | 98 | BEGIN_FW_FTR_SECTION |
| 99 | beq 33f | 99 | beq 33f |
| 100 | /* if from user, see if there are any DTL entries to process */ | 100 | /* if from user, see if there are any DTL entries to process */ |
| @@ -110,7 +110,7 @@ BEGIN_FW_FTR_SECTION | |||
| 110 | addi r9,r1,STACK_FRAME_OVERHEAD | 110 | addi r9,r1,STACK_FRAME_OVERHEAD |
| 111 | 33: | 111 | 33: |
| 112 | END_FW_FTR_SECTION_IFSET(FW_FEATURE_SPLPAR) | 112 | END_FW_FTR_SECTION_IFSET(FW_FEATURE_SPLPAR) |
| 113 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING && CONFIG_PPC_SPLPAR */ | 113 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE && CONFIG_PPC_SPLPAR */ |
| 114 | 114 | ||
| 115 | /* | 115 | /* |
| 116 | * A syscall should always be called with interrupts enabled | 116 | * A syscall should always be called with interrupts enabled |
diff --git a/arch/powerpc/kernel/time.c b/arch/powerpc/kernel/time.c index 127361e093f4..89b0f58194d1 100644 --- a/arch/powerpc/kernel/time.c +++ b/arch/powerpc/kernel/time.c | |||
| @@ -143,7 +143,7 @@ EXPORT_SYMBOL_GPL(ppc_proc_freq); | |||
| 143 | unsigned long ppc_tb_freq; | 143 | unsigned long ppc_tb_freq; |
| 144 | EXPORT_SYMBOL_GPL(ppc_tb_freq); | 144 | EXPORT_SYMBOL_GPL(ppc_tb_freq); |
| 145 | 145 | ||
| 146 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 146 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 147 | /* | 147 | /* |
| 148 | * Factors for converting from cputime_t (timebase ticks) to | 148 | * Factors for converting from cputime_t (timebase ticks) to |
| 149 | * jiffies, microseconds, seconds, and clock_t (1/USER_HZ seconds). | 149 | * jiffies, microseconds, seconds, and clock_t (1/USER_HZ seconds). |
| @@ -347,6 +347,7 @@ void vtime_account_system(struct task_struct *tsk) | |||
| 347 | if (stolen) | 347 | if (stolen) |
| 348 | account_steal_time(stolen); | 348 | account_steal_time(stolen); |
| 349 | } | 349 | } |
| 350 | EXPORT_SYMBOL_GPL(vtime_account_system); | ||
| 350 | 351 | ||
| 351 | void vtime_account_idle(struct task_struct *tsk) | 352 | void vtime_account_idle(struct task_struct *tsk) |
| 352 | { | 353 | { |
| @@ -377,7 +378,7 @@ void vtime_account_user(struct task_struct *tsk) | |||
| 377 | account_user_time(tsk, utime, utimescaled); | 378 | account_user_time(tsk, utime, utimescaled); |
| 378 | } | 379 | } |
| 379 | 380 | ||
| 380 | #else /* ! CONFIG_VIRT_CPU_ACCOUNTING */ | 381 | #else /* ! CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 381 | #define calc_cputime_factors() | 382 | #define calc_cputime_factors() |
| 382 | #endif | 383 | #endif |
| 383 | 384 | ||
diff --git a/arch/powerpc/platforms/cell/spufs/sched.c b/arch/powerpc/platforms/cell/spufs/sched.c index 25db92a8e1cf..49318385d4fa 100644 --- a/arch/powerpc/platforms/cell/spufs/sched.c +++ b/arch/powerpc/platforms/cell/spufs/sched.c | |||
| @@ -24,6 +24,7 @@ | |||
| 24 | 24 | ||
| 25 | #include <linux/errno.h> | 25 | #include <linux/errno.h> |
| 26 | #include <linux/sched.h> | 26 | #include <linux/sched.h> |
| 27 | #include <linux/sched/rt.h> | ||
| 27 | #include <linux/kernel.h> | 28 | #include <linux/kernel.h> |
| 28 | #include <linux/mm.h> | 29 | #include <linux/mm.h> |
| 29 | #include <linux/slab.h> | 30 | #include <linux/slab.h> |
diff --git a/arch/powerpc/platforms/pseries/dtl.c b/arch/powerpc/platforms/pseries/dtl.c index a7648543c59e..0cc0ac07a55d 100644 --- a/arch/powerpc/platforms/pseries/dtl.c +++ b/arch/powerpc/platforms/pseries/dtl.c | |||
| @@ -57,7 +57,7 @@ static u8 dtl_event_mask = 0x7; | |||
| 57 | */ | 57 | */ |
| 58 | static int dtl_buf_entries = N_DISPATCH_LOG; | 58 | static int dtl_buf_entries = N_DISPATCH_LOG; |
| 59 | 59 | ||
| 60 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 60 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 61 | struct dtl_ring { | 61 | struct dtl_ring { |
| 62 | u64 write_index; | 62 | u64 write_index; |
| 63 | struct dtl_entry *write_ptr; | 63 | struct dtl_entry *write_ptr; |
| @@ -142,7 +142,7 @@ static u64 dtl_current_index(struct dtl *dtl) | |||
| 142 | return per_cpu(dtl_rings, dtl->cpu).write_index; | 142 | return per_cpu(dtl_rings, dtl->cpu).write_index; |
| 143 | } | 143 | } |
| 144 | 144 | ||
| 145 | #else /* CONFIG_VIRT_CPU_ACCOUNTING */ | 145 | #else /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 146 | 146 | ||
| 147 | static int dtl_start(struct dtl *dtl) | 147 | static int dtl_start(struct dtl *dtl) |
| 148 | { | 148 | { |
| @@ -188,7 +188,7 @@ static u64 dtl_current_index(struct dtl *dtl) | |||
| 188 | { | 188 | { |
| 189 | return lppaca_of(dtl->cpu).dtl_idx; | 189 | return lppaca_of(dtl->cpu).dtl_idx; |
| 190 | } | 190 | } |
| 191 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 191 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 192 | 192 | ||
| 193 | static int dtl_enable(struct dtl *dtl) | 193 | static int dtl_enable(struct dtl *dtl) |
| 194 | { | 194 | { |
diff --git a/arch/powerpc/platforms/pseries/setup.c b/arch/powerpc/platforms/pseries/setup.c index ca55882465d6..527e12c9573b 100644 --- a/arch/powerpc/platforms/pseries/setup.c +++ b/arch/powerpc/platforms/pseries/setup.c | |||
| @@ -281,7 +281,7 @@ static struct notifier_block pci_dn_reconfig_nb = { | |||
| 281 | 281 | ||
| 282 | struct kmem_cache *dtl_cache; | 282 | struct kmem_cache *dtl_cache; |
| 283 | 283 | ||
| 284 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 284 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 285 | /* | 285 | /* |
| 286 | * Allocate space for the dispatch trace log for all possible cpus | 286 | * Allocate space for the dispatch trace log for all possible cpus |
| 287 | * and register the buffers with the hypervisor. This is used for | 287 | * and register the buffers with the hypervisor. This is used for |
| @@ -332,12 +332,12 @@ static int alloc_dispatch_logs(void) | |||
| 332 | 332 | ||
| 333 | return 0; | 333 | return 0; |
| 334 | } | 334 | } |
| 335 | #else /* !CONFIG_VIRT_CPU_ACCOUNTING */ | 335 | #else /* !CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 336 | static inline int alloc_dispatch_logs(void) | 336 | static inline int alloc_dispatch_logs(void) |
| 337 | { | 337 | { |
| 338 | return 0; | 338 | return 0; |
| 339 | } | 339 | } |
| 340 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING */ | 340 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ |
| 341 | 341 | ||
| 342 | static int alloc_dispatch_log_kmem_cache(void) | 342 | static int alloc_dispatch_log_kmem_cache(void) |
| 343 | { | 343 | { |
diff --git a/arch/s390/kernel/vtime.c b/arch/s390/kernel/vtime.c index e84b8b68444a..ce9cc5aa2033 100644 --- a/arch/s390/kernel/vtime.c +++ b/arch/s390/kernel/vtime.c | |||
| @@ -127,7 +127,7 @@ void vtime_account_user(struct task_struct *tsk) | |||
| 127 | * Update process times based on virtual cpu times stored by entry.S | 127 | * Update process times based on virtual cpu times stored by entry.S |
| 128 | * to the lowcore fields user_timer, system_timer & steal_clock. | 128 | * to the lowcore fields user_timer, system_timer & steal_clock. |
| 129 | */ | 129 | */ |
| 130 | void vtime_account(struct task_struct *tsk) | 130 | void vtime_account_irq_enter(struct task_struct *tsk) |
| 131 | { | 131 | { |
| 132 | struct thread_info *ti = task_thread_info(tsk); | 132 | struct thread_info *ti = task_thread_info(tsk); |
| 133 | u64 timer, system; | 133 | u64 timer, system; |
| @@ -145,10 +145,10 @@ void vtime_account(struct task_struct *tsk) | |||
| 145 | 145 | ||
| 146 | virt_timer_forward(system); | 146 | virt_timer_forward(system); |
| 147 | } | 147 | } |
| 148 | EXPORT_SYMBOL_GPL(vtime_account); | 148 | EXPORT_SYMBOL_GPL(vtime_account_irq_enter); |
| 149 | 149 | ||
| 150 | void vtime_account_system(struct task_struct *tsk) | 150 | void vtime_account_system(struct task_struct *tsk) |
| 151 | __attribute__((alias("vtime_account"))); | 151 | __attribute__((alias("vtime_account_irq_enter"))); |
| 152 | EXPORT_SYMBOL_GPL(vtime_account_system); | 152 | EXPORT_SYMBOL_GPL(vtime_account_system); |
| 153 | 153 | ||
| 154 | void __kprobes vtime_stop_cpu(void) | 154 | void __kprobes vtime_stop_cpu(void) |
diff --git a/arch/x86/kernel/apm_32.c b/arch/x86/kernel/apm_32.c index d65464e43503..8d7012b7f402 100644 --- a/arch/x86/kernel/apm_32.c +++ b/arch/x86/kernel/apm_32.c | |||
| @@ -899,6 +899,7 @@ static void apm_cpu_idle(void) | |||
| 899 | static int use_apm_idle; /* = 0 */ | 899 | static int use_apm_idle; /* = 0 */ |
| 900 | static unsigned int last_jiffies; /* = 0 */ | 900 | static unsigned int last_jiffies; /* = 0 */ |
| 901 | static unsigned int last_stime; /* = 0 */ | 901 | static unsigned int last_stime; /* = 0 */ |
| 902 | cputime_t stime; | ||
| 902 | 903 | ||
| 903 | int apm_idle_done = 0; | 904 | int apm_idle_done = 0; |
| 904 | unsigned int jiffies_since_last_check = jiffies - last_jiffies; | 905 | unsigned int jiffies_since_last_check = jiffies - last_jiffies; |
| @@ -906,23 +907,23 @@ static void apm_cpu_idle(void) | |||
| 906 | 907 | ||
| 907 | WARN_ONCE(1, "deprecated apm_cpu_idle will be deleted in 2012"); | 908 | WARN_ONCE(1, "deprecated apm_cpu_idle will be deleted in 2012"); |
| 908 | recalc: | 909 | recalc: |
| 910 | task_cputime(current, NULL, &stime); | ||
| 909 | if (jiffies_since_last_check > IDLE_CALC_LIMIT) { | 911 | if (jiffies_since_last_check > IDLE_CALC_LIMIT) { |
| 910 | use_apm_idle = 0; | 912 | use_apm_idle = 0; |
| 911 | last_jiffies = jiffies; | ||
| 912 | last_stime = current->stime; | ||
| 913 | } else if (jiffies_since_last_check > idle_period) { | 913 | } else if (jiffies_since_last_check > idle_period) { |
| 914 | unsigned int idle_percentage; | 914 | unsigned int idle_percentage; |
| 915 | 915 | ||
| 916 | idle_percentage = current->stime - last_stime; | 916 | idle_percentage = stime - last_stime; |
| 917 | idle_percentage *= 100; | 917 | idle_percentage *= 100; |
| 918 | idle_percentage /= jiffies_since_last_check; | 918 | idle_percentage /= jiffies_since_last_check; |
| 919 | use_apm_idle = (idle_percentage > idle_threshold); | 919 | use_apm_idle = (idle_percentage > idle_threshold); |
| 920 | if (apm_info.forbid_idle) | 920 | if (apm_info.forbid_idle) |
| 921 | use_apm_idle = 0; | 921 | use_apm_idle = 0; |
| 922 | last_jiffies = jiffies; | ||
| 923 | last_stime = current->stime; | ||
| 924 | } | 922 | } |
| 925 | 923 | ||
| 924 | last_jiffies = jiffies; | ||
| 925 | last_stime = stime; | ||
| 926 | |||
| 926 | bucket = IDLE_LEAKY_MAX; | 927 | bucket = IDLE_LEAKY_MAX; |
| 927 | 928 | ||
| 928 | while (!need_resched()) { | 929 | while (!need_resched()) { |
diff --git a/block/blk-exec.c b/block/blk-exec.c index 74638ec234c8..c88202f973d9 100644 --- a/block/blk-exec.c +++ b/block/blk-exec.c | |||
| @@ -5,6 +5,7 @@ | |||
| 5 | #include <linux/module.h> | 5 | #include <linux/module.h> |
| 6 | #include <linux/bio.h> | 6 | #include <linux/bio.h> |
| 7 | #include <linux/blkdev.h> | 7 | #include <linux/blkdev.h> |
| 8 | #include <linux/sched/sysctl.h> | ||
| 8 | 9 | ||
| 9 | #include "blk.h" | 10 | #include "blk.h" |
| 10 | 11 | ||
diff --git a/drivers/isdn/mISDN/stack.c b/drivers/isdn/mISDN/stack.c index 5f21f629b7ae..deda591f70b9 100644 --- a/drivers/isdn/mISDN/stack.c +++ b/drivers/isdn/mISDN/stack.c | |||
| @@ -18,6 +18,7 @@ | |||
| 18 | #include <linux/slab.h> | 18 | #include <linux/slab.h> |
| 19 | #include <linux/mISDNif.h> | 19 | #include <linux/mISDNif.h> |
| 20 | #include <linux/kthread.h> | 20 | #include <linux/kthread.h> |
| 21 | #include <linux/sched.h> | ||
| 21 | #include "core.h" | 22 | #include "core.h" |
| 22 | 23 | ||
| 23 | static u_int *debug; | 24 | static u_int *debug; |
| @@ -202,6 +203,9 @@ static int | |||
| 202 | mISDNStackd(void *data) | 203 | mISDNStackd(void *data) |
| 203 | { | 204 | { |
| 204 | struct mISDNstack *st = data; | 205 | struct mISDNstack *st = data; |
| 206 | #ifdef MISDN_MSG_STATS | ||
| 207 | cputime_t utime, stime; | ||
| 208 | #endif | ||
| 205 | int err = 0; | 209 | int err = 0; |
| 206 | 210 | ||
| 207 | sigfillset(¤t->blocked); | 211 | sigfillset(¤t->blocked); |
| @@ -303,9 +307,10 @@ mISDNStackd(void *data) | |||
| 303 | "msg %d sleep %d stopped\n", | 307 | "msg %d sleep %d stopped\n", |
| 304 | dev_name(&st->dev->dev), st->msg_cnt, st->sleep_cnt, | 308 | dev_name(&st->dev->dev), st->msg_cnt, st->sleep_cnt, |
| 305 | st->stopped_cnt); | 309 | st->stopped_cnt); |
| 310 | task_cputime(st->thread, &utime, &stime); | ||
| 306 | printk(KERN_DEBUG | 311 | printk(KERN_DEBUG |
| 307 | "mISDNStackd daemon for %s utime(%ld) stime(%ld)\n", | 312 | "mISDNStackd daemon for %s utime(%ld) stime(%ld)\n", |
| 308 | dev_name(&st->dev->dev), st->thread->utime, st->thread->stime); | 313 | dev_name(&st->dev->dev), utime, stime); |
| 309 | printk(KERN_DEBUG | 314 | printk(KERN_DEBUG |
| 310 | "mISDNStackd daemon for %s nvcsw(%ld) nivcsw(%ld)\n", | 315 | "mISDNStackd daemon for %s nvcsw(%ld) nivcsw(%ld)\n", |
| 311 | dev_name(&st->dev->dev), st->thread->nvcsw, st->thread->nivcsw); | 316 | dev_name(&st->dev->dev), st->thread->nvcsw, st->thread->nivcsw); |
diff --git a/drivers/spi/spi.c b/drivers/spi/spi.c index 19ee901577da..3a6083b386a1 100644 --- a/drivers/spi/spi.c +++ b/drivers/spi/spi.c | |||
| @@ -33,7 +33,7 @@ | |||
| 33 | #include <linux/of_gpio.h> | 33 | #include <linux/of_gpio.h> |
| 34 | #include <linux/pm_runtime.h> | 34 | #include <linux/pm_runtime.h> |
| 35 | #include <linux/export.h> | 35 | #include <linux/export.h> |
| 36 | #include <linux/sched.h> | 36 | #include <linux/sched/rt.h> |
| 37 | #include <linux/delay.h> | 37 | #include <linux/delay.h> |
| 38 | #include <linux/kthread.h> | 38 | #include <linux/kthread.h> |
| 39 | #include <linux/ioport.h> | 39 | #include <linux/ioport.h> |
diff --git a/drivers/staging/csr/bh.c b/drivers/staging/csr/bh.c index 1a1f5c79822a..7b133597e923 100644 --- a/drivers/staging/csr/bh.c +++ b/drivers/staging/csr/bh.c | |||
| @@ -15,7 +15,7 @@ | |||
| 15 | */ | 15 | */ |
| 16 | #include "csr_wifi_hip_unifi.h" | 16 | #include "csr_wifi_hip_unifi.h" |
| 17 | #include "unifi_priv.h" | 17 | #include "unifi_priv.h" |
| 18 | 18 | #include <linux/sched/rt.h> | |
| 19 | 19 | ||
| 20 | /* | 20 | /* |
| 21 | * --------------------------------------------------------------------------- | 21 | * --------------------------------------------------------------------------- |
diff --git a/drivers/staging/csr/unifi_sme.c b/drivers/staging/csr/unifi_sme.c index 7c6c4138fc76..49395da34b7f 100644 --- a/drivers/staging/csr/unifi_sme.c +++ b/drivers/staging/csr/unifi_sme.c | |||
| @@ -15,7 +15,7 @@ | |||
| 15 | #include "unifi_priv.h" | 15 | #include "unifi_priv.h" |
| 16 | #include "csr_wifi_hip_unifi.h" | 16 | #include "csr_wifi_hip_unifi.h" |
| 17 | #include "csr_wifi_hip_conversions.h" | 17 | #include "csr_wifi_hip_conversions.h" |
| 18 | 18 | #include <linux/sched/rt.h> | |
| 19 | 19 | ||
| 20 | 20 | ||
| 21 | 21 | ||
diff --git a/drivers/tty/sysrq.c b/drivers/tty/sysrq.c index b3c4a250ff86..40e5b3919e27 100644 --- a/drivers/tty/sysrq.c +++ b/drivers/tty/sysrq.c | |||
| @@ -15,6 +15,7 @@ | |||
| 15 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt | 15 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
| 16 | 16 | ||
| 17 | #include <linux/sched.h> | 17 | #include <linux/sched.h> |
| 18 | #include <linux/sched/rt.h> | ||
| 18 | #include <linux/interrupt.h> | 19 | #include <linux/interrupt.h> |
| 19 | #include <linux/mm.h> | 20 | #include <linux/mm.h> |
| 20 | #include <linux/fs.h> | 21 | #include <linux/fs.h> |
diff --git a/fs/binfmt_elf.c b/fs/binfmt_elf.c index 0c42cdbabecf..49d0b43458b7 100644 --- a/fs/binfmt_elf.c +++ b/fs/binfmt_elf.c | |||
| @@ -33,6 +33,7 @@ | |||
| 33 | #include <linux/elf.h> | 33 | #include <linux/elf.h> |
| 34 | #include <linux/utsname.h> | 34 | #include <linux/utsname.h> |
| 35 | #include <linux/coredump.h> | 35 | #include <linux/coredump.h> |
| 36 | #include <linux/sched.h> | ||
| 36 | #include <asm/uaccess.h> | 37 | #include <asm/uaccess.h> |
| 37 | #include <asm/param.h> | 38 | #include <asm/param.h> |
| 38 | #include <asm/page.h> | 39 | #include <asm/page.h> |
| @@ -1320,8 +1321,11 @@ static void fill_prstatus(struct elf_prstatus *prstatus, | |||
| 1320 | cputime_to_timeval(cputime.utime, &prstatus->pr_utime); | 1321 | cputime_to_timeval(cputime.utime, &prstatus->pr_utime); |
| 1321 | cputime_to_timeval(cputime.stime, &prstatus->pr_stime); | 1322 | cputime_to_timeval(cputime.stime, &prstatus->pr_stime); |
| 1322 | } else { | 1323 | } else { |
| 1323 | cputime_to_timeval(p->utime, &prstatus->pr_utime); | 1324 | cputime_t utime, stime; |
| 1324 | cputime_to_timeval(p->stime, &prstatus->pr_stime); | 1325 | |
| 1326 | task_cputime(p, &utime, &stime); | ||
| 1327 | cputime_to_timeval(utime, &prstatus->pr_utime); | ||
| 1328 | cputime_to_timeval(stime, &prstatus->pr_stime); | ||
| 1325 | } | 1329 | } |
| 1326 | cputime_to_timeval(p->signal->cutime, &prstatus->pr_cutime); | 1330 | cputime_to_timeval(p->signal->cutime, &prstatus->pr_cutime); |
| 1327 | cputime_to_timeval(p->signal->cstime, &prstatus->pr_cstime); | 1331 | cputime_to_timeval(p->signal->cstime, &prstatus->pr_cstime); |
diff --git a/fs/binfmt_elf_fdpic.c b/fs/binfmt_elf_fdpic.c index dc84732e554f..cb240dd3b402 100644 --- a/fs/binfmt_elf_fdpic.c +++ b/fs/binfmt_elf_fdpic.c | |||
| @@ -1375,8 +1375,11 @@ static void fill_prstatus(struct elf_prstatus *prstatus, | |||
| 1375 | cputime_to_timeval(cputime.utime, &prstatus->pr_utime); | 1375 | cputime_to_timeval(cputime.utime, &prstatus->pr_utime); |
| 1376 | cputime_to_timeval(cputime.stime, &prstatus->pr_stime); | 1376 | cputime_to_timeval(cputime.stime, &prstatus->pr_stime); |
| 1377 | } else { | 1377 | } else { |
| 1378 | cputime_to_timeval(p->utime, &prstatus->pr_utime); | 1378 | cputime_t utime, stime; |
| 1379 | cputime_to_timeval(p->stime, &prstatus->pr_stime); | 1379 | |
| 1380 | task_cputime(p, &utime, &stime); | ||
| 1381 | cputime_to_timeval(utime, &prstatus->pr_utime); | ||
| 1382 | cputime_to_timeval(stime, &prstatus->pr_stime); | ||
| 1380 | } | 1383 | } |
| 1381 | cputime_to_timeval(p->signal->cutime, &prstatus->pr_cutime); | 1384 | cputime_to_timeval(p->signal->cutime, &prstatus->pr_cutime); |
| 1382 | cputime_to_timeval(p->signal->cstime, &prstatus->pr_cstime); | 1385 | cputime_to_timeval(p->signal->cstime, &prstatus->pr_cstime); |
diff --git a/fs/proc/array.c b/fs/proc/array.c index 6a91e6ffbcbd..f7ed9ee46eb9 100644 --- a/fs/proc/array.c +++ b/fs/proc/array.c | |||
| @@ -449,7 +449,7 @@ static int do_task_stat(struct seq_file *m, struct pid_namespace *ns, | |||
| 449 | do { | 449 | do { |
| 450 | min_flt += t->min_flt; | 450 | min_flt += t->min_flt; |
| 451 | maj_flt += t->maj_flt; | 451 | maj_flt += t->maj_flt; |
| 452 | gtime += t->gtime; | 452 | gtime += task_gtime(t); |
| 453 | t = next_thread(t); | 453 | t = next_thread(t); |
| 454 | } while (t != task); | 454 | } while (t != task); |
| 455 | 455 | ||
| @@ -472,7 +472,7 @@ static int do_task_stat(struct seq_file *m, struct pid_namespace *ns, | |||
| 472 | min_flt = task->min_flt; | 472 | min_flt = task->min_flt; |
| 473 | maj_flt = task->maj_flt; | 473 | maj_flt = task->maj_flt; |
| 474 | task_cputime_adjusted(task, &utime, &stime); | 474 | task_cputime_adjusted(task, &utime, &stime); |
| 475 | gtime = task->gtime; | 475 | gtime = task_gtime(task); |
| 476 | } | 476 | } |
| 477 | 477 | ||
| 478 | /* scale priority and nice values from timeslices to -20..20 */ | 478 | /* scale priority and nice values from timeslices to -20..20 */ |
diff --git a/fs/select.c b/fs/select.c index 2ef72d965036..8c1c96c27062 100644 --- a/fs/select.c +++ b/fs/select.c | |||
| @@ -26,6 +26,7 @@ | |||
| 26 | #include <linux/fs.h> | 26 | #include <linux/fs.h> |
| 27 | #include <linux/rcupdate.h> | 27 | #include <linux/rcupdate.h> |
| 28 | #include <linux/hrtimer.h> | 28 | #include <linux/hrtimer.h> |
| 29 | #include <linux/sched/rt.h> | ||
| 29 | 30 | ||
| 30 | #include <asm/uaccess.h> | 31 | #include <asm/uaccess.h> |
| 31 | 32 | ||
diff --git a/include/asm-generic/cputime.h b/include/asm-generic/cputime.h index 9a62937c56ca..51969436b8b8 100644 --- a/include/asm-generic/cputime.h +++ b/include/asm-generic/cputime.h | |||
| @@ -4,66 +4,12 @@ | |||
| 4 | #include <linux/time.h> | 4 | #include <linux/time.h> |
| 5 | #include <linux/jiffies.h> | 5 | #include <linux/jiffies.h> |
| 6 | 6 | ||
| 7 | typedef unsigned long __nocast cputime_t; | 7 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING |
| 8 | 8 | # include <asm-generic/cputime_jiffies.h> | |
| 9 | #define cputime_one_jiffy jiffies_to_cputime(1) | 9 | #endif |
| 10 | #define cputime_to_jiffies(__ct) (__force unsigned long)(__ct) | ||
| 11 | #define cputime_to_scaled(__ct) (__ct) | ||
| 12 | #define jiffies_to_cputime(__hz) (__force cputime_t)(__hz) | ||
| 13 | |||
| 14 | typedef u64 __nocast cputime64_t; | ||
| 15 | |||
| 16 | #define cputime64_to_jiffies64(__ct) (__force u64)(__ct) | ||
| 17 | #define jiffies64_to_cputime64(__jif) (__force cputime64_t)(__jif) | ||
| 18 | |||
| 19 | #define nsecs_to_cputime64(__ct) \ | ||
| 20 | jiffies64_to_cputime64(nsecs_to_jiffies64(__ct)) | ||
| 21 | |||
| 22 | |||
| 23 | /* | ||
| 24 | * Convert cputime to microseconds and back. | ||
| 25 | */ | ||
| 26 | #define cputime_to_usecs(__ct) \ | ||
| 27 | jiffies_to_usecs(cputime_to_jiffies(__ct)) | ||
| 28 | #define usecs_to_cputime(__usec) \ | ||
| 29 | jiffies_to_cputime(usecs_to_jiffies(__usec)) | ||
| 30 | #define usecs_to_cputime64(__usec) \ | ||
| 31 | jiffies64_to_cputime64(nsecs_to_jiffies64((__usec) * 1000)) | ||
| 32 | |||
| 33 | /* | ||
| 34 | * Convert cputime to seconds and back. | ||
| 35 | */ | ||
| 36 | #define cputime_to_secs(jif) (cputime_to_jiffies(jif) / HZ) | ||
| 37 | #define secs_to_cputime(sec) jiffies_to_cputime((sec) * HZ) | ||
| 38 | |||
| 39 | /* | ||
| 40 | * Convert cputime to timespec and back. | ||
| 41 | */ | ||
| 42 | #define timespec_to_cputime(__val) \ | ||
| 43 | jiffies_to_cputime(timespec_to_jiffies(__val)) | ||
| 44 | #define cputime_to_timespec(__ct,__val) \ | ||
| 45 | jiffies_to_timespec(cputime_to_jiffies(__ct),__val) | ||
| 46 | |||
| 47 | /* | ||
| 48 | * Convert cputime to timeval and back. | ||
| 49 | */ | ||
| 50 | #define timeval_to_cputime(__val) \ | ||
| 51 | jiffies_to_cputime(timeval_to_jiffies(__val)) | ||
| 52 | #define cputime_to_timeval(__ct,__val) \ | ||
| 53 | jiffies_to_timeval(cputime_to_jiffies(__ct),__val) | ||
| 54 | |||
| 55 | /* | ||
| 56 | * Convert cputime to clock and back. | ||
| 57 | */ | ||
| 58 | #define cputime_to_clock_t(__ct) \ | ||
| 59 | jiffies_to_clock_t(cputime_to_jiffies(__ct)) | ||
| 60 | #define clock_t_to_cputime(__x) \ | ||
| 61 | jiffies_to_cputime(clock_t_to_jiffies(__x)) | ||
| 62 | 10 | ||
| 63 | /* | 11 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN |
| 64 | * Convert cputime64 to clock. | 12 | # include <asm-generic/cputime_nsecs.h> |
| 65 | */ | 13 | #endif |
| 66 | #define cputime64_to_clock_t(__ct) \ | ||
| 67 | jiffies_64_to_clock_t(cputime64_to_jiffies64(__ct)) | ||
| 68 | 14 | ||
| 69 | #endif | 15 | #endif |
diff --git a/include/asm-generic/cputime_jiffies.h b/include/asm-generic/cputime_jiffies.h new file mode 100644 index 000000000000..272ecba9f588 --- /dev/null +++ b/include/asm-generic/cputime_jiffies.h | |||
| @@ -0,0 +1,72 @@ | |||
| 1 | #ifndef _ASM_GENERIC_CPUTIME_JIFFIES_H | ||
| 2 | #define _ASM_GENERIC_CPUTIME_JIFFIES_H | ||
| 3 | |||
| 4 | typedef unsigned long __nocast cputime_t; | ||
| 5 | |||
| 6 | #define cputime_one_jiffy jiffies_to_cputime(1) | ||
| 7 | #define cputime_to_jiffies(__ct) (__force unsigned long)(__ct) | ||
| 8 | #define cputime_to_scaled(__ct) (__ct) | ||
| 9 | #define jiffies_to_cputime(__hz) (__force cputime_t)(__hz) | ||
| 10 | |||
| 11 | typedef u64 __nocast cputime64_t; | ||
| 12 | |||
| 13 | #define cputime64_to_jiffies64(__ct) (__force u64)(__ct) | ||
| 14 | #define jiffies64_to_cputime64(__jif) (__force cputime64_t)(__jif) | ||
| 15 | |||
| 16 | |||
| 17 | /* | ||
| 18 | * Convert nanoseconds to cputime | ||
| 19 | */ | ||
| 20 | #define nsecs_to_cputime64(__nsec) \ | ||
| 21 | jiffies64_to_cputime64(nsecs_to_jiffies64(__nsec)) | ||
| 22 | #define nsecs_to_cputime(__nsec) \ | ||
| 23 | jiffies_to_cputime(nsecs_to_jiffies(__nsec)) | ||
| 24 | |||
| 25 | |||
| 26 | /* | ||
| 27 | * Convert cputime to microseconds and back. | ||
| 28 | */ | ||
| 29 | #define cputime_to_usecs(__ct) \ | ||
| 30 | jiffies_to_usecs(cputime_to_jiffies(__ct)) | ||
| 31 | #define usecs_to_cputime(__usec) \ | ||
| 32 | jiffies_to_cputime(usecs_to_jiffies(__usec)) | ||
| 33 | #define usecs_to_cputime64(__usec) \ | ||
| 34 | jiffies64_to_cputime64(nsecs_to_jiffies64((__usec) * 1000)) | ||
| 35 | |||
| 36 | /* | ||
| 37 | * Convert cputime to seconds and back. | ||
| 38 | */ | ||
| 39 | #define cputime_to_secs(jif) (cputime_to_jiffies(jif) / HZ) | ||
| 40 | #define secs_to_cputime(sec) jiffies_to_cputime((sec) * HZ) | ||
| 41 | |||
| 42 | /* | ||
| 43 | * Convert cputime to timespec and back. | ||
| 44 | */ | ||
| 45 | #define timespec_to_cputime(__val) \ | ||
| 46 | jiffies_to_cputime(timespec_to_jiffies(__val)) | ||
| 47 | #define cputime_to_timespec(__ct,__val) \ | ||
| 48 | jiffies_to_timespec(cputime_to_jiffies(__ct),__val) | ||
| 49 | |||
| 50 | /* | ||
| 51 | * Convert cputime to timeval and back. | ||
| 52 | */ | ||
| 53 | #define timeval_to_cputime(__val) \ | ||
| 54 | jiffies_to_cputime(timeval_to_jiffies(__val)) | ||
| 55 | #define cputime_to_timeval(__ct,__val) \ | ||
| 56 | jiffies_to_timeval(cputime_to_jiffies(__ct),__val) | ||
| 57 | |||
| 58 | /* | ||
| 59 | * Convert cputime to clock and back. | ||
| 60 | */ | ||
| 61 | #define cputime_to_clock_t(__ct) \ | ||
| 62 | jiffies_to_clock_t(cputime_to_jiffies(__ct)) | ||
| 63 | #define clock_t_to_cputime(__x) \ | ||
| 64 | jiffies_to_cputime(clock_t_to_jiffies(__x)) | ||
| 65 | |||
| 66 | /* | ||
| 67 | * Convert cputime64 to clock. | ||
| 68 | */ | ||
| 69 | #define cputime64_to_clock_t(__ct) \ | ||
| 70 | jiffies_64_to_clock_t(cputime64_to_jiffies64(__ct)) | ||
| 71 | |||
| 72 | #endif | ||
diff --git a/include/asm-generic/cputime_nsecs.h b/include/asm-generic/cputime_nsecs.h new file mode 100644 index 000000000000..b6485cafb7bd --- /dev/null +++ b/include/asm-generic/cputime_nsecs.h | |||
| @@ -0,0 +1,104 @@ | |||
| 1 | /* | ||
| 2 | * Definitions for measuring cputime in nsecs resolution. | ||
| 3 | * | ||
| 4 | * Based on <arch/ia64/include/asm/cputime.h> | ||
| 5 | * | ||
| 6 | * Copyright (C) 2007 FUJITSU LIMITED | ||
| 7 | * Copyright (C) 2007 Hidetoshi Seto <seto.hidetoshi@jp.fujitsu.com> | ||
| 8 | * | ||
| 9 | * This program is free software; you can redistribute it and/or | ||
| 10 | * modify it under the terms of the GNU General Public License | ||
| 11 | * as published by the Free Software Foundation; either version | ||
| 12 | * 2 of the License, or (at your option) any later version. | ||
| 13 | * | ||
| 14 | */ | ||
| 15 | |||
| 16 | #ifndef _ASM_GENERIC_CPUTIME_NSECS_H | ||
| 17 | #define _ASM_GENERIC_CPUTIME_NSECS_H | ||
| 18 | |||
| 19 | typedef u64 __nocast cputime_t; | ||
| 20 | typedef u64 __nocast cputime64_t; | ||
| 21 | |||
| 22 | #define cputime_one_jiffy jiffies_to_cputime(1) | ||
| 23 | |||
| 24 | /* | ||
| 25 | * Convert cputime <-> jiffies (HZ) | ||
| 26 | */ | ||
| 27 | #define cputime_to_jiffies(__ct) \ | ||
| 28 | ((__force u64)(__ct) / (NSEC_PER_SEC / HZ)) | ||
| 29 | #define cputime_to_scaled(__ct) (__ct) | ||
| 30 | #define jiffies_to_cputime(__jif) \ | ||
| 31 | (__force cputime_t)((__jif) * (NSEC_PER_SEC / HZ)) | ||
| 32 | #define cputime64_to_jiffies64(__ct) \ | ||
| 33 | ((__force u64)(__ct) / (NSEC_PER_SEC / HZ)) | ||
| 34 | #define jiffies64_to_cputime64(__jif) \ | ||
| 35 | (__force cputime64_t)((__jif) * (NSEC_PER_SEC / HZ)) | ||
| 36 | |||
| 37 | |||
| 38 | /* | ||
| 39 | * Convert cputime <-> nanoseconds | ||
| 40 | */ | ||
| 41 | #define nsecs_to_cputime(__nsecs) ((__force u64)(__nsecs)) | ||
| 42 | |||
| 43 | |||
| 44 | /* | ||
| 45 | * Convert cputime <-> microseconds | ||
| 46 | */ | ||
| 47 | #define cputime_to_usecs(__ct) \ | ||
| 48 | ((__force u64)(__ct) / NSEC_PER_USEC) | ||
| 49 | #define usecs_to_cputime(__usecs) \ | ||
| 50 | (__force cputime_t)((__usecs) * NSEC_PER_USEC) | ||
| 51 | #define usecs_to_cputime64(__usecs) \ | ||
| 52 | (__force cputime64_t)((__usecs) * NSEC_PER_USEC) | ||
| 53 | |||
| 54 | /* | ||
| 55 | * Convert cputime <-> seconds | ||
| 56 | */ | ||
| 57 | #define cputime_to_secs(__ct) \ | ||
| 58 | ((__force u64)(__ct) / NSEC_PER_SEC) | ||
| 59 | #define secs_to_cputime(__secs) \ | ||
| 60 | (__force cputime_t)((__secs) * NSEC_PER_SEC) | ||
| 61 | |||
| 62 | /* | ||
| 63 | * Convert cputime <-> timespec (nsec) | ||
| 64 | */ | ||
| 65 | static inline cputime_t timespec_to_cputime(const struct timespec *val) | ||
| 66 | { | ||
| 67 | u64 ret = val->tv_sec * NSEC_PER_SEC + val->tv_nsec; | ||
| 68 | return (__force cputime_t) ret; | ||
| 69 | } | ||
| 70 | static inline void cputime_to_timespec(const cputime_t ct, struct timespec *val) | ||
| 71 | { | ||
| 72 | val->tv_sec = (__force u64) ct / NSEC_PER_SEC; | ||
| 73 | val->tv_nsec = (__force u64) ct % NSEC_PER_SEC; | ||
| 74 | } | ||
| 75 | |||
| 76 | /* | ||
| 77 | * Convert cputime <-> timeval (msec) | ||
| 78 | */ | ||
| 79 | static inline cputime_t timeval_to_cputime(struct timeval *val) | ||
| 80 | { | ||
| 81 | u64 ret = val->tv_sec * NSEC_PER_SEC + val->tv_usec * NSEC_PER_USEC; | ||
| 82 | return (__force cputime_t) ret; | ||
| 83 | } | ||
| 84 | static inline void cputime_to_timeval(const cputime_t ct, struct timeval *val) | ||
| 85 | { | ||
| 86 | val->tv_sec = (__force u64) ct / NSEC_PER_SEC; | ||
| 87 | val->tv_usec = ((__force u64) ct % NSEC_PER_SEC) / NSEC_PER_USEC; | ||
| 88 | } | ||
| 89 | |||
| 90 | /* | ||
| 91 | * Convert cputime <-> clock (USER_HZ) | ||
| 92 | */ | ||
| 93 | #define cputime_to_clock_t(__ct) \ | ||
| 94 | ((__force u64)(__ct) / (NSEC_PER_SEC / USER_HZ)) | ||
| 95 | #define clock_t_to_cputime(__x) \ | ||
| 96 | (__force cputime_t)((__x) * (NSEC_PER_SEC / USER_HZ)) | ||
| 97 | |||
| 98 | /* | ||
| 99 | * Convert cputime64 to clock. | ||
| 100 | */ | ||
| 101 | #define cputime64_to_clock_t(__ct) \ | ||
| 102 | cputime_to_clock_t((__force cputime_t)__ct) | ||
| 103 | |||
| 104 | #endif | ||
diff --git a/include/linux/context_tracking.h b/include/linux/context_tracking.h index e24339ccb7f0..b28d161c1091 100644 --- a/include/linux/context_tracking.h +++ b/include/linux/context_tracking.h | |||
| @@ -3,12 +3,40 @@ | |||
| 3 | 3 | ||
| 4 | #ifdef CONFIG_CONTEXT_TRACKING | 4 | #ifdef CONFIG_CONTEXT_TRACKING |
| 5 | #include <linux/sched.h> | 5 | #include <linux/sched.h> |
| 6 | #include <linux/percpu.h> | ||
| 7 | |||
| 8 | struct context_tracking { | ||
| 9 | /* | ||
| 10 | * When active is false, probes are unset in order | ||
| 11 | * to minimize overhead: TIF flags are cleared | ||
| 12 | * and calls to user_enter/exit are ignored. This | ||
| 13 | * may be further optimized using static keys. | ||
| 14 | */ | ||
| 15 | bool active; | ||
| 16 | enum { | ||
| 17 | IN_KERNEL = 0, | ||
| 18 | IN_USER, | ||
| 19 | } state; | ||
| 20 | }; | ||
| 21 | |||
| 22 | DECLARE_PER_CPU(struct context_tracking, context_tracking); | ||
| 23 | |||
| 24 | static inline bool context_tracking_in_user(void) | ||
| 25 | { | ||
| 26 | return __this_cpu_read(context_tracking.state) == IN_USER; | ||
| 27 | } | ||
| 28 | |||
| 29 | static inline bool context_tracking_active(void) | ||
| 30 | { | ||
| 31 | return __this_cpu_read(context_tracking.active); | ||
| 32 | } | ||
| 6 | 33 | ||
| 7 | extern void user_enter(void); | 34 | extern void user_enter(void); |
| 8 | extern void user_exit(void); | 35 | extern void user_exit(void); |
| 9 | extern void context_tracking_task_switch(struct task_struct *prev, | 36 | extern void context_tracking_task_switch(struct task_struct *prev, |
| 10 | struct task_struct *next); | 37 | struct task_struct *next); |
| 11 | #else | 38 | #else |
| 39 | static inline bool context_tracking_in_user(void) { return false; } | ||
| 12 | static inline void user_enter(void) { } | 40 | static inline void user_enter(void) { } |
| 13 | static inline void user_exit(void) { } | 41 | static inline void user_exit(void) { } |
| 14 | static inline void context_tracking_task_switch(struct task_struct *prev, | 42 | static inline void context_tracking_task_switch(struct task_struct *prev, |
diff --git a/include/linux/hardirq.h b/include/linux/hardirq.h index 57bfdce8fb90..29eb805ea4a6 100644 --- a/include/linux/hardirq.h +++ b/include/linux/hardirq.h | |||
| @@ -153,7 +153,7 @@ extern void rcu_nmi_exit(void); | |||
| 153 | */ | 153 | */ |
| 154 | #define __irq_enter() \ | 154 | #define __irq_enter() \ |
| 155 | do { \ | 155 | do { \ |
| 156 | vtime_account_irq_enter(current); \ | 156 | account_irq_enter_time(current); \ |
| 157 | add_preempt_count(HARDIRQ_OFFSET); \ | 157 | add_preempt_count(HARDIRQ_OFFSET); \ |
| 158 | trace_hardirq_enter(); \ | 158 | trace_hardirq_enter(); \ |
| 159 | } while (0) | 159 | } while (0) |
| @@ -169,7 +169,7 @@ extern void irq_enter(void); | |||
| 169 | #define __irq_exit() \ | 169 | #define __irq_exit() \ |
| 170 | do { \ | 170 | do { \ |
| 171 | trace_hardirq_exit(); \ | 171 | trace_hardirq_exit(); \ |
| 172 | vtime_account_irq_exit(current); \ | 172 | account_irq_exit_time(current); \ |
| 173 | sub_preempt_count(HARDIRQ_OFFSET); \ | 173 | sub_preempt_count(HARDIRQ_OFFSET); \ |
| 174 | } while (0) | 174 | } while (0) |
| 175 | 175 | ||
diff --git a/include/linux/init_task.h b/include/linux/init_task.h index 6d087c5f57f7..5cd0f0949927 100644 --- a/include/linux/init_task.h +++ b/include/linux/init_task.h | |||
| @@ -10,7 +10,9 @@ | |||
| 10 | #include <linux/pid_namespace.h> | 10 | #include <linux/pid_namespace.h> |
| 11 | #include <linux/user_namespace.h> | 11 | #include <linux/user_namespace.h> |
| 12 | #include <linux/securebits.h> | 12 | #include <linux/securebits.h> |
| 13 | #include <linux/seqlock.h> | ||
| 13 | #include <net/net_namespace.h> | 14 | #include <net/net_namespace.h> |
| 15 | #include <linux/sched/rt.h> | ||
| 14 | 16 | ||
| 15 | #ifdef CONFIG_SMP | 17 | #ifdef CONFIG_SMP |
| 16 | # define INIT_PUSHABLE_TASKS(tsk) \ | 18 | # define INIT_PUSHABLE_TASKS(tsk) \ |
| @@ -141,6 +143,15 @@ extern struct task_group root_task_group; | |||
| 141 | # define INIT_PERF_EVENTS(tsk) | 143 | # define INIT_PERF_EVENTS(tsk) |
| 142 | #endif | 144 | #endif |
| 143 | 145 | ||
| 146 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 147 | # define INIT_VTIME(tsk) \ | ||
| 148 | .vtime_seqlock = __SEQLOCK_UNLOCKED(tsk.vtime_seqlock), \ | ||
| 149 | .vtime_snap = 0, \ | ||
| 150 | .vtime_snap_whence = VTIME_SYS, | ||
| 151 | #else | ||
| 152 | # define INIT_VTIME(tsk) | ||
| 153 | #endif | ||
| 154 | |||
| 144 | #define INIT_TASK_COMM "swapper" | 155 | #define INIT_TASK_COMM "swapper" |
| 145 | 156 | ||
| 146 | /* | 157 | /* |
| @@ -210,6 +221,7 @@ extern struct task_group root_task_group; | |||
| 210 | INIT_TRACE_RECURSION \ | 221 | INIT_TRACE_RECURSION \ |
| 211 | INIT_TASK_RCU_PREEMPT(tsk) \ | 222 | INIT_TASK_RCU_PREEMPT(tsk) \ |
| 212 | INIT_CPUSET_SEQ \ | 223 | INIT_CPUSET_SEQ \ |
| 224 | INIT_VTIME(tsk) \ | ||
| 213 | } | 225 | } |
| 214 | 226 | ||
| 215 | 227 | ||
diff --git a/include/linux/kernel_stat.h b/include/linux/kernel_stat.h index 66b70780e910..ed5f6ed6eb77 100644 --- a/include/linux/kernel_stat.h +++ b/include/linux/kernel_stat.h | |||
| @@ -127,7 +127,7 @@ extern void account_system_time(struct task_struct *, int, cputime_t, cputime_t) | |||
| 127 | extern void account_steal_time(cputime_t); | 127 | extern void account_steal_time(cputime_t); |
| 128 | extern void account_idle_time(cputime_t); | 128 | extern void account_idle_time(cputime_t); |
| 129 | 129 | ||
| 130 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 130 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 131 | static inline void account_process_tick(struct task_struct *tsk, int user) | 131 | static inline void account_process_tick(struct task_struct *tsk, int user) |
| 132 | { | 132 | { |
| 133 | vtime_account_user(tsk); | 133 | vtime_account_user(tsk); |
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index 2c497ab0d03d..b7996a768eb2 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h | |||
| @@ -22,6 +22,7 @@ | |||
| 22 | #include <linux/rcupdate.h> | 22 | #include <linux/rcupdate.h> |
| 23 | #include <linux/ratelimit.h> | 23 | #include <linux/ratelimit.h> |
| 24 | #include <linux/err.h> | 24 | #include <linux/err.h> |
| 25 | #include <linux/irqflags.h> | ||
| 25 | #include <asm/signal.h> | 26 | #include <asm/signal.h> |
| 26 | 27 | ||
| 27 | #include <linux/kvm.h> | 28 | #include <linux/kvm.h> |
| @@ -740,15 +741,52 @@ static inline int kvm_deassign_device(struct kvm *kvm, | |||
| 740 | } | 741 | } |
| 741 | #endif /* CONFIG_IOMMU_API */ | 742 | #endif /* CONFIG_IOMMU_API */ |
| 742 | 743 | ||
| 743 | static inline void kvm_guest_enter(void) | 744 | static inline void __guest_enter(void) |
| 744 | { | 745 | { |
| 745 | BUG_ON(preemptible()); | ||
| 746 | /* | 746 | /* |
| 747 | * This is running in ioctl context so we can avoid | 747 | * This is running in ioctl context so we can avoid |
| 748 | * the call to vtime_account() with its unnecessary idle check. | 748 | * the call to vtime_account() with its unnecessary idle check. |
| 749 | */ | 749 | */ |
| 750 | vtime_account_system_irqsafe(current); | 750 | vtime_account_system(current); |
| 751 | current->flags |= PF_VCPU; | 751 | current->flags |= PF_VCPU; |
| 752 | } | ||
| 753 | |||
| 754 | static inline void __guest_exit(void) | ||
| 755 | { | ||
| 756 | /* | ||
| 757 | * This is running in ioctl context so we can avoid | ||
| 758 | * the call to vtime_account() with its unnecessary idle check. | ||
| 759 | */ | ||
| 760 | vtime_account_system(current); | ||
| 761 | current->flags &= ~PF_VCPU; | ||
| 762 | } | ||
| 763 | |||
| 764 | #ifdef CONFIG_CONTEXT_TRACKING | ||
| 765 | extern void guest_enter(void); | ||
| 766 | extern void guest_exit(void); | ||
| 767 | |||
| 768 | #else /* !CONFIG_CONTEXT_TRACKING */ | ||
| 769 | static inline void guest_enter(void) | ||
| 770 | { | ||
| 771 | __guest_enter(); | ||
| 772 | } | ||
| 773 | |||
| 774 | static inline void guest_exit(void) | ||
| 775 | { | ||
| 776 | __guest_exit(); | ||
| 777 | } | ||
| 778 | #endif /* !CONFIG_CONTEXT_TRACKING */ | ||
| 779 | |||
| 780 | static inline void kvm_guest_enter(void) | ||
| 781 | { | ||
| 782 | unsigned long flags; | ||
| 783 | |||
| 784 | BUG_ON(preemptible()); | ||
| 785 | |||
| 786 | local_irq_save(flags); | ||
| 787 | guest_enter(); | ||
| 788 | local_irq_restore(flags); | ||
| 789 | |||
| 752 | /* KVM does not hold any references to rcu protected data when it | 790 | /* KVM does not hold any references to rcu protected data when it |
| 753 | * switches CPU into a guest mode. In fact switching to a guest mode | 791 | * switches CPU into a guest mode. In fact switching to a guest mode |
| 754 | * is very similar to exiting to userspase from rcu point of view. In | 792 | * is very similar to exiting to userspase from rcu point of view. In |
| @@ -761,12 +799,11 @@ static inline void kvm_guest_enter(void) | |||
| 761 | 799 | ||
| 762 | static inline void kvm_guest_exit(void) | 800 | static inline void kvm_guest_exit(void) |
| 763 | { | 801 | { |
| 764 | /* | 802 | unsigned long flags; |
| 765 | * This is running in ioctl context so we can avoid | 803 | |
| 766 | * the call to vtime_account() with its unnecessary idle check. | 804 | local_irq_save(flags); |
| 767 | */ | 805 | guest_exit(); |
| 768 | vtime_account_system_irqsafe(current); | 806 | local_irq_restore(flags); |
| 769 | current->flags &= ~PF_VCPU; | ||
| 770 | } | 807 | } |
| 771 | 808 | ||
| 772 | /* | 809 | /* |
diff --git a/include/linux/sched.h b/include/linux/sched.h index d2112477ff5e..33cc42130371 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
| @@ -304,19 +304,6 @@ static inline void lockup_detector_init(void) | |||
| 304 | } | 304 | } |
| 305 | #endif | 305 | #endif |
| 306 | 306 | ||
| 307 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
| 308 | extern unsigned int sysctl_hung_task_panic; | ||
| 309 | extern unsigned long sysctl_hung_task_check_count; | ||
| 310 | extern unsigned long sysctl_hung_task_timeout_secs; | ||
| 311 | extern unsigned long sysctl_hung_task_warnings; | ||
| 312 | extern int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, | ||
| 313 | void __user *buffer, | ||
| 314 | size_t *lenp, loff_t *ppos); | ||
| 315 | #else | ||
| 316 | /* Avoid need for ifdefs elsewhere in the code */ | ||
| 317 | enum { sysctl_hung_task_timeout_secs = 0 }; | ||
| 318 | #endif | ||
| 319 | |||
| 320 | /* Attach to any functions which should be ignored in wchan output. */ | 307 | /* Attach to any functions which should be ignored in wchan output. */ |
| 321 | #define __sched __attribute__((__section__(".sched.text"))) | 308 | #define __sched __attribute__((__section__(".sched.text"))) |
| 322 | 309 | ||
| @@ -338,23 +325,6 @@ extern int mutex_spin_on_owner(struct mutex *lock, struct task_struct *owner); | |||
| 338 | struct nsproxy; | 325 | struct nsproxy; |
| 339 | struct user_namespace; | 326 | struct user_namespace; |
| 340 | 327 | ||
| 341 | /* | ||
| 342 | * Default maximum number of active map areas, this limits the number of vmas | ||
| 343 | * per mm struct. Users can overwrite this number by sysctl but there is a | ||
| 344 | * problem. | ||
| 345 | * | ||
| 346 | * When a program's coredump is generated as ELF format, a section is created | ||
| 347 | * per a vma. In ELF, the number of sections is represented in unsigned short. | ||
| 348 | * This means the number of sections should be smaller than 65535 at coredump. | ||
| 349 | * Because the kernel adds some informative sections to a image of program at | ||
| 350 | * generating coredump, we need some margin. The number of extra sections is | ||
| 351 | * 1-3 now and depends on arch. We use "5" as safe margin, here. | ||
| 352 | */ | ||
| 353 | #define MAPCOUNT_ELF_CORE_MARGIN (5) | ||
| 354 | #define DEFAULT_MAX_MAP_COUNT (USHRT_MAX - MAPCOUNT_ELF_CORE_MARGIN) | ||
| 355 | |||
| 356 | extern int sysctl_max_map_count; | ||
| 357 | |||
| 358 | #include <linux/aio.h> | 328 | #include <linux/aio.h> |
| 359 | 329 | ||
| 360 | #ifdef CONFIG_MMU | 330 | #ifdef CONFIG_MMU |
| @@ -1194,6 +1164,7 @@ struct sched_entity { | |||
| 1194 | /* rq "owned" by this entity/group: */ | 1164 | /* rq "owned" by this entity/group: */ |
| 1195 | struct cfs_rq *my_q; | 1165 | struct cfs_rq *my_q; |
| 1196 | #endif | 1166 | #endif |
| 1167 | |||
| 1197 | /* | 1168 | /* |
| 1198 | * Load-tracking only depends on SMP, FAIR_GROUP_SCHED dependency below may be | 1169 | * Load-tracking only depends on SMP, FAIR_GROUP_SCHED dependency below may be |
| 1199 | * removed when useful for applications beyond shares distribution (e.g. | 1170 | * removed when useful for applications beyond shares distribution (e.g. |
| @@ -1208,6 +1179,7 @@ struct sched_entity { | |||
| 1208 | struct sched_rt_entity { | 1179 | struct sched_rt_entity { |
| 1209 | struct list_head run_list; | 1180 | struct list_head run_list; |
| 1210 | unsigned long timeout; | 1181 | unsigned long timeout; |
| 1182 | unsigned long watchdog_stamp; | ||
| 1211 | unsigned int time_slice; | 1183 | unsigned int time_slice; |
| 1212 | 1184 | ||
| 1213 | struct sched_rt_entity *back; | 1185 | struct sched_rt_entity *back; |
| @@ -1220,11 +1192,6 @@ struct sched_rt_entity { | |||
| 1220 | #endif | 1192 | #endif |
| 1221 | }; | 1193 | }; |
| 1222 | 1194 | ||
| 1223 | /* | ||
| 1224 | * default timeslice is 100 msecs (used only for SCHED_RR tasks). | ||
| 1225 | * Timeslices get refilled after they expire. | ||
| 1226 | */ | ||
| 1227 | #define RR_TIMESLICE (100 * HZ / 1000) | ||
| 1228 | 1195 | ||
| 1229 | struct rcu_node; | 1196 | struct rcu_node; |
| 1230 | 1197 | ||
| @@ -1368,6 +1335,15 @@ struct task_struct { | |||
| 1368 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 1335 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING |
| 1369 | struct cputime prev_cputime; | 1336 | struct cputime prev_cputime; |
| 1370 | #endif | 1337 | #endif |
| 1338 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 1339 | seqlock_t vtime_seqlock; | ||
| 1340 | unsigned long long vtime_snap; | ||
| 1341 | enum { | ||
| 1342 | VTIME_SLEEPING = 0, | ||
| 1343 | VTIME_USER, | ||
| 1344 | VTIME_SYS, | ||
| 1345 | } vtime_snap_whence; | ||
| 1346 | #endif | ||
| 1371 | unsigned long nvcsw, nivcsw; /* context switch counts */ | 1347 | unsigned long nvcsw, nivcsw; /* context switch counts */ |
| 1372 | struct timespec start_time; /* monotonic time */ | 1348 | struct timespec start_time; /* monotonic time */ |
| 1373 | struct timespec real_start_time; /* boot based time */ | 1349 | struct timespec real_start_time; /* boot based time */ |
| @@ -1622,37 +1598,6 @@ static inline void set_numabalancing_state(bool enabled) | |||
| 1622 | } | 1598 | } |
| 1623 | #endif | 1599 | #endif |
| 1624 | 1600 | ||
| 1625 | /* | ||
| 1626 | * Priority of a process goes from 0..MAX_PRIO-1, valid RT | ||
| 1627 | * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH | ||
| 1628 | * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority | ||
| 1629 | * values are inverted: lower p->prio value means higher priority. | ||
| 1630 | * | ||
| 1631 | * The MAX_USER_RT_PRIO value allows the actual maximum | ||
| 1632 | * RT priority to be separate from the value exported to | ||
| 1633 | * user-space. This allows kernel threads to set their | ||
| 1634 | * priority to a value higher than any user task. Note: | ||
| 1635 | * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. | ||
| 1636 | */ | ||
| 1637 | |||
| 1638 | #define MAX_USER_RT_PRIO 100 | ||
| 1639 | #define MAX_RT_PRIO MAX_USER_RT_PRIO | ||
| 1640 | |||
| 1641 | #define MAX_PRIO (MAX_RT_PRIO + 40) | ||
| 1642 | #define DEFAULT_PRIO (MAX_RT_PRIO + 20) | ||
| 1643 | |||
| 1644 | static inline int rt_prio(int prio) | ||
| 1645 | { | ||
| 1646 | if (unlikely(prio < MAX_RT_PRIO)) | ||
| 1647 | return 1; | ||
| 1648 | return 0; | ||
| 1649 | } | ||
| 1650 | |||
| 1651 | static inline int rt_task(struct task_struct *p) | ||
| 1652 | { | ||
| 1653 | return rt_prio(p->prio); | ||
| 1654 | } | ||
| 1655 | |||
| 1656 | static inline struct pid *task_pid(struct task_struct *task) | 1601 | static inline struct pid *task_pid(struct task_struct *task) |
| 1657 | { | 1602 | { |
| 1658 | return task->pids[PIDTYPE_PID].pid; | 1603 | return task->pids[PIDTYPE_PID].pid; |
| @@ -1792,6 +1737,37 @@ static inline void put_task_struct(struct task_struct *t) | |||
| 1792 | __put_task_struct(t); | 1737 | __put_task_struct(t); |
| 1793 | } | 1738 | } |
| 1794 | 1739 | ||
| 1740 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 1741 | extern void task_cputime(struct task_struct *t, | ||
| 1742 | cputime_t *utime, cputime_t *stime); | ||
| 1743 | extern void task_cputime_scaled(struct task_struct *t, | ||
| 1744 | cputime_t *utimescaled, cputime_t *stimescaled); | ||
| 1745 | extern cputime_t task_gtime(struct task_struct *t); | ||
| 1746 | #else | ||
| 1747 | static inline void task_cputime(struct task_struct *t, | ||
| 1748 | cputime_t *utime, cputime_t *stime) | ||
| 1749 | { | ||
| 1750 | if (utime) | ||
| 1751 | *utime = t->utime; | ||
| 1752 | if (stime) | ||
| 1753 | *stime = t->stime; | ||
| 1754 | } | ||
| 1755 | |||
| 1756 | static inline void task_cputime_scaled(struct task_struct *t, | ||
| 1757 | cputime_t *utimescaled, | ||
| 1758 | cputime_t *stimescaled) | ||
| 1759 | { | ||
| 1760 | if (utimescaled) | ||
| 1761 | *utimescaled = t->utimescaled; | ||
| 1762 | if (stimescaled) | ||
| 1763 | *stimescaled = t->stimescaled; | ||
| 1764 | } | ||
| 1765 | |||
| 1766 | static inline cputime_t task_gtime(struct task_struct *t) | ||
| 1767 | { | ||
| 1768 | return t->gtime; | ||
| 1769 | } | ||
| 1770 | #endif | ||
| 1795 | extern void task_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st); | 1771 | extern void task_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st); |
| 1796 | extern void thread_group_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st); | 1772 | extern void thread_group_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st); |
| 1797 | 1773 | ||
| @@ -2033,58 +2009,7 @@ extern void wake_up_idle_cpu(int cpu); | |||
| 2033 | static inline void wake_up_idle_cpu(int cpu) { } | 2009 | static inline void wake_up_idle_cpu(int cpu) { } |
| 2034 | #endif | 2010 | #endif |
| 2035 | 2011 | ||
| 2036 | extern unsigned int sysctl_sched_latency; | ||
| 2037 | extern unsigned int sysctl_sched_min_granularity; | ||
| 2038 | extern unsigned int sysctl_sched_wakeup_granularity; | ||
| 2039 | extern unsigned int sysctl_sched_child_runs_first; | ||
| 2040 | |||
| 2041 | enum sched_tunable_scaling { | ||
| 2042 | SCHED_TUNABLESCALING_NONE, | ||
| 2043 | SCHED_TUNABLESCALING_LOG, | ||
| 2044 | SCHED_TUNABLESCALING_LINEAR, | ||
| 2045 | SCHED_TUNABLESCALING_END, | ||
| 2046 | }; | ||
| 2047 | extern enum sched_tunable_scaling sysctl_sched_tunable_scaling; | ||
| 2048 | |||
| 2049 | extern unsigned int sysctl_numa_balancing_scan_delay; | ||
| 2050 | extern unsigned int sysctl_numa_balancing_scan_period_min; | ||
| 2051 | extern unsigned int sysctl_numa_balancing_scan_period_max; | ||
| 2052 | extern unsigned int sysctl_numa_balancing_scan_period_reset; | ||
| 2053 | extern unsigned int sysctl_numa_balancing_scan_size; | ||
| 2054 | extern unsigned int sysctl_numa_balancing_settle_count; | ||
| 2055 | |||
| 2056 | #ifdef CONFIG_SCHED_DEBUG | ||
| 2057 | extern unsigned int sysctl_sched_migration_cost; | ||
| 2058 | extern unsigned int sysctl_sched_nr_migrate; | ||
| 2059 | extern unsigned int sysctl_sched_time_avg; | ||
| 2060 | extern unsigned int sysctl_timer_migration; | ||
| 2061 | extern unsigned int sysctl_sched_shares_window; | ||
| 2062 | |||
| 2063 | int sched_proc_update_handler(struct ctl_table *table, int write, | ||
| 2064 | void __user *buffer, size_t *length, | ||
| 2065 | loff_t *ppos); | ||
| 2066 | #endif | ||
| 2067 | #ifdef CONFIG_SCHED_DEBUG | ||
| 2068 | static inline unsigned int get_sysctl_timer_migration(void) | ||
| 2069 | { | ||
| 2070 | return sysctl_timer_migration; | ||
| 2071 | } | ||
| 2072 | #else | ||
| 2073 | static inline unsigned int get_sysctl_timer_migration(void) | ||
| 2074 | { | ||
| 2075 | return 1; | ||
| 2076 | } | ||
| 2077 | #endif | ||
| 2078 | extern unsigned int sysctl_sched_rt_period; | ||
| 2079 | extern int sysctl_sched_rt_runtime; | ||
| 2080 | |||
| 2081 | int sched_rt_handler(struct ctl_table *table, int write, | ||
| 2082 | void __user *buffer, size_t *lenp, | ||
| 2083 | loff_t *ppos); | ||
| 2084 | |||
| 2085 | #ifdef CONFIG_SCHED_AUTOGROUP | 2012 | #ifdef CONFIG_SCHED_AUTOGROUP |
| 2086 | extern unsigned int sysctl_sched_autogroup_enabled; | ||
| 2087 | |||
| 2088 | extern void sched_autogroup_create_attach(struct task_struct *p); | 2013 | extern void sched_autogroup_create_attach(struct task_struct *p); |
| 2089 | extern void sched_autogroup_detach(struct task_struct *p); | 2014 | extern void sched_autogroup_detach(struct task_struct *p); |
| 2090 | extern void sched_autogroup_fork(struct signal_struct *sig); | 2015 | extern void sched_autogroup_fork(struct signal_struct *sig); |
| @@ -2100,30 +2025,6 @@ static inline void sched_autogroup_fork(struct signal_struct *sig) { } | |||
| 2100 | static inline void sched_autogroup_exit(struct signal_struct *sig) { } | 2025 | static inline void sched_autogroup_exit(struct signal_struct *sig) { } |
| 2101 | #endif | 2026 | #endif |
| 2102 | 2027 | ||
| 2103 | #ifdef CONFIG_CFS_BANDWIDTH | ||
| 2104 | extern unsigned int sysctl_sched_cfs_bandwidth_slice; | ||
| 2105 | #endif | ||
| 2106 | |||
| 2107 | #ifdef CONFIG_RT_MUTEXES | ||
| 2108 | extern int rt_mutex_getprio(struct task_struct *p); | ||
| 2109 | extern void rt_mutex_setprio(struct task_struct *p, int prio); | ||
| 2110 | extern void rt_mutex_adjust_pi(struct task_struct *p); | ||
| 2111 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 2112 | { | ||
| 2113 | return tsk->pi_blocked_on != NULL; | ||
| 2114 | } | ||
| 2115 | #else | ||
| 2116 | static inline int rt_mutex_getprio(struct task_struct *p) | ||
| 2117 | { | ||
| 2118 | return p->normal_prio; | ||
| 2119 | } | ||
| 2120 | # define rt_mutex_adjust_pi(p) do { } while (0) | ||
| 2121 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 2122 | { | ||
| 2123 | return false; | ||
| 2124 | } | ||
| 2125 | #endif | ||
| 2126 | |||
| 2127 | extern bool yield_to(struct task_struct *p, bool preempt); | 2028 | extern bool yield_to(struct task_struct *p, bool preempt); |
| 2128 | extern void set_user_nice(struct task_struct *p, long nice); | 2029 | extern void set_user_nice(struct task_struct *p, long nice); |
| 2129 | extern int task_prio(const struct task_struct *p); | 2030 | extern int task_prio(const struct task_struct *p); |
| @@ -2753,8 +2654,6 @@ static inline void set_task_cpu(struct task_struct *p, unsigned int cpu) | |||
| 2753 | extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); | 2654 | extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); |
| 2754 | extern long sched_getaffinity(pid_t pid, struct cpumask *mask); | 2655 | extern long sched_getaffinity(pid_t pid, struct cpumask *mask); |
| 2755 | 2656 | ||
| 2756 | extern void normalize_rt_tasks(void); | ||
| 2757 | |||
| 2758 | #ifdef CONFIG_CGROUP_SCHED | 2657 | #ifdef CONFIG_CGROUP_SCHED |
| 2759 | 2658 | ||
| 2760 | extern struct task_group root_task_group; | 2659 | extern struct task_group root_task_group; |
diff --git a/include/linux/sched/rt.h b/include/linux/sched/rt.h new file mode 100644 index 000000000000..94e19ea28fc3 --- /dev/null +++ b/include/linux/sched/rt.h | |||
| @@ -0,0 +1,58 @@ | |||
| 1 | #ifndef _SCHED_RT_H | ||
| 2 | #define _SCHED_RT_H | ||
| 3 | |||
| 4 | /* | ||
| 5 | * Priority of a process goes from 0..MAX_PRIO-1, valid RT | ||
| 6 | * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH | ||
| 7 | * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority | ||
| 8 | * values are inverted: lower p->prio value means higher priority. | ||
| 9 | * | ||
| 10 | * The MAX_USER_RT_PRIO value allows the actual maximum | ||
| 11 | * RT priority to be separate from the value exported to | ||
| 12 | * user-space. This allows kernel threads to set their | ||
| 13 | * priority to a value higher than any user task. Note: | ||
| 14 | * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. | ||
| 15 | */ | ||
| 16 | |||
| 17 | #define MAX_USER_RT_PRIO 100 | ||
| 18 | #define MAX_RT_PRIO MAX_USER_RT_PRIO | ||
| 19 | |||
| 20 | #define MAX_PRIO (MAX_RT_PRIO + 40) | ||
| 21 | #define DEFAULT_PRIO (MAX_RT_PRIO + 20) | ||
| 22 | |||
| 23 | static inline int rt_prio(int prio) | ||
| 24 | { | ||
| 25 | if (unlikely(prio < MAX_RT_PRIO)) | ||
| 26 | return 1; | ||
| 27 | return 0; | ||
| 28 | } | ||
| 29 | |||
| 30 | static inline int rt_task(struct task_struct *p) | ||
| 31 | { | ||
| 32 | return rt_prio(p->prio); | ||
| 33 | } | ||
| 34 | |||
| 35 | #ifdef CONFIG_RT_MUTEXES | ||
| 36 | extern int rt_mutex_getprio(struct task_struct *p); | ||
| 37 | extern void rt_mutex_setprio(struct task_struct *p, int prio); | ||
| 38 | extern void rt_mutex_adjust_pi(struct task_struct *p); | ||
| 39 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 40 | { | ||
| 41 | return tsk->pi_blocked_on != NULL; | ||
| 42 | } | ||
| 43 | #else | ||
| 44 | static inline int rt_mutex_getprio(struct task_struct *p) | ||
| 45 | { | ||
| 46 | return p->normal_prio; | ||
| 47 | } | ||
| 48 | # define rt_mutex_adjust_pi(p) do { } while (0) | ||
| 49 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 50 | { | ||
| 51 | return false; | ||
| 52 | } | ||
| 53 | #endif | ||
| 54 | |||
| 55 | extern void normalize_rt_tasks(void); | ||
| 56 | |||
| 57 | |||
| 58 | #endif /* _SCHED_RT_H */ | ||
diff --git a/include/linux/sched/sysctl.h b/include/linux/sched/sysctl.h new file mode 100644 index 000000000000..d2bb0ae979d0 --- /dev/null +++ b/include/linux/sched/sysctl.h | |||
| @@ -0,0 +1,110 @@ | |||
| 1 | #ifndef _SCHED_SYSCTL_H | ||
| 2 | #define _SCHED_SYSCTL_H | ||
| 3 | |||
| 4 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
| 5 | extern unsigned int sysctl_hung_task_panic; | ||
| 6 | extern unsigned long sysctl_hung_task_check_count; | ||
| 7 | extern unsigned long sysctl_hung_task_timeout_secs; | ||
| 8 | extern unsigned long sysctl_hung_task_warnings; | ||
| 9 | extern int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, | ||
| 10 | void __user *buffer, | ||
| 11 | size_t *lenp, loff_t *ppos); | ||
| 12 | #else | ||
| 13 | /* Avoid need for ifdefs elsewhere in the code */ | ||
| 14 | enum { sysctl_hung_task_timeout_secs = 0 }; | ||
| 15 | #endif | ||
| 16 | |||
| 17 | /* | ||
| 18 | * Default maximum number of active map areas, this limits the number of vmas | ||
| 19 | * per mm struct. Users can overwrite this number by sysctl but there is a | ||
| 20 | * problem. | ||
| 21 | * | ||
| 22 | * When a program's coredump is generated as ELF format, a section is created | ||
| 23 | * per a vma. In ELF, the number of sections is represented in unsigned short. | ||
| 24 | * This means the number of sections should be smaller than 65535 at coredump. | ||
| 25 | * Because the kernel adds some informative sections to a image of program at | ||
| 26 | * generating coredump, we need some margin. The number of extra sections is | ||
| 27 | * 1-3 now and depends on arch. We use "5" as safe margin, here. | ||
| 28 | */ | ||
| 29 | #define MAPCOUNT_ELF_CORE_MARGIN (5) | ||
| 30 | #define DEFAULT_MAX_MAP_COUNT (USHRT_MAX - MAPCOUNT_ELF_CORE_MARGIN) | ||
| 31 | |||
| 32 | extern int sysctl_max_map_count; | ||
| 33 | |||
| 34 | extern unsigned int sysctl_sched_latency; | ||
| 35 | extern unsigned int sysctl_sched_min_granularity; | ||
| 36 | extern unsigned int sysctl_sched_wakeup_granularity; | ||
| 37 | extern unsigned int sysctl_sched_child_runs_first; | ||
| 38 | |||
| 39 | enum sched_tunable_scaling { | ||
| 40 | SCHED_TUNABLESCALING_NONE, | ||
| 41 | SCHED_TUNABLESCALING_LOG, | ||
| 42 | SCHED_TUNABLESCALING_LINEAR, | ||
| 43 | SCHED_TUNABLESCALING_END, | ||
| 44 | }; | ||
| 45 | extern enum sched_tunable_scaling sysctl_sched_tunable_scaling; | ||
| 46 | |||
| 47 | extern unsigned int sysctl_numa_balancing_scan_delay; | ||
| 48 | extern unsigned int sysctl_numa_balancing_scan_period_min; | ||
| 49 | extern unsigned int sysctl_numa_balancing_scan_period_max; | ||
| 50 | extern unsigned int sysctl_numa_balancing_scan_period_reset; | ||
| 51 | extern unsigned int sysctl_numa_balancing_scan_size; | ||
| 52 | extern unsigned int sysctl_numa_balancing_settle_count; | ||
| 53 | |||
| 54 | #ifdef CONFIG_SCHED_DEBUG | ||
| 55 | extern unsigned int sysctl_sched_migration_cost; | ||
| 56 | extern unsigned int sysctl_sched_nr_migrate; | ||
| 57 | extern unsigned int sysctl_sched_time_avg; | ||
| 58 | extern unsigned int sysctl_timer_migration; | ||
| 59 | extern unsigned int sysctl_sched_shares_window; | ||
| 60 | |||
| 61 | int sched_proc_update_handler(struct ctl_table *table, int write, | ||
| 62 | void __user *buffer, size_t *length, | ||
| 63 | loff_t *ppos); | ||
| 64 | #endif | ||
| 65 | #ifdef CONFIG_SCHED_DEBUG | ||
| 66 | static inline unsigned int get_sysctl_timer_migration(void) | ||
| 67 | { | ||
| 68 | return sysctl_timer_migration; | ||
| 69 | } | ||
| 70 | #else | ||
| 71 | static inline unsigned int get_sysctl_timer_migration(void) | ||
| 72 | { | ||
| 73 | return 1; | ||
| 74 | } | ||
| 75 | #endif | ||
| 76 | |||
| 77 | /* | ||
| 78 | * control realtime throttling: | ||
| 79 | * | ||
| 80 | * /proc/sys/kernel/sched_rt_period_us | ||
| 81 | * /proc/sys/kernel/sched_rt_runtime_us | ||
| 82 | */ | ||
| 83 | extern unsigned int sysctl_sched_rt_period; | ||
| 84 | extern int sysctl_sched_rt_runtime; | ||
| 85 | |||
| 86 | #ifdef CONFIG_CFS_BANDWIDTH | ||
| 87 | extern unsigned int sysctl_sched_cfs_bandwidth_slice; | ||
| 88 | #endif | ||
| 89 | |||
| 90 | #ifdef CONFIG_SCHED_AUTOGROUP | ||
| 91 | extern unsigned int sysctl_sched_autogroup_enabled; | ||
| 92 | #endif | ||
| 93 | |||
| 94 | /* | ||
| 95 | * default timeslice is 100 msecs (used only for SCHED_RR tasks). | ||
| 96 | * Timeslices get refilled after they expire. | ||
| 97 | */ | ||
| 98 | #define RR_TIMESLICE (100 * HZ / 1000) | ||
| 99 | |||
| 100 | extern int sched_rr_timeslice; | ||
| 101 | |||
| 102 | extern int sched_rr_handler(struct ctl_table *table, int write, | ||
| 103 | void __user *buffer, size_t *lenp, | ||
| 104 | loff_t *ppos); | ||
| 105 | |||
| 106 | extern int sched_rt_handler(struct ctl_table *table, int write, | ||
| 107 | void __user *buffer, size_t *lenp, | ||
| 108 | loff_t *ppos); | ||
| 109 | |||
| 110 | #endif /* _SCHED_SYSCTL_H */ | ||
diff --git a/include/linux/tsacct_kern.h b/include/linux/tsacct_kern.h index 44893e5ec8f7..3251965bf4cc 100644 --- a/include/linux/tsacct_kern.h +++ b/include/linux/tsacct_kern.h | |||
| @@ -23,12 +23,15 @@ static inline void bacct_add_tsk(struct user_namespace *user_ns, | |||
| 23 | #ifdef CONFIG_TASK_XACCT | 23 | #ifdef CONFIG_TASK_XACCT |
| 24 | extern void xacct_add_tsk(struct taskstats *stats, struct task_struct *p); | 24 | extern void xacct_add_tsk(struct taskstats *stats, struct task_struct *p); |
| 25 | extern void acct_update_integrals(struct task_struct *tsk); | 25 | extern void acct_update_integrals(struct task_struct *tsk); |
| 26 | extern void acct_account_cputime(struct task_struct *tsk); | ||
| 26 | extern void acct_clear_integrals(struct task_struct *tsk); | 27 | extern void acct_clear_integrals(struct task_struct *tsk); |
| 27 | #else | 28 | #else |
| 28 | static inline void xacct_add_tsk(struct taskstats *stats, struct task_struct *p) | 29 | static inline void xacct_add_tsk(struct taskstats *stats, struct task_struct *p) |
| 29 | {} | 30 | {} |
| 30 | static inline void acct_update_integrals(struct task_struct *tsk) | 31 | static inline void acct_update_integrals(struct task_struct *tsk) |
| 31 | {} | 32 | {} |
| 33 | static inline void acct_account_cputime(struct task_struct *tsk) | ||
| 34 | {} | ||
| 32 | static inline void acct_clear_integrals(struct task_struct *tsk) | 35 | static inline void acct_clear_integrals(struct task_struct *tsk) |
| 33 | {} | 36 | {} |
| 34 | #endif /* CONFIG_TASK_XACCT */ | 37 | #endif /* CONFIG_TASK_XACCT */ |
diff --git a/include/linux/vtime.h b/include/linux/vtime.h index ae30ab58431a..71a5782d8c59 100644 --- a/include/linux/vtime.h +++ b/include/linux/vtime.h | |||
| @@ -6,15 +6,46 @@ struct task_struct; | |||
| 6 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING | 6 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING |
| 7 | extern void vtime_task_switch(struct task_struct *prev); | 7 | extern void vtime_task_switch(struct task_struct *prev); |
| 8 | extern void vtime_account_system(struct task_struct *tsk); | 8 | extern void vtime_account_system(struct task_struct *tsk); |
| 9 | extern void vtime_account_system_irqsafe(struct task_struct *tsk); | ||
| 10 | extern void vtime_account_idle(struct task_struct *tsk); | 9 | extern void vtime_account_idle(struct task_struct *tsk); |
| 11 | extern void vtime_account_user(struct task_struct *tsk); | 10 | extern void vtime_account_user(struct task_struct *tsk); |
| 12 | extern void vtime_account(struct task_struct *tsk); | 11 | extern void vtime_account_irq_enter(struct task_struct *tsk); |
| 13 | #else | 12 | |
| 13 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE | ||
| 14 | static inline bool vtime_accounting_enabled(void) { return true; } | ||
| 15 | #endif | ||
| 16 | |||
| 17 | #else /* !CONFIG_VIRT_CPU_ACCOUNTING */ | ||
| 18 | |||
| 14 | static inline void vtime_task_switch(struct task_struct *prev) { } | 19 | static inline void vtime_task_switch(struct task_struct *prev) { } |
| 15 | static inline void vtime_account_system(struct task_struct *tsk) { } | 20 | static inline void vtime_account_system(struct task_struct *tsk) { } |
| 16 | static inline void vtime_account_system_irqsafe(struct task_struct *tsk) { } | 21 | static inline void vtime_account_user(struct task_struct *tsk) { } |
| 17 | static inline void vtime_account(struct task_struct *tsk) { } | 22 | static inline void vtime_account_irq_enter(struct task_struct *tsk) { } |
| 23 | static inline bool vtime_accounting_enabled(void) { return false; } | ||
| 24 | #endif | ||
| 25 | |||
| 26 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 27 | extern void arch_vtime_task_switch(struct task_struct *tsk); | ||
| 28 | extern void vtime_account_irq_exit(struct task_struct *tsk); | ||
| 29 | extern bool vtime_accounting_enabled(void); | ||
| 30 | extern void vtime_user_enter(struct task_struct *tsk); | ||
| 31 | static inline void vtime_user_exit(struct task_struct *tsk) | ||
| 32 | { | ||
| 33 | vtime_account_user(tsk); | ||
| 34 | } | ||
| 35 | extern void vtime_guest_enter(struct task_struct *tsk); | ||
| 36 | extern void vtime_guest_exit(struct task_struct *tsk); | ||
| 37 | extern void vtime_init_idle(struct task_struct *tsk); | ||
| 38 | #else | ||
| 39 | static inline void vtime_account_irq_exit(struct task_struct *tsk) | ||
| 40 | { | ||
| 41 | /* On hard|softirq exit we always account to hard|softirq cputime */ | ||
| 42 | vtime_account_system(tsk); | ||
| 43 | } | ||
| 44 | static inline void vtime_user_enter(struct task_struct *tsk) { } | ||
| 45 | static inline void vtime_user_exit(struct task_struct *tsk) { } | ||
| 46 | static inline void vtime_guest_enter(struct task_struct *tsk) { } | ||
| 47 | static inline void vtime_guest_exit(struct task_struct *tsk) { } | ||
| 48 | static inline void vtime_init_idle(struct task_struct *tsk) { } | ||
| 18 | #endif | 49 | #endif |
| 19 | 50 | ||
| 20 | #ifdef CONFIG_IRQ_TIME_ACCOUNTING | 51 | #ifdef CONFIG_IRQ_TIME_ACCOUNTING |
| @@ -23,25 +54,15 @@ extern void irqtime_account_irq(struct task_struct *tsk); | |||
| 23 | static inline void irqtime_account_irq(struct task_struct *tsk) { } | 54 | static inline void irqtime_account_irq(struct task_struct *tsk) { } |
| 24 | #endif | 55 | #endif |
| 25 | 56 | ||
| 26 | static inline void vtime_account_irq_enter(struct task_struct *tsk) | 57 | static inline void account_irq_enter_time(struct task_struct *tsk) |
| 27 | { | 58 | { |
| 28 | /* | 59 | vtime_account_irq_enter(tsk); |
| 29 | * Hardirq can interrupt idle task anytime. So we need vtime_account() | ||
| 30 | * that performs the idle check in CONFIG_VIRT_CPU_ACCOUNTING. | ||
| 31 | * Softirq can also interrupt idle task directly if it calls | ||
| 32 | * local_bh_enable(). Such case probably don't exist but we never know. | ||
| 33 | * Ksoftirqd is not concerned because idle time is flushed on context | ||
| 34 | * switch. Softirqs in the end of hardirqs are also not a problem because | ||
| 35 | * the idle time is flushed on hardirq time already. | ||
| 36 | */ | ||
| 37 | vtime_account(tsk); | ||
| 38 | irqtime_account_irq(tsk); | 60 | irqtime_account_irq(tsk); |
| 39 | } | 61 | } |
| 40 | 62 | ||
| 41 | static inline void vtime_account_irq_exit(struct task_struct *tsk) | 63 | static inline void account_irq_exit_time(struct task_struct *tsk) |
| 42 | { | 64 | { |
| 43 | /* On hard|softirq exit we always account to hard|softirq cputime */ | 65 | vtime_account_irq_exit(tsk); |
| 44 | vtime_account_system(tsk); | ||
| 45 | irqtime_account_irq(tsk); | 66 | irqtime_account_irq(tsk); |
| 46 | } | 67 | } |
| 47 | 68 | ||
diff --git a/init/Kconfig b/init/Kconfig index dcb68ac42b78..7000d9657402 100644 --- a/init/Kconfig +++ b/init/Kconfig | |||
| @@ -322,10 +322,13 @@ source "kernel/time/Kconfig" | |||
| 322 | 322 | ||
| 323 | menu "CPU/Task time and stats accounting" | 323 | menu "CPU/Task time and stats accounting" |
| 324 | 324 | ||
| 325 | config VIRT_CPU_ACCOUNTING | ||
| 326 | bool | ||
| 327 | |||
| 325 | choice | 328 | choice |
| 326 | prompt "Cputime accounting" | 329 | prompt "Cputime accounting" |
| 327 | default TICK_CPU_ACCOUNTING if !PPC64 | 330 | default TICK_CPU_ACCOUNTING if !PPC64 |
| 328 | default VIRT_CPU_ACCOUNTING if PPC64 | 331 | default VIRT_CPU_ACCOUNTING_NATIVE if PPC64 |
| 329 | 332 | ||
| 330 | # Kind of a stub config for the pure tick based cputime accounting | 333 | # Kind of a stub config for the pure tick based cputime accounting |
| 331 | config TICK_CPU_ACCOUNTING | 334 | config TICK_CPU_ACCOUNTING |
| @@ -338,9 +341,10 @@ config TICK_CPU_ACCOUNTING | |||
| 338 | 341 | ||
| 339 | If unsure, say Y. | 342 | If unsure, say Y. |
| 340 | 343 | ||
| 341 | config VIRT_CPU_ACCOUNTING | 344 | config VIRT_CPU_ACCOUNTING_NATIVE |
| 342 | bool "Deterministic task and CPU time accounting" | 345 | bool "Deterministic task and CPU time accounting" |
| 343 | depends on HAVE_VIRT_CPU_ACCOUNTING | 346 | depends on HAVE_VIRT_CPU_ACCOUNTING |
| 347 | select VIRT_CPU_ACCOUNTING | ||
| 344 | help | 348 | help |
| 345 | Select this option to enable more accurate task and CPU time | 349 | Select this option to enable more accurate task and CPU time |
| 346 | accounting. This is done by reading a CPU counter on each | 350 | accounting. This is done by reading a CPU counter on each |
| @@ -350,6 +354,23 @@ config VIRT_CPU_ACCOUNTING | |||
| 350 | this also enables accounting of stolen time on logically-partitioned | 354 | this also enables accounting of stolen time on logically-partitioned |
| 351 | systems. | 355 | systems. |
| 352 | 356 | ||
| 357 | config VIRT_CPU_ACCOUNTING_GEN | ||
| 358 | bool "Full dynticks CPU time accounting" | ||
| 359 | depends on HAVE_CONTEXT_TRACKING && 64BIT | ||
| 360 | select VIRT_CPU_ACCOUNTING | ||
| 361 | select CONTEXT_TRACKING | ||
| 362 | help | ||
| 363 | Select this option to enable task and CPU time accounting on full | ||
| 364 | dynticks systems. This accounting is implemented by watching every | ||
| 365 | kernel-user boundaries using the context tracking subsystem. | ||
| 366 | The accounting is thus performed at the expense of some significant | ||
| 367 | overhead. | ||
| 368 | |||
| 369 | For now this is only useful if you are working on the full | ||
| 370 | dynticks subsystem development. | ||
| 371 | |||
| 372 | If unsure, say N. | ||
| 373 | |||
| 353 | config IRQ_TIME_ACCOUNTING | 374 | config IRQ_TIME_ACCOUNTING |
| 354 | bool "Fine granularity task level IRQ time accounting" | 375 | bool "Fine granularity task level IRQ time accounting" |
| 355 | depends on HAVE_IRQ_TIME_ACCOUNTING | 376 | depends on HAVE_IRQ_TIME_ACCOUNTING |
diff --git a/init/init_task.c b/init/init_task.c index 8b2f3996b035..ba0a7f362d9e 100644 --- a/init/init_task.c +++ b/init/init_task.c | |||
| @@ -2,6 +2,8 @@ | |||
| 2 | #include <linux/export.h> | 2 | #include <linux/export.h> |
| 3 | #include <linux/mqueue.h> | 3 | #include <linux/mqueue.h> |
| 4 | #include <linux/sched.h> | 4 | #include <linux/sched.h> |
| 5 | #include <linux/sched/sysctl.h> | ||
| 6 | #include <linux/sched/rt.h> | ||
| 5 | #include <linux/init.h> | 7 | #include <linux/init.h> |
| 6 | #include <linux/fs.h> | 8 | #include <linux/fs.h> |
| 7 | #include <linux/mm.h> | 9 | #include <linux/mm.h> |
diff --git a/kernel/acct.c b/kernel/acct.c index 051e071a06e7..e8b1627ab9c7 100644 --- a/kernel/acct.c +++ b/kernel/acct.c | |||
| @@ -566,6 +566,7 @@ out: | |||
| 566 | void acct_collect(long exitcode, int group_dead) | 566 | void acct_collect(long exitcode, int group_dead) |
| 567 | { | 567 | { |
| 568 | struct pacct_struct *pacct = ¤t->signal->pacct; | 568 | struct pacct_struct *pacct = ¤t->signal->pacct; |
| 569 | cputime_t utime, stime; | ||
| 569 | unsigned long vsize = 0; | 570 | unsigned long vsize = 0; |
| 570 | 571 | ||
| 571 | if (group_dead && current->mm) { | 572 | if (group_dead && current->mm) { |
| @@ -593,8 +594,9 @@ void acct_collect(long exitcode, int group_dead) | |||
| 593 | pacct->ac_flag |= ACORE; | 594 | pacct->ac_flag |= ACORE; |
| 594 | if (current->flags & PF_SIGNALED) | 595 | if (current->flags & PF_SIGNALED) |
| 595 | pacct->ac_flag |= AXSIG; | 596 | pacct->ac_flag |= AXSIG; |
| 596 | pacct->ac_utime += current->utime; | 597 | task_cputime(current, &utime, &stime); |
| 597 | pacct->ac_stime += current->stime; | 598 | pacct->ac_utime += utime; |
| 599 | pacct->ac_stime += stime; | ||
| 598 | pacct->ac_minflt += current->min_flt; | 600 | pacct->ac_minflt += current->min_flt; |
| 599 | pacct->ac_majflt += current->maj_flt; | 601 | pacct->ac_majflt += current->maj_flt; |
| 600 | spin_unlock_irq(¤t->sighand->siglock); | 602 | spin_unlock_irq(¤t->sighand->siglock); |
diff --git a/kernel/context_tracking.c b/kernel/context_tracking.c index d566aba7e801..65349f07b878 100644 --- a/kernel/context_tracking.c +++ b/kernel/context_tracking.c | |||
| @@ -15,26 +15,13 @@ | |||
| 15 | */ | 15 | */ |
| 16 | 16 | ||
| 17 | #include <linux/context_tracking.h> | 17 | #include <linux/context_tracking.h> |
| 18 | #include <linux/kvm_host.h> | ||
| 18 | #include <linux/rcupdate.h> | 19 | #include <linux/rcupdate.h> |
| 19 | #include <linux/sched.h> | 20 | #include <linux/sched.h> |
| 20 | #include <linux/percpu.h> | ||
| 21 | #include <linux/hardirq.h> | 21 | #include <linux/hardirq.h> |
| 22 | #include <linux/export.h> | ||
| 22 | 23 | ||
| 23 | struct context_tracking { | 24 | DEFINE_PER_CPU(struct context_tracking, context_tracking) = { |
| 24 | /* | ||
| 25 | * When active is false, probes are unset in order | ||
| 26 | * to minimize overhead: TIF flags are cleared | ||
| 27 | * and calls to user_enter/exit are ignored. This | ||
| 28 | * may be further optimized using static keys. | ||
| 29 | */ | ||
| 30 | bool active; | ||
| 31 | enum { | ||
| 32 | IN_KERNEL = 0, | ||
| 33 | IN_USER, | ||
| 34 | } state; | ||
| 35 | }; | ||
| 36 | |||
| 37 | static DEFINE_PER_CPU(struct context_tracking, context_tracking) = { | ||
| 38 | #ifdef CONFIG_CONTEXT_TRACKING_FORCE | 25 | #ifdef CONFIG_CONTEXT_TRACKING_FORCE |
| 39 | .active = true, | 26 | .active = true, |
| 40 | #endif | 27 | #endif |
| @@ -70,7 +57,6 @@ void user_enter(void) | |||
| 70 | local_irq_save(flags); | 57 | local_irq_save(flags); |
| 71 | if (__this_cpu_read(context_tracking.active) && | 58 | if (__this_cpu_read(context_tracking.active) && |
| 72 | __this_cpu_read(context_tracking.state) != IN_USER) { | 59 | __this_cpu_read(context_tracking.state) != IN_USER) { |
| 73 | __this_cpu_write(context_tracking.state, IN_USER); | ||
| 74 | /* | 60 | /* |
| 75 | * At this stage, only low level arch entry code remains and | 61 | * At this stage, only low level arch entry code remains and |
| 76 | * then we'll run in userspace. We can assume there won't be | 62 | * then we'll run in userspace. We can assume there won't be |
| @@ -78,7 +64,9 @@ void user_enter(void) | |||
| 78 | * user_exit() or rcu_irq_enter(). Let's remove RCU's dependency | 64 | * user_exit() or rcu_irq_enter(). Let's remove RCU's dependency |
| 79 | * on the tick. | 65 | * on the tick. |
| 80 | */ | 66 | */ |
| 67 | vtime_user_enter(current); | ||
| 81 | rcu_user_enter(); | 68 | rcu_user_enter(); |
| 69 | __this_cpu_write(context_tracking.state, IN_USER); | ||
| 82 | } | 70 | } |
| 83 | local_irq_restore(flags); | 71 | local_irq_restore(flags); |
| 84 | } | 72 | } |
| @@ -104,16 +92,35 @@ void user_exit(void) | |||
| 104 | 92 | ||
| 105 | local_irq_save(flags); | 93 | local_irq_save(flags); |
| 106 | if (__this_cpu_read(context_tracking.state) == IN_USER) { | 94 | if (__this_cpu_read(context_tracking.state) == IN_USER) { |
| 107 | __this_cpu_write(context_tracking.state, IN_KERNEL); | ||
| 108 | /* | 95 | /* |
| 109 | * We are going to run code that may use RCU. Inform | 96 | * We are going to run code that may use RCU. Inform |
| 110 | * RCU core about that (ie: we may need the tick again). | 97 | * RCU core about that (ie: we may need the tick again). |
| 111 | */ | 98 | */ |
| 112 | rcu_user_exit(); | 99 | rcu_user_exit(); |
| 100 | vtime_user_exit(current); | ||
| 101 | __this_cpu_write(context_tracking.state, IN_KERNEL); | ||
| 113 | } | 102 | } |
| 114 | local_irq_restore(flags); | 103 | local_irq_restore(flags); |
| 115 | } | 104 | } |
| 116 | 105 | ||
| 106 | void guest_enter(void) | ||
| 107 | { | ||
| 108 | if (vtime_accounting_enabled()) | ||
| 109 | vtime_guest_enter(current); | ||
| 110 | else | ||
| 111 | __guest_enter(); | ||
| 112 | } | ||
| 113 | EXPORT_SYMBOL_GPL(guest_enter); | ||
| 114 | |||
| 115 | void guest_exit(void) | ||
| 116 | { | ||
| 117 | if (vtime_accounting_enabled()) | ||
| 118 | vtime_guest_exit(current); | ||
| 119 | else | ||
| 120 | __guest_exit(); | ||
| 121 | } | ||
| 122 | EXPORT_SYMBOL_GPL(guest_exit); | ||
| 123 | |||
| 117 | 124 | ||
| 118 | /** | 125 | /** |
| 119 | * context_tracking_task_switch - context switch the syscall callbacks | 126 | * context_tracking_task_switch - context switch the syscall callbacks |
diff --git a/kernel/cpu.c b/kernel/cpu.c index 3046a503242c..e5d5e8e1e030 100644 --- a/kernel/cpu.c +++ b/kernel/cpu.c | |||
| @@ -224,11 +224,13 @@ void clear_tasks_mm_cpumask(int cpu) | |||
| 224 | static inline void check_for_tasks(int cpu) | 224 | static inline void check_for_tasks(int cpu) |
| 225 | { | 225 | { |
| 226 | struct task_struct *p; | 226 | struct task_struct *p; |
| 227 | cputime_t utime, stime; | ||
| 227 | 228 | ||
| 228 | write_lock_irq(&tasklist_lock); | 229 | write_lock_irq(&tasklist_lock); |
| 229 | for_each_process(p) { | 230 | for_each_process(p) { |
| 231 | task_cputime(p, &utime, &stime); | ||
| 230 | if (task_cpu(p) == cpu && p->state == TASK_RUNNING && | 232 | if (task_cpu(p) == cpu && p->state == TASK_RUNNING && |
| 231 | (p->utime || p->stime)) | 233 | (utime || stime)) |
| 232 | printk(KERN_WARNING "Task %s (pid = %d) is on cpu %d " | 234 | printk(KERN_WARNING "Task %s (pid = %d) is on cpu %d " |
| 233 | "(state = %ld, flags = %x)\n", | 235 | "(state = %ld, flags = %x)\n", |
| 234 | p->comm, task_pid_nr(p), cpu, | 236 | p->comm, task_pid_nr(p), cpu, |
diff --git a/kernel/delayacct.c b/kernel/delayacct.c index 418b3f7053aa..d473988c1d0b 100644 --- a/kernel/delayacct.c +++ b/kernel/delayacct.c | |||
| @@ -106,6 +106,7 @@ int __delayacct_add_tsk(struct taskstats *d, struct task_struct *tsk) | |||
| 106 | unsigned long long t2, t3; | 106 | unsigned long long t2, t3; |
| 107 | unsigned long flags; | 107 | unsigned long flags; |
| 108 | struct timespec ts; | 108 | struct timespec ts; |
| 109 | cputime_t utime, stime, stimescaled, utimescaled; | ||
| 109 | 110 | ||
| 110 | /* Though tsk->delays accessed later, early exit avoids | 111 | /* Though tsk->delays accessed later, early exit avoids |
| 111 | * unnecessary returning of other data | 112 | * unnecessary returning of other data |
| @@ -114,12 +115,14 @@ int __delayacct_add_tsk(struct taskstats *d, struct task_struct *tsk) | |||
| 114 | goto done; | 115 | goto done; |
| 115 | 116 | ||
| 116 | tmp = (s64)d->cpu_run_real_total; | 117 | tmp = (s64)d->cpu_run_real_total; |
| 117 | cputime_to_timespec(tsk->utime + tsk->stime, &ts); | 118 | task_cputime(tsk, &utime, &stime); |
| 119 | cputime_to_timespec(utime + stime, &ts); | ||
| 118 | tmp += timespec_to_ns(&ts); | 120 | tmp += timespec_to_ns(&ts); |
| 119 | d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; | 121 | d->cpu_run_real_total = (tmp < (s64)d->cpu_run_real_total) ? 0 : tmp; |
| 120 | 122 | ||
| 121 | tmp = (s64)d->cpu_scaled_run_real_total; | 123 | tmp = (s64)d->cpu_scaled_run_real_total; |
| 122 | cputime_to_timespec(tsk->utimescaled + tsk->stimescaled, &ts); | 124 | task_cputime_scaled(tsk, &utimescaled, &stimescaled); |
| 125 | cputime_to_timespec(utimescaled + stimescaled, &ts); | ||
| 123 | tmp += timespec_to_ns(&ts); | 126 | tmp += timespec_to_ns(&ts); |
| 124 | d->cpu_scaled_run_real_total = | 127 | d->cpu_scaled_run_real_total = |
| 125 | (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; | 128 | (tmp < (s64)d->cpu_scaled_run_real_total) ? 0 : tmp; |
diff --git a/kernel/exit.c b/kernel/exit.c index b4df21937216..7dd20408707c 100644 --- a/kernel/exit.c +++ b/kernel/exit.c | |||
| @@ -85,6 +85,7 @@ static void __exit_signal(struct task_struct *tsk) | |||
| 85 | bool group_dead = thread_group_leader(tsk); | 85 | bool group_dead = thread_group_leader(tsk); |
| 86 | struct sighand_struct *sighand; | 86 | struct sighand_struct *sighand; |
| 87 | struct tty_struct *uninitialized_var(tty); | 87 | struct tty_struct *uninitialized_var(tty); |
| 88 | cputime_t utime, stime; | ||
| 88 | 89 | ||
| 89 | sighand = rcu_dereference_check(tsk->sighand, | 90 | sighand = rcu_dereference_check(tsk->sighand, |
| 90 | lockdep_tasklist_lock_is_held()); | 91 | lockdep_tasklist_lock_is_held()); |
| @@ -123,9 +124,10 @@ static void __exit_signal(struct task_struct *tsk) | |||
| 123 | * We won't ever get here for the group leader, since it | 124 | * We won't ever get here for the group leader, since it |
| 124 | * will have been the last reference on the signal_struct. | 125 | * will have been the last reference on the signal_struct. |
| 125 | */ | 126 | */ |
| 126 | sig->utime += tsk->utime; | 127 | task_cputime(tsk, &utime, &stime); |
| 127 | sig->stime += tsk->stime; | 128 | sig->utime += utime; |
| 128 | sig->gtime += tsk->gtime; | 129 | sig->stime += stime; |
| 130 | sig->gtime += task_gtime(tsk); | ||
| 129 | sig->min_flt += tsk->min_flt; | 131 | sig->min_flt += tsk->min_flt; |
| 130 | sig->maj_flt += tsk->maj_flt; | 132 | sig->maj_flt += tsk->maj_flt; |
| 131 | sig->nvcsw += tsk->nvcsw; | 133 | sig->nvcsw += tsk->nvcsw; |
| @@ -1092,7 +1094,7 @@ static int wait_task_zombie(struct wait_opts *wo, struct task_struct *p) | |||
| 1092 | sig = p->signal; | 1094 | sig = p->signal; |
| 1093 | psig->cutime += tgutime + sig->cutime; | 1095 | psig->cutime += tgutime + sig->cutime; |
| 1094 | psig->cstime += tgstime + sig->cstime; | 1096 | psig->cstime += tgstime + sig->cstime; |
| 1095 | psig->cgtime += p->gtime + sig->gtime + sig->cgtime; | 1097 | psig->cgtime += task_gtime(p) + sig->gtime + sig->cgtime; |
| 1096 | psig->cmin_flt += | 1098 | psig->cmin_flt += |
| 1097 | p->min_flt + sig->min_flt + sig->cmin_flt; | 1099 | p->min_flt + sig->min_flt + sig->cmin_flt; |
| 1098 | psig->cmaj_flt += | 1100 | psig->cmaj_flt += |
diff --git a/kernel/fork.c b/kernel/fork.c index c535f33bbb9c..4133876d8cd2 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
| @@ -1233,6 +1233,12 @@ static struct task_struct *copy_process(unsigned long clone_flags, | |||
| 1233 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 1233 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING |
| 1234 | p->prev_cputime.utime = p->prev_cputime.stime = 0; | 1234 | p->prev_cputime.utime = p->prev_cputime.stime = 0; |
| 1235 | #endif | 1235 | #endif |
| 1236 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 1237 | seqlock_init(&p->vtime_seqlock); | ||
| 1238 | p->vtime_snap = 0; | ||
| 1239 | p->vtime_snap_whence = VTIME_SLEEPING; | ||
| 1240 | #endif | ||
| 1241 | |||
| 1236 | #if defined(SPLIT_RSS_COUNTING) | 1242 | #if defined(SPLIT_RSS_COUNTING) |
| 1237 | memset(&p->rss_stat, 0, sizeof(p->rss_stat)); | 1243 | memset(&p->rss_stat, 0, sizeof(p->rss_stat)); |
| 1238 | #endif | 1244 | #endif |
diff --git a/kernel/futex.c b/kernel/futex.c index 19eb089ca003..9618b6e9fb36 100644 --- a/kernel/futex.c +++ b/kernel/futex.c | |||
| @@ -60,6 +60,7 @@ | |||
| 60 | #include <linux/pid.h> | 60 | #include <linux/pid.h> |
| 61 | #include <linux/nsproxy.h> | 61 | #include <linux/nsproxy.h> |
| 62 | #include <linux/ptrace.h> | 62 | #include <linux/ptrace.h> |
| 63 | #include <linux/sched/rt.h> | ||
| 63 | 64 | ||
| 64 | #include <asm/futex.h> | 65 | #include <asm/futex.h> |
| 65 | 66 | ||
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 6db7a5ed52b5..c5dde988c0ce 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c | |||
| @@ -44,6 +44,8 @@ | |||
| 44 | #include <linux/err.h> | 44 | #include <linux/err.h> |
| 45 | #include <linux/debugobjects.h> | 45 | #include <linux/debugobjects.h> |
| 46 | #include <linux/sched.h> | 46 | #include <linux/sched.h> |
| 47 | #include <linux/sched/sysctl.h> | ||
| 48 | #include <linux/sched/rt.h> | ||
| 47 | #include <linux/timer.h> | 49 | #include <linux/timer.h> |
| 48 | 50 | ||
| 49 | #include <asm/uaccess.h> | 51 | #include <asm/uaccess.h> |
diff --git a/kernel/irq/manage.c b/kernel/irq/manage.c index 88e7bed62711..fa17855ca65a 100644 --- a/kernel/irq/manage.c +++ b/kernel/irq/manage.c | |||
| @@ -16,6 +16,7 @@ | |||
| 16 | #include <linux/interrupt.h> | 16 | #include <linux/interrupt.h> |
| 17 | #include <linux/slab.h> | 17 | #include <linux/slab.h> |
| 18 | #include <linux/sched.h> | 18 | #include <linux/sched.h> |
| 19 | #include <linux/sched/rt.h> | ||
| 19 | #include <linux/task_work.h> | 20 | #include <linux/task_work.h> |
| 20 | 21 | ||
| 21 | #include "internals.h" | 22 | #include "internals.h" |
diff --git a/kernel/mutex.c b/kernel/mutex.c index a307cc9c9526..52f23011b6e0 100644 --- a/kernel/mutex.c +++ b/kernel/mutex.c | |||
| @@ -19,6 +19,7 @@ | |||
| 19 | */ | 19 | */ |
| 20 | #include <linux/mutex.h> | 20 | #include <linux/mutex.h> |
| 21 | #include <linux/sched.h> | 21 | #include <linux/sched.h> |
| 22 | #include <linux/sched/rt.h> | ||
| 22 | #include <linux/export.h> | 23 | #include <linux/export.h> |
| 23 | #include <linux/spinlock.h> | 24 | #include <linux/spinlock.h> |
| 24 | #include <linux/interrupt.h> | 25 | #include <linux/interrupt.h> |
diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c index a278cad1d5d6..165d47698477 100644 --- a/kernel/posix-cpu-timers.c +++ b/kernel/posix-cpu-timers.c | |||
| @@ -155,11 +155,19 @@ static void bump_cpu_timer(struct k_itimer *timer, | |||
| 155 | 155 | ||
| 156 | static inline cputime_t prof_ticks(struct task_struct *p) | 156 | static inline cputime_t prof_ticks(struct task_struct *p) |
| 157 | { | 157 | { |
| 158 | return p->utime + p->stime; | 158 | cputime_t utime, stime; |
| 159 | |||
| 160 | task_cputime(p, &utime, &stime); | ||
| 161 | |||
| 162 | return utime + stime; | ||
| 159 | } | 163 | } |
| 160 | static inline cputime_t virt_ticks(struct task_struct *p) | 164 | static inline cputime_t virt_ticks(struct task_struct *p) |
| 161 | { | 165 | { |
| 162 | return p->utime; | 166 | cputime_t utime; |
| 167 | |||
| 168 | task_cputime(p, &utime, NULL); | ||
| 169 | |||
| 170 | return utime; | ||
| 163 | } | 171 | } |
| 164 | 172 | ||
| 165 | static int | 173 | static int |
| @@ -471,18 +479,23 @@ static void cleanup_timers(struct list_head *head, | |||
| 471 | */ | 479 | */ |
| 472 | void posix_cpu_timers_exit(struct task_struct *tsk) | 480 | void posix_cpu_timers_exit(struct task_struct *tsk) |
| 473 | { | 481 | { |
| 482 | cputime_t utime, stime; | ||
| 483 | |||
| 474 | add_device_randomness((const void*) &tsk->se.sum_exec_runtime, | 484 | add_device_randomness((const void*) &tsk->se.sum_exec_runtime, |
| 475 | sizeof(unsigned long long)); | 485 | sizeof(unsigned long long)); |
| 486 | task_cputime(tsk, &utime, &stime); | ||
| 476 | cleanup_timers(tsk->cpu_timers, | 487 | cleanup_timers(tsk->cpu_timers, |
| 477 | tsk->utime, tsk->stime, tsk->se.sum_exec_runtime); | 488 | utime, stime, tsk->se.sum_exec_runtime); |
| 478 | 489 | ||
| 479 | } | 490 | } |
| 480 | void posix_cpu_timers_exit_group(struct task_struct *tsk) | 491 | void posix_cpu_timers_exit_group(struct task_struct *tsk) |
| 481 | { | 492 | { |
| 482 | struct signal_struct *const sig = tsk->signal; | 493 | struct signal_struct *const sig = tsk->signal; |
| 494 | cputime_t utime, stime; | ||
| 483 | 495 | ||
| 496 | task_cputime(tsk, &utime, &stime); | ||
| 484 | cleanup_timers(tsk->signal->cpu_timers, | 497 | cleanup_timers(tsk->signal->cpu_timers, |
| 485 | tsk->utime + sig->utime, tsk->stime + sig->stime, | 498 | utime + sig->utime, stime + sig->stime, |
| 486 | tsk->se.sum_exec_runtime + sig->sum_sched_runtime); | 499 | tsk->se.sum_exec_runtime + sig->sum_sched_runtime); |
| 487 | } | 500 | } |
| 488 | 501 | ||
| @@ -1226,11 +1239,14 @@ static inline int task_cputime_expired(const struct task_cputime *sample, | |||
| 1226 | static inline int fastpath_timer_check(struct task_struct *tsk) | 1239 | static inline int fastpath_timer_check(struct task_struct *tsk) |
| 1227 | { | 1240 | { |
| 1228 | struct signal_struct *sig; | 1241 | struct signal_struct *sig; |
| 1242 | cputime_t utime, stime; | ||
| 1243 | |||
| 1244 | task_cputime(tsk, &utime, &stime); | ||
| 1229 | 1245 | ||
| 1230 | if (!task_cputime_zero(&tsk->cputime_expires)) { | 1246 | if (!task_cputime_zero(&tsk->cputime_expires)) { |
| 1231 | struct task_cputime task_sample = { | 1247 | struct task_cputime task_sample = { |
| 1232 | .utime = tsk->utime, | 1248 | .utime = utime, |
| 1233 | .stime = tsk->stime, | 1249 | .stime = stime, |
| 1234 | .sum_exec_runtime = tsk->se.sum_exec_runtime | 1250 | .sum_exec_runtime = tsk->se.sum_exec_runtime |
| 1235 | }; | 1251 | }; |
| 1236 | 1252 | ||
diff --git a/kernel/rtmutex-debug.c b/kernel/rtmutex-debug.c index 16502d3a71c8..13b243a323fa 100644 --- a/kernel/rtmutex-debug.c +++ b/kernel/rtmutex-debug.c | |||
| @@ -17,6 +17,7 @@ | |||
| 17 | * See rt.c in preempt-rt for proper credits and further information | 17 | * See rt.c in preempt-rt for proper credits and further information |
| 18 | */ | 18 | */ |
| 19 | #include <linux/sched.h> | 19 | #include <linux/sched.h> |
| 20 | #include <linux/sched/rt.h> | ||
| 20 | #include <linux/delay.h> | 21 | #include <linux/delay.h> |
| 21 | #include <linux/export.h> | 22 | #include <linux/export.h> |
| 22 | #include <linux/spinlock.h> | 23 | #include <linux/spinlock.h> |
diff --git a/kernel/rtmutex-tester.c b/kernel/rtmutex-tester.c index 98ec49475460..7890b10084a7 100644 --- a/kernel/rtmutex-tester.c +++ b/kernel/rtmutex-tester.c | |||
| @@ -10,6 +10,7 @@ | |||
| 10 | #include <linux/kthread.h> | 10 | #include <linux/kthread.h> |
| 11 | #include <linux/export.h> | 11 | #include <linux/export.h> |
| 12 | #include <linux/sched.h> | 12 | #include <linux/sched.h> |
| 13 | #include <linux/sched/rt.h> | ||
| 13 | #include <linux/spinlock.h> | 14 | #include <linux/spinlock.h> |
| 14 | #include <linux/timer.h> | 15 | #include <linux/timer.h> |
| 15 | #include <linux/freezer.h> | 16 | #include <linux/freezer.h> |
diff --git a/kernel/rtmutex.c b/kernel/rtmutex.c index a242e691c993..1e09308bf2a1 100644 --- a/kernel/rtmutex.c +++ b/kernel/rtmutex.c | |||
| @@ -13,6 +13,7 @@ | |||
| 13 | #include <linux/spinlock.h> | 13 | #include <linux/spinlock.h> |
| 14 | #include <linux/export.h> | 14 | #include <linux/export.h> |
| 15 | #include <linux/sched.h> | 15 | #include <linux/sched.h> |
| 16 | #include <linux/sched/rt.h> | ||
| 16 | #include <linux/timer.h> | 17 | #include <linux/timer.h> |
| 17 | 18 | ||
| 18 | #include "rtmutex_common.h" | 19 | #include "rtmutex_common.h" |
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 26058d0bebba..4a88f1d51563 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c | |||
| @@ -4371,7 +4371,7 @@ bool __sched yield_to(struct task_struct *p, bool preempt) | |||
| 4371 | struct task_struct *curr = current; | 4371 | struct task_struct *curr = current; |
| 4372 | struct rq *rq, *p_rq; | 4372 | struct rq *rq, *p_rq; |
| 4373 | unsigned long flags; | 4373 | unsigned long flags; |
| 4374 | bool yielded = 0; | 4374 | int yielded = 0; |
| 4375 | 4375 | ||
| 4376 | local_irq_save(flags); | 4376 | local_irq_save(flags); |
| 4377 | rq = this_rq(); | 4377 | rq = this_rq(); |
| @@ -4667,6 +4667,7 @@ void __cpuinit init_idle(struct task_struct *idle, int cpu) | |||
| 4667 | */ | 4667 | */ |
| 4668 | idle->sched_class = &idle_sched_class; | 4668 | idle->sched_class = &idle_sched_class; |
| 4669 | ftrace_graph_init_idle_task(idle, cpu); | 4669 | ftrace_graph_init_idle_task(idle, cpu); |
| 4670 | vtime_init_idle(idle); | ||
| 4670 | #if defined(CONFIG_SMP) | 4671 | #if defined(CONFIG_SMP) |
| 4671 | sprintf(idle->comm, "%s/%d", INIT_TASK_COMM, cpu); | 4672 | sprintf(idle->comm, "%s/%d", INIT_TASK_COMM, cpu); |
| 4672 | #endif | 4673 | #endif |
| @@ -7508,6 +7509,25 @@ static int sched_rt_global_constraints(void) | |||
| 7508 | } | 7509 | } |
| 7509 | #endif /* CONFIG_RT_GROUP_SCHED */ | 7510 | #endif /* CONFIG_RT_GROUP_SCHED */ |
| 7510 | 7511 | ||
| 7512 | int sched_rr_handler(struct ctl_table *table, int write, | ||
| 7513 | void __user *buffer, size_t *lenp, | ||
| 7514 | loff_t *ppos) | ||
| 7515 | { | ||
| 7516 | int ret; | ||
| 7517 | static DEFINE_MUTEX(mutex); | ||
| 7518 | |||
| 7519 | mutex_lock(&mutex); | ||
| 7520 | ret = proc_dointvec(table, write, buffer, lenp, ppos); | ||
| 7521 | /* make sure that internally we keep jiffies */ | ||
| 7522 | /* also, writing zero resets timeslice to default */ | ||
| 7523 | if (!ret && write) { | ||
| 7524 | sched_rr_timeslice = sched_rr_timeslice <= 0 ? | ||
| 7525 | RR_TIMESLICE : msecs_to_jiffies(sched_rr_timeslice); | ||
| 7526 | } | ||
| 7527 | mutex_unlock(&mutex); | ||
| 7528 | return ret; | ||
| 7529 | } | ||
| 7530 | |||
| 7511 | int sched_rt_handler(struct ctl_table *table, int write, | 7531 | int sched_rt_handler(struct ctl_table *table, int write, |
| 7512 | void __user *buffer, size_t *lenp, | 7532 | void __user *buffer, size_t *lenp, |
| 7513 | loff_t *ppos) | 7533 | loff_t *ppos) |
diff --git a/kernel/sched/cpupri.c b/kernel/sched/cpupri.c index 23aa789c53ee..1095e878a46f 100644 --- a/kernel/sched/cpupri.c +++ b/kernel/sched/cpupri.c | |||
| @@ -28,6 +28,8 @@ | |||
| 28 | */ | 28 | */ |
| 29 | 29 | ||
| 30 | #include <linux/gfp.h> | 30 | #include <linux/gfp.h> |
| 31 | #include <linux/sched.h> | ||
| 32 | #include <linux/sched/rt.h> | ||
| 31 | #include "cpupri.h" | 33 | #include "cpupri.h" |
| 32 | 34 | ||
| 33 | /* Convert between a 140 based task->prio, and our 102 based cpupri */ | 35 | /* Convert between a 140 based task->prio, and our 102 based cpupri */ |
diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c index 293b202fcf79..9857329ed280 100644 --- a/kernel/sched/cputime.c +++ b/kernel/sched/cputime.c | |||
| @@ -3,6 +3,7 @@ | |||
| 3 | #include <linux/tsacct_kern.h> | 3 | #include <linux/tsacct_kern.h> |
| 4 | #include <linux/kernel_stat.h> | 4 | #include <linux/kernel_stat.h> |
| 5 | #include <linux/static_key.h> | 5 | #include <linux/static_key.h> |
| 6 | #include <linux/context_tracking.h> | ||
| 6 | #include "sched.h" | 7 | #include "sched.h" |
| 7 | 8 | ||
| 8 | 9 | ||
| @@ -163,7 +164,7 @@ void account_user_time(struct task_struct *p, cputime_t cputime, | |||
| 163 | task_group_account_field(p, index, (__force u64) cputime); | 164 | task_group_account_field(p, index, (__force u64) cputime); |
| 164 | 165 | ||
| 165 | /* Account for user time used */ | 166 | /* Account for user time used */ |
| 166 | acct_update_integrals(p); | 167 | acct_account_cputime(p); |
| 167 | } | 168 | } |
| 168 | 169 | ||
| 169 | /* | 170 | /* |
| @@ -213,7 +214,7 @@ void __account_system_time(struct task_struct *p, cputime_t cputime, | |||
| 213 | task_group_account_field(p, index, (__force u64) cputime); | 214 | task_group_account_field(p, index, (__force u64) cputime); |
| 214 | 215 | ||
| 215 | /* Account for system time used */ | 216 | /* Account for system time used */ |
| 216 | acct_update_integrals(p); | 217 | acct_account_cputime(p); |
| 217 | } | 218 | } |
| 218 | 219 | ||
| 219 | /* | 220 | /* |
| @@ -295,6 +296,7 @@ static __always_inline bool steal_account_process_tick(void) | |||
| 295 | void thread_group_cputime(struct task_struct *tsk, struct task_cputime *times) | 296 | void thread_group_cputime(struct task_struct *tsk, struct task_cputime *times) |
| 296 | { | 297 | { |
| 297 | struct signal_struct *sig = tsk->signal; | 298 | struct signal_struct *sig = tsk->signal; |
| 299 | cputime_t utime, stime; | ||
| 298 | struct task_struct *t; | 300 | struct task_struct *t; |
| 299 | 301 | ||
| 300 | times->utime = sig->utime; | 302 | times->utime = sig->utime; |
| @@ -308,16 +310,15 @@ void thread_group_cputime(struct task_struct *tsk, struct task_cputime *times) | |||
| 308 | 310 | ||
| 309 | t = tsk; | 311 | t = tsk; |
| 310 | do { | 312 | do { |
| 311 | times->utime += t->utime; | 313 | task_cputime(tsk, &utime, &stime); |
| 312 | times->stime += t->stime; | 314 | times->utime += utime; |
| 315 | times->stime += stime; | ||
| 313 | times->sum_exec_runtime += task_sched_runtime(t); | 316 | times->sum_exec_runtime += task_sched_runtime(t); |
| 314 | } while_each_thread(tsk, t); | 317 | } while_each_thread(tsk, t); |
| 315 | out: | 318 | out: |
| 316 | rcu_read_unlock(); | 319 | rcu_read_unlock(); |
| 317 | } | 320 | } |
| 318 | 321 | ||
| 319 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | ||
| 320 | |||
| 321 | #ifdef CONFIG_IRQ_TIME_ACCOUNTING | 322 | #ifdef CONFIG_IRQ_TIME_ACCOUNTING |
| 322 | /* | 323 | /* |
| 323 | * Account a tick to a process and cpustat | 324 | * Account a tick to a process and cpustat |
| @@ -382,11 +383,12 @@ static void irqtime_account_idle_ticks(int ticks) | |||
| 382 | irqtime_account_process_tick(current, 0, rq); | 383 | irqtime_account_process_tick(current, 0, rq); |
| 383 | } | 384 | } |
| 384 | #else /* CONFIG_IRQ_TIME_ACCOUNTING */ | 385 | #else /* CONFIG_IRQ_TIME_ACCOUNTING */ |
| 385 | static void irqtime_account_idle_ticks(int ticks) {} | 386 | static inline void irqtime_account_idle_ticks(int ticks) {} |
| 386 | static void irqtime_account_process_tick(struct task_struct *p, int user_tick, | 387 | static inline void irqtime_account_process_tick(struct task_struct *p, int user_tick, |
| 387 | struct rq *rq) {} | 388 | struct rq *rq) {} |
| 388 | #endif /* CONFIG_IRQ_TIME_ACCOUNTING */ | 389 | #endif /* CONFIG_IRQ_TIME_ACCOUNTING */ |
| 389 | 390 | ||
| 391 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE | ||
| 390 | /* | 392 | /* |
| 391 | * Account a single tick of cpu time. | 393 | * Account a single tick of cpu time. |
| 392 | * @p: the process that the cpu time gets accounted to | 394 | * @p: the process that the cpu time gets accounted to |
| @@ -397,6 +399,9 @@ void account_process_tick(struct task_struct *p, int user_tick) | |||
| 397 | cputime_t one_jiffy_scaled = cputime_to_scaled(cputime_one_jiffy); | 399 | cputime_t one_jiffy_scaled = cputime_to_scaled(cputime_one_jiffy); |
| 398 | struct rq *rq = this_rq(); | 400 | struct rq *rq = this_rq(); |
| 399 | 401 | ||
| 402 | if (vtime_accounting_enabled()) | ||
| 403 | return; | ||
| 404 | |||
| 400 | if (sched_clock_irqtime) { | 405 | if (sched_clock_irqtime) { |
| 401 | irqtime_account_process_tick(p, user_tick, rq); | 406 | irqtime_account_process_tick(p, user_tick, rq); |
| 402 | return; | 407 | return; |
| @@ -438,8 +443,7 @@ void account_idle_ticks(unsigned long ticks) | |||
| 438 | 443 | ||
| 439 | account_idle_time(jiffies_to_cputime(ticks)); | 444 | account_idle_time(jiffies_to_cputime(ticks)); |
| 440 | } | 445 | } |
| 441 | 446 | #endif /* !CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */ | |
| 442 | #endif | ||
| 443 | 447 | ||
| 444 | /* | 448 | /* |
| 445 | * Use precise platform statistics if available: | 449 | * Use precise platform statistics if available: |
| @@ -461,25 +465,20 @@ void thread_group_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime | |||
| 461 | *st = cputime.stime; | 465 | *st = cputime.stime; |
| 462 | } | 466 | } |
| 463 | 467 | ||
| 464 | void vtime_account_system_irqsafe(struct task_struct *tsk) | ||
| 465 | { | ||
| 466 | unsigned long flags; | ||
| 467 | |||
| 468 | local_irq_save(flags); | ||
| 469 | vtime_account_system(tsk); | ||
| 470 | local_irq_restore(flags); | ||
| 471 | } | ||
| 472 | EXPORT_SYMBOL_GPL(vtime_account_system_irqsafe); | ||
| 473 | |||
| 474 | #ifndef __ARCH_HAS_VTIME_TASK_SWITCH | 468 | #ifndef __ARCH_HAS_VTIME_TASK_SWITCH |
| 475 | void vtime_task_switch(struct task_struct *prev) | 469 | void vtime_task_switch(struct task_struct *prev) |
| 476 | { | 470 | { |
| 471 | if (!vtime_accounting_enabled()) | ||
| 472 | return; | ||
| 473 | |||
| 477 | if (is_idle_task(prev)) | 474 | if (is_idle_task(prev)) |
| 478 | vtime_account_idle(prev); | 475 | vtime_account_idle(prev); |
| 479 | else | 476 | else |
| 480 | vtime_account_system(prev); | 477 | vtime_account_system(prev); |
| 481 | 478 | ||
| 479 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE | ||
| 482 | vtime_account_user(prev); | 480 | vtime_account_user(prev); |
| 481 | #endif | ||
| 483 | arch_vtime_task_switch(prev); | 482 | arch_vtime_task_switch(prev); |
| 484 | } | 483 | } |
| 485 | #endif | 484 | #endif |
| @@ -493,27 +492,40 @@ void vtime_task_switch(struct task_struct *prev) | |||
| 493 | * vtime_account(). | 492 | * vtime_account(). |
| 494 | */ | 493 | */ |
| 495 | #ifndef __ARCH_HAS_VTIME_ACCOUNT | 494 | #ifndef __ARCH_HAS_VTIME_ACCOUNT |
| 496 | void vtime_account(struct task_struct *tsk) | 495 | void vtime_account_irq_enter(struct task_struct *tsk) |
| 497 | { | 496 | { |
| 498 | if (in_interrupt() || !is_idle_task(tsk)) | 497 | if (!vtime_accounting_enabled()) |
| 499 | vtime_account_system(tsk); | 498 | return; |
| 500 | else | 499 | |
| 501 | vtime_account_idle(tsk); | 500 | if (!in_interrupt()) { |
| 501 | /* | ||
| 502 | * If we interrupted user, context_tracking_in_user() | ||
| 503 | * is 1 because the context tracking don't hook | ||
| 504 | * on irq entry/exit. This way we know if | ||
| 505 | * we need to flush user time on kernel entry. | ||
| 506 | */ | ||
| 507 | if (context_tracking_in_user()) { | ||
| 508 | vtime_account_user(tsk); | ||
| 509 | return; | ||
| 510 | } | ||
| 511 | |||
| 512 | if (is_idle_task(tsk)) { | ||
| 513 | vtime_account_idle(tsk); | ||
| 514 | return; | ||
| 515 | } | ||
| 516 | } | ||
| 517 | vtime_account_system(tsk); | ||
| 502 | } | 518 | } |
| 503 | EXPORT_SYMBOL_GPL(vtime_account); | 519 | EXPORT_SYMBOL_GPL(vtime_account_irq_enter); |
| 504 | #endif /* __ARCH_HAS_VTIME_ACCOUNT */ | 520 | #endif /* __ARCH_HAS_VTIME_ACCOUNT */ |
| 505 | 521 | ||
| 506 | #else | 522 | #else /* !CONFIG_VIRT_CPU_ACCOUNTING */ |
| 507 | |||
| 508 | #ifndef nsecs_to_cputime | ||
| 509 | # define nsecs_to_cputime(__nsecs) nsecs_to_jiffies(__nsecs) | ||
| 510 | #endif | ||
| 511 | 523 | ||
| 512 | static cputime_t scale_utime(cputime_t utime, cputime_t rtime, cputime_t total) | 524 | static cputime_t scale_stime(cputime_t stime, cputime_t rtime, cputime_t total) |
| 513 | { | 525 | { |
| 514 | u64 temp = (__force u64) rtime; | 526 | u64 temp = (__force u64) rtime; |
| 515 | 527 | ||
| 516 | temp *= (__force u64) utime; | 528 | temp *= (__force u64) stime; |
| 517 | 529 | ||
| 518 | if (sizeof(cputime_t) == 4) | 530 | if (sizeof(cputime_t) == 4) |
| 519 | temp = div_u64(temp, (__force u32) total); | 531 | temp = div_u64(temp, (__force u32) total); |
| @@ -531,10 +543,10 @@ static void cputime_adjust(struct task_cputime *curr, | |||
| 531 | struct cputime *prev, | 543 | struct cputime *prev, |
| 532 | cputime_t *ut, cputime_t *st) | 544 | cputime_t *ut, cputime_t *st) |
| 533 | { | 545 | { |
| 534 | cputime_t rtime, utime, total; | 546 | cputime_t rtime, stime, total; |
| 535 | 547 | ||
| 536 | utime = curr->utime; | 548 | stime = curr->stime; |
| 537 | total = utime + curr->stime; | 549 | total = stime + curr->utime; |
| 538 | 550 | ||
| 539 | /* | 551 | /* |
| 540 | * Tick based cputime accounting depend on random scheduling | 552 | * Tick based cputime accounting depend on random scheduling |
| @@ -549,17 +561,17 @@ static void cputime_adjust(struct task_cputime *curr, | |||
| 549 | rtime = nsecs_to_cputime(curr->sum_exec_runtime); | 561 | rtime = nsecs_to_cputime(curr->sum_exec_runtime); |
| 550 | 562 | ||
| 551 | if (total) | 563 | if (total) |
| 552 | utime = scale_utime(utime, rtime, total); | 564 | stime = scale_stime(stime, rtime, total); |
| 553 | else | 565 | else |
| 554 | utime = rtime; | 566 | stime = rtime; |
| 555 | 567 | ||
| 556 | /* | 568 | /* |
| 557 | * If the tick based count grows faster than the scheduler one, | 569 | * If the tick based count grows faster than the scheduler one, |
| 558 | * the result of the scaling may go backward. | 570 | * the result of the scaling may go backward. |
| 559 | * Let's enforce monotonicity. | 571 | * Let's enforce monotonicity. |
| 560 | */ | 572 | */ |
| 561 | prev->utime = max(prev->utime, utime); | 573 | prev->stime = max(prev->stime, stime); |
| 562 | prev->stime = max(prev->stime, rtime - prev->utime); | 574 | prev->utime = max(prev->utime, rtime - prev->stime); |
| 563 | 575 | ||
| 564 | *ut = prev->utime; | 576 | *ut = prev->utime; |
| 565 | *st = prev->stime; | 577 | *st = prev->stime; |
| @@ -568,11 +580,10 @@ static void cputime_adjust(struct task_cputime *curr, | |||
| 568 | void task_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st) | 580 | void task_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime_t *st) |
| 569 | { | 581 | { |
| 570 | struct task_cputime cputime = { | 582 | struct task_cputime cputime = { |
| 571 | .utime = p->utime, | ||
| 572 | .stime = p->stime, | ||
| 573 | .sum_exec_runtime = p->se.sum_exec_runtime, | 583 | .sum_exec_runtime = p->se.sum_exec_runtime, |
| 574 | }; | 584 | }; |
| 575 | 585 | ||
| 586 | task_cputime(p, &cputime.utime, &cputime.stime); | ||
| 576 | cputime_adjust(&cputime, &p->prev_cputime, ut, st); | 587 | cputime_adjust(&cputime, &p->prev_cputime, ut, st); |
| 577 | } | 588 | } |
| 578 | 589 | ||
| @@ -586,4 +597,221 @@ void thread_group_cputime_adjusted(struct task_struct *p, cputime_t *ut, cputime | |||
| 586 | thread_group_cputime(p, &cputime); | 597 | thread_group_cputime(p, &cputime); |
| 587 | cputime_adjust(&cputime, &p->signal->prev_cputime, ut, st); | 598 | cputime_adjust(&cputime, &p->signal->prev_cputime, ut, st); |
| 588 | } | 599 | } |
| 589 | #endif | 600 | #endif /* !CONFIG_VIRT_CPU_ACCOUNTING */ |
| 601 | |||
| 602 | #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN | ||
| 603 | static unsigned long long vtime_delta(struct task_struct *tsk) | ||
| 604 | { | ||
| 605 | unsigned long long clock; | ||
| 606 | |||
| 607 | clock = sched_clock(); | ||
| 608 | if (clock < tsk->vtime_snap) | ||
| 609 | return 0; | ||
| 610 | |||
| 611 | return clock - tsk->vtime_snap; | ||
| 612 | } | ||
| 613 | |||
| 614 | static cputime_t get_vtime_delta(struct task_struct *tsk) | ||
| 615 | { | ||
| 616 | unsigned long long delta = vtime_delta(tsk); | ||
| 617 | |||
| 618 | WARN_ON_ONCE(tsk->vtime_snap_whence == VTIME_SLEEPING); | ||
| 619 | tsk->vtime_snap += delta; | ||
| 620 | |||
| 621 | /* CHECKME: always safe to convert nsecs to cputime? */ | ||
| 622 | return nsecs_to_cputime(delta); | ||
| 623 | } | ||
| 624 | |||
| 625 | static void __vtime_account_system(struct task_struct *tsk) | ||
| 626 | { | ||
| 627 | cputime_t delta_cpu = get_vtime_delta(tsk); | ||
| 628 | |||
| 629 | account_system_time(tsk, irq_count(), delta_cpu, cputime_to_scaled(delta_cpu)); | ||
| 630 | } | ||
| 631 | |||
| 632 | void vtime_account_system(struct task_struct *tsk) | ||
| 633 | { | ||
| 634 | if (!vtime_accounting_enabled()) | ||
| 635 | return; | ||
| 636 | |||
| 637 | write_seqlock(&tsk->vtime_seqlock); | ||
| 638 | __vtime_account_system(tsk); | ||
| 639 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 640 | } | ||
| 641 | |||
| 642 | void vtime_account_irq_exit(struct task_struct *tsk) | ||
| 643 | { | ||
| 644 | if (!vtime_accounting_enabled()) | ||
| 645 | return; | ||
| 646 | |||
| 647 | write_seqlock(&tsk->vtime_seqlock); | ||
| 648 | if (context_tracking_in_user()) | ||
| 649 | tsk->vtime_snap_whence = VTIME_USER; | ||
| 650 | __vtime_account_system(tsk); | ||
| 651 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 652 | } | ||
| 653 | |||
| 654 | void vtime_account_user(struct task_struct *tsk) | ||
| 655 | { | ||
| 656 | cputime_t delta_cpu; | ||
| 657 | |||
| 658 | if (!vtime_accounting_enabled()) | ||
| 659 | return; | ||
| 660 | |||
| 661 | delta_cpu = get_vtime_delta(tsk); | ||
| 662 | |||
| 663 | write_seqlock(&tsk->vtime_seqlock); | ||
| 664 | tsk->vtime_snap_whence = VTIME_SYS; | ||
| 665 | account_user_time(tsk, delta_cpu, cputime_to_scaled(delta_cpu)); | ||
| 666 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 667 | } | ||
| 668 | |||
| 669 | void vtime_user_enter(struct task_struct *tsk) | ||
| 670 | { | ||
| 671 | if (!vtime_accounting_enabled()) | ||
| 672 | return; | ||
| 673 | |||
| 674 | write_seqlock(&tsk->vtime_seqlock); | ||
| 675 | tsk->vtime_snap_whence = VTIME_USER; | ||
| 676 | __vtime_account_system(tsk); | ||
| 677 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 678 | } | ||
| 679 | |||
| 680 | void vtime_guest_enter(struct task_struct *tsk) | ||
| 681 | { | ||
| 682 | write_seqlock(&tsk->vtime_seqlock); | ||
| 683 | __vtime_account_system(tsk); | ||
| 684 | current->flags |= PF_VCPU; | ||
| 685 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 686 | } | ||
| 687 | |||
| 688 | void vtime_guest_exit(struct task_struct *tsk) | ||
| 689 | { | ||
| 690 | write_seqlock(&tsk->vtime_seqlock); | ||
| 691 | __vtime_account_system(tsk); | ||
| 692 | current->flags &= ~PF_VCPU; | ||
| 693 | write_sequnlock(&tsk->vtime_seqlock); | ||
| 694 | } | ||
| 695 | |||
| 696 | void vtime_account_idle(struct task_struct *tsk) | ||
| 697 | { | ||
| 698 | cputime_t delta_cpu = get_vtime_delta(tsk); | ||
| 699 | |||
| 700 | account_idle_time(delta_cpu); | ||
| 701 | } | ||
| 702 | |||
| 703 | bool vtime_accounting_enabled(void) | ||
| 704 | { | ||
| 705 | return context_tracking_active(); | ||
| 706 | } | ||
| 707 | |||
| 708 | void arch_vtime_task_switch(struct task_struct *prev) | ||
| 709 | { | ||
| 710 | write_seqlock(&prev->vtime_seqlock); | ||
| 711 | prev->vtime_snap_whence = VTIME_SLEEPING; | ||
| 712 | write_sequnlock(&prev->vtime_seqlock); | ||
| 713 | |||
| 714 | write_seqlock(¤t->vtime_seqlock); | ||
| 715 | current->vtime_snap_whence = VTIME_SYS; | ||
| 716 | current->vtime_snap = sched_clock(); | ||
| 717 | write_sequnlock(¤t->vtime_seqlock); | ||
| 718 | } | ||
| 719 | |||
| 720 | void vtime_init_idle(struct task_struct *t) | ||
| 721 | { | ||
| 722 | unsigned long flags; | ||
| 723 | |||
| 724 | write_seqlock_irqsave(&t->vtime_seqlock, flags); | ||
| 725 | t->vtime_snap_whence = VTIME_SYS; | ||
| 726 | t->vtime_snap = sched_clock(); | ||
| 727 | write_sequnlock_irqrestore(&t->vtime_seqlock, flags); | ||
| 728 | } | ||
| 729 | |||
| 730 | cputime_t task_gtime(struct task_struct *t) | ||
| 731 | { | ||
| 732 | unsigned int seq; | ||
| 733 | cputime_t gtime; | ||
| 734 | |||
| 735 | do { | ||
| 736 | seq = read_seqbegin(&t->vtime_seqlock); | ||
| 737 | |||
| 738 | gtime = t->gtime; | ||
| 739 | if (t->flags & PF_VCPU) | ||
| 740 | gtime += vtime_delta(t); | ||
| 741 | |||
| 742 | } while (read_seqretry(&t->vtime_seqlock, seq)); | ||
| 743 | |||
| 744 | return gtime; | ||
| 745 | } | ||
| 746 | |||
| 747 | /* | ||
| 748 | * Fetch cputime raw values from fields of task_struct and | ||
| 749 | * add up the pending nohz execution time since the last | ||
| 750 | * cputime snapshot. | ||
| 751 | */ | ||
| 752 | static void | ||
| 753 | fetch_task_cputime(struct task_struct *t, | ||
| 754 | cputime_t *u_dst, cputime_t *s_dst, | ||
| 755 | cputime_t *u_src, cputime_t *s_src, | ||
| 756 | cputime_t *udelta, cputime_t *sdelta) | ||
| 757 | { | ||
| 758 | unsigned int seq; | ||
| 759 | unsigned long long delta; | ||
| 760 | |||
| 761 | do { | ||
| 762 | *udelta = 0; | ||
| 763 | *sdelta = 0; | ||
| 764 | |||
| 765 | seq = read_seqbegin(&t->vtime_seqlock); | ||
| 766 | |||
| 767 | if (u_dst) | ||
| 768 | *u_dst = *u_src; | ||
| 769 | if (s_dst) | ||
| 770 | *s_dst = *s_src; | ||
| 771 | |||
| 772 | /* Task is sleeping, nothing to add */ | ||
| 773 | if (t->vtime_snap_whence == VTIME_SLEEPING || | ||
| 774 | is_idle_task(t)) | ||
| 775 | continue; | ||
| 776 | |||
| 777 | delta = vtime_delta(t); | ||
| 778 | |||
| 779 | /* | ||
| 780 | * Task runs either in user or kernel space, add pending nohz time to | ||
| 781 | * the right place. | ||
| 782 | */ | ||
| 783 | if (t->vtime_snap_whence == VTIME_USER || t->flags & PF_VCPU) { | ||
| 784 | *udelta = delta; | ||
| 785 | } else { | ||
| 786 | if (t->vtime_snap_whence == VTIME_SYS) | ||
| 787 | *sdelta = delta; | ||
| 788 | } | ||
| 789 | } while (read_seqretry(&t->vtime_seqlock, seq)); | ||
| 790 | } | ||
| 791 | |||
| 792 | |||
| 793 | void task_cputime(struct task_struct *t, cputime_t *utime, cputime_t *stime) | ||
| 794 | { | ||
| 795 | cputime_t udelta, sdelta; | ||
| 796 | |||
| 797 | fetch_task_cputime(t, utime, stime, &t->utime, | ||
| 798 | &t->stime, &udelta, &sdelta); | ||
| 799 | if (utime) | ||
| 800 | *utime += udelta; | ||
| 801 | if (stime) | ||
| 802 | *stime += sdelta; | ||
| 803 | } | ||
| 804 | |||
| 805 | void task_cputime_scaled(struct task_struct *t, | ||
| 806 | cputime_t *utimescaled, cputime_t *stimescaled) | ||
| 807 | { | ||
| 808 | cputime_t udelta, sdelta; | ||
| 809 | |||
| 810 | fetch_task_cputime(t, utimescaled, stimescaled, | ||
| 811 | &t->utimescaled, &t->stimescaled, &udelta, &sdelta); | ||
| 812 | if (utimescaled) | ||
| 813 | *utimescaled += cputime_to_scaled(udelta); | ||
| 814 | if (stimescaled) | ||
| 815 | *stimescaled += cputime_to_scaled(sdelta); | ||
| 816 | } | ||
| 817 | #endif /* CONFIG_VIRT_CPU_ACCOUNTING_GEN */ | ||
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 81fa53643409..7a33e5986fc5 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
| @@ -1680,9 +1680,7 @@ place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int initial) | |||
| 1680 | } | 1680 | } |
| 1681 | 1681 | ||
| 1682 | /* ensure we never gain time by being placed backwards. */ | 1682 | /* ensure we never gain time by being placed backwards. */ |
| 1683 | vruntime = max_vruntime(se->vruntime, vruntime); | 1683 | se->vruntime = max_vruntime(se->vruntime, vruntime); |
| 1684 | |||
| 1685 | se->vruntime = vruntime; | ||
| 1686 | } | 1684 | } |
| 1687 | 1685 | ||
| 1688 | static void check_enqueue_throttle(struct cfs_rq *cfs_rq); | 1686 | static void check_enqueue_throttle(struct cfs_rq *cfs_rq); |
| @@ -3254,25 +3252,18 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) | |||
| 3254 | */ | 3252 | */ |
| 3255 | static int select_idle_sibling(struct task_struct *p, int target) | 3253 | static int select_idle_sibling(struct task_struct *p, int target) |
| 3256 | { | 3254 | { |
| 3257 | int cpu = smp_processor_id(); | ||
| 3258 | int prev_cpu = task_cpu(p); | ||
| 3259 | struct sched_domain *sd; | 3255 | struct sched_domain *sd; |
| 3260 | struct sched_group *sg; | 3256 | struct sched_group *sg; |
| 3261 | int i; | 3257 | int i = task_cpu(p); |
| 3262 | 3258 | ||
| 3263 | /* | 3259 | if (idle_cpu(target)) |
| 3264 | * If the task is going to be woken-up on this cpu and if it is | 3260 | return target; |
| 3265 | * already idle, then it is the right target. | ||
| 3266 | */ | ||
| 3267 | if (target == cpu && idle_cpu(cpu)) | ||
| 3268 | return cpu; | ||
| 3269 | 3261 | ||
| 3270 | /* | 3262 | /* |
| 3271 | * If the task is going to be woken-up on the cpu where it previously | 3263 | * If the prevous cpu is cache affine and idle, don't be stupid. |
| 3272 | * ran and if it is currently idle, then it the right target. | ||
| 3273 | */ | 3264 | */ |
| 3274 | if (target == prev_cpu && idle_cpu(prev_cpu)) | 3265 | if (i != target && cpus_share_cache(i, target) && idle_cpu(i)) |
| 3275 | return prev_cpu; | 3266 | return i; |
| 3276 | 3267 | ||
| 3277 | /* | 3268 | /* |
| 3278 | * Otherwise, iterate the domains and find an elegible idle cpu. | 3269 | * Otherwise, iterate the domains and find an elegible idle cpu. |
| @@ -3286,7 +3277,7 @@ static int select_idle_sibling(struct task_struct *p, int target) | |||
| 3286 | goto next; | 3277 | goto next; |
| 3287 | 3278 | ||
| 3288 | for_each_cpu(i, sched_group_cpus(sg)) { | 3279 | for_each_cpu(i, sched_group_cpus(sg)) { |
| 3289 | if (!idle_cpu(i)) | 3280 | if (i == target || !idle_cpu(i)) |
| 3290 | goto next; | 3281 | goto next; |
| 3291 | } | 3282 | } |
| 3292 | 3283 | ||
| @@ -6101,7 +6092,7 @@ static unsigned int get_rr_interval_fair(struct rq *rq, struct task_struct *task | |||
| 6101 | * idle runqueue: | 6092 | * idle runqueue: |
| 6102 | */ | 6093 | */ |
| 6103 | if (rq->cfs.load.weight) | 6094 | if (rq->cfs.load.weight) |
| 6104 | rr_interval = NS_TO_JIFFIES(sched_slice(&rq->cfs, se)); | 6095 | rr_interval = NS_TO_JIFFIES(sched_slice(cfs_rq_of(se), se)); |
| 6105 | 6096 | ||
| 6106 | return rr_interval; | 6097 | return rr_interval; |
| 6107 | } | 6098 | } |
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c index 4f02b2847357..127a2c4cf4ab 100644 --- a/kernel/sched/rt.c +++ b/kernel/sched/rt.c | |||
| @@ -7,6 +7,8 @@ | |||
| 7 | 7 | ||
| 8 | #include <linux/slab.h> | 8 | #include <linux/slab.h> |
| 9 | 9 | ||
| 10 | int sched_rr_timeslice = RR_TIMESLICE; | ||
| 11 | |||
| 10 | static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun); | 12 | static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun); |
| 11 | 13 | ||
| 12 | struct rt_bandwidth def_rt_bandwidth; | 14 | struct rt_bandwidth def_rt_bandwidth; |
| @@ -925,8 +927,8 @@ static void update_curr_rt(struct rq *rq) | |||
| 925 | return; | 927 | return; |
| 926 | 928 | ||
| 927 | delta_exec = rq->clock_task - curr->se.exec_start; | 929 | delta_exec = rq->clock_task - curr->se.exec_start; |
| 928 | if (unlikely((s64)delta_exec < 0)) | 930 | if (unlikely((s64)delta_exec <= 0)) |
| 929 | delta_exec = 0; | 931 | return; |
| 930 | 932 | ||
| 931 | schedstat_set(curr->se.statistics.exec_max, | 933 | schedstat_set(curr->se.statistics.exec_max, |
| 932 | max(curr->se.statistics.exec_max, delta_exec)); | 934 | max(curr->se.statistics.exec_max, delta_exec)); |
| @@ -1427,8 +1429,7 @@ static void put_prev_task_rt(struct rq *rq, struct task_struct *p) | |||
| 1427 | static int pick_rt_task(struct rq *rq, struct task_struct *p, int cpu) | 1429 | static int pick_rt_task(struct rq *rq, struct task_struct *p, int cpu) |
| 1428 | { | 1430 | { |
| 1429 | if (!task_running(rq, p) && | 1431 | if (!task_running(rq, p) && |
| 1430 | (cpu < 0 || cpumask_test_cpu(cpu, tsk_cpus_allowed(p))) && | 1432 | cpumask_test_cpu(cpu, tsk_cpus_allowed(p))) |
| 1431 | (p->nr_cpus_allowed > 1)) | ||
| 1432 | return 1; | 1433 | return 1; |
| 1433 | return 0; | 1434 | return 0; |
| 1434 | } | 1435 | } |
| @@ -1889,8 +1890,11 @@ static void switched_from_rt(struct rq *rq, struct task_struct *p) | |||
| 1889 | * we may need to handle the pulling of RT tasks | 1890 | * we may need to handle the pulling of RT tasks |
| 1890 | * now. | 1891 | * now. |
| 1891 | */ | 1892 | */ |
| 1892 | if (p->on_rq && !rq->rt.rt_nr_running) | 1893 | if (!p->on_rq || rq->rt.rt_nr_running) |
| 1893 | pull_rt_task(rq); | 1894 | return; |
| 1895 | |||
| 1896 | if (pull_rt_task(rq)) | ||
| 1897 | resched_task(rq->curr); | ||
| 1894 | } | 1898 | } |
| 1895 | 1899 | ||
| 1896 | void init_sched_rt_class(void) | 1900 | void init_sched_rt_class(void) |
| @@ -1985,7 +1989,11 @@ static void watchdog(struct rq *rq, struct task_struct *p) | |||
| 1985 | if (soft != RLIM_INFINITY) { | 1989 | if (soft != RLIM_INFINITY) { |
| 1986 | unsigned long next; | 1990 | unsigned long next; |
| 1987 | 1991 | ||
| 1988 | p->rt.timeout++; | 1992 | if (p->rt.watchdog_stamp != jiffies) { |
| 1993 | p->rt.timeout++; | ||
| 1994 | p->rt.watchdog_stamp = jiffies; | ||
| 1995 | } | ||
| 1996 | |||
| 1989 | next = DIV_ROUND_UP(min(soft, hard), USEC_PER_SEC/HZ); | 1997 | next = DIV_ROUND_UP(min(soft, hard), USEC_PER_SEC/HZ); |
| 1990 | if (p->rt.timeout > next) | 1998 | if (p->rt.timeout > next) |
| 1991 | p->cputime_expires.sched_exp = p->se.sum_exec_runtime; | 1999 | p->cputime_expires.sched_exp = p->se.sum_exec_runtime; |
| @@ -2010,7 +2018,7 @@ static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued) | |||
| 2010 | if (--p->rt.time_slice) | 2018 | if (--p->rt.time_slice) |
| 2011 | return; | 2019 | return; |
| 2012 | 2020 | ||
| 2013 | p->rt.time_slice = RR_TIMESLICE; | 2021 | p->rt.time_slice = sched_rr_timeslice; |
| 2014 | 2022 | ||
| 2015 | /* | 2023 | /* |
| 2016 | * Requeue to the end of queue if we (and all of our ancestors) are the | 2024 | * Requeue to the end of queue if we (and all of our ancestors) are the |
| @@ -2041,7 +2049,7 @@ static unsigned int get_rr_interval_rt(struct rq *rq, struct task_struct *task) | |||
| 2041 | * Time slice is 0 for SCHED_FIFO tasks | 2049 | * Time slice is 0 for SCHED_FIFO tasks |
| 2042 | */ | 2050 | */ |
| 2043 | if (task->policy == SCHED_RR) | 2051 | if (task->policy == SCHED_RR) |
| 2044 | return RR_TIMESLICE; | 2052 | return sched_rr_timeslice; |
| 2045 | else | 2053 | else |
| 2046 | return 0; | 2054 | return 0; |
| 2047 | } | 2055 | } |
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index fc886441436a..cc03cfdf469f 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h | |||
| @@ -1,5 +1,7 @@ | |||
| 1 | 1 | ||
| 2 | #include <linux/sched.h> | 2 | #include <linux/sched.h> |
| 3 | #include <linux/sched/sysctl.h> | ||
| 4 | #include <linux/sched/rt.h> | ||
| 3 | #include <linux/mutex.h> | 5 | #include <linux/mutex.h> |
| 4 | #include <linux/spinlock.h> | 6 | #include <linux/spinlock.h> |
| 5 | #include <linux/stop_machine.h> | 7 | #include <linux/stop_machine.h> |
diff --git a/kernel/signal.c b/kernel/signal.c index 3d09cf6cde75..7f82adbad480 100644 --- a/kernel/signal.c +++ b/kernel/signal.c | |||
| @@ -1632,6 +1632,7 @@ bool do_notify_parent(struct task_struct *tsk, int sig) | |||
| 1632 | unsigned long flags; | 1632 | unsigned long flags; |
| 1633 | struct sighand_struct *psig; | 1633 | struct sighand_struct *psig; |
| 1634 | bool autoreap = false; | 1634 | bool autoreap = false; |
| 1635 | cputime_t utime, stime; | ||
| 1635 | 1636 | ||
| 1636 | BUG_ON(sig == -1); | 1637 | BUG_ON(sig == -1); |
| 1637 | 1638 | ||
| @@ -1669,8 +1670,9 @@ bool do_notify_parent(struct task_struct *tsk, int sig) | |||
| 1669 | task_uid(tsk)); | 1670 | task_uid(tsk)); |
| 1670 | rcu_read_unlock(); | 1671 | rcu_read_unlock(); |
| 1671 | 1672 | ||
| 1672 | info.si_utime = cputime_to_clock_t(tsk->utime + tsk->signal->utime); | 1673 | task_cputime(tsk, &utime, &stime); |
| 1673 | info.si_stime = cputime_to_clock_t(tsk->stime + tsk->signal->stime); | 1674 | info.si_utime = cputime_to_clock_t(utime + tsk->signal->utime); |
| 1675 | info.si_stime = cputime_to_clock_t(stime + tsk->signal->stime); | ||
| 1674 | 1676 | ||
| 1675 | info.si_status = tsk->exit_code & 0x7f; | 1677 | info.si_status = tsk->exit_code & 0x7f; |
| 1676 | if (tsk->exit_code & 0x80) | 1678 | if (tsk->exit_code & 0x80) |
| @@ -1734,6 +1736,7 @@ static void do_notify_parent_cldstop(struct task_struct *tsk, | |||
| 1734 | unsigned long flags; | 1736 | unsigned long flags; |
| 1735 | struct task_struct *parent; | 1737 | struct task_struct *parent; |
| 1736 | struct sighand_struct *sighand; | 1738 | struct sighand_struct *sighand; |
| 1739 | cputime_t utime, stime; | ||
| 1737 | 1740 | ||
| 1738 | if (for_ptracer) { | 1741 | if (for_ptracer) { |
| 1739 | parent = tsk->parent; | 1742 | parent = tsk->parent; |
| @@ -1752,8 +1755,9 @@ static void do_notify_parent_cldstop(struct task_struct *tsk, | |||
| 1752 | info.si_uid = from_kuid_munged(task_cred_xxx(parent, user_ns), task_uid(tsk)); | 1755 | info.si_uid = from_kuid_munged(task_cred_xxx(parent, user_ns), task_uid(tsk)); |
| 1753 | rcu_read_unlock(); | 1756 | rcu_read_unlock(); |
| 1754 | 1757 | ||
| 1755 | info.si_utime = cputime_to_clock_t(tsk->utime); | 1758 | task_cputime(tsk, &utime, &stime); |
| 1756 | info.si_stime = cputime_to_clock_t(tsk->stime); | 1759 | info.si_utime = cputime_to_clock_t(utime); |
| 1760 | info.si_stime = cputime_to_clock_t(stime); | ||
| 1757 | 1761 | ||
| 1758 | info.si_code = why; | 1762 | info.si_code = why; |
| 1759 | switch (why) { | 1763 | switch (why) { |
diff --git a/kernel/softirq.c b/kernel/softirq.c index ed567babe789..f5cc25f147a6 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c | |||
| @@ -221,7 +221,7 @@ asmlinkage void __do_softirq(void) | |||
| 221 | current->flags &= ~PF_MEMALLOC; | 221 | current->flags &= ~PF_MEMALLOC; |
| 222 | 222 | ||
| 223 | pending = local_softirq_pending(); | 223 | pending = local_softirq_pending(); |
| 224 | vtime_account_irq_enter(current); | 224 | account_irq_enter_time(current); |
| 225 | 225 | ||
| 226 | __local_bh_disable((unsigned long)__builtin_return_address(0), | 226 | __local_bh_disable((unsigned long)__builtin_return_address(0), |
| 227 | SOFTIRQ_OFFSET); | 227 | SOFTIRQ_OFFSET); |
| @@ -272,7 +272,7 @@ restart: | |||
| 272 | 272 | ||
| 273 | lockdep_softirq_exit(); | 273 | lockdep_softirq_exit(); |
| 274 | 274 | ||
| 275 | vtime_account_irq_exit(current); | 275 | account_irq_exit_time(current); |
| 276 | __local_bh_enable(SOFTIRQ_OFFSET); | 276 | __local_bh_enable(SOFTIRQ_OFFSET); |
| 277 | tsk_restore_flags(current, old_flags, PF_MEMALLOC); | 277 | tsk_restore_flags(current, old_flags, PF_MEMALLOC); |
| 278 | } | 278 | } |
| @@ -341,7 +341,7 @@ static inline void invoke_softirq(void) | |||
| 341 | */ | 341 | */ |
| 342 | void irq_exit(void) | 342 | void irq_exit(void) |
| 343 | { | 343 | { |
| 344 | vtime_account_irq_exit(current); | 344 | account_irq_exit_time(current); |
| 345 | trace_hardirq_exit(); | 345 | trace_hardirq_exit(); |
| 346 | sub_preempt_count(IRQ_EXIT_OFFSET); | 346 | sub_preempt_count(IRQ_EXIT_OFFSET); |
| 347 | if (!in_interrupt() && local_softirq_pending()) | 347 | if (!in_interrupt() && local_softirq_pending()) |
diff --git a/kernel/sysctl.c b/kernel/sysctl.c index c88878db491e..4fc9be955c71 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c | |||
| @@ -61,6 +61,7 @@ | |||
| 61 | #include <linux/kmod.h> | 61 | #include <linux/kmod.h> |
| 62 | #include <linux/capability.h> | 62 | #include <linux/capability.h> |
| 63 | #include <linux/binfmts.h> | 63 | #include <linux/binfmts.h> |
| 64 | #include <linux/sched/sysctl.h> | ||
| 64 | 65 | ||
| 65 | #include <asm/uaccess.h> | 66 | #include <asm/uaccess.h> |
| 66 | #include <asm/processor.h> | 67 | #include <asm/processor.h> |
| @@ -403,6 +404,13 @@ static struct ctl_table kern_table[] = { | |||
| 403 | .mode = 0644, | 404 | .mode = 0644, |
| 404 | .proc_handler = sched_rt_handler, | 405 | .proc_handler = sched_rt_handler, |
| 405 | }, | 406 | }, |
| 407 | { | ||
| 408 | .procname = "sched_rr_timeslice_ms", | ||
| 409 | .data = &sched_rr_timeslice, | ||
| 410 | .maxlen = sizeof(int), | ||
| 411 | .mode = 0644, | ||
| 412 | .proc_handler = sched_rr_handler, | ||
| 413 | }, | ||
| 406 | #ifdef CONFIG_SCHED_AUTOGROUP | 414 | #ifdef CONFIG_SCHED_AUTOGROUP |
| 407 | { | 415 | { |
| 408 | .procname = "sched_autogroup_enabled", | 416 | .procname = "sched_autogroup_enabled", |
diff --git a/kernel/time/tick-sched.c b/kernel/time/tick-sched.c index fb8e5e469d1c..314b9ee07edf 100644 --- a/kernel/time/tick-sched.c +++ b/kernel/time/tick-sched.c | |||
| @@ -632,8 +632,11 @@ static void tick_nohz_restart_sched_tick(struct tick_sched *ts, ktime_t now) | |||
| 632 | 632 | ||
| 633 | static void tick_nohz_account_idle_ticks(struct tick_sched *ts) | 633 | static void tick_nohz_account_idle_ticks(struct tick_sched *ts) |
| 634 | { | 634 | { |
| 635 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING | 635 | #ifndef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE |
| 636 | unsigned long ticks; | 636 | unsigned long ticks; |
| 637 | |||
| 638 | if (vtime_accounting_enabled()) | ||
| 639 | return; | ||
| 637 | /* | 640 | /* |
| 638 | * We stopped the tick in idle. Update process times would miss the | 641 | * We stopped the tick in idle. Update process times would miss the |
| 639 | * time we slept as update_process_times does only a 1 tick | 642 | * time we slept as update_process_times does only a 1 tick |
diff --git a/kernel/timer.c b/kernel/timer.c index ff3b5165737b..dbf7a78a1ef1 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
| @@ -39,6 +39,7 @@ | |||
| 39 | #include <linux/kallsyms.h> | 39 | #include <linux/kallsyms.h> |
| 40 | #include <linux/irq_work.h> | 40 | #include <linux/irq_work.h> |
| 41 | #include <linux/sched.h> | 41 | #include <linux/sched.h> |
| 42 | #include <linux/sched/sysctl.h> | ||
| 42 | #include <linux/slab.h> | 43 | #include <linux/slab.h> |
| 43 | 44 | ||
| 44 | #include <asm/uaccess.h> | 45 | #include <asm/uaccess.h> |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 5d520b7bb4c5..c2e2c2310374 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -39,6 +39,7 @@ | |||
| 39 | #include <linux/poll.h> | 39 | #include <linux/poll.h> |
| 40 | #include <linux/nmi.h> | 40 | #include <linux/nmi.h> |
| 41 | #include <linux/fs.h> | 41 | #include <linux/fs.h> |
| 42 | #include <linux/sched/rt.h> | ||
| 42 | 43 | ||
| 43 | #include "trace.h" | 44 | #include "trace.h" |
| 44 | #include "trace_output.h" | 45 | #include "trace_output.h" |
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index 9fe45fcefca0..75aa97fbe1a1 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c | |||
| @@ -15,8 +15,8 @@ | |||
| 15 | #include <linux/kallsyms.h> | 15 | #include <linux/kallsyms.h> |
| 16 | #include <linux/uaccess.h> | 16 | #include <linux/uaccess.h> |
| 17 | #include <linux/ftrace.h> | 17 | #include <linux/ftrace.h> |
| 18 | #include <linux/sched/rt.h> | ||
| 18 | #include <trace/events/sched.h> | 19 | #include <trace/events/sched.h> |
| 19 | |||
| 20 | #include "trace.h" | 20 | #include "trace.h" |
| 21 | 21 | ||
| 22 | static struct trace_array *wakeup_trace; | 22 | static struct trace_array *wakeup_trace; |
diff --git a/kernel/tsacct.c b/kernel/tsacct.c index 625df0b44690..a1dd9a1b1327 100644 --- a/kernel/tsacct.c +++ b/kernel/tsacct.c | |||
| @@ -32,6 +32,7 @@ void bacct_add_tsk(struct user_namespace *user_ns, | |||
| 32 | { | 32 | { |
| 33 | const struct cred *tcred; | 33 | const struct cred *tcred; |
| 34 | struct timespec uptime, ts; | 34 | struct timespec uptime, ts; |
| 35 | cputime_t utime, stime, utimescaled, stimescaled; | ||
| 35 | u64 ac_etime; | 36 | u64 ac_etime; |
| 36 | 37 | ||
| 37 | BUILD_BUG_ON(TS_COMM_LEN < TASK_COMM_LEN); | 38 | BUILD_BUG_ON(TS_COMM_LEN < TASK_COMM_LEN); |
| @@ -65,10 +66,15 @@ void bacct_add_tsk(struct user_namespace *user_ns, | |||
| 65 | stats->ac_ppid = pid_alive(tsk) ? | 66 | stats->ac_ppid = pid_alive(tsk) ? |
| 66 | task_tgid_nr_ns(rcu_dereference(tsk->real_parent), pid_ns) : 0; | 67 | task_tgid_nr_ns(rcu_dereference(tsk->real_parent), pid_ns) : 0; |
| 67 | rcu_read_unlock(); | 68 | rcu_read_unlock(); |
| 68 | stats->ac_utime = cputime_to_usecs(tsk->utime); | 69 | |
| 69 | stats->ac_stime = cputime_to_usecs(tsk->stime); | 70 | task_cputime(tsk, &utime, &stime); |
| 70 | stats->ac_utimescaled = cputime_to_usecs(tsk->utimescaled); | 71 | stats->ac_utime = cputime_to_usecs(utime); |
| 71 | stats->ac_stimescaled = cputime_to_usecs(tsk->stimescaled); | 72 | stats->ac_stime = cputime_to_usecs(stime); |
| 73 | |||
| 74 | task_cputime_scaled(tsk, &utimescaled, &stimescaled); | ||
| 75 | stats->ac_utimescaled = cputime_to_usecs(utimescaled); | ||
| 76 | stats->ac_stimescaled = cputime_to_usecs(stimescaled); | ||
| 77 | |||
| 72 | stats->ac_minflt = tsk->min_flt; | 78 | stats->ac_minflt = tsk->min_flt; |
| 73 | stats->ac_majflt = tsk->maj_flt; | 79 | stats->ac_majflt = tsk->maj_flt; |
| 74 | 80 | ||
| @@ -115,11 +121,8 @@ void xacct_add_tsk(struct taskstats *stats, struct task_struct *p) | |||
| 115 | #undef KB | 121 | #undef KB |
| 116 | #undef MB | 122 | #undef MB |
| 117 | 123 | ||
| 118 | /** | 124 | static void __acct_update_integrals(struct task_struct *tsk, |
| 119 | * acct_update_integrals - update mm integral fields in task_struct | 125 | cputime_t utime, cputime_t stime) |
| 120 | * @tsk: task_struct for accounting | ||
| 121 | */ | ||
| 122 | void acct_update_integrals(struct task_struct *tsk) | ||
| 123 | { | 126 | { |
| 124 | if (likely(tsk->mm)) { | 127 | if (likely(tsk->mm)) { |
| 125 | cputime_t time, dtime; | 128 | cputime_t time, dtime; |
| @@ -128,7 +131,7 @@ void acct_update_integrals(struct task_struct *tsk) | |||
| 128 | u64 delta; | 131 | u64 delta; |
| 129 | 132 | ||
| 130 | local_irq_save(flags); | 133 | local_irq_save(flags); |
| 131 | time = tsk->stime + tsk->utime; | 134 | time = stime + utime; |
| 132 | dtime = time - tsk->acct_timexpd; | 135 | dtime = time - tsk->acct_timexpd; |
| 133 | jiffies_to_timeval(cputime_to_jiffies(dtime), &value); | 136 | jiffies_to_timeval(cputime_to_jiffies(dtime), &value); |
| 134 | delta = value.tv_sec; | 137 | delta = value.tv_sec; |
| @@ -145,6 +148,27 @@ void acct_update_integrals(struct task_struct *tsk) | |||
| 145 | } | 148 | } |
| 146 | 149 | ||
| 147 | /** | 150 | /** |
| 151 | * acct_update_integrals - update mm integral fields in task_struct | ||
| 152 | * @tsk: task_struct for accounting | ||
| 153 | */ | ||
| 154 | void acct_update_integrals(struct task_struct *tsk) | ||
| 155 | { | ||
| 156 | cputime_t utime, stime; | ||
| 157 | |||
| 158 | task_cputime(tsk, &utime, &stime); | ||
| 159 | __acct_update_integrals(tsk, utime, stime); | ||
| 160 | } | ||
| 161 | |||
| 162 | /** | ||
| 163 | * acct_account_cputime - update mm integral after cputime update | ||
| 164 | * @tsk: task_struct for accounting | ||
| 165 | */ | ||
| 166 | void acct_account_cputime(struct task_struct *tsk) | ||
| 167 | { | ||
| 168 | __acct_update_integrals(tsk, tsk->utime, tsk->stime); | ||
| 169 | } | ||
| 170 | |||
| 171 | /** | ||
| 148 | * acct_clear_integrals - clear the mm integral fields in task_struct | 172 | * acct_clear_integrals - clear the mm integral fields in task_struct |
| 149 | * @tsk: task_struct whose accounting fields are cleared | 173 | * @tsk: task_struct whose accounting fields are cleared |
| 150 | */ | 174 | */ |
diff --git a/kernel/watchdog.c b/kernel/watchdog.c index 75a2ab3d0b02..27689422aa92 100644 --- a/kernel/watchdog.c +++ b/kernel/watchdog.c | |||
| @@ -23,6 +23,7 @@ | |||
| 23 | #include <linux/module.h> | 23 | #include <linux/module.h> |
| 24 | #include <linux/sysctl.h> | 24 | #include <linux/sysctl.h> |
| 25 | #include <linux/smpboot.h> | 25 | #include <linux/smpboot.h> |
| 26 | #include <linux/sched/rt.h> | ||
| 26 | 27 | ||
| 27 | #include <asm/irq_regs.h> | 28 | #include <asm/irq_regs.h> |
| 28 | #include <linux/kvm_para.h> | 29 | #include <linux/kvm_para.h> |
| @@ -32,6 +32,7 @@ | |||
| 32 | #include <linux/khugepaged.h> | 32 | #include <linux/khugepaged.h> |
| 33 | #include <linux/uprobes.h> | 33 | #include <linux/uprobes.h> |
| 34 | #include <linux/rbtree_augmented.h> | 34 | #include <linux/rbtree_augmented.h> |
| 35 | #include <linux/sched/sysctl.h> | ||
| 35 | 36 | ||
| 36 | #include <asm/uaccess.h> | 37 | #include <asm/uaccess.h> |
| 37 | #include <asm/cacheflush.h> | 38 | #include <asm/cacheflush.h> |
diff --git a/mm/mremap.c b/mm/mremap.c index e1031e1f6a61..f9766f460299 100644 --- a/mm/mremap.c +++ b/mm/mremap.c | |||
| @@ -19,6 +19,7 @@ | |||
| 19 | #include <linux/security.h> | 19 | #include <linux/security.h> |
| 20 | #include <linux/syscalls.h> | 20 | #include <linux/syscalls.h> |
| 21 | #include <linux/mmu_notifier.h> | 21 | #include <linux/mmu_notifier.h> |
| 22 | #include <linux/sched/sysctl.h> | ||
| 22 | 23 | ||
| 23 | #include <asm/uaccess.h> | 24 | #include <asm/uaccess.h> |
| 24 | #include <asm/cacheflush.h> | 25 | #include <asm/cacheflush.h> |
diff --git a/mm/nommu.c b/mm/nommu.c index 79c3cac87afa..b20db4e22263 100644 --- a/mm/nommu.c +++ b/mm/nommu.c | |||
| @@ -29,6 +29,7 @@ | |||
| 29 | #include <linux/security.h> | 29 | #include <linux/security.h> |
| 30 | #include <linux/syscalls.h> | 30 | #include <linux/syscalls.h> |
| 31 | #include <linux/audit.h> | 31 | #include <linux/audit.h> |
| 32 | #include <linux/sched/sysctl.h> | ||
| 32 | 33 | ||
| 33 | #include <asm/uaccess.h> | 34 | #include <asm/uaccess.h> |
| 34 | #include <asm/tlb.h> | 35 | #include <asm/tlb.h> |
diff --git a/mm/page-writeback.c b/mm/page-writeback.c index 0713bfbf0954..66a0024becd9 100644 --- a/mm/page-writeback.c +++ b/mm/page-writeback.c | |||
| @@ -35,6 +35,7 @@ | |||
| 35 | #include <linux/buffer_head.h> /* __set_page_dirty_buffers */ | 35 | #include <linux/buffer_head.h> /* __set_page_dirty_buffers */ |
| 36 | #include <linux/pagevec.h> | 36 | #include <linux/pagevec.h> |
| 37 | #include <linux/timer.h> | 37 | #include <linux/timer.h> |
| 38 | #include <linux/sched/rt.h> | ||
| 38 | #include <trace/events/writeback.h> | 39 | #include <trace/events/writeback.h> |
| 39 | 40 | ||
| 40 | /* | 41 | /* |
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 6a83cd35cfde..d1107adf174a 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
| @@ -58,6 +58,7 @@ | |||
| 58 | #include <linux/prefetch.h> | 58 | #include <linux/prefetch.h> |
| 59 | #include <linux/migrate.h> | 59 | #include <linux/migrate.h> |
| 60 | #include <linux/page-debug-flags.h> | 60 | #include <linux/page-debug-flags.h> |
| 61 | #include <linux/sched/rt.h> | ||
| 61 | 62 | ||
| 62 | #include <asm/tlbflush.h> | 63 | #include <asm/tlbflush.h> |
| 63 | #include <asm/div64.h> | 64 | #include <asm/div64.h> |
