diff options
-rw-r--r-- | Documentation/cpu-hotplug.txt | 6 | ||||
-rw-r--r-- | include/linux/cpu.h | 2 | ||||
-rw-r--r-- | include/linux/perf_event.h | 2 | ||||
-rw-r--r-- | init/calibrate.c | 13 | ||||
-rw-r--r-- | kernel/cpu.c | 6 | ||||
-rw-r--r-- | kernel/events/core.c | 4 | ||||
-rw-r--r-- | kernel/fork.c | 2 | ||||
-rw-r--r-- | kernel/hrtimer.c | 6 | ||||
-rw-r--r-- | kernel/printk.c | 2 | ||||
-rw-r--r-- | kernel/profile.c | 2 | ||||
-rw-r--r-- | kernel/relay.c | 2 | ||||
-rw-r--r-- | kernel/sched/core.c | 12 | ||||
-rw-r--r-- | kernel/sched/fair.c | 2 | ||||
-rw-r--r-- | kernel/smp.c | 2 | ||||
-rw-r--r-- | kernel/smpboot.c | 2 | ||||
-rw-r--r-- | kernel/softirq.c | 8 | ||||
-rw-r--r-- | kernel/time/tick-sched.c | 2 | ||||
-rw-r--r-- | kernel/timer.c | 10 | ||||
-rw-r--r-- | kernel/workqueue.c | 4 | ||||
-rw-r--r-- | lib/Kconfig.debug | 2 | ||||
-rw-r--r-- | lib/earlycpio.c | 2 | ||||
-rw-r--r-- | lib/percpu_counter.c | 2 | ||||
-rw-r--r-- | mm/memcontrol.c | 2 | ||||
-rw-r--r-- | mm/page-writeback.c | 4 | ||||
-rw-r--r-- | mm/slab.c | 10 | ||||
-rw-r--r-- | mm/slub.c | 4 | ||||
-rw-r--r-- | mm/vmstat.c | 6 |
27 files changed, 62 insertions, 59 deletions
diff --git a/Documentation/cpu-hotplug.txt b/Documentation/cpu-hotplug.txt index edd4b4df3932..786dc82f98ce 100644 --- a/Documentation/cpu-hotplug.txt +++ b/Documentation/cpu-hotplug.txt | |||
@@ -267,8 +267,8 @@ Q: If i have some kernel code that needs to be aware of CPU arrival and | |||
267 | A: This is what you would need in your kernel code to receive notifications. | 267 | A: This is what you would need in your kernel code to receive notifications. |
268 | 268 | ||
269 | #include <linux/cpu.h> | 269 | #include <linux/cpu.h> |
270 | static int __cpuinit foobar_cpu_callback(struct notifier_block *nfb, | 270 | static int foobar_cpu_callback(struct notifier_block *nfb, |
271 | unsigned long action, void *hcpu) | 271 | unsigned long action, void *hcpu) |
272 | { | 272 | { |
273 | unsigned int cpu = (unsigned long)hcpu; | 273 | unsigned int cpu = (unsigned long)hcpu; |
274 | 274 | ||
@@ -285,7 +285,7 @@ A: This is what you would need in your kernel code to receive notifications. | |||
285 | return NOTIFY_OK; | 285 | return NOTIFY_OK; |
286 | } | 286 | } |
287 | 287 | ||
288 | static struct notifier_block __cpuinitdata foobar_cpu_notifer = | 288 | static struct notifier_block foobar_cpu_notifer = |
289 | { | 289 | { |
290 | .notifier_call = foobar_cpu_callback, | 290 | .notifier_call = foobar_cpu_callback, |
291 | }; | 291 | }; |
diff --git a/include/linux/cpu.h b/include/linux/cpu.h index 944f283f01c4..ab0eade73039 100644 --- a/include/linux/cpu.h +++ b/include/linux/cpu.h | |||
@@ -114,7 +114,7 @@ enum { | |||
114 | /* Need to know about CPUs going up/down? */ | 114 | /* Need to know about CPUs going up/down? */ |
115 | #if defined(CONFIG_HOTPLUG_CPU) || !defined(MODULE) | 115 | #if defined(CONFIG_HOTPLUG_CPU) || !defined(MODULE) |
116 | #define cpu_notifier(fn, pri) { \ | 116 | #define cpu_notifier(fn, pri) { \ |
117 | static struct notifier_block fn##_nb __cpuinitdata = \ | 117 | static struct notifier_block fn##_nb = \ |
118 | { .notifier_call = fn, .priority = pri }; \ | 118 | { .notifier_call = fn, .priority = pri }; \ |
119 | register_cpu_notifier(&fn##_nb); \ | 119 | register_cpu_notifier(&fn##_nb); \ |
120 | } | 120 | } |
diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h index 8873f82c7baa..c43f6eabad5b 100644 --- a/include/linux/perf_event.h +++ b/include/linux/perf_event.h | |||
@@ -826,7 +826,7 @@ static inline void perf_restore_debug_store(void) { } | |||
826 | */ | 826 | */ |
827 | #define perf_cpu_notifier(fn) \ | 827 | #define perf_cpu_notifier(fn) \ |
828 | do { \ | 828 | do { \ |
829 | static struct notifier_block fn##_nb __cpuinitdata = \ | 829 | static struct notifier_block fn##_nb = \ |
830 | { .notifier_call = fn, .priority = CPU_PRI_PERF }; \ | 830 | { .notifier_call = fn, .priority = CPU_PRI_PERF }; \ |
831 | unsigned long cpu = smp_processor_id(); \ | 831 | unsigned long cpu = smp_processor_id(); \ |
832 | unsigned long flags; \ | 832 | unsigned long flags; \ |
diff --git a/init/calibrate.c b/init/calibrate.c index fda0a7b0f06c..520702db9acc 100644 --- a/init/calibrate.c +++ b/init/calibrate.c | |||
@@ -31,7 +31,7 @@ __setup("lpj=", lpj_setup); | |||
31 | #define DELAY_CALIBRATION_TICKS ((HZ < 100) ? 1 : (HZ/100)) | 31 | #define DELAY_CALIBRATION_TICKS ((HZ < 100) ? 1 : (HZ/100)) |
32 | #define MAX_DIRECT_CALIBRATION_RETRIES 5 | 32 | #define MAX_DIRECT_CALIBRATION_RETRIES 5 |
33 | 33 | ||
34 | static unsigned long __cpuinit calibrate_delay_direct(void) | 34 | static unsigned long calibrate_delay_direct(void) |
35 | { | 35 | { |
36 | unsigned long pre_start, start, post_start; | 36 | unsigned long pre_start, start, post_start; |
37 | unsigned long pre_end, end, post_end; | 37 | unsigned long pre_end, end, post_end; |
@@ -166,7 +166,10 @@ static unsigned long __cpuinit calibrate_delay_direct(void) | |||
166 | return 0; | 166 | return 0; |
167 | } | 167 | } |
168 | #else | 168 | #else |
169 | static unsigned long __cpuinit calibrate_delay_direct(void) {return 0;} | 169 | static unsigned long calibrate_delay_direct(void) |
170 | { | ||
171 | return 0; | ||
172 | } | ||
170 | #endif | 173 | #endif |
171 | 174 | ||
172 | /* | 175 | /* |
@@ -180,7 +183,7 @@ static unsigned long __cpuinit calibrate_delay_direct(void) {return 0;} | |||
180 | */ | 183 | */ |
181 | #define LPS_PREC 8 | 184 | #define LPS_PREC 8 |
182 | 185 | ||
183 | static unsigned long __cpuinit calibrate_delay_converge(void) | 186 | static unsigned long calibrate_delay_converge(void) |
184 | { | 187 | { |
185 | /* First stage - slowly accelerate to find initial bounds */ | 188 | /* First stage - slowly accelerate to find initial bounds */ |
186 | unsigned long lpj, lpj_base, ticks, loopadd, loopadd_base, chop_limit; | 189 | unsigned long lpj, lpj_base, ticks, loopadd, loopadd_base, chop_limit; |
@@ -254,12 +257,12 @@ static DEFINE_PER_CPU(unsigned long, cpu_loops_per_jiffy) = { 0 }; | |||
254 | * Architectures should override this function if a faster calibration | 257 | * Architectures should override this function if a faster calibration |
255 | * method is available. | 258 | * method is available. |
256 | */ | 259 | */ |
257 | unsigned long __attribute__((weak)) __cpuinit calibrate_delay_is_known(void) | 260 | unsigned long __attribute__((weak)) calibrate_delay_is_known(void) |
258 | { | 261 | { |
259 | return 0; | 262 | return 0; |
260 | } | 263 | } |
261 | 264 | ||
262 | void __cpuinit calibrate_delay(void) | 265 | void calibrate_delay(void) |
263 | { | 266 | { |
264 | unsigned long lpj; | 267 | unsigned long lpj; |
265 | static bool printed; | 268 | static bool printed; |
diff --git a/kernel/cpu.c b/kernel/cpu.c index 198a38883e64..b2b227b82123 100644 --- a/kernel/cpu.c +++ b/kernel/cpu.c | |||
@@ -366,7 +366,7 @@ EXPORT_SYMBOL(cpu_down); | |||
366 | #endif /*CONFIG_HOTPLUG_CPU*/ | 366 | #endif /*CONFIG_HOTPLUG_CPU*/ |
367 | 367 | ||
368 | /* Requires cpu_add_remove_lock to be held */ | 368 | /* Requires cpu_add_remove_lock to be held */ |
369 | static int __cpuinit _cpu_up(unsigned int cpu, int tasks_frozen) | 369 | static int _cpu_up(unsigned int cpu, int tasks_frozen) |
370 | { | 370 | { |
371 | int ret, nr_calls = 0; | 371 | int ret, nr_calls = 0; |
372 | void *hcpu = (void *)(long)cpu; | 372 | void *hcpu = (void *)(long)cpu; |
@@ -419,7 +419,7 @@ out: | |||
419 | return ret; | 419 | return ret; |
420 | } | 420 | } |
421 | 421 | ||
422 | int __cpuinit cpu_up(unsigned int cpu) | 422 | int cpu_up(unsigned int cpu) |
423 | { | 423 | { |
424 | int err = 0; | 424 | int err = 0; |
425 | 425 | ||
@@ -618,7 +618,7 @@ core_initcall(cpu_hotplug_pm_sync_init); | |||
618 | * It must be called by the arch code on the new cpu, before the new cpu | 618 | * It must be called by the arch code on the new cpu, before the new cpu |
619 | * enables interrupts and before the "boot" cpu returns from __cpu_up(). | 619 | * enables interrupts and before the "boot" cpu returns from __cpu_up(). |
620 | */ | 620 | */ |
621 | void __cpuinit notify_cpu_starting(unsigned int cpu) | 621 | void notify_cpu_starting(unsigned int cpu) |
622 | { | 622 | { |
623 | unsigned long val = CPU_STARTING; | 623 | unsigned long val = CPU_STARTING; |
624 | 624 | ||
diff --git a/kernel/events/core.c b/kernel/events/core.c index eba8fb5834ae..f3e9dce39bc9 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c | |||
@@ -7630,7 +7630,7 @@ static void __init perf_event_init_all_cpus(void) | |||
7630 | } | 7630 | } |
7631 | } | 7631 | } |
7632 | 7632 | ||
7633 | static void __cpuinit perf_event_init_cpu(int cpu) | 7633 | static void perf_event_init_cpu(int cpu) |
7634 | { | 7634 | { |
7635 | struct swevent_htable *swhash = &per_cpu(swevent_htable, cpu); | 7635 | struct swevent_htable *swhash = &per_cpu(swevent_htable, cpu); |
7636 | 7636 | ||
@@ -7719,7 +7719,7 @@ static struct notifier_block perf_reboot_notifier = { | |||
7719 | .priority = INT_MIN, | 7719 | .priority = INT_MIN, |
7720 | }; | 7720 | }; |
7721 | 7721 | ||
7722 | static int __cpuinit | 7722 | static int |
7723 | perf_cpu_notify(struct notifier_block *self, unsigned long action, void *hcpu) | 7723 | perf_cpu_notify(struct notifier_block *self, unsigned long action, void *hcpu) |
7724 | { | 7724 | { |
7725 | unsigned int cpu = (long)hcpu; | 7725 | unsigned int cpu = (long)hcpu; |
diff --git a/kernel/fork.c b/kernel/fork.c index 66635c80a813..403d2bb8a968 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
@@ -1546,7 +1546,7 @@ static inline void init_idle_pids(struct pid_link *links) | |||
1546 | } | 1546 | } |
1547 | } | 1547 | } |
1548 | 1548 | ||
1549 | struct task_struct * __cpuinit fork_idle(int cpu) | 1549 | struct task_struct *fork_idle(int cpu) |
1550 | { | 1550 | { |
1551 | struct task_struct *task; | 1551 | struct task_struct *task; |
1552 | task = copy_process(CLONE_VM, 0, 0, NULL, &init_struct_pid, 0); | 1552 | task = copy_process(CLONE_VM, 0, 0, NULL, &init_struct_pid, 0); |
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index f0f4fe29cd21..383319bae3f7 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c | |||
@@ -1659,7 +1659,7 @@ SYSCALL_DEFINE2(nanosleep, struct timespec __user *, rqtp, | |||
1659 | /* | 1659 | /* |
1660 | * Functions related to boot-time initialization: | 1660 | * Functions related to boot-time initialization: |
1661 | */ | 1661 | */ |
1662 | static void __cpuinit init_hrtimers_cpu(int cpu) | 1662 | static void init_hrtimers_cpu(int cpu) |
1663 | { | 1663 | { |
1664 | struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu); | 1664 | struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu); |
1665 | int i; | 1665 | int i; |
@@ -1740,7 +1740,7 @@ static void migrate_hrtimers(int scpu) | |||
1740 | 1740 | ||
1741 | #endif /* CONFIG_HOTPLUG_CPU */ | 1741 | #endif /* CONFIG_HOTPLUG_CPU */ |
1742 | 1742 | ||
1743 | static int __cpuinit hrtimer_cpu_notify(struct notifier_block *self, | 1743 | static int hrtimer_cpu_notify(struct notifier_block *self, |
1744 | unsigned long action, void *hcpu) | 1744 | unsigned long action, void *hcpu) |
1745 | { | 1745 | { |
1746 | int scpu = (long)hcpu; | 1746 | int scpu = (long)hcpu; |
@@ -1773,7 +1773,7 @@ static int __cpuinit hrtimer_cpu_notify(struct notifier_block *self, | |||
1773 | return NOTIFY_OK; | 1773 | return NOTIFY_OK; |
1774 | } | 1774 | } |
1775 | 1775 | ||
1776 | static struct notifier_block __cpuinitdata hrtimers_nb = { | 1776 | static struct notifier_block hrtimers_nb = { |
1777 | .notifier_call = hrtimer_cpu_notify, | 1777 | .notifier_call = hrtimer_cpu_notify, |
1778 | }; | 1778 | }; |
1779 | 1779 | ||
diff --git a/kernel/printk.c b/kernel/printk.c index d37d45c90ae6..69b0890ed7e5 100644 --- a/kernel/printk.c +++ b/kernel/printk.c | |||
@@ -1921,7 +1921,7 @@ void resume_console(void) | |||
1921 | * called when a new CPU comes online (or fails to come up), and ensures | 1921 | * called when a new CPU comes online (or fails to come up), and ensures |
1922 | * that any such output gets printed. | 1922 | * that any such output gets printed. |
1923 | */ | 1923 | */ |
1924 | static int __cpuinit console_cpu_notify(struct notifier_block *self, | 1924 | static int console_cpu_notify(struct notifier_block *self, |
1925 | unsigned long action, void *hcpu) | 1925 | unsigned long action, void *hcpu) |
1926 | { | 1926 | { |
1927 | switch (action) { | 1927 | switch (action) { |
diff --git a/kernel/profile.c b/kernel/profile.c index 0bf400737660..6631e1ef55ab 100644 --- a/kernel/profile.c +++ b/kernel/profile.c | |||
@@ -331,7 +331,7 @@ out: | |||
331 | put_cpu(); | 331 | put_cpu(); |
332 | } | 332 | } |
333 | 333 | ||
334 | static int __cpuinit profile_cpu_callback(struct notifier_block *info, | 334 | static int profile_cpu_callback(struct notifier_block *info, |
335 | unsigned long action, void *__cpu) | 335 | unsigned long action, void *__cpu) |
336 | { | 336 | { |
337 | int node, cpu = (unsigned long)__cpu; | 337 | int node, cpu = (unsigned long)__cpu; |
diff --git a/kernel/relay.c b/kernel/relay.c index b91488ba2e5a..5001c9887db1 100644 --- a/kernel/relay.c +++ b/kernel/relay.c | |||
@@ -516,7 +516,7 @@ static void setup_callbacks(struct rchan *chan, | |||
516 | * | 516 | * |
517 | * Returns the success/failure of the operation. (%NOTIFY_OK, %NOTIFY_BAD) | 517 | * Returns the success/failure of the operation. (%NOTIFY_OK, %NOTIFY_BAD) |
518 | */ | 518 | */ |
519 | static int __cpuinit relay_hotcpu_callback(struct notifier_block *nb, | 519 | static int relay_hotcpu_callback(struct notifier_block *nb, |
520 | unsigned long action, | 520 | unsigned long action, |
521 | void *hcpu) | 521 | void *hcpu) |
522 | { | 522 | { |
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 0d8eb4525e76..b7c32cb7bfeb 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c | |||
@@ -4133,7 +4133,7 @@ void show_state_filter(unsigned long state_filter) | |||
4133 | debug_show_all_locks(); | 4133 | debug_show_all_locks(); |
4134 | } | 4134 | } |
4135 | 4135 | ||
4136 | void __cpuinit init_idle_bootup_task(struct task_struct *idle) | 4136 | void init_idle_bootup_task(struct task_struct *idle) |
4137 | { | 4137 | { |
4138 | idle->sched_class = &idle_sched_class; | 4138 | idle->sched_class = &idle_sched_class; |
4139 | } | 4139 | } |
@@ -4146,7 +4146,7 @@ void __cpuinit init_idle_bootup_task(struct task_struct *idle) | |||
4146 | * NOTE: this function does not set the idle thread's NEED_RESCHED | 4146 | * NOTE: this function does not set the idle thread's NEED_RESCHED |
4147 | * flag, to make booting more robust. | 4147 | * flag, to make booting more robust. |
4148 | */ | 4148 | */ |
4149 | void __cpuinit init_idle(struct task_struct *idle, int cpu) | 4149 | void init_idle(struct task_struct *idle, int cpu) |
4150 | { | 4150 | { |
4151 | struct rq *rq = cpu_rq(cpu); | 4151 | struct rq *rq = cpu_rq(cpu); |
4152 | unsigned long flags; | 4152 | unsigned long flags; |
@@ -4630,7 +4630,7 @@ static void set_rq_offline(struct rq *rq) | |||
4630 | * migration_call - callback that gets triggered when a CPU is added. | 4630 | * migration_call - callback that gets triggered when a CPU is added. |
4631 | * Here we can start up the necessary migration thread for the new CPU. | 4631 | * Here we can start up the necessary migration thread for the new CPU. |
4632 | */ | 4632 | */ |
4633 | static int __cpuinit | 4633 | static int |
4634 | migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu) | 4634 | migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu) |
4635 | { | 4635 | { |
4636 | int cpu = (long)hcpu; | 4636 | int cpu = (long)hcpu; |
@@ -4684,12 +4684,12 @@ migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu) | |||
4684 | * happens before everything else. This has to be lower priority than | 4684 | * happens before everything else. This has to be lower priority than |
4685 | * the notifier in the perf_event subsystem, though. | 4685 | * the notifier in the perf_event subsystem, though. |
4686 | */ | 4686 | */ |
4687 | static struct notifier_block __cpuinitdata migration_notifier = { | 4687 | static struct notifier_block migration_notifier = { |
4688 | .notifier_call = migration_call, | 4688 | .notifier_call = migration_call, |
4689 | .priority = CPU_PRI_MIGRATION, | 4689 | .priority = CPU_PRI_MIGRATION, |
4690 | }; | 4690 | }; |
4691 | 4691 | ||
4692 | static int __cpuinit sched_cpu_active(struct notifier_block *nfb, | 4692 | static int sched_cpu_active(struct notifier_block *nfb, |
4693 | unsigned long action, void *hcpu) | 4693 | unsigned long action, void *hcpu) |
4694 | { | 4694 | { |
4695 | switch (action & ~CPU_TASKS_FROZEN) { | 4695 | switch (action & ~CPU_TASKS_FROZEN) { |
@@ -4702,7 +4702,7 @@ static int __cpuinit sched_cpu_active(struct notifier_block *nfb, | |||
4702 | } | 4702 | } |
4703 | } | 4703 | } |
4704 | 4704 | ||
4705 | static int __cpuinit sched_cpu_inactive(struct notifier_block *nfb, | 4705 | static int sched_cpu_inactive(struct notifier_block *nfb, |
4706 | unsigned long action, void *hcpu) | 4706 | unsigned long action, void *hcpu) |
4707 | { | 4707 | { |
4708 | switch (action & ~CPU_TASKS_FROZEN) { | 4708 | switch (action & ~CPU_TASKS_FROZEN) { |
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index f77f9c527449..bb456f44b7b1 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -5506,7 +5506,7 @@ void nohz_balance_enter_idle(int cpu) | |||
5506 | set_bit(NOHZ_TICK_STOPPED, nohz_flags(cpu)); | 5506 | set_bit(NOHZ_TICK_STOPPED, nohz_flags(cpu)); |
5507 | } | 5507 | } |
5508 | 5508 | ||
5509 | static int __cpuinit sched_ilb_notifier(struct notifier_block *nfb, | 5509 | static int sched_ilb_notifier(struct notifier_block *nfb, |
5510 | unsigned long action, void *hcpu) | 5510 | unsigned long action, void *hcpu) |
5511 | { | 5511 | { |
5512 | switch (action & ~CPU_TASKS_FROZEN) { | 5512 | switch (action & ~CPU_TASKS_FROZEN) { |
diff --git a/kernel/smp.c b/kernel/smp.c index 4dba0f7b72ad..fe9f773d7114 100644 --- a/kernel/smp.c +++ b/kernel/smp.c | |||
@@ -73,7 +73,7 @@ hotplug_cfd(struct notifier_block *nfb, unsigned long action, void *hcpu) | |||
73 | return NOTIFY_OK; | 73 | return NOTIFY_OK; |
74 | } | 74 | } |
75 | 75 | ||
76 | static struct notifier_block __cpuinitdata hotplug_cfd_notifier = { | 76 | static struct notifier_block hotplug_cfd_notifier = { |
77 | .notifier_call = hotplug_cfd, | 77 | .notifier_call = hotplug_cfd, |
78 | }; | 78 | }; |
79 | 79 | ||
diff --git a/kernel/smpboot.c b/kernel/smpboot.c index 02fc5c933673..eb89e1807408 100644 --- a/kernel/smpboot.c +++ b/kernel/smpboot.c | |||
@@ -24,7 +24,7 @@ | |||
24 | */ | 24 | */ |
25 | static DEFINE_PER_CPU(struct task_struct *, idle_threads); | 25 | static DEFINE_PER_CPU(struct task_struct *, idle_threads); |
26 | 26 | ||
27 | struct task_struct * __cpuinit idle_thread_get(unsigned int cpu) | 27 | struct task_struct *idle_thread_get(unsigned int cpu) |
28 | { | 28 | { |
29 | struct task_struct *tsk = per_cpu(idle_threads, cpu); | 29 | struct task_struct *tsk = per_cpu(idle_threads, cpu); |
30 | 30 | ||
diff --git a/kernel/softirq.c b/kernel/softirq.c index ca25e6e704a2..be3d3514c325 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c | |||
@@ -699,7 +699,7 @@ void send_remote_softirq(struct call_single_data *cp, int cpu, int softirq) | |||
699 | } | 699 | } |
700 | EXPORT_SYMBOL(send_remote_softirq); | 700 | EXPORT_SYMBOL(send_remote_softirq); |
701 | 701 | ||
702 | static int __cpuinit remote_softirq_cpu_notify(struct notifier_block *self, | 702 | static int remote_softirq_cpu_notify(struct notifier_block *self, |
703 | unsigned long action, void *hcpu) | 703 | unsigned long action, void *hcpu) |
704 | { | 704 | { |
705 | /* | 705 | /* |
@@ -728,7 +728,7 @@ static int __cpuinit remote_softirq_cpu_notify(struct notifier_block *self, | |||
728 | return NOTIFY_OK; | 728 | return NOTIFY_OK; |
729 | } | 729 | } |
730 | 730 | ||
731 | static struct notifier_block __cpuinitdata remote_softirq_cpu_notifier = { | 731 | static struct notifier_block remote_softirq_cpu_notifier = { |
732 | .notifier_call = remote_softirq_cpu_notify, | 732 | .notifier_call = remote_softirq_cpu_notify, |
733 | }; | 733 | }; |
734 | 734 | ||
@@ -830,7 +830,7 @@ static void takeover_tasklets(unsigned int cpu) | |||
830 | } | 830 | } |
831 | #endif /* CONFIG_HOTPLUG_CPU */ | 831 | #endif /* CONFIG_HOTPLUG_CPU */ |
832 | 832 | ||
833 | static int __cpuinit cpu_callback(struct notifier_block *nfb, | 833 | static int cpu_callback(struct notifier_block *nfb, |
834 | unsigned long action, | 834 | unsigned long action, |
835 | void *hcpu) | 835 | void *hcpu) |
836 | { | 836 | { |
@@ -845,7 +845,7 @@ static int __cpuinit cpu_callback(struct notifier_block *nfb, | |||
845 | return NOTIFY_OK; | 845 | return NOTIFY_OK; |
846 | } | 846 | } |
847 | 847 | ||
848 | static struct notifier_block __cpuinitdata cpu_nfb = { | 848 | static struct notifier_block cpu_nfb = { |
849 | .notifier_call = cpu_callback | 849 | .notifier_call = cpu_callback |
850 | }; | 850 | }; |
851 | 851 | ||
diff --git a/kernel/time/tick-sched.c b/kernel/time/tick-sched.c index 69601726a745..e80183f4a6c4 100644 --- a/kernel/time/tick-sched.c +++ b/kernel/time/tick-sched.c | |||
@@ -298,7 +298,7 @@ static int __init tick_nohz_full_setup(char *str) | |||
298 | } | 298 | } |
299 | __setup("nohz_full=", tick_nohz_full_setup); | 299 | __setup("nohz_full=", tick_nohz_full_setup); |
300 | 300 | ||
301 | static int __cpuinit tick_nohz_cpu_down_callback(struct notifier_block *nfb, | 301 | static int tick_nohz_cpu_down_callback(struct notifier_block *nfb, |
302 | unsigned long action, | 302 | unsigned long action, |
303 | void *hcpu) | 303 | void *hcpu) |
304 | { | 304 | { |
diff --git a/kernel/timer.c b/kernel/timer.c index 15bc1b41021d..4296d13db3d1 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
@@ -1505,11 +1505,11 @@ signed long __sched schedule_timeout_uninterruptible(signed long timeout) | |||
1505 | } | 1505 | } |
1506 | EXPORT_SYMBOL(schedule_timeout_uninterruptible); | 1506 | EXPORT_SYMBOL(schedule_timeout_uninterruptible); |
1507 | 1507 | ||
1508 | static int __cpuinit init_timers_cpu(int cpu) | 1508 | static int init_timers_cpu(int cpu) |
1509 | { | 1509 | { |
1510 | int j; | 1510 | int j; |
1511 | struct tvec_base *base; | 1511 | struct tvec_base *base; |
1512 | static char __cpuinitdata tvec_base_done[NR_CPUS]; | 1512 | static char tvec_base_done[NR_CPUS]; |
1513 | 1513 | ||
1514 | if (!tvec_base_done[cpu]) { | 1514 | if (!tvec_base_done[cpu]) { |
1515 | static char boot_done; | 1515 | static char boot_done; |
@@ -1577,7 +1577,7 @@ static void migrate_timer_list(struct tvec_base *new_base, struct list_head *hea | |||
1577 | } | 1577 | } |
1578 | } | 1578 | } |
1579 | 1579 | ||
1580 | static void __cpuinit migrate_timers(int cpu) | 1580 | static void migrate_timers(int cpu) |
1581 | { | 1581 | { |
1582 | struct tvec_base *old_base; | 1582 | struct tvec_base *old_base; |
1583 | struct tvec_base *new_base; | 1583 | struct tvec_base *new_base; |
@@ -1610,7 +1610,7 @@ static void __cpuinit migrate_timers(int cpu) | |||
1610 | } | 1610 | } |
1611 | #endif /* CONFIG_HOTPLUG_CPU */ | 1611 | #endif /* CONFIG_HOTPLUG_CPU */ |
1612 | 1612 | ||
1613 | static int __cpuinit timer_cpu_notify(struct notifier_block *self, | 1613 | static int timer_cpu_notify(struct notifier_block *self, |
1614 | unsigned long action, void *hcpu) | 1614 | unsigned long action, void *hcpu) |
1615 | { | 1615 | { |
1616 | long cpu = (long)hcpu; | 1616 | long cpu = (long)hcpu; |
@@ -1635,7 +1635,7 @@ static int __cpuinit timer_cpu_notify(struct notifier_block *self, | |||
1635 | return NOTIFY_OK; | 1635 | return NOTIFY_OK; |
1636 | } | 1636 | } |
1637 | 1637 | ||
1638 | static struct notifier_block __cpuinitdata timers_nb = { | 1638 | static struct notifier_block timers_nb = { |
1639 | .notifier_call = timer_cpu_notify, | 1639 | .notifier_call = timer_cpu_notify, |
1640 | }; | 1640 | }; |
1641 | 1641 | ||
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index f02c4a4a0c3c..0b72e816b8d0 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
@@ -4644,7 +4644,7 @@ static void restore_unbound_workers_cpumask(struct worker_pool *pool, int cpu) | |||
4644 | * Workqueues should be brought up before normal priority CPU notifiers. | 4644 | * Workqueues should be brought up before normal priority CPU notifiers. |
4645 | * This will be registered high priority CPU notifier. | 4645 | * This will be registered high priority CPU notifier. |
4646 | */ | 4646 | */ |
4647 | static int __cpuinit workqueue_cpu_up_callback(struct notifier_block *nfb, | 4647 | static int workqueue_cpu_up_callback(struct notifier_block *nfb, |
4648 | unsigned long action, | 4648 | unsigned long action, |
4649 | void *hcpu) | 4649 | void *hcpu) |
4650 | { | 4650 | { |
@@ -4697,7 +4697,7 @@ static int __cpuinit workqueue_cpu_up_callback(struct notifier_block *nfb, | |||
4697 | * Workqueues should be brought down after normal priority CPU notifiers. | 4697 | * Workqueues should be brought down after normal priority CPU notifiers. |
4698 | * This will be registered as low priority CPU notifier. | 4698 | * This will be registered as low priority CPU notifier. |
4699 | */ | 4699 | */ |
4700 | static int __cpuinit workqueue_cpu_down_callback(struct notifier_block *nfb, | 4700 | static int workqueue_cpu_down_callback(struct notifier_block *nfb, |
4701 | unsigned long action, | 4701 | unsigned long action, |
4702 | void *hcpu) | 4702 | void *hcpu) |
4703 | { | 4703 | { |
diff --git a/lib/Kconfig.debug b/lib/Kconfig.debug index 98ac17ed6222..1501aa553221 100644 --- a/lib/Kconfig.debug +++ b/lib/Kconfig.debug | |||
@@ -238,7 +238,7 @@ config DEBUG_SECTION_MISMATCH | |||
238 | any use of code/data previously in these sections would | 238 | any use of code/data previously in these sections would |
239 | most likely result in an oops. | 239 | most likely result in an oops. |
240 | In the code, functions and variables are annotated with | 240 | In the code, functions and variables are annotated with |
241 | __init, __cpuinit, etc. (see the full list in include/linux/init.h), | 241 | __init,, etc. (see the full list in include/linux/init.h), |
242 | which results in the code/data being placed in specific sections. | 242 | which results in the code/data being placed in specific sections. |
243 | The section mismatch analysis is always performed after a full | 243 | The section mismatch analysis is always performed after a full |
244 | kernel build, and enabling this option causes the following | 244 | kernel build, and enabling this option causes the following |
diff --git a/lib/earlycpio.c b/lib/earlycpio.c index 8078ef49cb79..7aa7ce250c94 100644 --- a/lib/earlycpio.c +++ b/lib/earlycpio.c | |||
@@ -63,7 +63,7 @@ enum cpio_fields { | |||
63 | * the match returned an empty filename string. | 63 | * the match returned an empty filename string. |
64 | */ | 64 | */ |
65 | 65 | ||
66 | struct cpio_data __cpuinit find_cpio_data(const char *path, void *data, | 66 | struct cpio_data find_cpio_data(const char *path, void *data, |
67 | size_t len, long *offset) | 67 | size_t len, long *offset) |
68 | { | 68 | { |
69 | const size_t cpio_header_len = 8*C_NFIELDS - 2; | 69 | const size_t cpio_header_len = 8*C_NFIELDS - 2; |
diff --git a/lib/percpu_counter.c b/lib/percpu_counter.c index 1fc23a3277e1..93c5d5ecff4e 100644 --- a/lib/percpu_counter.c +++ b/lib/percpu_counter.c | |||
@@ -158,7 +158,7 @@ static void compute_batch_value(void) | |||
158 | percpu_counter_batch = max(32, nr*2); | 158 | percpu_counter_batch = max(32, nr*2); |
159 | } | 159 | } |
160 | 160 | ||
161 | static int __cpuinit percpu_counter_hotcpu_callback(struct notifier_block *nb, | 161 | static int percpu_counter_hotcpu_callback(struct notifier_block *nb, |
162 | unsigned long action, void *hcpu) | 162 | unsigned long action, void *hcpu) |
163 | { | 163 | { |
164 | #ifdef CONFIG_HOTPLUG_CPU | 164 | #ifdef CONFIG_HOTPLUG_CPU |
diff --git a/mm/memcontrol.c b/mm/memcontrol.c index d12ca6f3c293..00a7a664b9c1 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c | |||
@@ -2522,7 +2522,7 @@ static void mem_cgroup_drain_pcp_counter(struct mem_cgroup *memcg, int cpu) | |||
2522 | spin_unlock(&memcg->pcp_counter_lock); | 2522 | spin_unlock(&memcg->pcp_counter_lock); |
2523 | } | 2523 | } |
2524 | 2524 | ||
2525 | static int __cpuinit memcg_cpu_hotplug_callback(struct notifier_block *nb, | 2525 | static int memcg_cpu_hotplug_callback(struct notifier_block *nb, |
2526 | unsigned long action, | 2526 | unsigned long action, |
2527 | void *hcpu) | 2527 | void *hcpu) |
2528 | { | 2528 | { |
diff --git a/mm/page-writeback.c b/mm/page-writeback.c index 4514ad7415c3..3f0c895c71fe 100644 --- a/mm/page-writeback.c +++ b/mm/page-writeback.c | |||
@@ -1619,7 +1619,7 @@ void writeback_set_ratelimit(void) | |||
1619 | ratelimit_pages = 16; | 1619 | ratelimit_pages = 16; |
1620 | } | 1620 | } |
1621 | 1621 | ||
1622 | static int __cpuinit | 1622 | static int |
1623 | ratelimit_handler(struct notifier_block *self, unsigned long action, | 1623 | ratelimit_handler(struct notifier_block *self, unsigned long action, |
1624 | void *hcpu) | 1624 | void *hcpu) |
1625 | { | 1625 | { |
@@ -1634,7 +1634,7 @@ ratelimit_handler(struct notifier_block *self, unsigned long action, | |||
1634 | } | 1634 | } |
1635 | } | 1635 | } |
1636 | 1636 | ||
1637 | static struct notifier_block __cpuinitdata ratelimit_nb = { | 1637 | static struct notifier_block ratelimit_nb = { |
1638 | .notifier_call = ratelimit_handler, | 1638 | .notifier_call = ratelimit_handler, |
1639 | .next = NULL, | 1639 | .next = NULL, |
1640 | }; | 1640 | }; |
@@ -787,7 +787,7 @@ static void next_reap_node(void) | |||
787 | * the CPUs getting into lockstep and contending for the global cache chain | 787 | * the CPUs getting into lockstep and contending for the global cache chain |
788 | * lock. | 788 | * lock. |
789 | */ | 789 | */ |
790 | static void __cpuinit start_cpu_timer(int cpu) | 790 | static void start_cpu_timer(int cpu) |
791 | { | 791 | { |
792 | struct delayed_work *reap_work = &per_cpu(slab_reap_work, cpu); | 792 | struct delayed_work *reap_work = &per_cpu(slab_reap_work, cpu); |
793 | 793 | ||
@@ -1186,7 +1186,7 @@ static inline int slabs_tofree(struct kmem_cache *cachep, | |||
1186 | return (n->free_objects + cachep->num - 1) / cachep->num; | 1186 | return (n->free_objects + cachep->num - 1) / cachep->num; |
1187 | } | 1187 | } |
1188 | 1188 | ||
1189 | static void __cpuinit cpuup_canceled(long cpu) | 1189 | static void cpuup_canceled(long cpu) |
1190 | { | 1190 | { |
1191 | struct kmem_cache *cachep; | 1191 | struct kmem_cache *cachep; |
1192 | struct kmem_cache_node *n = NULL; | 1192 | struct kmem_cache_node *n = NULL; |
@@ -1251,7 +1251,7 @@ free_array_cache: | |||
1251 | } | 1251 | } |
1252 | } | 1252 | } |
1253 | 1253 | ||
1254 | static int __cpuinit cpuup_prepare(long cpu) | 1254 | static int cpuup_prepare(long cpu) |
1255 | { | 1255 | { |
1256 | struct kmem_cache *cachep; | 1256 | struct kmem_cache *cachep; |
1257 | struct kmem_cache_node *n = NULL; | 1257 | struct kmem_cache_node *n = NULL; |
@@ -1334,7 +1334,7 @@ bad: | |||
1334 | return -ENOMEM; | 1334 | return -ENOMEM; |
1335 | } | 1335 | } |
1336 | 1336 | ||
1337 | static int __cpuinit cpuup_callback(struct notifier_block *nfb, | 1337 | static int cpuup_callback(struct notifier_block *nfb, |
1338 | unsigned long action, void *hcpu) | 1338 | unsigned long action, void *hcpu) |
1339 | { | 1339 | { |
1340 | long cpu = (long)hcpu; | 1340 | long cpu = (long)hcpu; |
@@ -1390,7 +1390,7 @@ static int __cpuinit cpuup_callback(struct notifier_block *nfb, | |||
1390 | return notifier_from_errno(err); | 1390 | return notifier_from_errno(err); |
1391 | } | 1391 | } |
1392 | 1392 | ||
1393 | static struct notifier_block __cpuinitdata cpucache_notifier = { | 1393 | static struct notifier_block cpucache_notifier = { |
1394 | &cpuup_callback, NULL, 0 | 1394 | &cpuup_callback, NULL, 0 |
1395 | }; | 1395 | }; |
1396 | 1396 | ||
@@ -3773,7 +3773,7 @@ int __kmem_cache_create(struct kmem_cache *s, unsigned long flags) | |||
3773 | * Use the cpu notifier to insure that the cpu slabs are flushed when | 3773 | * Use the cpu notifier to insure that the cpu slabs are flushed when |
3774 | * necessary. | 3774 | * necessary. |
3775 | */ | 3775 | */ |
3776 | static int __cpuinit slab_cpuup_callback(struct notifier_block *nfb, | 3776 | static int slab_cpuup_callback(struct notifier_block *nfb, |
3777 | unsigned long action, void *hcpu) | 3777 | unsigned long action, void *hcpu) |
3778 | { | 3778 | { |
3779 | long cpu = (long)hcpu; | 3779 | long cpu = (long)hcpu; |
@@ -3799,7 +3799,7 @@ static int __cpuinit slab_cpuup_callback(struct notifier_block *nfb, | |||
3799 | return NOTIFY_OK; | 3799 | return NOTIFY_OK; |
3800 | } | 3800 | } |
3801 | 3801 | ||
3802 | static struct notifier_block __cpuinitdata slab_notifier = { | 3802 | static struct notifier_block slab_notifier = { |
3803 | .notifier_call = slab_cpuup_callback | 3803 | .notifier_call = slab_cpuup_callback |
3804 | }; | 3804 | }; |
3805 | 3805 | ||
diff --git a/mm/vmstat.c b/mm/vmstat.c index f42745e65780..20c2ef4458fa 100644 --- a/mm/vmstat.c +++ b/mm/vmstat.c | |||
@@ -1182,7 +1182,7 @@ static void vmstat_update(struct work_struct *w) | |||
1182 | round_jiffies_relative(sysctl_stat_interval)); | 1182 | round_jiffies_relative(sysctl_stat_interval)); |
1183 | } | 1183 | } |
1184 | 1184 | ||
1185 | static void __cpuinit start_cpu_timer(int cpu) | 1185 | static void start_cpu_timer(int cpu) |
1186 | { | 1186 | { |
1187 | struct delayed_work *work = &per_cpu(vmstat_work, cpu); | 1187 | struct delayed_work *work = &per_cpu(vmstat_work, cpu); |
1188 | 1188 | ||
@@ -1194,7 +1194,7 @@ static void __cpuinit start_cpu_timer(int cpu) | |||
1194 | * Use the cpu notifier to insure that the thresholds are recalculated | 1194 | * Use the cpu notifier to insure that the thresholds are recalculated |
1195 | * when necessary. | 1195 | * when necessary. |
1196 | */ | 1196 | */ |
1197 | static int __cpuinit vmstat_cpuup_callback(struct notifier_block *nfb, | 1197 | static int vmstat_cpuup_callback(struct notifier_block *nfb, |
1198 | unsigned long action, | 1198 | unsigned long action, |
1199 | void *hcpu) | 1199 | void *hcpu) |
1200 | { | 1200 | { |
@@ -1226,7 +1226,7 @@ static int __cpuinit vmstat_cpuup_callback(struct notifier_block *nfb, | |||
1226 | return NOTIFY_OK; | 1226 | return NOTIFY_OK; |
1227 | } | 1227 | } |
1228 | 1228 | ||
1229 | static struct notifier_block __cpuinitdata vmstat_notifier = | 1229 | static struct notifier_block vmstat_notifier = |
1230 | { &vmstat_cpuup_callback, NULL, 0 }; | 1230 | { &vmstat_cpuup_callback, NULL, 0 }; |
1231 | #endif | 1231 | #endif |
1232 | 1232 | ||