diff options
Diffstat (limited to 'kernel/power')
| -rw-r--r-- | kernel/power/disk.c | 50 | ||||
| -rw-r--r-- | kernel/power/main.c | 16 | ||||
| -rw-r--r-- | kernel/power/process.c | 97 | ||||
| -rw-r--r-- | kernel/power/user.c | 71 |
4 files changed, 135 insertions, 99 deletions
diff --git a/kernel/power/disk.c b/kernel/power/disk.c index 14a656cdc652..f011e0870b52 100644 --- a/kernel/power/disk.c +++ b/kernel/power/disk.c | |||
| @@ -180,6 +180,17 @@ static void platform_restore_cleanup(int platform_mode) | |||
| 180 | } | 180 | } |
| 181 | 181 | ||
| 182 | /** | 182 | /** |
| 183 | * platform_recover - recover the platform from a failure to suspend | ||
| 184 | * devices. | ||
| 185 | */ | ||
| 186 | |||
| 187 | static void platform_recover(int platform_mode) | ||
| 188 | { | ||
| 189 | if (platform_mode && hibernation_ops && hibernation_ops->recover) | ||
| 190 | hibernation_ops->recover(); | ||
| 191 | } | ||
| 192 | |||
| 193 | /** | ||
| 183 | * create_image - freeze devices that need to be frozen with interrupts | 194 | * create_image - freeze devices that need to be frozen with interrupts |
| 184 | * off, create the hibernation image and thaw those devices. Control | 195 | * off, create the hibernation image and thaw those devices. Control |
| 185 | * reappears in this routine after a restore. | 196 | * reappears in this routine after a restore. |
| @@ -193,6 +204,7 @@ static int create_image(int platform_mode) | |||
| 193 | if (error) | 204 | if (error) |
| 194 | return error; | 205 | return error; |
| 195 | 206 | ||
| 207 | device_pm_lock(); | ||
| 196 | local_irq_disable(); | 208 | local_irq_disable(); |
| 197 | /* At this point, device_suspend() has been called, but *not* | 209 | /* At this point, device_suspend() has been called, but *not* |
| 198 | * device_power_down(). We *must* call device_power_down() now. | 210 | * device_power_down(). We *must* call device_power_down() now. |
| @@ -224,9 +236,11 @@ static int create_image(int platform_mode) | |||
| 224 | /* NOTE: device_power_up() is just a resume() for devices | 236 | /* NOTE: device_power_up() is just a resume() for devices |
| 225 | * that suspended with irqs off ... no overall powerup. | 237 | * that suspended with irqs off ... no overall powerup. |
| 226 | */ | 238 | */ |
| 227 | device_power_up(); | 239 | device_power_up(in_suspend ? |
| 240 | (error ? PMSG_RECOVER : PMSG_THAW) : PMSG_RESTORE); | ||
| 228 | Enable_irqs: | 241 | Enable_irqs: |
| 229 | local_irq_enable(); | 242 | local_irq_enable(); |
| 243 | device_pm_unlock(); | ||
| 230 | return error; | 244 | return error; |
| 231 | } | 245 | } |
| 232 | 246 | ||
| @@ -255,10 +269,10 @@ int hibernation_snapshot(int platform_mode) | |||
| 255 | suspend_console(); | 269 | suspend_console(); |
| 256 | error = device_suspend(PMSG_FREEZE); | 270 | error = device_suspend(PMSG_FREEZE); |
| 257 | if (error) | 271 | if (error) |
| 258 | goto Resume_console; | 272 | goto Recover_platform; |
| 259 | 273 | ||
| 260 | if (hibernation_test(TEST_DEVICES)) | 274 | if (hibernation_test(TEST_DEVICES)) |
| 261 | goto Resume_devices; | 275 | goto Recover_platform; |
| 262 | 276 | ||
| 263 | error = platform_pre_snapshot(platform_mode); | 277 | error = platform_pre_snapshot(platform_mode); |
| 264 | if (error || hibernation_test(TEST_PLATFORM)) | 278 | if (error || hibernation_test(TEST_PLATFORM)) |
| @@ -280,12 +294,16 @@ int hibernation_snapshot(int platform_mode) | |||
| 280 | Finish: | 294 | Finish: |
| 281 | platform_finish(platform_mode); | 295 | platform_finish(platform_mode); |
| 282 | Resume_devices: | 296 | Resume_devices: |
| 283 | device_resume(); | 297 | device_resume(in_suspend ? |
| 284 | Resume_console: | 298 | (error ? PMSG_RECOVER : PMSG_THAW) : PMSG_RESTORE); |
| 285 | resume_console(); | 299 | resume_console(); |
| 286 | Close: | 300 | Close: |
| 287 | platform_end(platform_mode); | 301 | platform_end(platform_mode); |
| 288 | return error; | 302 | return error; |
| 303 | |||
| 304 | Recover_platform: | ||
| 305 | platform_recover(platform_mode); | ||
| 306 | goto Resume_devices; | ||
| 289 | } | 307 | } |
| 290 | 308 | ||
| 291 | /** | 309 | /** |
| @@ -300,8 +318,9 @@ static int resume_target_kernel(void) | |||
| 300 | { | 318 | { |
| 301 | int error; | 319 | int error; |
| 302 | 320 | ||
| 321 | device_pm_lock(); | ||
| 303 | local_irq_disable(); | 322 | local_irq_disable(); |
| 304 | error = device_power_down(PMSG_PRETHAW); | 323 | error = device_power_down(PMSG_QUIESCE); |
| 305 | if (error) { | 324 | if (error) { |
| 306 | printk(KERN_ERR "PM: Some devices failed to power down, " | 325 | printk(KERN_ERR "PM: Some devices failed to power down, " |
| 307 | "aborting resume\n"); | 326 | "aborting resume\n"); |
| @@ -329,9 +348,10 @@ static int resume_target_kernel(void) | |||
| 329 | swsusp_free(); | 348 | swsusp_free(); |
| 330 | restore_processor_state(); | 349 | restore_processor_state(); |
| 331 | touch_softlockup_watchdog(); | 350 | touch_softlockup_watchdog(); |
| 332 | device_power_up(); | 351 | device_power_up(PMSG_RECOVER); |
| 333 | Enable_irqs: | 352 | Enable_irqs: |
| 334 | local_irq_enable(); | 353 | local_irq_enable(); |
| 354 | device_pm_unlock(); | ||
| 335 | return error; | 355 | return error; |
| 336 | } | 356 | } |
| 337 | 357 | ||
| @@ -350,7 +370,7 @@ int hibernation_restore(int platform_mode) | |||
| 350 | 370 | ||
| 351 | pm_prepare_console(); | 371 | pm_prepare_console(); |
| 352 | suspend_console(); | 372 | suspend_console(); |
| 353 | error = device_suspend(PMSG_PRETHAW); | 373 | error = device_suspend(PMSG_QUIESCE); |
| 354 | if (error) | 374 | if (error) |
| 355 | goto Finish; | 375 | goto Finish; |
| 356 | 376 | ||
| @@ -362,7 +382,7 @@ int hibernation_restore(int platform_mode) | |||
| 362 | enable_nonboot_cpus(); | 382 | enable_nonboot_cpus(); |
| 363 | } | 383 | } |
| 364 | platform_restore_cleanup(platform_mode); | 384 | platform_restore_cleanup(platform_mode); |
| 365 | device_resume(); | 385 | device_resume(PMSG_RECOVER); |
| 366 | Finish: | 386 | Finish: |
| 367 | resume_console(); | 387 | resume_console(); |
| 368 | pm_restore_console(); | 388 | pm_restore_console(); |
| @@ -392,8 +412,11 @@ int hibernation_platform_enter(void) | |||
| 392 | 412 | ||
| 393 | suspend_console(); | 413 | suspend_console(); |
| 394 | error = device_suspend(PMSG_HIBERNATE); | 414 | error = device_suspend(PMSG_HIBERNATE); |
| 395 | if (error) | 415 | if (error) { |
| 396 | goto Resume_console; | 416 | if (hibernation_ops->recover) |
| 417 | hibernation_ops->recover(); | ||
| 418 | goto Resume_devices; | ||
| 419 | } | ||
| 397 | 420 | ||
| 398 | error = hibernation_ops->prepare(); | 421 | error = hibernation_ops->prepare(); |
| 399 | if (error) | 422 | if (error) |
| @@ -403,6 +426,7 @@ int hibernation_platform_enter(void) | |||
| 403 | if (error) | 426 | if (error) |
| 404 | goto Finish; | 427 | goto Finish; |
| 405 | 428 | ||
| 429 | device_pm_lock(); | ||
| 406 | local_irq_disable(); | 430 | local_irq_disable(); |
| 407 | error = device_power_down(PMSG_HIBERNATE); | 431 | error = device_power_down(PMSG_HIBERNATE); |
| 408 | if (!error) { | 432 | if (!error) { |
| @@ -411,6 +435,7 @@ int hibernation_platform_enter(void) | |||
| 411 | while (1); | 435 | while (1); |
| 412 | } | 436 | } |
| 413 | local_irq_enable(); | 437 | local_irq_enable(); |
| 438 | device_pm_unlock(); | ||
| 414 | 439 | ||
| 415 | /* | 440 | /* |
| 416 | * We don't need to reenable the nonboot CPUs or resume consoles, since | 441 | * We don't need to reenable the nonboot CPUs or resume consoles, since |
| @@ -419,8 +444,7 @@ int hibernation_platform_enter(void) | |||
| 419 | Finish: | 444 | Finish: |
| 420 | hibernation_ops->finish(); | 445 | hibernation_ops->finish(); |
| 421 | Resume_devices: | 446 | Resume_devices: |
| 422 | device_resume(); | 447 | device_resume(PMSG_RESTORE); |
| 423 | Resume_console: | ||
| 424 | resume_console(); | 448 | resume_console(); |
| 425 | Close: | 449 | Close: |
| 426 | hibernation_ops->end(); | 450 | hibernation_ops->end(); |
diff --git a/kernel/power/main.c b/kernel/power/main.c index 6a6d5eb3524e..3398f4651aa1 100644 --- a/kernel/power/main.c +++ b/kernel/power/main.c | |||
| @@ -228,6 +228,7 @@ static int suspend_enter(suspend_state_t state) | |||
| 228 | { | 228 | { |
| 229 | int error = 0; | 229 | int error = 0; |
| 230 | 230 | ||
| 231 | device_pm_lock(); | ||
| 231 | arch_suspend_disable_irqs(); | 232 | arch_suspend_disable_irqs(); |
| 232 | BUG_ON(!irqs_disabled()); | 233 | BUG_ON(!irqs_disabled()); |
| 233 | 234 | ||
| @@ -239,10 +240,11 @@ static int suspend_enter(suspend_state_t state) | |||
| 239 | if (!suspend_test(TEST_CORE)) | 240 | if (!suspend_test(TEST_CORE)) |
| 240 | error = suspend_ops->enter(state); | 241 | error = suspend_ops->enter(state); |
| 241 | 242 | ||
| 242 | device_power_up(); | 243 | device_power_up(PMSG_RESUME); |
| 243 | Done: | 244 | Done: |
| 244 | arch_suspend_enable_irqs(); | 245 | arch_suspend_enable_irqs(); |
| 245 | BUG_ON(irqs_disabled()); | 246 | BUG_ON(irqs_disabled()); |
| 247 | device_pm_unlock(); | ||
| 246 | return error; | 248 | return error; |
| 247 | } | 249 | } |
| 248 | 250 | ||
| @@ -267,11 +269,11 @@ int suspend_devices_and_enter(suspend_state_t state) | |||
| 267 | error = device_suspend(PMSG_SUSPEND); | 269 | error = device_suspend(PMSG_SUSPEND); |
| 268 | if (error) { | 270 | if (error) { |
| 269 | printk(KERN_ERR "PM: Some devices failed to suspend\n"); | 271 | printk(KERN_ERR "PM: Some devices failed to suspend\n"); |
| 270 | goto Resume_console; | 272 | goto Recover_platform; |
| 271 | } | 273 | } |
| 272 | 274 | ||
| 273 | if (suspend_test(TEST_DEVICES)) | 275 | if (suspend_test(TEST_DEVICES)) |
| 274 | goto Resume_devices; | 276 | goto Recover_platform; |
| 275 | 277 | ||
| 276 | if (suspend_ops->prepare) { | 278 | if (suspend_ops->prepare) { |
| 277 | error = suspend_ops->prepare(); | 279 | error = suspend_ops->prepare(); |
| @@ -291,13 +293,17 @@ int suspend_devices_and_enter(suspend_state_t state) | |||
| 291 | if (suspend_ops->finish) | 293 | if (suspend_ops->finish) |
| 292 | suspend_ops->finish(); | 294 | suspend_ops->finish(); |
| 293 | Resume_devices: | 295 | Resume_devices: |
| 294 | device_resume(); | 296 | device_resume(PMSG_RESUME); |
| 295 | Resume_console: | ||
| 296 | resume_console(); | 297 | resume_console(); |
| 297 | Close: | 298 | Close: |
| 298 | if (suspend_ops->end) | 299 | if (suspend_ops->end) |
| 299 | suspend_ops->end(); | 300 | suspend_ops->end(); |
| 300 | return error; | 301 | return error; |
| 302 | |||
| 303 | Recover_platform: | ||
| 304 | if (suspend_ops->recover) | ||
| 305 | suspend_ops->recover(); | ||
| 306 | goto Resume_devices; | ||
| 301 | } | 307 | } |
| 302 | 308 | ||
| 303 | /** | 309 | /** |
diff --git a/kernel/power/process.c b/kernel/power/process.c index f1d0b345c9ba..5fb87652f214 100644 --- a/kernel/power/process.c +++ b/kernel/power/process.c | |||
| @@ -19,9 +19,6 @@ | |||
| 19 | */ | 19 | */ |
| 20 | #define TIMEOUT (20 * HZ) | 20 | #define TIMEOUT (20 * HZ) |
| 21 | 21 | ||
| 22 | #define FREEZER_KERNEL_THREADS 0 | ||
| 23 | #define FREEZER_USER_SPACE 1 | ||
| 24 | |||
| 25 | static inline int freezeable(struct task_struct * p) | 22 | static inline int freezeable(struct task_struct * p) |
| 26 | { | 23 | { |
| 27 | if ((p == current) || | 24 | if ((p == current) || |
| @@ -84,63 +81,53 @@ static void fake_signal_wake_up(struct task_struct *p) | |||
| 84 | spin_unlock_irqrestore(&p->sighand->siglock, flags); | 81 | spin_unlock_irqrestore(&p->sighand->siglock, flags); |
| 85 | } | 82 | } |
| 86 | 83 | ||
| 87 | static int has_mm(struct task_struct *p) | 84 | static inline bool should_send_signal(struct task_struct *p) |
| 88 | { | 85 | { |
| 89 | return (p->mm && !(p->flags & PF_BORROWED_MM)); | 86 | return !(p->flags & PF_FREEZER_NOSIG); |
| 90 | } | 87 | } |
| 91 | 88 | ||
| 92 | /** | 89 | /** |
| 93 | * freeze_task - send a freeze request to given task | 90 | * freeze_task - send a freeze request to given task |
| 94 | * @p: task to send the request to | 91 | * @p: task to send the request to |
| 95 | * @with_mm_only: if set, the request will only be sent if the task has its | 92 | * @sig_only: if set, the request will only be sent if the task has the |
| 96 | * own mm | 93 | * PF_FREEZER_NOSIG flag unset |
| 97 | * Return value: 0, if @with_mm_only is set and the task has no mm of its | 94 | * Return value: 'false', if @sig_only is set and the task has |
| 98 | * own or the task is frozen, 1, otherwise | 95 | * PF_FREEZER_NOSIG set or the task is frozen, 'true', otherwise |
| 99 | * | 96 | * |
| 100 | * The freeze request is sent by seting the tasks's TIF_FREEZE flag and | 97 | * The freeze request is sent by setting the tasks's TIF_FREEZE flag and |
| 101 | * either sending a fake signal to it or waking it up, depending on whether | 98 | * either sending a fake signal to it or waking it up, depending on whether |
| 102 | * or not it has its own mm (ie. it is a user land task). If @with_mm_only | 99 | * or not it has PF_FREEZER_NOSIG set. If @sig_only is set and the task |
| 103 | * is set and the task has no mm of its own (ie. it is a kernel thread), | 100 | * has PF_FREEZER_NOSIG set (ie. it is a typical kernel thread), its |
| 104 | * its TIF_FREEZE flag should not be set. | 101 | * TIF_FREEZE flag will not be set. |
| 105 | * | ||
| 106 | * The task_lock() is necessary to prevent races with exit_mm() or | ||
| 107 | * use_mm()/unuse_mm() from occuring. | ||
| 108 | */ | 102 | */ |
| 109 | static int freeze_task(struct task_struct *p, int with_mm_only) | 103 | static bool freeze_task(struct task_struct *p, bool sig_only) |
| 110 | { | 104 | { |
| 111 | int ret = 1; | 105 | /* |
| 106 | * We first check if the task is freezing and next if it has already | ||
| 107 | * been frozen to avoid the race with frozen_process() which first marks | ||
| 108 | * the task as frozen and next clears its TIF_FREEZE. | ||
| 109 | */ | ||
| 110 | if (!freezing(p)) { | ||
| 111 | rmb(); | ||
| 112 | if (frozen(p)) | ||
| 113 | return false; | ||
| 112 | 114 | ||
| 113 | task_lock(p); | 115 | if (!sig_only || should_send_signal(p)) |
| 114 | if (freezing(p)) { | 116 | set_freeze_flag(p); |
| 115 | if (has_mm(p)) { | 117 | else |
| 116 | if (!signal_pending(p)) | 118 | return false; |
| 117 | fake_signal_wake_up(p); | 119 | } |
| 118 | } else { | 120 | |
| 119 | if (with_mm_only) | 121 | if (should_send_signal(p)) { |
| 120 | ret = 0; | 122 | if (!signal_pending(p)) |
| 121 | else | 123 | fake_signal_wake_up(p); |
| 122 | wake_up_state(p, TASK_INTERRUPTIBLE); | 124 | } else if (sig_only) { |
| 123 | } | 125 | return false; |
| 124 | } else { | 126 | } else { |
| 125 | rmb(); | 127 | wake_up_state(p, TASK_INTERRUPTIBLE); |
| 126 | if (frozen(p)) { | ||
| 127 | ret = 0; | ||
| 128 | } else { | ||
| 129 | if (has_mm(p)) { | ||
| 130 | set_freeze_flag(p); | ||
| 131 | fake_signal_wake_up(p); | ||
| 132 | } else { | ||
| 133 | if (with_mm_only) { | ||
| 134 | ret = 0; | ||
| 135 | } else { | ||
| 136 | set_freeze_flag(p); | ||
| 137 | wake_up_state(p, TASK_INTERRUPTIBLE); | ||
| 138 | } | ||
| 139 | } | ||
| 140 | } | ||
| 141 | } | 128 | } |
| 142 | task_unlock(p); | 129 | |
| 143 | return ret; | 130 | return true; |
| 144 | } | 131 | } |
| 145 | 132 | ||
| 146 | static void cancel_freezing(struct task_struct *p) | 133 | static void cancel_freezing(struct task_struct *p) |
| @@ -156,7 +143,7 @@ static void cancel_freezing(struct task_struct *p) | |||
| 156 | } | 143 | } |
| 157 | } | 144 | } |
| 158 | 145 | ||
| 159 | static int try_to_freeze_tasks(int freeze_user_space) | 146 | static int try_to_freeze_tasks(bool sig_only) |
| 160 | { | 147 | { |
| 161 | struct task_struct *g, *p; | 148 | struct task_struct *g, *p; |
| 162 | unsigned long end_time; | 149 | unsigned long end_time; |
| @@ -175,7 +162,7 @@ static int try_to_freeze_tasks(int freeze_user_space) | |||
| 175 | if (frozen(p) || !freezeable(p)) | 162 | if (frozen(p) || !freezeable(p)) |
| 176 | continue; | 163 | continue; |
| 177 | 164 | ||
| 178 | if (!freeze_task(p, freeze_user_space)) | 165 | if (!freeze_task(p, sig_only)) |
| 179 | continue; | 166 | continue; |
| 180 | 167 | ||
| 181 | /* | 168 | /* |
| @@ -235,13 +222,13 @@ int freeze_processes(void) | |||
| 235 | int error; | 222 | int error; |
| 236 | 223 | ||
| 237 | printk("Freezing user space processes ... "); | 224 | printk("Freezing user space processes ... "); |
| 238 | error = try_to_freeze_tasks(FREEZER_USER_SPACE); | 225 | error = try_to_freeze_tasks(true); |
| 239 | if (error) | 226 | if (error) |
| 240 | goto Exit; | 227 | goto Exit; |
| 241 | printk("done.\n"); | 228 | printk("done.\n"); |
| 242 | 229 | ||
| 243 | printk("Freezing remaining freezable tasks ... "); | 230 | printk("Freezing remaining freezable tasks ... "); |
| 244 | error = try_to_freeze_tasks(FREEZER_KERNEL_THREADS); | 231 | error = try_to_freeze_tasks(false); |
| 245 | if (error) | 232 | if (error) |
| 246 | goto Exit; | 233 | goto Exit; |
| 247 | printk("done."); | 234 | printk("done."); |
| @@ -251,7 +238,7 @@ int freeze_processes(void) | |||
| 251 | return error; | 238 | return error; |
| 252 | } | 239 | } |
| 253 | 240 | ||
| 254 | static void thaw_tasks(int thaw_user_space) | 241 | static void thaw_tasks(bool nosig_only) |
| 255 | { | 242 | { |
| 256 | struct task_struct *g, *p; | 243 | struct task_struct *g, *p; |
| 257 | 244 | ||
| @@ -260,7 +247,7 @@ static void thaw_tasks(int thaw_user_space) | |||
| 260 | if (!freezeable(p)) | 247 | if (!freezeable(p)) |
| 261 | continue; | 248 | continue; |
| 262 | 249 | ||
| 263 | if (!p->mm == thaw_user_space) | 250 | if (nosig_only && should_send_signal(p)) |
| 264 | continue; | 251 | continue; |
| 265 | 252 | ||
| 266 | thaw_process(p); | 253 | thaw_process(p); |
| @@ -271,8 +258,8 @@ static void thaw_tasks(int thaw_user_space) | |||
| 271 | void thaw_processes(void) | 258 | void thaw_processes(void) |
| 272 | { | 259 | { |
| 273 | printk("Restarting tasks ... "); | 260 | printk("Restarting tasks ... "); |
| 274 | thaw_tasks(FREEZER_KERNEL_THREADS); | 261 | thaw_tasks(true); |
| 275 | thaw_tasks(FREEZER_USER_SPACE); | 262 | thaw_tasks(false); |
| 276 | schedule(); | 263 | schedule(); |
| 277 | printk("done.\n"); | 264 | printk("done.\n"); |
| 278 | } | 265 | } |
diff --git a/kernel/power/user.c b/kernel/power/user.c index f5512cb3aa86..a6332a313262 100644 --- a/kernel/power/user.c +++ b/kernel/power/user.c | |||
| @@ -23,6 +23,7 @@ | |||
| 23 | #include <linux/console.h> | 23 | #include <linux/console.h> |
| 24 | #include <linux/cpu.h> | 24 | #include <linux/cpu.h> |
| 25 | #include <linux/freezer.h> | 25 | #include <linux/freezer.h> |
| 26 | #include <linux/smp_lock.h> | ||
| 26 | 27 | ||
| 27 | #include <asm/uaccess.h> | 28 | #include <asm/uaccess.h> |
| 28 | 29 | ||
| @@ -69,16 +70,22 @@ static int snapshot_open(struct inode *inode, struct file *filp) | |||
| 69 | struct snapshot_data *data; | 70 | struct snapshot_data *data; |
| 70 | int error; | 71 | int error; |
| 71 | 72 | ||
| 72 | if (!atomic_add_unless(&snapshot_device_available, -1, 0)) | 73 | mutex_lock(&pm_mutex); |
| 73 | return -EBUSY; | 74 | |
| 75 | if (!atomic_add_unless(&snapshot_device_available, -1, 0)) { | ||
| 76 | error = -EBUSY; | ||
| 77 | goto Unlock; | ||
| 78 | } | ||
| 74 | 79 | ||
| 75 | if ((filp->f_flags & O_ACCMODE) == O_RDWR) { | 80 | if ((filp->f_flags & O_ACCMODE) == O_RDWR) { |
| 76 | atomic_inc(&snapshot_device_available); | 81 | atomic_inc(&snapshot_device_available); |
| 77 | return -ENOSYS; | 82 | error = -ENOSYS; |
| 83 | goto Unlock; | ||
| 78 | } | 84 | } |
| 79 | if(create_basic_memory_bitmaps()) { | 85 | if(create_basic_memory_bitmaps()) { |
| 80 | atomic_inc(&snapshot_device_available); | 86 | atomic_inc(&snapshot_device_available); |
| 81 | return -ENOMEM; | 87 | error = -ENOMEM; |
| 88 | goto Unlock; | ||
| 82 | } | 89 | } |
| 83 | nonseekable_open(inode, filp); | 90 | nonseekable_open(inode, filp); |
| 84 | data = &snapshot_state; | 91 | data = &snapshot_state; |
| @@ -98,33 +105,36 @@ static int snapshot_open(struct inode *inode, struct file *filp) | |||
| 98 | if (error) | 105 | if (error) |
| 99 | pm_notifier_call_chain(PM_POST_HIBERNATION); | 106 | pm_notifier_call_chain(PM_POST_HIBERNATION); |
| 100 | } | 107 | } |
| 101 | if (error) { | 108 | if (error) |
| 102 | atomic_inc(&snapshot_device_available); | 109 | atomic_inc(&snapshot_device_available); |
| 103 | return error; | ||
| 104 | } | ||
| 105 | data->frozen = 0; | 110 | data->frozen = 0; |
| 106 | data->ready = 0; | 111 | data->ready = 0; |
| 107 | data->platform_support = 0; | 112 | data->platform_support = 0; |
| 108 | 113 | ||
| 109 | return 0; | 114 | Unlock: |
| 115 | mutex_unlock(&pm_mutex); | ||
| 116 | |||
| 117 | return error; | ||
| 110 | } | 118 | } |
| 111 | 119 | ||
| 112 | static int snapshot_release(struct inode *inode, struct file *filp) | 120 | static int snapshot_release(struct inode *inode, struct file *filp) |
| 113 | { | 121 | { |
| 114 | struct snapshot_data *data; | 122 | struct snapshot_data *data; |
| 115 | 123 | ||
| 124 | mutex_lock(&pm_mutex); | ||
| 125 | |||
| 116 | swsusp_free(); | 126 | swsusp_free(); |
| 117 | free_basic_memory_bitmaps(); | 127 | free_basic_memory_bitmaps(); |
| 118 | data = filp->private_data; | 128 | data = filp->private_data; |
| 119 | free_all_swap_pages(data->swap); | 129 | free_all_swap_pages(data->swap); |
| 120 | if (data->frozen) { | 130 | if (data->frozen) |
| 121 | mutex_lock(&pm_mutex); | ||
| 122 | thaw_processes(); | 131 | thaw_processes(); |
| 123 | mutex_unlock(&pm_mutex); | ||
| 124 | } | ||
| 125 | pm_notifier_call_chain(data->mode == O_WRONLY ? | 132 | pm_notifier_call_chain(data->mode == O_WRONLY ? |
| 126 | PM_POST_HIBERNATION : PM_POST_RESTORE); | 133 | PM_POST_HIBERNATION : PM_POST_RESTORE); |
| 127 | atomic_inc(&snapshot_device_available); | 134 | atomic_inc(&snapshot_device_available); |
| 135 | |||
| 136 | mutex_unlock(&pm_mutex); | ||
| 137 | |||
| 128 | return 0; | 138 | return 0; |
| 129 | } | 139 | } |
| 130 | 140 | ||
| @@ -134,9 +144,13 @@ static ssize_t snapshot_read(struct file *filp, char __user *buf, | |||
| 134 | struct snapshot_data *data; | 144 | struct snapshot_data *data; |
| 135 | ssize_t res; | 145 | ssize_t res; |
| 136 | 146 | ||
| 147 | mutex_lock(&pm_mutex); | ||
| 148 | |||
| 137 | data = filp->private_data; | 149 | data = filp->private_data; |
| 138 | if (!data->ready) | 150 | if (!data->ready) { |
| 139 | return -ENODATA; | 151 | res = -ENODATA; |
| 152 | goto Unlock; | ||
| 153 | } | ||
| 140 | res = snapshot_read_next(&data->handle, count); | 154 | res = snapshot_read_next(&data->handle, count); |
| 141 | if (res > 0) { | 155 | if (res > 0) { |
| 142 | if (copy_to_user(buf, data_of(data->handle), res)) | 156 | if (copy_to_user(buf, data_of(data->handle), res)) |
| @@ -144,6 +158,10 @@ static ssize_t snapshot_read(struct file *filp, char __user *buf, | |||
| 144 | else | 158 | else |
| 145 | *offp = data->handle.offset; | 159 | *offp = data->handle.offset; |
| 146 | } | 160 | } |
| 161 | |||
| 162 | Unlock: | ||
| 163 | mutex_unlock(&pm_mutex); | ||
| 164 | |||
| 147 | return res; | 165 | return res; |
| 148 | } | 166 | } |
| 149 | 167 | ||
| @@ -153,6 +171,8 @@ static ssize_t snapshot_write(struct file *filp, const char __user *buf, | |||
| 153 | struct snapshot_data *data; | 171 | struct snapshot_data *data; |
| 154 | ssize_t res; | 172 | ssize_t res; |
| 155 | 173 | ||
| 174 | mutex_lock(&pm_mutex); | ||
| 175 | |||
| 156 | data = filp->private_data; | 176 | data = filp->private_data; |
| 157 | res = snapshot_write_next(&data->handle, count); | 177 | res = snapshot_write_next(&data->handle, count); |
| 158 | if (res > 0) { | 178 | if (res > 0) { |
| @@ -161,11 +181,14 @@ static ssize_t snapshot_write(struct file *filp, const char __user *buf, | |||
| 161 | else | 181 | else |
| 162 | *offp = data->handle.offset; | 182 | *offp = data->handle.offset; |
| 163 | } | 183 | } |
| 184 | |||
| 185 | mutex_unlock(&pm_mutex); | ||
| 186 | |||
| 164 | return res; | 187 | return res; |
| 165 | } | 188 | } |
| 166 | 189 | ||
| 167 | static int snapshot_ioctl(struct inode *inode, struct file *filp, | 190 | static long snapshot_ioctl(struct file *filp, unsigned int cmd, |
| 168 | unsigned int cmd, unsigned long arg) | 191 | unsigned long arg) |
| 169 | { | 192 | { |
| 170 | int error = 0; | 193 | int error = 0; |
| 171 | struct snapshot_data *data; | 194 | struct snapshot_data *data; |
| @@ -179,6 +202,9 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 179 | if (!capable(CAP_SYS_ADMIN)) | 202 | if (!capable(CAP_SYS_ADMIN)) |
| 180 | return -EPERM; | 203 | return -EPERM; |
| 181 | 204 | ||
| 205 | if (!mutex_trylock(&pm_mutex)) | ||
| 206 | return -EBUSY; | ||
| 207 | |||
| 182 | data = filp->private_data; | 208 | data = filp->private_data; |
| 183 | 209 | ||
| 184 | switch (cmd) { | 210 | switch (cmd) { |
| @@ -186,7 +212,6 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 186 | case SNAPSHOT_FREEZE: | 212 | case SNAPSHOT_FREEZE: |
| 187 | if (data->frozen) | 213 | if (data->frozen) |
| 188 | break; | 214 | break; |
| 189 | mutex_lock(&pm_mutex); | ||
| 190 | printk("Syncing filesystems ... "); | 215 | printk("Syncing filesystems ... "); |
| 191 | sys_sync(); | 216 | sys_sync(); |
| 192 | printk("done.\n"); | 217 | printk("done.\n"); |
| @@ -194,7 +219,6 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 194 | error = freeze_processes(); | 219 | error = freeze_processes(); |
| 195 | if (error) | 220 | if (error) |
| 196 | thaw_processes(); | 221 | thaw_processes(); |
| 197 | mutex_unlock(&pm_mutex); | ||
| 198 | if (!error) | 222 | if (!error) |
| 199 | data->frozen = 1; | 223 | data->frozen = 1; |
| 200 | break; | 224 | break; |
| @@ -202,9 +226,7 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 202 | case SNAPSHOT_UNFREEZE: | 226 | case SNAPSHOT_UNFREEZE: |
| 203 | if (!data->frozen || data->ready) | 227 | if (!data->frozen || data->ready) |
| 204 | break; | 228 | break; |
| 205 | mutex_lock(&pm_mutex); | ||
| 206 | thaw_processes(); | 229 | thaw_processes(); |
| 207 | mutex_unlock(&pm_mutex); | ||
| 208 | data->frozen = 0; | 230 | data->frozen = 0; |
| 209 | break; | 231 | break; |
| 210 | 232 | ||
| @@ -307,16 +329,11 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 307 | error = -EPERM; | 329 | error = -EPERM; |
| 308 | break; | 330 | break; |
| 309 | } | 331 | } |
| 310 | if (!mutex_trylock(&pm_mutex)) { | ||
| 311 | error = -EBUSY; | ||
| 312 | break; | ||
| 313 | } | ||
| 314 | /* | 332 | /* |
| 315 | * Tasks are frozen and the notifiers have been called with | 333 | * Tasks are frozen and the notifiers have been called with |
| 316 | * PM_HIBERNATION_PREPARE | 334 | * PM_HIBERNATION_PREPARE |
| 317 | */ | 335 | */ |
| 318 | error = suspend_devices_and_enter(PM_SUSPEND_MEM); | 336 | error = suspend_devices_and_enter(PM_SUSPEND_MEM); |
| 319 | mutex_unlock(&pm_mutex); | ||
| 320 | break; | 337 | break; |
| 321 | 338 | ||
| 322 | case SNAPSHOT_PLATFORM_SUPPORT: | 339 | case SNAPSHOT_PLATFORM_SUPPORT: |
| @@ -390,6 +407,8 @@ static int snapshot_ioctl(struct inode *inode, struct file *filp, | |||
| 390 | 407 | ||
| 391 | } | 408 | } |
| 392 | 409 | ||
| 410 | mutex_unlock(&pm_mutex); | ||
| 411 | |||
| 393 | return error; | 412 | return error; |
| 394 | } | 413 | } |
| 395 | 414 | ||
| @@ -399,7 +418,7 @@ static const struct file_operations snapshot_fops = { | |||
| 399 | .read = snapshot_read, | 418 | .read = snapshot_read, |
| 400 | .write = snapshot_write, | 419 | .write = snapshot_write, |
| 401 | .llseek = no_llseek, | 420 | .llseek = no_llseek, |
| 402 | .ioctl = snapshot_ioctl, | 421 | .unlocked_ioctl = snapshot_ioctl, |
| 403 | }; | 422 | }; |
| 404 | 423 | ||
| 405 | static struct miscdevice snapshot_device = { | 424 | static struct miscdevice snapshot_device = { |
