diff options
Diffstat (limited to 'kernel/power')
-rw-r--r-- | kernel/power/Kconfig | 14 | ||||
-rw-r--r-- | kernel/power/hibernate.c | 21 | ||||
-rw-r--r-- | kernel/power/main.c | 17 | ||||
-rw-r--r-- | kernel/power/power.h | 2 | ||||
-rw-r--r-- | kernel/power/snapshot.c | 412 |
5 files changed, 359 insertions, 107 deletions
diff --git a/kernel/power/Kconfig b/kernel/power/Kconfig index 72067cbdb37f..91e09d3b2eb2 100644 --- a/kernel/power/Kconfig +++ b/kernel/power/Kconfig | |||
@@ -208,3 +208,17 @@ config APM_EMULATION | |||
208 | random kernel OOPSes or reboots that don't seem to be related to | 208 | random kernel OOPSes or reboots that don't seem to be related to |
209 | anything, try disabling/enabling this option (or disabling/enabling | 209 | anything, try disabling/enabling this option (or disabling/enabling |
210 | APM in your BIOS). | 210 | APM in your BIOS). |
211 | |||
212 | config PM_RUNTIME | ||
213 | bool "Run-time PM core functionality" | ||
214 | depends on PM | ||
215 | ---help--- | ||
216 | Enable functionality allowing I/O devices to be put into energy-saving | ||
217 | (low power) states at run time (or autosuspended) after a specified | ||
218 | period of inactivity and woken up in response to a hardware-generated | ||
219 | wake-up event or a driver's request. | ||
220 | |||
221 | Hardware support is generally required for this functionality to work | ||
222 | and the bus type drivers of the buses the devices are on are | ||
223 | responsible for the actual handling of the autosuspend requests and | ||
224 | wake-up events. | ||
diff --git a/kernel/power/hibernate.c b/kernel/power/hibernate.c index 81d2e7464893..04b3a83d686f 100644 --- a/kernel/power/hibernate.c +++ b/kernel/power/hibernate.c | |||
@@ -298,8 +298,8 @@ int hibernation_snapshot(int platform_mode) | |||
298 | if (error) | 298 | if (error) |
299 | return error; | 299 | return error; |
300 | 300 | ||
301 | /* Free memory before shutting down devices. */ | 301 | /* Preallocate image memory before shutting down devices. */ |
302 | error = swsusp_shrink_memory(); | 302 | error = hibernate_preallocate_memory(); |
303 | if (error) | 303 | if (error) |
304 | goto Close; | 304 | goto Close; |
305 | 305 | ||
@@ -315,6 +315,10 @@ int hibernation_snapshot(int platform_mode) | |||
315 | /* Control returns here after successful restore */ | 315 | /* Control returns here after successful restore */ |
316 | 316 | ||
317 | Resume_devices: | 317 | Resume_devices: |
318 | /* We may need to release the preallocated image pages here. */ | ||
319 | if (error || !in_suspend) | ||
320 | swsusp_free(); | ||
321 | |||
318 | dpm_resume_end(in_suspend ? | 322 | dpm_resume_end(in_suspend ? |
319 | (error ? PMSG_RECOVER : PMSG_THAW) : PMSG_RESTORE); | 323 | (error ? PMSG_RECOVER : PMSG_THAW) : PMSG_RESTORE); |
320 | resume_console(); | 324 | resume_console(); |
@@ -460,11 +464,11 @@ int hibernation_platform_enter(void) | |||
460 | 464 | ||
461 | error = hibernation_ops->prepare(); | 465 | error = hibernation_ops->prepare(); |
462 | if (error) | 466 | if (error) |
463 | goto Platofrm_finish; | 467 | goto Platform_finish; |
464 | 468 | ||
465 | error = disable_nonboot_cpus(); | 469 | error = disable_nonboot_cpus(); |
466 | if (error) | 470 | if (error) |
467 | goto Platofrm_finish; | 471 | goto Platform_finish; |
468 | 472 | ||
469 | local_irq_disable(); | 473 | local_irq_disable(); |
470 | sysdev_suspend(PMSG_HIBERNATE); | 474 | sysdev_suspend(PMSG_HIBERNATE); |
@@ -476,7 +480,7 @@ int hibernation_platform_enter(void) | |||
476 | * We don't need to reenable the nonboot CPUs or resume consoles, since | 480 | * We don't need to reenable the nonboot CPUs or resume consoles, since |
477 | * the system is going to be halted anyway. | 481 | * the system is going to be halted anyway. |
478 | */ | 482 | */ |
479 | Platofrm_finish: | 483 | Platform_finish: |
480 | hibernation_ops->finish(); | 484 | hibernation_ops->finish(); |
481 | 485 | ||
482 | dpm_suspend_noirq(PMSG_RESTORE); | 486 | dpm_suspend_noirq(PMSG_RESTORE); |
@@ -578,7 +582,10 @@ int hibernate(void) | |||
578 | goto Thaw; | 582 | goto Thaw; |
579 | 583 | ||
580 | error = hibernation_snapshot(hibernation_mode == HIBERNATION_PLATFORM); | 584 | error = hibernation_snapshot(hibernation_mode == HIBERNATION_PLATFORM); |
581 | if (in_suspend && !error) { | 585 | if (error) |
586 | goto Thaw; | ||
587 | |||
588 | if (in_suspend) { | ||
582 | unsigned int flags = 0; | 589 | unsigned int flags = 0; |
583 | 590 | ||
584 | if (hibernation_mode == HIBERNATION_PLATFORM) | 591 | if (hibernation_mode == HIBERNATION_PLATFORM) |
@@ -590,8 +597,8 @@ int hibernate(void) | |||
590 | power_down(); | 597 | power_down(); |
591 | } else { | 598 | } else { |
592 | pr_debug("PM: Image restored successfully.\n"); | 599 | pr_debug("PM: Image restored successfully.\n"); |
593 | swsusp_free(); | ||
594 | } | 600 | } |
601 | |||
595 | Thaw: | 602 | Thaw: |
596 | thaw_processes(); | 603 | thaw_processes(); |
597 | Finish: | 604 | Finish: |
diff --git a/kernel/power/main.c b/kernel/power/main.c index f710e36930cc..347d2cc88cd0 100644 --- a/kernel/power/main.c +++ b/kernel/power/main.c | |||
@@ -11,6 +11,7 @@ | |||
11 | #include <linux/kobject.h> | 11 | #include <linux/kobject.h> |
12 | #include <linux/string.h> | 12 | #include <linux/string.h> |
13 | #include <linux/resume-trace.h> | 13 | #include <linux/resume-trace.h> |
14 | #include <linux/workqueue.h> | ||
14 | 15 | ||
15 | #include "power.h" | 16 | #include "power.h" |
16 | 17 | ||
@@ -217,8 +218,24 @@ static struct attribute_group attr_group = { | |||
217 | .attrs = g, | 218 | .attrs = g, |
218 | }; | 219 | }; |
219 | 220 | ||
221 | #ifdef CONFIG_PM_RUNTIME | ||
222 | struct workqueue_struct *pm_wq; | ||
223 | |||
224 | static int __init pm_start_workqueue(void) | ||
225 | { | ||
226 | pm_wq = create_freezeable_workqueue("pm"); | ||
227 | |||
228 | return pm_wq ? 0 : -ENOMEM; | ||
229 | } | ||
230 | #else | ||
231 | static inline int pm_start_workqueue(void) { return 0; } | ||
232 | #endif | ||
233 | |||
220 | static int __init pm_init(void) | 234 | static int __init pm_init(void) |
221 | { | 235 | { |
236 | int error = pm_start_workqueue(); | ||
237 | if (error) | ||
238 | return error; | ||
222 | power_kobj = kobject_create_and_add("power", NULL); | 239 | power_kobj = kobject_create_and_add("power", NULL); |
223 | if (!power_kobj) | 240 | if (!power_kobj) |
224 | return -ENOMEM; | 241 | return -ENOMEM; |
diff --git a/kernel/power/power.h b/kernel/power/power.h index 26d5a26f82e3..46c5a26630a3 100644 --- a/kernel/power/power.h +++ b/kernel/power/power.h | |||
@@ -74,7 +74,7 @@ extern asmlinkage int swsusp_arch_resume(void); | |||
74 | 74 | ||
75 | extern int create_basic_memory_bitmaps(void); | 75 | extern int create_basic_memory_bitmaps(void); |
76 | extern void free_basic_memory_bitmaps(void); | 76 | extern void free_basic_memory_bitmaps(void); |
77 | extern int swsusp_shrink_memory(void); | 77 | extern int hibernate_preallocate_memory(void); |
78 | 78 | ||
79 | /** | 79 | /** |
80 | * Auxiliary structure used for reading the snapshot image data and | 80 | * Auxiliary structure used for reading the snapshot image data and |
diff --git a/kernel/power/snapshot.c b/kernel/power/snapshot.c index 523a451b45d3..97955b0e44f4 100644 --- a/kernel/power/snapshot.c +++ b/kernel/power/snapshot.c | |||
@@ -233,7 +233,7 @@ static void *chain_alloc(struct chain_allocator *ca, unsigned int size) | |||
233 | 233 | ||
234 | #define BM_END_OF_MAP (~0UL) | 234 | #define BM_END_OF_MAP (~0UL) |
235 | 235 | ||
236 | #define BM_BITS_PER_BLOCK (PAGE_SIZE << 3) | 236 | #define BM_BITS_PER_BLOCK (PAGE_SIZE * BITS_PER_BYTE) |
237 | 237 | ||
238 | struct bm_block { | 238 | struct bm_block { |
239 | struct list_head hook; /* hook into a list of bitmap blocks */ | 239 | struct list_head hook; /* hook into a list of bitmap blocks */ |
@@ -275,7 +275,7 @@ static void memory_bm_free(struct memory_bitmap *bm, int clear_nosave_free); | |||
275 | 275 | ||
276 | /** | 276 | /** |
277 | * create_bm_block_list - create a list of block bitmap objects | 277 | * create_bm_block_list - create a list of block bitmap objects |
278 | * @nr_blocks - number of blocks to allocate | 278 | * @pages - number of pages to track |
279 | * @list - list to put the allocated blocks into | 279 | * @list - list to put the allocated blocks into |
280 | * @ca - chain allocator to be used for allocating memory | 280 | * @ca - chain allocator to be used for allocating memory |
281 | */ | 281 | */ |
@@ -853,7 +853,7 @@ static unsigned int count_highmem_pages(void) | |||
853 | struct zone *zone; | 853 | struct zone *zone; |
854 | unsigned int n = 0; | 854 | unsigned int n = 0; |
855 | 855 | ||
856 | for_each_zone(zone) { | 856 | for_each_populated_zone(zone) { |
857 | unsigned long pfn, max_zone_pfn; | 857 | unsigned long pfn, max_zone_pfn; |
858 | 858 | ||
859 | if (!is_highmem(zone)) | 859 | if (!is_highmem(zone)) |
@@ -916,7 +916,7 @@ static unsigned int count_data_pages(void) | |||
916 | unsigned long pfn, max_zone_pfn; | 916 | unsigned long pfn, max_zone_pfn; |
917 | unsigned int n = 0; | 917 | unsigned int n = 0; |
918 | 918 | ||
919 | for_each_zone(zone) { | 919 | for_each_populated_zone(zone) { |
920 | if (is_highmem(zone)) | 920 | if (is_highmem(zone)) |
921 | continue; | 921 | continue; |
922 | 922 | ||
@@ -1010,7 +1010,7 @@ copy_data_pages(struct memory_bitmap *copy_bm, struct memory_bitmap *orig_bm) | |||
1010 | struct zone *zone; | 1010 | struct zone *zone; |
1011 | unsigned long pfn; | 1011 | unsigned long pfn; |
1012 | 1012 | ||
1013 | for_each_zone(zone) { | 1013 | for_each_populated_zone(zone) { |
1014 | unsigned long max_zone_pfn; | 1014 | unsigned long max_zone_pfn; |
1015 | 1015 | ||
1016 | mark_free_pages(zone); | 1016 | mark_free_pages(zone); |
@@ -1033,6 +1033,25 @@ copy_data_pages(struct memory_bitmap *copy_bm, struct memory_bitmap *orig_bm) | |||
1033 | static unsigned int nr_copy_pages; | 1033 | static unsigned int nr_copy_pages; |
1034 | /* Number of pages needed for saving the original pfns of the image pages */ | 1034 | /* Number of pages needed for saving the original pfns of the image pages */ |
1035 | static unsigned int nr_meta_pages; | 1035 | static unsigned int nr_meta_pages; |
1036 | /* | ||
1037 | * Numbers of normal and highmem page frames allocated for hibernation image | ||
1038 | * before suspending devices. | ||
1039 | */ | ||
1040 | unsigned int alloc_normal, alloc_highmem; | ||
1041 | /* | ||
1042 | * Memory bitmap used for marking saveable pages (during hibernation) or | ||
1043 | * hibernation image pages (during restore) | ||
1044 | */ | ||
1045 | static struct memory_bitmap orig_bm; | ||
1046 | /* | ||
1047 | * Memory bitmap used during hibernation for marking allocated page frames that | ||
1048 | * will contain copies of saveable pages. During restore it is initially used | ||
1049 | * for marking hibernation image pages, but then the set bits from it are | ||
1050 | * duplicated in @orig_bm and it is released. On highmem systems it is next | ||
1051 | * used for marking "safe" highmem pages, but it has to be reinitialized for | ||
1052 | * this purpose. | ||
1053 | */ | ||
1054 | static struct memory_bitmap copy_bm; | ||
1036 | 1055 | ||
1037 | /** | 1056 | /** |
1038 | * swsusp_free - free pages allocated for the suspend. | 1057 | * swsusp_free - free pages allocated for the suspend. |
@@ -1046,7 +1065,7 @@ void swsusp_free(void) | |||
1046 | struct zone *zone; | 1065 | struct zone *zone; |
1047 | unsigned long pfn, max_zone_pfn; | 1066 | unsigned long pfn, max_zone_pfn; |
1048 | 1067 | ||
1049 | for_each_zone(zone) { | 1068 | for_each_populated_zone(zone) { |
1050 | max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; | 1069 | max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; |
1051 | for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) | 1070 | for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) |
1052 | if (pfn_valid(pfn)) { | 1071 | if (pfn_valid(pfn)) { |
@@ -1064,74 +1083,286 @@ void swsusp_free(void) | |||
1064 | nr_meta_pages = 0; | 1083 | nr_meta_pages = 0; |
1065 | restore_pblist = NULL; | 1084 | restore_pblist = NULL; |
1066 | buffer = NULL; | 1085 | buffer = NULL; |
1086 | alloc_normal = 0; | ||
1087 | alloc_highmem = 0; | ||
1067 | } | 1088 | } |
1068 | 1089 | ||
1090 | /* Helper functions used for the shrinking of memory. */ | ||
1091 | |||
1092 | #define GFP_IMAGE (GFP_KERNEL | __GFP_NOWARN) | ||
1093 | |||
1069 | /** | 1094 | /** |
1070 | * swsusp_shrink_memory - Try to free as much memory as needed | 1095 | * preallocate_image_pages - Allocate a number of pages for hibernation image |
1071 | * | 1096 | * @nr_pages: Number of page frames to allocate. |
1072 | * ... but do not OOM-kill anyone | 1097 | * @mask: GFP flags to use for the allocation. |
1073 | * | 1098 | * |
1074 | * Notice: all userland should be stopped before it is called, or | 1099 | * Return value: Number of page frames actually allocated |
1075 | * livelock is possible. | 1100 | */ |
1101 | static unsigned long preallocate_image_pages(unsigned long nr_pages, gfp_t mask) | ||
1102 | { | ||
1103 | unsigned long nr_alloc = 0; | ||
1104 | |||
1105 | while (nr_pages > 0) { | ||
1106 | struct page *page; | ||
1107 | |||
1108 | page = alloc_image_page(mask); | ||
1109 | if (!page) | ||
1110 | break; | ||
1111 | memory_bm_set_bit(©_bm, page_to_pfn(page)); | ||
1112 | if (PageHighMem(page)) | ||
1113 | alloc_highmem++; | ||
1114 | else | ||
1115 | alloc_normal++; | ||
1116 | nr_pages--; | ||
1117 | nr_alloc++; | ||
1118 | } | ||
1119 | |||
1120 | return nr_alloc; | ||
1121 | } | ||
1122 | |||
1123 | static unsigned long preallocate_image_memory(unsigned long nr_pages) | ||
1124 | { | ||
1125 | return preallocate_image_pages(nr_pages, GFP_IMAGE); | ||
1126 | } | ||
1127 | |||
1128 | #ifdef CONFIG_HIGHMEM | ||
1129 | static unsigned long preallocate_image_highmem(unsigned long nr_pages) | ||
1130 | { | ||
1131 | return preallocate_image_pages(nr_pages, GFP_IMAGE | __GFP_HIGHMEM); | ||
1132 | } | ||
1133 | |||
1134 | /** | ||
1135 | * __fraction - Compute (an approximation of) x * (multiplier / base) | ||
1076 | */ | 1136 | */ |
1137 | static unsigned long __fraction(u64 x, u64 multiplier, u64 base) | ||
1138 | { | ||
1139 | x *= multiplier; | ||
1140 | do_div(x, base); | ||
1141 | return (unsigned long)x; | ||
1142 | } | ||
1143 | |||
1144 | static unsigned long preallocate_highmem_fraction(unsigned long nr_pages, | ||
1145 | unsigned long highmem, | ||
1146 | unsigned long total) | ||
1147 | { | ||
1148 | unsigned long alloc = __fraction(nr_pages, highmem, total); | ||
1077 | 1149 | ||
1078 | #define SHRINK_BITE 10000 | 1150 | return preallocate_image_pages(alloc, GFP_IMAGE | __GFP_HIGHMEM); |
1079 | static inline unsigned long __shrink_memory(long tmp) | 1151 | } |
1152 | #else /* CONFIG_HIGHMEM */ | ||
1153 | static inline unsigned long preallocate_image_highmem(unsigned long nr_pages) | ||
1080 | { | 1154 | { |
1081 | if (tmp > SHRINK_BITE) | 1155 | return 0; |
1082 | tmp = SHRINK_BITE; | ||
1083 | return shrink_all_memory(tmp); | ||
1084 | } | 1156 | } |
1085 | 1157 | ||
1086 | int swsusp_shrink_memory(void) | 1158 | static inline unsigned long preallocate_highmem_fraction(unsigned long nr_pages, |
1159 | unsigned long highmem, | ||
1160 | unsigned long total) | ||
1161 | { | ||
1162 | return 0; | ||
1163 | } | ||
1164 | #endif /* CONFIG_HIGHMEM */ | ||
1165 | |||
1166 | /** | ||
1167 | * free_unnecessary_pages - Release preallocated pages not needed for the image | ||
1168 | */ | ||
1169 | static void free_unnecessary_pages(void) | ||
1170 | { | ||
1171 | unsigned long save_highmem, to_free_normal, to_free_highmem; | ||
1172 | |||
1173 | to_free_normal = alloc_normal - count_data_pages(); | ||
1174 | save_highmem = count_highmem_pages(); | ||
1175 | if (alloc_highmem > save_highmem) { | ||
1176 | to_free_highmem = alloc_highmem - save_highmem; | ||
1177 | } else { | ||
1178 | to_free_highmem = 0; | ||
1179 | to_free_normal -= save_highmem - alloc_highmem; | ||
1180 | } | ||
1181 | |||
1182 | memory_bm_position_reset(©_bm); | ||
1183 | |||
1184 | while (to_free_normal > 0 && to_free_highmem > 0) { | ||
1185 | unsigned long pfn = memory_bm_next_pfn(©_bm); | ||
1186 | struct page *page = pfn_to_page(pfn); | ||
1187 | |||
1188 | if (PageHighMem(page)) { | ||
1189 | if (!to_free_highmem) | ||
1190 | continue; | ||
1191 | to_free_highmem--; | ||
1192 | alloc_highmem--; | ||
1193 | } else { | ||
1194 | if (!to_free_normal) | ||
1195 | continue; | ||
1196 | to_free_normal--; | ||
1197 | alloc_normal--; | ||
1198 | } | ||
1199 | memory_bm_clear_bit(©_bm, pfn); | ||
1200 | swsusp_unset_page_forbidden(page); | ||
1201 | swsusp_unset_page_free(page); | ||
1202 | __free_page(page); | ||
1203 | } | ||
1204 | } | ||
1205 | |||
1206 | /** | ||
1207 | * minimum_image_size - Estimate the minimum acceptable size of an image | ||
1208 | * @saveable: Number of saveable pages in the system. | ||
1209 | * | ||
1210 | * We want to avoid attempting to free too much memory too hard, so estimate the | ||
1211 | * minimum acceptable size of a hibernation image to use as the lower limit for | ||
1212 | * preallocating memory. | ||
1213 | * | ||
1214 | * We assume that the minimum image size should be proportional to | ||
1215 | * | ||
1216 | * [number of saveable pages] - [number of pages that can be freed in theory] | ||
1217 | * | ||
1218 | * where the second term is the sum of (1) reclaimable slab pages, (2) active | ||
1219 | * and (3) inactive anonymouns pages, (4) active and (5) inactive file pages, | ||
1220 | * minus mapped file pages. | ||
1221 | */ | ||
1222 | static unsigned long minimum_image_size(unsigned long saveable) | ||
1223 | { | ||
1224 | unsigned long size; | ||
1225 | |||
1226 | size = global_page_state(NR_SLAB_RECLAIMABLE) | ||
1227 | + global_page_state(NR_ACTIVE_ANON) | ||
1228 | + global_page_state(NR_INACTIVE_ANON) | ||
1229 | + global_page_state(NR_ACTIVE_FILE) | ||
1230 | + global_page_state(NR_INACTIVE_FILE) | ||
1231 | - global_page_state(NR_FILE_MAPPED); | ||
1232 | |||
1233 | return saveable <= size ? 0 : saveable - size; | ||
1234 | } | ||
1235 | |||
1236 | /** | ||
1237 | * hibernate_preallocate_memory - Preallocate memory for hibernation image | ||
1238 | * | ||
1239 | * To create a hibernation image it is necessary to make a copy of every page | ||
1240 | * frame in use. We also need a number of page frames to be free during | ||
1241 | * hibernation for allocations made while saving the image and for device | ||
1242 | * drivers, in case they need to allocate memory from their hibernation | ||
1243 | * callbacks (these two numbers are given by PAGES_FOR_IO and SPARE_PAGES, | ||
1244 | * respectively, both of which are rough estimates). To make this happen, we | ||
1245 | * compute the total number of available page frames and allocate at least | ||
1246 | * | ||
1247 | * ([page frames total] + PAGES_FOR_IO + [metadata pages]) / 2 + 2 * SPARE_PAGES | ||
1248 | * | ||
1249 | * of them, which corresponds to the maximum size of a hibernation image. | ||
1250 | * | ||
1251 | * If image_size is set below the number following from the above formula, | ||
1252 | * the preallocation of memory is continued until the total number of saveable | ||
1253 | * pages in the system is below the requested image size or the minimum | ||
1254 | * acceptable image size returned by minimum_image_size(), whichever is greater. | ||
1255 | */ | ||
1256 | int hibernate_preallocate_memory(void) | ||
1087 | { | 1257 | { |
1088 | long tmp; | ||
1089 | struct zone *zone; | 1258 | struct zone *zone; |
1090 | unsigned long pages = 0; | 1259 | unsigned long saveable, size, max_size, count, highmem, pages = 0; |
1091 | unsigned int i = 0; | 1260 | unsigned long alloc, save_highmem, pages_highmem; |
1092 | char *p = "-\\|/"; | ||
1093 | struct timeval start, stop; | 1261 | struct timeval start, stop; |
1262 | int error; | ||
1094 | 1263 | ||
1095 | printk(KERN_INFO "PM: Shrinking memory... "); | 1264 | printk(KERN_INFO "PM: Preallocating image memory... "); |
1096 | do_gettimeofday(&start); | 1265 | do_gettimeofday(&start); |
1097 | do { | ||
1098 | long size, highmem_size; | ||
1099 | |||
1100 | highmem_size = count_highmem_pages(); | ||
1101 | size = count_data_pages() + PAGES_FOR_IO + SPARE_PAGES; | ||
1102 | tmp = size; | ||
1103 | size += highmem_size; | ||
1104 | for_each_populated_zone(zone) { | ||
1105 | tmp += snapshot_additional_pages(zone); | ||
1106 | if (is_highmem(zone)) { | ||
1107 | highmem_size -= | ||
1108 | zone_page_state(zone, NR_FREE_PAGES); | ||
1109 | } else { | ||
1110 | tmp -= zone_page_state(zone, NR_FREE_PAGES); | ||
1111 | tmp += zone->lowmem_reserve[ZONE_NORMAL]; | ||
1112 | } | ||
1113 | } | ||
1114 | 1266 | ||
1115 | if (highmem_size < 0) | 1267 | error = memory_bm_create(&orig_bm, GFP_IMAGE, PG_ANY); |
1116 | highmem_size = 0; | 1268 | if (error) |
1269 | goto err_out; | ||
1117 | 1270 | ||
1118 | tmp += highmem_size; | 1271 | error = memory_bm_create(©_bm, GFP_IMAGE, PG_ANY); |
1119 | if (tmp > 0) { | 1272 | if (error) |
1120 | tmp = __shrink_memory(tmp); | 1273 | goto err_out; |
1121 | if (!tmp) | 1274 | |
1122 | return -ENOMEM; | 1275 | alloc_normal = 0; |
1123 | pages += tmp; | 1276 | alloc_highmem = 0; |
1124 | } else if (size > image_size / PAGE_SIZE) { | 1277 | |
1125 | tmp = __shrink_memory(size - (image_size / PAGE_SIZE)); | 1278 | /* Count the number of saveable data pages. */ |
1126 | pages += tmp; | 1279 | save_highmem = count_highmem_pages(); |
1127 | } | 1280 | saveable = count_data_pages(); |
1128 | printk("\b%c", p[i++%4]); | 1281 | |
1129 | } while (tmp > 0); | 1282 | /* |
1283 | * Compute the total number of page frames we can use (count) and the | ||
1284 | * number of pages needed for image metadata (size). | ||
1285 | */ | ||
1286 | count = saveable; | ||
1287 | saveable += save_highmem; | ||
1288 | highmem = save_highmem; | ||
1289 | size = 0; | ||
1290 | for_each_populated_zone(zone) { | ||
1291 | size += snapshot_additional_pages(zone); | ||
1292 | if (is_highmem(zone)) | ||
1293 | highmem += zone_page_state(zone, NR_FREE_PAGES); | ||
1294 | else | ||
1295 | count += zone_page_state(zone, NR_FREE_PAGES); | ||
1296 | } | ||
1297 | count += highmem; | ||
1298 | count -= totalreserve_pages; | ||
1299 | |||
1300 | /* Compute the maximum number of saveable pages to leave in memory. */ | ||
1301 | max_size = (count - (size + PAGES_FOR_IO)) / 2 - 2 * SPARE_PAGES; | ||
1302 | size = DIV_ROUND_UP(image_size, PAGE_SIZE); | ||
1303 | if (size > max_size) | ||
1304 | size = max_size; | ||
1305 | /* | ||
1306 | * If the maximum is not less than the current number of saveable pages | ||
1307 | * in memory, allocate page frames for the image and we're done. | ||
1308 | */ | ||
1309 | if (size >= saveable) { | ||
1310 | pages = preallocate_image_highmem(save_highmem); | ||
1311 | pages += preallocate_image_memory(saveable - pages); | ||
1312 | goto out; | ||
1313 | } | ||
1314 | |||
1315 | /* Estimate the minimum size of the image. */ | ||
1316 | pages = minimum_image_size(saveable); | ||
1317 | if (size < pages) | ||
1318 | size = min_t(unsigned long, pages, max_size); | ||
1319 | |||
1320 | /* | ||
1321 | * Let the memory management subsystem know that we're going to need a | ||
1322 | * large number of page frames to allocate and make it free some memory. | ||
1323 | * NOTE: If this is not done, performance will be hurt badly in some | ||
1324 | * test cases. | ||
1325 | */ | ||
1326 | shrink_all_memory(saveable - size); | ||
1327 | |||
1328 | /* | ||
1329 | * The number of saveable pages in memory was too high, so apply some | ||
1330 | * pressure to decrease it. First, make room for the largest possible | ||
1331 | * image and fail if that doesn't work. Next, try to decrease the size | ||
1332 | * of the image as much as indicated by 'size' using allocations from | ||
1333 | * highmem and non-highmem zones separately. | ||
1334 | */ | ||
1335 | pages_highmem = preallocate_image_highmem(highmem / 2); | ||
1336 | alloc = (count - max_size) - pages_highmem; | ||
1337 | pages = preallocate_image_memory(alloc); | ||
1338 | if (pages < alloc) | ||
1339 | goto err_out; | ||
1340 | size = max_size - size; | ||
1341 | alloc = size; | ||
1342 | size = preallocate_highmem_fraction(size, highmem, count); | ||
1343 | pages_highmem += size; | ||
1344 | alloc -= size; | ||
1345 | pages += preallocate_image_memory(alloc); | ||
1346 | pages += pages_highmem; | ||
1347 | |||
1348 | /* | ||
1349 | * We only need as many page frames for the image as there are saveable | ||
1350 | * pages in memory, but we have allocated more. Release the excessive | ||
1351 | * ones now. | ||
1352 | */ | ||
1353 | free_unnecessary_pages(); | ||
1354 | |||
1355 | out: | ||
1130 | do_gettimeofday(&stop); | 1356 | do_gettimeofday(&stop); |
1131 | printk("\bdone (%lu pages freed)\n", pages); | 1357 | printk(KERN_CONT "done (allocated %lu pages)\n", pages); |
1132 | swsusp_show_speed(&start, &stop, pages, "Freed"); | 1358 | swsusp_show_speed(&start, &stop, pages, "Allocated"); |
1133 | 1359 | ||
1134 | return 0; | 1360 | return 0; |
1361 | |||
1362 | err_out: | ||
1363 | printk(KERN_CONT "\n"); | ||
1364 | swsusp_free(); | ||
1365 | return -ENOMEM; | ||
1135 | } | 1366 | } |
1136 | 1367 | ||
1137 | #ifdef CONFIG_HIGHMEM | 1368 | #ifdef CONFIG_HIGHMEM |
@@ -1142,7 +1373,7 @@ int swsusp_shrink_memory(void) | |||
1142 | 1373 | ||
1143 | static unsigned int count_pages_for_highmem(unsigned int nr_highmem) | 1374 | static unsigned int count_pages_for_highmem(unsigned int nr_highmem) |
1144 | { | 1375 | { |
1145 | unsigned int free_highmem = count_free_highmem_pages(); | 1376 | unsigned int free_highmem = count_free_highmem_pages() + alloc_highmem; |
1146 | 1377 | ||
1147 | if (free_highmem >= nr_highmem) | 1378 | if (free_highmem >= nr_highmem) |
1148 | nr_highmem = 0; | 1379 | nr_highmem = 0; |
@@ -1164,19 +1395,17 @@ count_pages_for_highmem(unsigned int nr_highmem) { return 0; } | |||
1164 | static int enough_free_mem(unsigned int nr_pages, unsigned int nr_highmem) | 1395 | static int enough_free_mem(unsigned int nr_pages, unsigned int nr_highmem) |
1165 | { | 1396 | { |
1166 | struct zone *zone; | 1397 | struct zone *zone; |
1167 | unsigned int free = 0, meta = 0; | 1398 | unsigned int free = alloc_normal; |
1168 | 1399 | ||
1169 | for_each_zone(zone) { | 1400 | for_each_populated_zone(zone) |
1170 | meta += snapshot_additional_pages(zone); | ||
1171 | if (!is_highmem(zone)) | 1401 | if (!is_highmem(zone)) |
1172 | free += zone_page_state(zone, NR_FREE_PAGES); | 1402 | free += zone_page_state(zone, NR_FREE_PAGES); |
1173 | } | ||
1174 | 1403 | ||
1175 | nr_pages += count_pages_for_highmem(nr_highmem); | 1404 | nr_pages += count_pages_for_highmem(nr_highmem); |
1176 | pr_debug("PM: Normal pages needed: %u + %u + %u, available pages: %u\n", | 1405 | pr_debug("PM: Normal pages needed: %u + %u, available pages: %u\n", |
1177 | nr_pages, PAGES_FOR_IO, meta, free); | 1406 | nr_pages, PAGES_FOR_IO, free); |
1178 | 1407 | ||
1179 | return free > nr_pages + PAGES_FOR_IO + meta; | 1408 | return free > nr_pages + PAGES_FOR_IO; |
1180 | } | 1409 | } |
1181 | 1410 | ||
1182 | #ifdef CONFIG_HIGHMEM | 1411 | #ifdef CONFIG_HIGHMEM |
@@ -1198,7 +1427,7 @@ static inline int get_highmem_buffer(int safe_needed) | |||
1198 | */ | 1427 | */ |
1199 | 1428 | ||
1200 | static inline unsigned int | 1429 | static inline unsigned int |
1201 | alloc_highmem_image_pages(struct memory_bitmap *bm, unsigned int nr_highmem) | 1430 | alloc_highmem_pages(struct memory_bitmap *bm, unsigned int nr_highmem) |
1202 | { | 1431 | { |
1203 | unsigned int to_alloc = count_free_highmem_pages(); | 1432 | unsigned int to_alloc = count_free_highmem_pages(); |
1204 | 1433 | ||
@@ -1218,7 +1447,7 @@ alloc_highmem_image_pages(struct memory_bitmap *bm, unsigned int nr_highmem) | |||
1218 | static inline int get_highmem_buffer(int safe_needed) { return 0; } | 1447 | static inline int get_highmem_buffer(int safe_needed) { return 0; } |
1219 | 1448 | ||
1220 | static inline unsigned int | 1449 | static inline unsigned int |
1221 | alloc_highmem_image_pages(struct memory_bitmap *bm, unsigned int n) { return 0; } | 1450 | alloc_highmem_pages(struct memory_bitmap *bm, unsigned int n) { return 0; } |
1222 | #endif /* CONFIG_HIGHMEM */ | 1451 | #endif /* CONFIG_HIGHMEM */ |
1223 | 1452 | ||
1224 | /** | 1453 | /** |
@@ -1237,51 +1466,36 @@ static int | |||
1237 | swsusp_alloc(struct memory_bitmap *orig_bm, struct memory_bitmap *copy_bm, | 1466 | swsusp_alloc(struct memory_bitmap *orig_bm, struct memory_bitmap *copy_bm, |
1238 | unsigned int nr_pages, unsigned int nr_highmem) | 1467 | unsigned int nr_pages, unsigned int nr_highmem) |
1239 | { | 1468 | { |
1240 | int error; | 1469 | int error = 0; |
1241 | |||
1242 | error = memory_bm_create(orig_bm, GFP_ATOMIC | __GFP_COLD, PG_ANY); | ||
1243 | if (error) | ||
1244 | goto Free; | ||
1245 | |||
1246 | error = memory_bm_create(copy_bm, GFP_ATOMIC | __GFP_COLD, PG_ANY); | ||
1247 | if (error) | ||
1248 | goto Free; | ||
1249 | 1470 | ||
1250 | if (nr_highmem > 0) { | 1471 | if (nr_highmem > 0) { |
1251 | error = get_highmem_buffer(PG_ANY); | 1472 | error = get_highmem_buffer(PG_ANY); |
1252 | if (error) | 1473 | if (error) |
1253 | goto Free; | 1474 | goto err_out; |
1254 | 1475 | if (nr_highmem > alloc_highmem) { | |
1255 | nr_pages += alloc_highmem_image_pages(copy_bm, nr_highmem); | 1476 | nr_highmem -= alloc_highmem; |
1477 | nr_pages += alloc_highmem_pages(copy_bm, nr_highmem); | ||
1478 | } | ||
1256 | } | 1479 | } |
1257 | while (nr_pages-- > 0) { | 1480 | if (nr_pages > alloc_normal) { |
1258 | struct page *page = alloc_image_page(GFP_ATOMIC | __GFP_COLD); | 1481 | nr_pages -= alloc_normal; |
1259 | 1482 | while (nr_pages-- > 0) { | |
1260 | if (!page) | 1483 | struct page *page; |
1261 | goto Free; | ||
1262 | 1484 | ||
1263 | memory_bm_set_bit(copy_bm, page_to_pfn(page)); | 1485 | page = alloc_image_page(GFP_ATOMIC | __GFP_COLD); |
1486 | if (!page) | ||
1487 | goto err_out; | ||
1488 | memory_bm_set_bit(copy_bm, page_to_pfn(page)); | ||
1489 | } | ||
1264 | } | 1490 | } |
1491 | |||
1265 | return 0; | 1492 | return 0; |
1266 | 1493 | ||
1267 | Free: | 1494 | err_out: |
1268 | swsusp_free(); | 1495 | swsusp_free(); |
1269 | return -ENOMEM; | 1496 | return error; |
1270 | } | 1497 | } |
1271 | 1498 | ||
1272 | /* Memory bitmap used for marking saveable pages (during suspend) or the | ||
1273 | * suspend image pages (during resume) | ||
1274 | */ | ||
1275 | static struct memory_bitmap orig_bm; | ||
1276 | /* Memory bitmap used on suspend for marking allocated pages that will contain | ||
1277 | * the copies of saveable pages. During resume it is initially used for | ||
1278 | * marking the suspend image pages, but then its set bits are duplicated in | ||
1279 | * @orig_bm and it is released. Next, on systems with high memory, it may be | ||
1280 | * used for marking "safe" highmem pages, but it has to be reinitialized for | ||
1281 | * this purpose. | ||
1282 | */ | ||
1283 | static struct memory_bitmap copy_bm; | ||
1284 | |||
1285 | asmlinkage int swsusp_save(void) | 1499 | asmlinkage int swsusp_save(void) |
1286 | { | 1500 | { |
1287 | unsigned int nr_pages, nr_highmem; | 1501 | unsigned int nr_pages, nr_highmem; |
@@ -1474,7 +1688,7 @@ static int mark_unsafe_pages(struct memory_bitmap *bm) | |||
1474 | unsigned long pfn, max_zone_pfn; | 1688 | unsigned long pfn, max_zone_pfn; |
1475 | 1689 | ||
1476 | /* Clear page flags */ | 1690 | /* Clear page flags */ |
1477 | for_each_zone(zone) { | 1691 | for_each_populated_zone(zone) { |
1478 | max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; | 1692 | max_zone_pfn = zone->zone_start_pfn + zone->spanned_pages; |
1479 | for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) | 1693 | for (pfn = zone->zone_start_pfn; pfn < max_zone_pfn; pfn++) |
1480 | if (pfn_valid(pfn)) | 1694 | if (pfn_valid(pfn)) |