diff options
Diffstat (limited to 'mm/oom_kill.c')
-rw-r--r-- | mm/oom_kill.c | 103 |
1 files changed, 71 insertions, 32 deletions
diff --git a/mm/oom_kill.c b/mm/oom_kill.c index ea2147dabba6..f52481b1c1e5 100644 --- a/mm/oom_kill.c +++ b/mm/oom_kill.c | |||
@@ -196,27 +196,46 @@ unsigned long badness(struct task_struct *p, unsigned long uptime) | |||
196 | /* | 196 | /* |
197 | * Determine the type of allocation constraint. | 197 | * Determine the type of allocation constraint. |
198 | */ | 198 | */ |
199 | static inline enum oom_constraint constrained_alloc(struct zonelist *zonelist, | ||
200 | gfp_t gfp_mask) | ||
201 | { | ||
202 | #ifdef CONFIG_NUMA | 199 | #ifdef CONFIG_NUMA |
200 | static enum oom_constraint constrained_alloc(struct zonelist *zonelist, | ||
201 | gfp_t gfp_mask, nodemask_t *nodemask) | ||
202 | { | ||
203 | struct zone *zone; | 203 | struct zone *zone; |
204 | struct zoneref *z; | 204 | struct zoneref *z; |
205 | enum zone_type high_zoneidx = gfp_zone(gfp_mask); | 205 | enum zone_type high_zoneidx = gfp_zone(gfp_mask); |
206 | nodemask_t nodes = node_states[N_HIGH_MEMORY]; | ||
207 | 206 | ||
208 | for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) | 207 | /* |
209 | if (cpuset_zone_allowed_softwall(zone, gfp_mask)) | 208 | * Reach here only when __GFP_NOFAIL is used. So, we should avoid |
210 | node_clear(zone_to_nid(zone), nodes); | 209 | * to kill current.We have to random task kill in this case. |
211 | else | 210 | * Hopefully, CONSTRAINT_THISNODE...but no way to handle it, now. |
212 | return CONSTRAINT_CPUSET; | 211 | */ |
212 | if (gfp_mask & __GFP_THISNODE) | ||
213 | return CONSTRAINT_NONE; | ||
213 | 214 | ||
214 | if (!nodes_empty(nodes)) | 215 | /* |
216 | * The nodemask here is a nodemask passed to alloc_pages(). Now, | ||
217 | * cpuset doesn't use this nodemask for its hardwall/softwall/hierarchy | ||
218 | * feature. mempolicy is an only user of nodemask here. | ||
219 | * check mempolicy's nodemask contains all N_HIGH_MEMORY | ||
220 | */ | ||
221 | if (nodemask && !nodes_subset(node_states[N_HIGH_MEMORY], *nodemask)) | ||
215 | return CONSTRAINT_MEMORY_POLICY; | 222 | return CONSTRAINT_MEMORY_POLICY; |
216 | #endif | ||
217 | 223 | ||
224 | /* Check this allocation failure is caused by cpuset's wall function */ | ||
225 | for_each_zone_zonelist_nodemask(zone, z, zonelist, | ||
226 | high_zoneidx, nodemask) | ||
227 | if (!cpuset_zone_allowed_softwall(zone, gfp_mask)) | ||
228 | return CONSTRAINT_CPUSET; | ||
229 | |||
230 | return CONSTRAINT_NONE; | ||
231 | } | ||
232 | #else | ||
233 | static enum oom_constraint constrained_alloc(struct zonelist *zonelist, | ||
234 | gfp_t gfp_mask, nodemask_t *nodemask) | ||
235 | { | ||
218 | return CONSTRAINT_NONE; | 236 | return CONSTRAINT_NONE; |
219 | } | 237 | } |
238 | #endif | ||
220 | 239 | ||
221 | /* | 240 | /* |
222 | * Simple selection loop. We chose the process with the highest | 241 | * Simple selection loop. We chose the process with the highest |
@@ -337,6 +356,24 @@ static void dump_tasks(const struct mem_cgroup *mem) | |||
337 | } while_each_thread(g, p); | 356 | } while_each_thread(g, p); |
338 | } | 357 | } |
339 | 358 | ||
359 | static void dump_header(struct task_struct *p, gfp_t gfp_mask, int order, | ||
360 | struct mem_cgroup *mem) | ||
361 | { | ||
362 | pr_warning("%s invoked oom-killer: gfp_mask=0x%x, order=%d, " | ||
363 | "oom_adj=%d\n", | ||
364 | current->comm, gfp_mask, order, current->signal->oom_adj); | ||
365 | task_lock(current); | ||
366 | cpuset_print_task_mems_allowed(current); | ||
367 | task_unlock(current); | ||
368 | dump_stack(); | ||
369 | mem_cgroup_print_oom_info(mem, p); | ||
370 | show_mem(); | ||
371 | if (sysctl_oom_dump_tasks) | ||
372 | dump_tasks(mem); | ||
373 | } | ||
374 | |||
375 | #define K(x) ((x) << (PAGE_SHIFT-10)) | ||
376 | |||
340 | /* | 377 | /* |
341 | * Send SIGKILL to the selected process irrespective of CAP_SYS_RAW_IO | 378 | * Send SIGKILL to the selected process irrespective of CAP_SYS_RAW_IO |
342 | * flag though it's unlikely that we select a process with CAP_SYS_RAW_IO | 379 | * flag though it's unlikely that we select a process with CAP_SYS_RAW_IO |
@@ -350,15 +387,23 @@ static void __oom_kill_task(struct task_struct *p, int verbose) | |||
350 | return; | 387 | return; |
351 | } | 388 | } |
352 | 389 | ||
390 | task_lock(p); | ||
353 | if (!p->mm) { | 391 | if (!p->mm) { |
354 | WARN_ON(1); | 392 | WARN_ON(1); |
355 | printk(KERN_WARNING "tried to kill an mm-less task!\n"); | 393 | printk(KERN_WARNING "tried to kill an mm-less task %d (%s)!\n", |
394 | task_pid_nr(p), p->comm); | ||
395 | task_unlock(p); | ||
356 | return; | 396 | return; |
357 | } | 397 | } |
358 | 398 | ||
359 | if (verbose) | 399 | if (verbose) |
360 | printk(KERN_ERR "Killed process %d (%s)\n", | 400 | printk(KERN_ERR "Killed process %d (%s) " |
361 | task_pid_nr(p), p->comm); | 401 | "vsz:%lukB, anon-rss:%lukB, file-rss:%lukB\n", |
402 | task_pid_nr(p), p->comm, | ||
403 | K(p->mm->total_vm), | ||
404 | K(get_mm_counter(p->mm, anon_rss)), | ||
405 | K(get_mm_counter(p->mm, file_rss))); | ||
406 | task_unlock(p); | ||
362 | 407 | ||
363 | /* | 408 | /* |
364 | * We give our sacrificial lamb high priority and access to | 409 | * We give our sacrificial lamb high priority and access to |
@@ -395,20 +440,8 @@ static int oom_kill_process(struct task_struct *p, gfp_t gfp_mask, int order, | |||
395 | { | 440 | { |
396 | struct task_struct *c; | 441 | struct task_struct *c; |
397 | 442 | ||
398 | if (printk_ratelimit()) { | 443 | if (printk_ratelimit()) |
399 | printk(KERN_WARNING "%s invoked oom-killer: " | 444 | dump_header(p, gfp_mask, order, mem); |
400 | "gfp_mask=0x%x, order=%d, oom_adj=%d\n", | ||
401 | current->comm, gfp_mask, order, | ||
402 | current->signal->oom_adj); | ||
403 | task_lock(current); | ||
404 | cpuset_print_task_mems_allowed(current); | ||
405 | task_unlock(current); | ||
406 | dump_stack(); | ||
407 | mem_cgroup_print_oom_info(mem, current); | ||
408 | show_mem(); | ||
409 | if (sysctl_oom_dump_tasks) | ||
410 | dump_tasks(mem); | ||
411 | } | ||
412 | 445 | ||
413 | /* | 446 | /* |
414 | * If the task is already exiting, don't alarm the sysadmin or kill | 447 | * If the task is already exiting, don't alarm the sysadmin or kill |
@@ -544,6 +577,7 @@ retry: | |||
544 | /* Found nothing?!?! Either we hang forever, or we panic. */ | 577 | /* Found nothing?!?! Either we hang forever, or we panic. */ |
545 | if (!p) { | 578 | if (!p) { |
546 | read_unlock(&tasklist_lock); | 579 | read_unlock(&tasklist_lock); |
580 | dump_header(NULL, gfp_mask, order, NULL); | ||
547 | panic("Out of memory and no killable processes...\n"); | 581 | panic("Out of memory and no killable processes...\n"); |
548 | } | 582 | } |
549 | 583 | ||
@@ -599,7 +633,8 @@ rest_and_return: | |||
599 | * OR try to be smart about which process to kill. Note that we | 633 | * OR try to be smart about which process to kill. Note that we |
600 | * don't have to be perfect here, we just have to be good. | 634 | * don't have to be perfect here, we just have to be good. |
601 | */ | 635 | */ |
602 | void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order) | 636 | void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, |
637 | int order, nodemask_t *nodemask) | ||
603 | { | 638 | { |
604 | unsigned long freed = 0; | 639 | unsigned long freed = 0; |
605 | enum oom_constraint constraint; | 640 | enum oom_constraint constraint; |
@@ -609,14 +644,16 @@ void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order) | |||
609 | /* Got some memory back in the last second. */ | 644 | /* Got some memory back in the last second. */ |
610 | return; | 645 | return; |
611 | 646 | ||
612 | if (sysctl_panic_on_oom == 2) | 647 | if (sysctl_panic_on_oom == 2) { |
648 | dump_header(NULL, gfp_mask, order, NULL); | ||
613 | panic("out of memory. Compulsory panic_on_oom is selected.\n"); | 649 | panic("out of memory. Compulsory panic_on_oom is selected.\n"); |
650 | } | ||
614 | 651 | ||
615 | /* | 652 | /* |
616 | * Check if there were limitations on the allocation (only relevant for | 653 | * Check if there were limitations on the allocation (only relevant for |
617 | * NUMA) that may require different handling. | 654 | * NUMA) that may require different handling. |
618 | */ | 655 | */ |
619 | constraint = constrained_alloc(zonelist, gfp_mask); | 656 | constraint = constrained_alloc(zonelist, gfp_mask, nodemask); |
620 | read_lock(&tasklist_lock); | 657 | read_lock(&tasklist_lock); |
621 | 658 | ||
622 | switch (constraint) { | 659 | switch (constraint) { |
@@ -626,8 +663,10 @@ void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order) | |||
626 | break; | 663 | break; |
627 | 664 | ||
628 | case CONSTRAINT_NONE: | 665 | case CONSTRAINT_NONE: |
629 | if (sysctl_panic_on_oom) | 666 | if (sysctl_panic_on_oom) { |
667 | dump_header(NULL, gfp_mask, order, NULL); | ||
630 | panic("out of memory. panic_on_oom is selected\n"); | 668 | panic("out of memory. panic_on_oom is selected\n"); |
669 | } | ||
631 | /* Fall-through */ | 670 | /* Fall-through */ |
632 | case CONSTRAINT_CPUSET: | 671 | case CONSTRAINT_CPUSET: |
633 | __out_of_memory(gfp_mask, order); | 672 | __out_of_memory(gfp_mask, order); |