diff options
Diffstat (limited to 'fs/proc/task_mmu.c')
-rw-r--r-- | fs/proc/task_mmu.c | 131 |
1 files changed, 55 insertions, 76 deletions
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c index 47c03f4336b8..070553427dd5 100644 --- a/fs/proc/task_mmu.c +++ b/fs/proc/task_mmu.c | |||
@@ -4,6 +4,7 @@ | |||
4 | #include <linux/seq_file.h> | 4 | #include <linux/seq_file.h> |
5 | #include <linux/highmem.h> | 5 | #include <linux/highmem.h> |
6 | #include <linux/ptrace.h> | 6 | #include <linux/ptrace.h> |
7 | #include <linux/slab.h> | ||
7 | #include <linux/pagemap.h> | 8 | #include <linux/pagemap.h> |
8 | #include <linux/mempolicy.h> | 9 | #include <linux/mempolicy.h> |
9 | #include <linux/swap.h> | 10 | #include <linux/swap.h> |
@@ -16,7 +17,7 @@ | |||
16 | 17 | ||
17 | void task_mem(struct seq_file *m, struct mm_struct *mm) | 18 | void task_mem(struct seq_file *m, struct mm_struct *mm) |
18 | { | 19 | { |
19 | unsigned long data, text, lib; | 20 | unsigned long data, text, lib, swap; |
20 | unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss; | 21 | unsigned long hiwater_vm, total_vm, hiwater_rss, total_rss; |
21 | 22 | ||
22 | /* | 23 | /* |
@@ -36,6 +37,7 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | |||
36 | data = mm->total_vm - mm->shared_vm - mm->stack_vm; | 37 | data = mm->total_vm - mm->shared_vm - mm->stack_vm; |
37 | text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; | 38 | text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; |
38 | lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; | 39 | lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; |
40 | swap = get_mm_counter(mm, MM_SWAPENTS); | ||
39 | seq_printf(m, | 41 | seq_printf(m, |
40 | "VmPeak:\t%8lu kB\n" | 42 | "VmPeak:\t%8lu kB\n" |
41 | "VmSize:\t%8lu kB\n" | 43 | "VmSize:\t%8lu kB\n" |
@@ -46,7 +48,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | |||
46 | "VmStk:\t%8lu kB\n" | 48 | "VmStk:\t%8lu kB\n" |
47 | "VmExe:\t%8lu kB\n" | 49 | "VmExe:\t%8lu kB\n" |
48 | "VmLib:\t%8lu kB\n" | 50 | "VmLib:\t%8lu kB\n" |
49 | "VmPTE:\t%8lu kB\n", | 51 | "VmPTE:\t%8lu kB\n" |
52 | "VmSwap:\t%8lu kB\n", | ||
50 | hiwater_vm << (PAGE_SHIFT-10), | 53 | hiwater_vm << (PAGE_SHIFT-10), |
51 | (total_vm - mm->reserved_vm) << (PAGE_SHIFT-10), | 54 | (total_vm - mm->reserved_vm) << (PAGE_SHIFT-10), |
52 | mm->locked_vm << (PAGE_SHIFT-10), | 55 | mm->locked_vm << (PAGE_SHIFT-10), |
@@ -54,7 +57,8 @@ void task_mem(struct seq_file *m, struct mm_struct *mm) | |||
54 | total_rss << (PAGE_SHIFT-10), | 57 | total_rss << (PAGE_SHIFT-10), |
55 | data << (PAGE_SHIFT-10), | 58 | data << (PAGE_SHIFT-10), |
56 | mm->stack_vm << (PAGE_SHIFT-10), text, lib, | 59 | mm->stack_vm << (PAGE_SHIFT-10), text, lib, |
57 | (PTRS_PER_PTE*sizeof(pte_t)*mm->nr_ptes) >> 10); | 60 | (PTRS_PER_PTE*sizeof(pte_t)*mm->nr_ptes) >> 10, |
61 | swap << (PAGE_SHIFT-10)); | ||
58 | } | 62 | } |
59 | 63 | ||
60 | unsigned long task_vsize(struct mm_struct *mm) | 64 | unsigned long task_vsize(struct mm_struct *mm) |
@@ -65,11 +69,11 @@ unsigned long task_vsize(struct mm_struct *mm) | |||
65 | int task_statm(struct mm_struct *mm, int *shared, int *text, | 69 | int task_statm(struct mm_struct *mm, int *shared, int *text, |
66 | int *data, int *resident) | 70 | int *data, int *resident) |
67 | { | 71 | { |
68 | *shared = get_mm_counter(mm, file_rss); | 72 | *shared = get_mm_counter(mm, MM_FILEPAGES); |
69 | *text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) | 73 | *text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) |
70 | >> PAGE_SHIFT; | 74 | >> PAGE_SHIFT; |
71 | *data = mm->total_vm - mm->shared_vm; | 75 | *data = mm->total_vm - mm->shared_vm; |
72 | *resident = *shared + get_mm_counter(mm, anon_rss); | 76 | *resident = *shared + get_mm_counter(mm, MM_ANONPAGES); |
73 | return mm->total_vm; | 77 | return mm->total_vm; |
74 | } | 78 | } |
75 | 79 | ||
@@ -361,12 +365,11 @@ static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, | |||
361 | if (!pte_present(ptent)) | 365 | if (!pte_present(ptent)) |
362 | continue; | 366 | continue; |
363 | 367 | ||
364 | mss->resident += PAGE_SIZE; | ||
365 | |||
366 | page = vm_normal_page(vma, addr, ptent); | 368 | page = vm_normal_page(vma, addr, ptent); |
367 | if (!page) | 369 | if (!page) |
368 | continue; | 370 | continue; |
369 | 371 | ||
372 | mss->resident += PAGE_SIZE; | ||
370 | /* Accumulate the size in pages that have been accessed. */ | 373 | /* Accumulate the size in pages that have been accessed. */ |
371 | if (pte_young(ptent) || PageReferenced(page)) | 374 | if (pte_young(ptent) || PageReferenced(page)) |
372 | mss->referenced += PAGE_SIZE; | 375 | mss->referenced += PAGE_SIZE; |
@@ -404,6 +407,7 @@ static int show_smap(struct seq_file *m, void *v) | |||
404 | 407 | ||
405 | memset(&mss, 0, sizeof mss); | 408 | memset(&mss, 0, sizeof mss); |
406 | mss.vma = vma; | 409 | mss.vma = vma; |
410 | /* mmap_sem is held in m_start */ | ||
407 | if (vma->vm_mm && !is_vm_hugetlb_page(vma)) | 411 | if (vma->vm_mm && !is_vm_hugetlb_page(vma)) |
408 | walk_page_range(vma->vm_start, vma->vm_end, &smaps_walk); | 412 | walk_page_range(vma->vm_start, vma->vm_end, &smaps_walk); |
409 | 413 | ||
@@ -550,7 +554,8 @@ const struct file_operations proc_clear_refs_operations = { | |||
550 | }; | 554 | }; |
551 | 555 | ||
552 | struct pagemapread { | 556 | struct pagemapread { |
553 | u64 __user *out, *end; | 557 | int pos, len; |
558 | u64 *buffer; | ||
554 | }; | 559 | }; |
555 | 560 | ||
556 | #define PM_ENTRY_BYTES sizeof(u64) | 561 | #define PM_ENTRY_BYTES sizeof(u64) |
@@ -573,10 +578,8 @@ struct pagemapread { | |||
573 | static int add_to_pagemap(unsigned long addr, u64 pfn, | 578 | static int add_to_pagemap(unsigned long addr, u64 pfn, |
574 | struct pagemapread *pm) | 579 | struct pagemapread *pm) |
575 | { | 580 | { |
576 | if (put_user(pfn, pm->out)) | 581 | pm->buffer[pm->pos++] = pfn; |
577 | return -EFAULT; | 582 | if (pm->pos >= pm->len) |
578 | pm->out++; | ||
579 | if (pm->out >= pm->end) | ||
580 | return PM_END_OF_BUFFER; | 583 | return PM_END_OF_BUFFER; |
581 | return 0; | 584 | return 0; |
582 | } | 585 | } |
@@ -659,31 +662,18 @@ static u64 huge_pte_to_pagemap_entry(pte_t pte, int offset) | |||
659 | return pme; | 662 | return pme; |
660 | } | 663 | } |
661 | 664 | ||
662 | static int pagemap_hugetlb_range(pte_t *pte, unsigned long addr, | 665 | /* This function walks within one hugetlb entry in the single call */ |
663 | unsigned long end, struct mm_walk *walk) | 666 | static int pagemap_hugetlb_range(pte_t *pte, unsigned long hmask, |
667 | unsigned long addr, unsigned long end, | ||
668 | struct mm_walk *walk) | ||
664 | { | 669 | { |
665 | struct vm_area_struct *vma; | ||
666 | struct pagemapread *pm = walk->private; | 670 | struct pagemapread *pm = walk->private; |
667 | struct hstate *hs = NULL; | ||
668 | int err = 0; | 671 | int err = 0; |
672 | u64 pfn; | ||
669 | 673 | ||
670 | vma = find_vma(walk->mm, addr); | ||
671 | if (vma) | ||
672 | hs = hstate_vma(vma); | ||
673 | for (; addr != end; addr += PAGE_SIZE) { | 674 | for (; addr != end; addr += PAGE_SIZE) { |
674 | u64 pfn = PM_NOT_PRESENT; | 675 | int offset = (addr & ~hmask) >> PAGE_SHIFT; |
675 | 676 | pfn = huge_pte_to_pagemap_entry(*pte, offset); | |
676 | if (vma && (addr >= vma->vm_end)) { | ||
677 | vma = find_vma(walk->mm, addr); | ||
678 | if (vma) | ||
679 | hs = hstate_vma(vma); | ||
680 | } | ||
681 | |||
682 | if (vma && (vma->vm_start <= addr) && is_vm_hugetlb_page(vma)) { | ||
683 | /* calculate pfn of the "raw" page in the hugepage. */ | ||
684 | int offset = (addr & ~huge_page_mask(hs)) >> PAGE_SHIFT; | ||
685 | pfn = huge_pte_to_pagemap_entry(*pte, offset); | ||
686 | } | ||
687 | err = add_to_pagemap(addr, pfn, pm); | 677 | err = add_to_pagemap(addr, pfn, pm); |
688 | if (err) | 678 | if (err) |
689 | return err; | 679 | return err; |
@@ -718,21 +708,20 @@ static int pagemap_hugetlb_range(pte_t *pte, unsigned long addr, | |||
718 | * determine which areas of memory are actually mapped and llseek to | 708 | * determine which areas of memory are actually mapped and llseek to |
719 | * skip over unmapped regions. | 709 | * skip over unmapped regions. |
720 | */ | 710 | */ |
711 | #define PAGEMAP_WALK_SIZE (PMD_SIZE) | ||
721 | static ssize_t pagemap_read(struct file *file, char __user *buf, | 712 | static ssize_t pagemap_read(struct file *file, char __user *buf, |
722 | size_t count, loff_t *ppos) | 713 | size_t count, loff_t *ppos) |
723 | { | 714 | { |
724 | struct task_struct *task = get_proc_task(file->f_path.dentry->d_inode); | 715 | struct task_struct *task = get_proc_task(file->f_path.dentry->d_inode); |
725 | struct page **pages, *page; | ||
726 | unsigned long uaddr, uend; | ||
727 | struct mm_struct *mm; | 716 | struct mm_struct *mm; |
728 | struct pagemapread pm; | 717 | struct pagemapread pm; |
729 | int pagecount; | ||
730 | int ret = -ESRCH; | 718 | int ret = -ESRCH; |
731 | struct mm_walk pagemap_walk = {}; | 719 | struct mm_walk pagemap_walk = {}; |
732 | unsigned long src; | 720 | unsigned long src; |
733 | unsigned long svpfn; | 721 | unsigned long svpfn; |
734 | unsigned long start_vaddr; | 722 | unsigned long start_vaddr; |
735 | unsigned long end_vaddr; | 723 | unsigned long end_vaddr; |
724 | int copied = 0; | ||
736 | 725 | ||
737 | if (!task) | 726 | if (!task) |
738 | goto out; | 727 | goto out; |
@@ -755,35 +744,12 @@ static ssize_t pagemap_read(struct file *file, char __user *buf, | |||
755 | if (!mm) | 744 | if (!mm) |
756 | goto out_task; | 745 | goto out_task; |
757 | 746 | ||
758 | 747 | pm.len = PM_ENTRY_BYTES * (PAGEMAP_WALK_SIZE >> PAGE_SHIFT); | |
759 | uaddr = (unsigned long)buf & PAGE_MASK; | 748 | pm.buffer = kmalloc(pm.len, GFP_TEMPORARY); |
760 | uend = (unsigned long)(buf + count); | ||
761 | pagecount = (PAGE_ALIGN(uend) - uaddr) / PAGE_SIZE; | ||
762 | ret = 0; | ||
763 | if (pagecount == 0) | ||
764 | goto out_mm; | ||
765 | pages = kcalloc(pagecount, sizeof(struct page *), GFP_KERNEL); | ||
766 | ret = -ENOMEM; | 749 | ret = -ENOMEM; |
767 | if (!pages) | 750 | if (!pm.buffer) |
768 | goto out_mm; | 751 | goto out_mm; |
769 | 752 | ||
770 | down_read(¤t->mm->mmap_sem); | ||
771 | ret = get_user_pages(current, current->mm, uaddr, pagecount, | ||
772 | 1, 0, pages, NULL); | ||
773 | up_read(¤t->mm->mmap_sem); | ||
774 | |||
775 | if (ret < 0) | ||
776 | goto out_free; | ||
777 | |||
778 | if (ret != pagecount) { | ||
779 | pagecount = ret; | ||
780 | ret = -EFAULT; | ||
781 | goto out_pages; | ||
782 | } | ||
783 | |||
784 | pm.out = (u64 __user *)buf; | ||
785 | pm.end = (u64 __user *)(buf + count); | ||
786 | |||
787 | pagemap_walk.pmd_entry = pagemap_pte_range; | 753 | pagemap_walk.pmd_entry = pagemap_pte_range; |
788 | pagemap_walk.pte_hole = pagemap_pte_hole; | 754 | pagemap_walk.pte_hole = pagemap_pte_hole; |
789 | pagemap_walk.hugetlb_entry = pagemap_hugetlb_range; | 755 | pagemap_walk.hugetlb_entry = pagemap_hugetlb_range; |
@@ -805,23 +771,36 @@ static ssize_t pagemap_read(struct file *file, char __user *buf, | |||
805 | * user buffer is tracked in "pm", and the walk | 771 | * user buffer is tracked in "pm", and the walk |
806 | * will stop when we hit the end of the buffer. | 772 | * will stop when we hit the end of the buffer. |
807 | */ | 773 | */ |
808 | ret = walk_page_range(start_vaddr, end_vaddr, &pagemap_walk); | 774 | ret = 0; |
809 | if (ret == PM_END_OF_BUFFER) | 775 | while (count && (start_vaddr < end_vaddr)) { |
810 | ret = 0; | 776 | int len; |
811 | /* don't need mmap_sem for these, but this looks cleaner */ | 777 | unsigned long end; |
812 | *ppos += (char __user *)pm.out - buf; | 778 | |
813 | if (!ret) | 779 | pm.pos = 0; |
814 | ret = (char __user *)pm.out - buf; | 780 | end = start_vaddr + PAGEMAP_WALK_SIZE; |
815 | 781 | /* overflow ? */ | |
816 | out_pages: | 782 | if (end < start_vaddr || end > end_vaddr) |
817 | for (; pagecount; pagecount--) { | 783 | end = end_vaddr; |
818 | page = pages[pagecount-1]; | 784 | down_read(&mm->mmap_sem); |
819 | if (!PageReserved(page)) | 785 | ret = walk_page_range(start_vaddr, end, &pagemap_walk); |
820 | SetPageDirty(page); | 786 | up_read(&mm->mmap_sem); |
821 | page_cache_release(page); | 787 | start_vaddr = end; |
788 | |||
789 | len = min(count, PM_ENTRY_BYTES * pm.pos); | ||
790 | if (copy_to_user(buf, pm.buffer, len)) { | ||
791 | ret = -EFAULT; | ||
792 | goto out_free; | ||
793 | } | ||
794 | copied += len; | ||
795 | buf += len; | ||
796 | count -= len; | ||
822 | } | 797 | } |
798 | *ppos += copied; | ||
799 | if (!ret || ret == PM_END_OF_BUFFER) | ||
800 | ret = copied; | ||
801 | |||
823 | out_free: | 802 | out_free: |
824 | kfree(pages); | 803 | kfree(pm.buffer); |
825 | out_mm: | 804 | out_mm: |
826 | mmput(mm); | 805 | mmput(mm); |
827 | out_task: | 806 | out_task: |