diff options
Diffstat (limited to 'fs/proc/page.c')
| -rw-r--r-- | fs/proc/page.c | 162 |
1 files changed, 127 insertions, 35 deletions
diff --git a/fs/proc/page.c b/fs/proc/page.c index e9983837d08d..2707c6c7a20f 100644 --- a/fs/proc/page.c +++ b/fs/proc/page.c | |||
| @@ -6,11 +6,13 @@ | |||
| 6 | #include <linux/mmzone.h> | 6 | #include <linux/mmzone.h> |
| 7 | #include <linux/proc_fs.h> | 7 | #include <linux/proc_fs.h> |
| 8 | #include <linux/seq_file.h> | 8 | #include <linux/seq_file.h> |
| 9 | #include <linux/hugetlb.h> | ||
| 9 | #include <asm/uaccess.h> | 10 | #include <asm/uaccess.h> |
| 10 | #include "internal.h" | 11 | #include "internal.h" |
| 11 | 12 | ||
| 12 | #define KPMSIZE sizeof(u64) | 13 | #define KPMSIZE sizeof(u64) |
| 13 | #define KPMMASK (KPMSIZE - 1) | 14 | #define KPMMASK (KPMSIZE - 1) |
| 15 | |||
| 14 | /* /proc/kpagecount - an array exposing page counts | 16 | /* /proc/kpagecount - an array exposing page counts |
| 15 | * | 17 | * |
| 16 | * Each entry is a u64 representing the corresponding | 18 | * Each entry is a u64 representing the corresponding |
| @@ -32,20 +34,22 @@ static ssize_t kpagecount_read(struct file *file, char __user *buf, | |||
| 32 | return -EINVAL; | 34 | return -EINVAL; |
| 33 | 35 | ||
| 34 | while (count > 0) { | 36 | while (count > 0) { |
| 35 | ppage = NULL; | ||
| 36 | if (pfn_valid(pfn)) | 37 | if (pfn_valid(pfn)) |
| 37 | ppage = pfn_to_page(pfn); | 38 | ppage = pfn_to_page(pfn); |
| 38 | pfn++; | 39 | else |
| 40 | ppage = NULL; | ||
| 39 | if (!ppage) | 41 | if (!ppage) |
| 40 | pcount = 0; | 42 | pcount = 0; |
| 41 | else | 43 | else |
| 42 | pcount = page_mapcount(ppage); | 44 | pcount = page_mapcount(ppage); |
| 43 | 45 | ||
| 44 | if (put_user(pcount, out++)) { | 46 | if (put_user(pcount, out)) { |
| 45 | ret = -EFAULT; | 47 | ret = -EFAULT; |
| 46 | break; | 48 | break; |
| 47 | } | 49 | } |
| 48 | 50 | ||
| 51 | pfn++; | ||
| 52 | out++; | ||
| 49 | count -= KPMSIZE; | 53 | count -= KPMSIZE; |
| 50 | } | 54 | } |
| 51 | 55 | ||
| @@ -68,19 +72,122 @@ static const struct file_operations proc_kpagecount_operations = { | |||
| 68 | 72 | ||
| 69 | /* These macros are used to decouple internal flags from exported ones */ | 73 | /* These macros are used to decouple internal flags from exported ones */ |
| 70 | 74 | ||
| 71 | #define KPF_LOCKED 0 | 75 | #define KPF_LOCKED 0 |
| 72 | #define KPF_ERROR 1 | 76 | #define KPF_ERROR 1 |
| 73 | #define KPF_REFERENCED 2 | 77 | #define KPF_REFERENCED 2 |
| 74 | #define KPF_UPTODATE 3 | 78 | #define KPF_UPTODATE 3 |
| 75 | #define KPF_DIRTY 4 | 79 | #define KPF_DIRTY 4 |
| 76 | #define KPF_LRU 5 | 80 | #define KPF_LRU 5 |
| 77 | #define KPF_ACTIVE 6 | 81 | #define KPF_ACTIVE 6 |
| 78 | #define KPF_SLAB 7 | 82 | #define KPF_SLAB 7 |
| 79 | #define KPF_WRITEBACK 8 | 83 | #define KPF_WRITEBACK 8 |
| 80 | #define KPF_RECLAIM 9 | 84 | #define KPF_RECLAIM 9 |
| 81 | #define KPF_BUDDY 10 | 85 | #define KPF_BUDDY 10 |
| 86 | |||
| 87 | /* 11-20: new additions in 2.6.31 */ | ||
| 88 | #define KPF_MMAP 11 | ||
| 89 | #define KPF_ANON 12 | ||
| 90 | #define KPF_SWAPCACHE 13 | ||
| 91 | #define KPF_SWAPBACKED 14 | ||
| 92 | #define KPF_COMPOUND_HEAD 15 | ||
| 93 | #define KPF_COMPOUND_TAIL 16 | ||
| 94 | #define KPF_HUGE 17 | ||
| 95 | #define KPF_UNEVICTABLE 18 | ||
| 96 | #define KPF_NOPAGE 20 | ||
| 97 | |||
| 98 | /* kernel hacking assistances | ||
| 99 | * WARNING: subject to change, never rely on them! | ||
| 100 | */ | ||
| 101 | #define KPF_RESERVED 32 | ||
| 102 | #define KPF_MLOCKED 33 | ||
| 103 | #define KPF_MAPPEDTODISK 34 | ||
| 104 | #define KPF_PRIVATE 35 | ||
| 105 | #define KPF_PRIVATE_2 36 | ||
| 106 | #define KPF_OWNER_PRIVATE 37 | ||
| 107 | #define KPF_ARCH 38 | ||
| 108 | #define KPF_UNCACHED 39 | ||
| 109 | |||
| 110 | static inline u64 kpf_copy_bit(u64 kflags, int ubit, int kbit) | ||
| 111 | { | ||
| 112 | return ((kflags >> kbit) & 1) << ubit; | ||
| 113 | } | ||
| 82 | 114 | ||
| 83 | #define kpf_copy_bit(flags, dstpos, srcpos) (((flags >> srcpos) & 1) << dstpos) | 115 | static u64 get_uflags(struct page *page) |
| 116 | { | ||
| 117 | u64 k; | ||
| 118 | u64 u; | ||
| 119 | |||
| 120 | /* | ||
| 121 | * pseudo flag: KPF_NOPAGE | ||
| 122 | * it differentiates a memory hole from a page with no flags | ||
| 123 | */ | ||
| 124 | if (!page) | ||
| 125 | return 1 << KPF_NOPAGE; | ||
| 126 | |||
| 127 | k = page->flags; | ||
| 128 | u = 0; | ||
| 129 | |||
| 130 | /* | ||
| 131 | * pseudo flags for the well known (anonymous) memory mapped pages | ||
| 132 | * | ||
| 133 | * Note that page->_mapcount is overloaded in SLOB/SLUB/SLQB, so the | ||
| 134 | * simple test in page_mapped() is not enough. | ||
| 135 | */ | ||
| 136 | if (!PageSlab(page) && page_mapped(page)) | ||
| 137 | u |= 1 << KPF_MMAP; | ||
| 138 | if (PageAnon(page)) | ||
| 139 | u |= 1 << KPF_ANON; | ||
| 140 | |||
| 141 | /* | ||
| 142 | * compound pages: export both head/tail info | ||
| 143 | * they together define a compound page's start/end pos and order | ||
| 144 | */ | ||
| 145 | if (PageHead(page)) | ||
| 146 | u |= 1 << KPF_COMPOUND_HEAD; | ||
| 147 | if (PageTail(page)) | ||
| 148 | u |= 1 << KPF_COMPOUND_TAIL; | ||
| 149 | if (PageHuge(page)) | ||
| 150 | u |= 1 << KPF_HUGE; | ||
| 151 | |||
| 152 | u |= kpf_copy_bit(k, KPF_LOCKED, PG_locked); | ||
| 153 | |||
| 154 | /* | ||
| 155 | * Caveats on high order pages: | ||
| 156 | * PG_buddy will only be set on the head page; SLUB/SLQB do the same | ||
| 157 | * for PG_slab; SLOB won't set PG_slab at all on compound pages. | ||
| 158 | */ | ||
| 159 | u |= kpf_copy_bit(k, KPF_SLAB, PG_slab); | ||
| 160 | u |= kpf_copy_bit(k, KPF_BUDDY, PG_buddy); | ||
| 161 | |||
| 162 | u |= kpf_copy_bit(k, KPF_ERROR, PG_error); | ||
| 163 | u |= kpf_copy_bit(k, KPF_DIRTY, PG_dirty); | ||
| 164 | u |= kpf_copy_bit(k, KPF_UPTODATE, PG_uptodate); | ||
| 165 | u |= kpf_copy_bit(k, KPF_WRITEBACK, PG_writeback); | ||
| 166 | |||
| 167 | u |= kpf_copy_bit(k, KPF_LRU, PG_lru); | ||
| 168 | u |= kpf_copy_bit(k, KPF_REFERENCED, PG_referenced); | ||
| 169 | u |= kpf_copy_bit(k, KPF_ACTIVE, PG_active); | ||
| 170 | u |= kpf_copy_bit(k, KPF_RECLAIM, PG_reclaim); | ||
| 171 | |||
| 172 | u |= kpf_copy_bit(k, KPF_SWAPCACHE, PG_swapcache); | ||
| 173 | u |= kpf_copy_bit(k, KPF_SWAPBACKED, PG_swapbacked); | ||
| 174 | |||
| 175 | u |= kpf_copy_bit(k, KPF_UNEVICTABLE, PG_unevictable); | ||
| 176 | u |= kpf_copy_bit(k, KPF_MLOCKED, PG_mlocked); | ||
| 177 | |||
| 178 | #ifdef CONFIG_IA64_UNCACHED_ALLOCATOR | ||
| 179 | u |= kpf_copy_bit(k, KPF_UNCACHED, PG_uncached); | ||
| 180 | #endif | ||
| 181 | |||
| 182 | u |= kpf_copy_bit(k, KPF_RESERVED, PG_reserved); | ||
| 183 | u |= kpf_copy_bit(k, KPF_MAPPEDTODISK, PG_mappedtodisk); | ||
| 184 | u |= kpf_copy_bit(k, KPF_PRIVATE, PG_private); | ||
| 185 | u |= kpf_copy_bit(k, KPF_PRIVATE_2, PG_private_2); | ||
| 186 | u |= kpf_copy_bit(k, KPF_OWNER_PRIVATE, PG_owner_priv_1); | ||
| 187 | u |= kpf_copy_bit(k, KPF_ARCH, PG_arch_1); | ||
| 188 | |||
| 189 | return u; | ||
| 190 | }; | ||
| 84 | 191 | ||
| 85 | static ssize_t kpageflags_read(struct file *file, char __user *buf, | 192 | static ssize_t kpageflags_read(struct file *file, char __user *buf, |
| 86 | size_t count, loff_t *ppos) | 193 | size_t count, loff_t *ppos) |
| @@ -90,7 +197,6 @@ static ssize_t kpageflags_read(struct file *file, char __user *buf, | |||
| 90 | unsigned long src = *ppos; | 197 | unsigned long src = *ppos; |
| 91 | unsigned long pfn; | 198 | unsigned long pfn; |
| 92 | ssize_t ret = 0; | 199 | ssize_t ret = 0; |
| 93 | u64 kflags, uflags; | ||
| 94 | 200 | ||
| 95 | pfn = src / KPMSIZE; | 201 | pfn = src / KPMSIZE; |
| 96 | count = min_t(unsigned long, count, (max_pfn * KPMSIZE) - src); | 202 | count = min_t(unsigned long, count, (max_pfn * KPMSIZE) - src); |
| @@ -98,32 +204,18 @@ static ssize_t kpageflags_read(struct file *file, char __user *buf, | |||
| 98 | return -EINVAL; | 204 | return -EINVAL; |
| 99 | 205 | ||
| 100 | while (count > 0) { | 206 | while (count > 0) { |
| 101 | ppage = NULL; | ||
| 102 | if (pfn_valid(pfn)) | 207 | if (pfn_valid(pfn)) |
| 103 | ppage = pfn_to_page(pfn); | 208 | ppage = pfn_to_page(pfn); |
| 104 | pfn++; | ||
| 105 | if (!ppage) | ||
| 106 | kflags = 0; | ||
| 107 | else | 209 | else |
| 108 | kflags = ppage->flags; | 210 | ppage = NULL; |
| 109 | 211 | ||
| 110 | uflags = kpf_copy_bit(kflags, KPF_LOCKED, PG_locked) | | 212 | if (put_user(get_uflags(ppage), out)) { |
| 111 | kpf_copy_bit(kflags, KPF_ERROR, PG_error) | | ||
| 112 | kpf_copy_bit(kflags, KPF_REFERENCED, PG_referenced) | | ||
| 113 | kpf_copy_bit(kflags, KPF_UPTODATE, PG_uptodate) | | ||
| 114 | kpf_copy_bit(kflags, KPF_DIRTY, PG_dirty) | | ||
| 115 | kpf_copy_bit(kflags, KPF_LRU, PG_lru) | | ||
| 116 | kpf_copy_bit(kflags, KPF_ACTIVE, PG_active) | | ||
| 117 | kpf_copy_bit(kflags, KPF_SLAB, PG_slab) | | ||
| 118 | kpf_copy_bit(kflags, KPF_WRITEBACK, PG_writeback) | | ||
| 119 | kpf_copy_bit(kflags, KPF_RECLAIM, PG_reclaim) | | ||
| 120 | kpf_copy_bit(kflags, KPF_BUDDY, PG_buddy); | ||
| 121 | |||
| 122 | if (put_user(uflags, out++)) { | ||
| 123 | ret = -EFAULT; | 213 | ret = -EFAULT; |
| 124 | break; | 214 | break; |
| 125 | } | 215 | } |
| 126 | 216 | ||
| 217 | pfn++; | ||
| 218 | out++; | ||
| 127 | count -= KPMSIZE; | 219 | count -= KPMSIZE; |
| 128 | } | 220 | } |
| 129 | 221 | ||
