aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMel Gorman <mel@csn.ul.ie>2007-07-17 07:03:13 -0400
committerLinus Torvalds <torvalds@woody.linux-foundation.org>2007-07-17 13:22:59 -0400
commit396faf0303d273219db5d7eb4a2879ad977ed185 (patch)
tree96cb64fd6713ef7a924f4f878e259aea781f079a
parent2a1e274acf0b1c192face19a4be7c12d4503eaaf (diff)
Allow huge page allocations to use GFP_HIGH_MOVABLE
Huge pages are not movable so are not allocated from ZONE_MOVABLE. However, as ZONE_MOVABLE will always have pages that can be migrated or reclaimed, it can be used to satisfy hugepage allocations even when the system has been running a long time. This allows an administrator to resize the hugepage pool at runtime depending on the size of ZONE_MOVABLE. This patch adds a new sysctl called hugepages_treat_as_movable. When a non-zero value is written to it, future allocations for the huge page pool will use ZONE_MOVABLE. Despite huge pages being non-movable, we do not introduce additional external fragmentation of note as huge pages are always the largest contiguous block we care about. [akpm@linux-foundation.org: various fixes] Signed-off-by: Mel Gorman <mel@csn.ul.ie> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r--include/linux/hugetlb.h2
-rw-r--r--include/linux/mempolicy.h6
-rw-r--r--kernel/sysctl.c8
-rw-r--r--mm/hugetlb.c23
-rw-r--r--mm/mempolicy.c5
5 files changed, 36 insertions, 8 deletions
diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h
index 2c13715e9dde..49b7053043ad 100644
--- a/include/linux/hugetlb.h
+++ b/include/linux/hugetlb.h
@@ -15,6 +15,7 @@ static inline int is_vm_hugetlb_page(struct vm_area_struct *vma)
15} 15}
16 16
17int hugetlb_sysctl_handler(struct ctl_table *, int, struct file *, void __user *, size_t *, loff_t *); 17int hugetlb_sysctl_handler(struct ctl_table *, int, struct file *, void __user *, size_t *, loff_t *);
18int hugetlb_treat_movable_handler(struct ctl_table *, int, struct file *, void __user *, size_t *, loff_t *);
18int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *); 19int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *);
19int follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *, struct page **, struct vm_area_struct **, unsigned long *, int *, int); 20int follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *, struct page **, struct vm_area_struct **, unsigned long *, int *, int);
20void unmap_hugepage_range(struct vm_area_struct *, unsigned long, unsigned long); 21void unmap_hugepage_range(struct vm_area_struct *, unsigned long, unsigned long);
@@ -29,6 +30,7 @@ int hugetlb_reserve_pages(struct inode *inode, long from, long to);
29void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed); 30void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed);
30 31
31extern unsigned long max_huge_pages; 32extern unsigned long max_huge_pages;
33extern unsigned long hugepages_treat_as_movable;
32extern const unsigned long hugetlb_zero, hugetlb_infinity; 34extern const unsigned long hugetlb_zero, hugetlb_infinity;
33extern int sysctl_hugetlb_shm_group; 35extern int sysctl_hugetlb_shm_group;
34 36
diff --git a/include/linux/mempolicy.h b/include/linux/mempolicy.h
index daabb3aa1ec6..e147cf50529f 100644
--- a/include/linux/mempolicy.h
+++ b/include/linux/mempolicy.h
@@ -159,7 +159,7 @@ extern void mpol_fix_fork_child_flag(struct task_struct *p);
159 159
160extern struct mempolicy default_policy; 160extern struct mempolicy default_policy;
161extern struct zonelist *huge_zonelist(struct vm_area_struct *vma, 161extern struct zonelist *huge_zonelist(struct vm_area_struct *vma,
162 unsigned long addr); 162 unsigned long addr, gfp_t gfp_flags);
163extern unsigned slab_node(struct mempolicy *policy); 163extern unsigned slab_node(struct mempolicy *policy);
164 164
165extern enum zone_type policy_zone; 165extern enum zone_type policy_zone;
@@ -256,9 +256,9 @@ static inline void mpol_fix_fork_child_flag(struct task_struct *p)
256#define set_cpuset_being_rebound(x) do {} while (0) 256#define set_cpuset_being_rebound(x) do {} while (0)
257 257
258static inline struct zonelist *huge_zonelist(struct vm_area_struct *vma, 258static inline struct zonelist *huge_zonelist(struct vm_area_struct *vma,
259 unsigned long addr) 259 unsigned long addr, gfp_t gfp_flags)
260{ 260{
261 return NODE_DATA(0)->node_zonelists + gfp_zone(GFP_HIGHUSER); 261 return NODE_DATA(0)->node_zonelists + gfp_zone(gfp_flags);
262} 262}
263 263
264static inline int do_migrate_pages(struct mm_struct *mm, 264static inline int do_migrate_pages(struct mm_struct *mm,
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
index 2ce7acf841ae..48dae075d5c2 100644
--- a/kernel/sysctl.c
+++ b/kernel/sysctl.c
@@ -826,6 +826,14 @@ static ctl_table vm_table[] = {
826 .mode = 0644, 826 .mode = 0644,
827 .proc_handler = &proc_dointvec, 827 .proc_handler = &proc_dointvec,
828 }, 828 },
829 {
830 .ctl_name = CTL_UNNUMBERED,
831 .procname = "hugepages_treat_as_movable",
832 .data = &hugepages_treat_as_movable,
833 .maxlen = sizeof(int),
834 .mode = 0644,
835 .proc_handler = &hugetlb_treat_movable_handler,
836 },
829#endif 837#endif
830 { 838 {
831 .ctl_name = VM_LOWMEM_RESERVE_RATIO, 839 .ctl_name = VM_LOWMEM_RESERVE_RATIO,
diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index acc0fb3cf067..58980676b842 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -27,6 +27,9 @@ unsigned long max_huge_pages;
27static struct list_head hugepage_freelists[MAX_NUMNODES]; 27static struct list_head hugepage_freelists[MAX_NUMNODES];
28static unsigned int nr_huge_pages_node[MAX_NUMNODES]; 28static unsigned int nr_huge_pages_node[MAX_NUMNODES];
29static unsigned int free_huge_pages_node[MAX_NUMNODES]; 29static unsigned int free_huge_pages_node[MAX_NUMNODES];
30static gfp_t htlb_alloc_mask = GFP_HIGHUSER;
31unsigned long hugepages_treat_as_movable;
32
30/* 33/*
31 * Protects updates to hugepage_freelists, nr_huge_pages, and free_huge_pages 34 * Protects updates to hugepage_freelists, nr_huge_pages, and free_huge_pages
32 */ 35 */
@@ -68,12 +71,13 @@ static struct page *dequeue_huge_page(struct vm_area_struct *vma,
68{ 71{
69 int nid; 72 int nid;
70 struct page *page = NULL; 73 struct page *page = NULL;
71 struct zonelist *zonelist = huge_zonelist(vma, address); 74 struct zonelist *zonelist = huge_zonelist(vma, address,
75 htlb_alloc_mask);
72 struct zone **z; 76 struct zone **z;
73 77
74 for (z = zonelist->zones; *z; z++) { 78 for (z = zonelist->zones; *z; z++) {
75 nid = zone_to_nid(*z); 79 nid = zone_to_nid(*z);
76 if (cpuset_zone_allowed_softwall(*z, GFP_HIGHUSER) && 80 if (cpuset_zone_allowed_softwall(*z, htlb_alloc_mask) &&
77 !list_empty(&hugepage_freelists[nid])) 81 !list_empty(&hugepage_freelists[nid]))
78 break; 82 break;
79 } 83 }
@@ -113,7 +117,7 @@ static int alloc_fresh_huge_page(void)
113 prev_nid = nid; 117 prev_nid = nid;
114 spin_unlock(&nid_lock); 118 spin_unlock(&nid_lock);
115 119
116 page = alloc_pages_node(nid, GFP_HIGHUSER|__GFP_COMP|__GFP_NOWARN, 120 page = alloc_pages_node(nid, htlb_alloc_mask|__GFP_COMP|__GFP_NOWARN,
117 HUGETLB_PAGE_ORDER); 121 HUGETLB_PAGE_ORDER);
118 if (page) { 122 if (page) {
119 set_compound_page_dtor(page, free_huge_page); 123 set_compound_page_dtor(page, free_huge_page);
@@ -263,6 +267,19 @@ int hugetlb_sysctl_handler(struct ctl_table *table, int write,
263 max_huge_pages = set_max_huge_pages(max_huge_pages); 267 max_huge_pages = set_max_huge_pages(max_huge_pages);
264 return 0; 268 return 0;
265} 269}
270
271int hugetlb_treat_movable_handler(struct ctl_table *table, int write,
272 struct file *file, void __user *buffer,
273 size_t *length, loff_t *ppos)
274{
275 proc_dointvec(table, write, file, buffer, length, ppos);
276 if (hugepages_treat_as_movable)
277 htlb_alloc_mask = GFP_HIGHUSER_MOVABLE;
278 else
279 htlb_alloc_mask = GFP_HIGHUSER;
280 return 0;
281}
282
266#endif /* CONFIG_SYSCTL */ 283#endif /* CONFIG_SYSCTL */
267 284
268int hugetlb_report_meminfo(char *buf) 285int hugetlb_report_meminfo(char *buf)
diff --git a/mm/mempolicy.c b/mm/mempolicy.c
index 4c0f99996811..9f4e9b95e8f2 100644
--- a/mm/mempolicy.c
+++ b/mm/mempolicy.c
@@ -1203,7 +1203,8 @@ static inline unsigned interleave_nid(struct mempolicy *pol,
1203 1203
1204#ifdef CONFIG_HUGETLBFS 1204#ifdef CONFIG_HUGETLBFS
1205/* Return a zonelist suitable for a huge page allocation. */ 1205/* Return a zonelist suitable for a huge page allocation. */
1206struct zonelist *huge_zonelist(struct vm_area_struct *vma, unsigned long addr) 1206struct zonelist *huge_zonelist(struct vm_area_struct *vma, unsigned long addr,
1207 gfp_t gfp_flags)
1207{ 1208{
1208 struct mempolicy *pol = get_vma_policy(current, vma, addr); 1209 struct mempolicy *pol = get_vma_policy(current, vma, addr);
1209 1210
@@ -1211,7 +1212,7 @@ struct zonelist *huge_zonelist(struct vm_area_struct *vma, unsigned long addr)
1211 unsigned nid; 1212 unsigned nid;
1212 1213
1213 nid = interleave_nid(pol, vma, addr, HPAGE_SHIFT); 1214 nid = interleave_nid(pol, vma, addr, HPAGE_SHIFT);
1214 return NODE_DATA(nid)->node_zonelists + gfp_zone(GFP_HIGHUSER); 1215 return NODE_DATA(nid)->node_zonelists + gfp_zone(gfp_flags);
1215 } 1216 }
1216 return zonelist_policy(GFP_HIGHUSER, pol); 1217 return zonelist_policy(GFP_HIGHUSER, pol);
1217} 1218}