aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/mm
diff options
context:
space:
mode:
Diffstat (limited to 'arch/x86/mm')
-rw-r--r--arch/x86/mm/fault.c2
-rw-r--r--arch/x86/mm/init_64.c36
-rw-r--r--arch/x86/mm/numa.c34
-rw-r--r--arch/x86/mm/pgtable_32.c35
4 files changed, 39 insertions, 68 deletions
diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c
index 9c5b32e2bdc0..d973e61e450d 100644
--- a/arch/x86/mm/fault.c
+++ b/arch/x86/mm/fault.c
@@ -349,7 +349,7 @@ out:
349 349
350void vmalloc_sync_all(void) 350void vmalloc_sync_all(void)
351{ 351{
352 sync_global_pgds(VMALLOC_START & PGDIR_MASK, VMALLOC_END); 352 sync_global_pgds(VMALLOC_START & PGDIR_MASK, VMALLOC_END, 0);
353} 353}
354 354
355/* 355/*
diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c
index 5d984769cbd8..4cb8763868fc 100644
--- a/arch/x86/mm/init_64.c
+++ b/arch/x86/mm/init_64.c
@@ -178,7 +178,7 @@ __setup("noexec32=", nonx32_setup);
178 * When memory was added/removed make sure all the processes MM have 178 * When memory was added/removed make sure all the processes MM have
179 * suitable PGD entries in the local PGD level page. 179 * suitable PGD entries in the local PGD level page.
180 */ 180 */
181void sync_global_pgds(unsigned long start, unsigned long end) 181void sync_global_pgds(unsigned long start, unsigned long end, int removed)
182{ 182{
183 unsigned long address; 183 unsigned long address;
184 184
@@ -186,7 +186,12 @@ void sync_global_pgds(unsigned long start, unsigned long end)
186 const pgd_t *pgd_ref = pgd_offset_k(address); 186 const pgd_t *pgd_ref = pgd_offset_k(address);
187 struct page *page; 187 struct page *page;
188 188
189 if (pgd_none(*pgd_ref)) 189 /*
190 * When it is called after memory hot remove, pgd_none()
191 * returns true. In this case (removed == 1), we must clear
192 * the PGD entries in the local PGD level page.
193 */
194 if (pgd_none(*pgd_ref) && !removed)
190 continue; 195 continue;
191 196
192 spin_lock(&pgd_lock); 197 spin_lock(&pgd_lock);
@@ -199,12 +204,18 @@ void sync_global_pgds(unsigned long start, unsigned long end)
199 pgt_lock = &pgd_page_get_mm(page)->page_table_lock; 204 pgt_lock = &pgd_page_get_mm(page)->page_table_lock;
200 spin_lock(pgt_lock); 205 spin_lock(pgt_lock);
201 206
202 if (pgd_none(*pgd)) 207 if (!pgd_none(*pgd_ref) && !pgd_none(*pgd))
203 set_pgd(pgd, *pgd_ref);
204 else
205 BUG_ON(pgd_page_vaddr(*pgd) 208 BUG_ON(pgd_page_vaddr(*pgd)
206 != pgd_page_vaddr(*pgd_ref)); 209 != pgd_page_vaddr(*pgd_ref));
207 210
211 if (removed) {
212 if (pgd_none(*pgd_ref) && !pgd_none(*pgd))
213 pgd_clear(pgd);
214 } else {
215 if (pgd_none(*pgd))
216 set_pgd(pgd, *pgd_ref);
217 }
218
208 spin_unlock(pgt_lock); 219 spin_unlock(pgt_lock);
209 } 220 }
210 spin_unlock(&pgd_lock); 221 spin_unlock(&pgd_lock);
@@ -633,7 +644,7 @@ kernel_physical_mapping_init(unsigned long start,
633 } 644 }
634 645
635 if (pgd_changed) 646 if (pgd_changed)
636 sync_global_pgds(addr, end - 1); 647 sync_global_pgds(addr, end - 1, 0);
637 648
638 __flush_tlb_all(); 649 __flush_tlb_all();
639 650
@@ -976,25 +987,26 @@ static void __meminit
976remove_pagetable(unsigned long start, unsigned long end, bool direct) 987remove_pagetable(unsigned long start, unsigned long end, bool direct)
977{ 988{
978 unsigned long next; 989 unsigned long next;
990 unsigned long addr;
979 pgd_t *pgd; 991 pgd_t *pgd;
980 pud_t *pud; 992 pud_t *pud;
981 bool pgd_changed = false; 993 bool pgd_changed = false;
982 994
983 for (; start < end; start = next) { 995 for (addr = start; addr < end; addr = next) {
984 next = pgd_addr_end(start, end); 996 next = pgd_addr_end(addr, end);
985 997
986 pgd = pgd_offset_k(start); 998 pgd = pgd_offset_k(addr);
987 if (!pgd_present(*pgd)) 999 if (!pgd_present(*pgd))
988 continue; 1000 continue;
989 1001
990 pud = (pud_t *)pgd_page_vaddr(*pgd); 1002 pud = (pud_t *)pgd_page_vaddr(*pgd);
991 remove_pud_table(pud, start, next, direct); 1003 remove_pud_table(pud, addr, next, direct);
992 if (free_pud_table(pud, pgd)) 1004 if (free_pud_table(pud, pgd))
993 pgd_changed = true; 1005 pgd_changed = true;
994 } 1006 }
995 1007
996 if (pgd_changed) 1008 if (pgd_changed)
997 sync_global_pgds(start, end - 1); 1009 sync_global_pgds(start, end - 1, 1);
998 1010
999 flush_tlb_all(); 1011 flush_tlb_all();
1000} 1012}
@@ -1341,7 +1353,7 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
1341 else 1353 else
1342 err = vmemmap_populate_basepages(start, end, node); 1354 err = vmemmap_populate_basepages(start, end, node);
1343 if (!err) 1355 if (!err)
1344 sync_global_pgds(start, end - 1); 1356 sync_global_pgds(start, end - 1, 0);
1345 return err; 1357 return err;
1346} 1358}
1347 1359
diff --git a/arch/x86/mm/numa.c b/arch/x86/mm/numa.c
index 4e1e5709fe17..1a883705a12a 100644
--- a/arch/x86/mm/numa.c
+++ b/arch/x86/mm/numa.c
@@ -185,8 +185,8 @@ int __init numa_add_memblk(int nid, u64 start, u64 end)
185 return numa_add_memblk_to(nid, start, end, &numa_meminfo); 185 return numa_add_memblk_to(nid, start, end, &numa_meminfo);
186} 186}
187 187
188/* Initialize NODE_DATA for a node on the local memory */ 188/* Allocate NODE_DATA for a node on the local memory */
189static void __init setup_node_data(int nid, u64 start, u64 end) 189static void __init alloc_node_data(int nid)
190{ 190{
191 const size_t nd_size = roundup(sizeof(pg_data_t), PAGE_SIZE); 191 const size_t nd_size = roundup(sizeof(pg_data_t), PAGE_SIZE);
192 u64 nd_pa; 192 u64 nd_pa;
@@ -194,18 +194,6 @@ static void __init setup_node_data(int nid, u64 start, u64 end)
194 int tnid; 194 int tnid;
195 195
196 /* 196 /*
197 * Don't confuse VM with a node that doesn't have the
198 * minimum amount of memory:
199 */
200 if (end && (end - start) < NODE_MIN_SIZE)
201 return;
202
203 start = roundup(start, ZONE_ALIGN);
204
205 printk(KERN_INFO "Initmem setup node %d [mem %#010Lx-%#010Lx]\n",
206 nid, start, end - 1);
207
208 /*
209 * Allocate node data. Try node-local memory and then any node. 197 * Allocate node data. Try node-local memory and then any node.
210 * Never allocate in DMA zone. 198 * Never allocate in DMA zone.
211 */ 199 */
@@ -222,7 +210,7 @@ static void __init setup_node_data(int nid, u64 start, u64 end)
222 nd = __va(nd_pa); 210 nd = __va(nd_pa);
223 211
224 /* report and initialize */ 212 /* report and initialize */
225 printk(KERN_INFO " NODE_DATA [mem %#010Lx-%#010Lx]\n", 213 printk(KERN_INFO "NODE_DATA(%d) allocated [mem %#010Lx-%#010Lx]\n", nid,
226 nd_pa, nd_pa + nd_size - 1); 214 nd_pa, nd_pa + nd_size - 1);
227 tnid = early_pfn_to_nid(nd_pa >> PAGE_SHIFT); 215 tnid = early_pfn_to_nid(nd_pa >> PAGE_SHIFT);
228 if (tnid != nid) 216 if (tnid != nid)
@@ -230,9 +218,6 @@ static void __init setup_node_data(int nid, u64 start, u64 end)
230 218
231 node_data[nid] = nd; 219 node_data[nid] = nd;
232 memset(NODE_DATA(nid), 0, sizeof(pg_data_t)); 220 memset(NODE_DATA(nid), 0, sizeof(pg_data_t));
233 NODE_DATA(nid)->node_id = nid;
234 NODE_DATA(nid)->node_start_pfn = start >> PAGE_SHIFT;
235 NODE_DATA(nid)->node_spanned_pages = (end - start) >> PAGE_SHIFT;
236 221
237 node_set_online(nid); 222 node_set_online(nid);
238} 223}
@@ -568,8 +553,17 @@ static int __init numa_register_memblks(struct numa_meminfo *mi)
568 end = max(mi->blk[i].end, end); 553 end = max(mi->blk[i].end, end);
569 } 554 }
570 555
571 if (start < end) 556 if (start >= end)
572 setup_node_data(nid, start, end); 557 continue;
558
559 /*
560 * Don't confuse VM with a node that doesn't have the
561 * minimum amount of memory:
562 */
563 if (end && (end - start) < NODE_MIN_SIZE)
564 continue;
565
566 alloc_node_data(nid);
573 } 567 }
574 568
575 /* Dump memblock with node info and return. */ 569 /* Dump memblock with node info and return. */
diff --git a/arch/x86/mm/pgtable_32.c b/arch/x86/mm/pgtable_32.c
index 4dd8cf652579..75cc0978d45d 100644
--- a/arch/x86/mm/pgtable_32.c
+++ b/arch/x86/mm/pgtable_32.c
@@ -59,41 +59,6 @@ void set_pte_vaddr(unsigned long vaddr, pte_t pteval)
59 __flush_tlb_one(vaddr); 59 __flush_tlb_one(vaddr);
60} 60}
61 61
62/*
63 * Associate a large virtual page frame with a given physical page frame
64 * and protection flags for that frame. pfn is for the base of the page,
65 * vaddr is what the page gets mapped to - both must be properly aligned.
66 * The pmd must already be instantiated. Assumes PAE mode.
67 */
68void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags)
69{
70 pgd_t *pgd;
71 pud_t *pud;
72 pmd_t *pmd;
73
74 if (vaddr & (PMD_SIZE-1)) { /* vaddr is misaligned */
75 printk(KERN_WARNING "set_pmd_pfn: vaddr misaligned\n");
76 return; /* BUG(); */
77 }
78 if (pfn & (PTRS_PER_PTE-1)) { /* pfn is misaligned */
79 printk(KERN_WARNING "set_pmd_pfn: pfn misaligned\n");
80 return; /* BUG(); */
81 }
82 pgd = swapper_pg_dir + pgd_index(vaddr);
83 if (pgd_none(*pgd)) {
84 printk(KERN_WARNING "set_pmd_pfn: pgd_none\n");
85 return; /* BUG(); */
86 }
87 pud = pud_offset(pgd, vaddr);
88 pmd = pmd_offset(pud, vaddr);
89 set_pmd(pmd, pfn_pmd(pfn, flags));
90 /*
91 * It's enough to flush this one mapping.
92 * (PGE mappings get flushed as well)
93 */
94 __flush_tlb_one(vaddr);
95}
96
97unsigned long __FIXADDR_TOP = 0xfffff000; 62unsigned long __FIXADDR_TOP = 0xfffff000;
98EXPORT_SYMBOL(__FIXADDR_TOP); 63EXPORT_SYMBOL(__FIXADDR_TOP);
99 64