aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorYinghai Lu <yhlu.kernel@gmail.com>2008-06-23 19:41:30 -0400
committerIngo Molnar <mingo@elte.hu>2008-07-08 06:50:23 -0400
commitc09434571d4b1d8abf530ba4ce28cb868b45f2e5 (patch)
treee6943ba46c4a36e0d091defbaae2b189ac33393a /arch
parent6a07a0edacba397205ff97308b22c6b6aab9f791 (diff)
x86: numa32 pfn print out using hex instead
Signed-off-by: Yinghai Lu <yhlu.kernel@gmail.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r--arch/x86/kernel/srat_32.c31
-rw-r--r--arch/x86/mm/discontig_32.c29
2 files changed, 34 insertions, 26 deletions
diff --git a/arch/x86/kernel/srat_32.c b/arch/x86/kernel/srat_32.c
index 5978023b799b..f41d67f8f831 100644
--- a/arch/x86/kernel/srat_32.c
+++ b/arch/x86/kernel/srat_32.c
@@ -93,7 +93,7 @@ acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *cpu_affinity)
93 93
94 apicid_to_pxm[cpu_affinity->apic_id] = cpu_affinity->proximity_domain_lo; 94 apicid_to_pxm[cpu_affinity->apic_id] = cpu_affinity->proximity_domain_lo;
95 95
96 printk("CPU 0x%02X in proximity domain 0x%02X\n", 96 printk(KERN_DEBUG "CPU %02x in proximity domain %02x\n",
97 cpu_affinity->apic_id, cpu_affinity->proximity_domain_lo); 97 cpu_affinity->apic_id, cpu_affinity->proximity_domain_lo);
98} 98}
99 99
@@ -134,7 +134,8 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *memory_affinity)
134 134
135 135
136 if (num_memory_chunks >= MAXCHUNKS) { 136 if (num_memory_chunks >= MAXCHUNKS) {
137 printk("Too many mem chunks in SRAT. Ignoring %lld MBytes at %llx\n", 137 printk(KERN_WARNING "Too many mem chunks in SRAT."
138 " Ignoring %lld MBytes at %llx\n",
138 size/(1024*1024), paddr); 139 size/(1024*1024), paddr);
139 return; 140 return;
140 } 141 }
@@ -155,7 +156,8 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *memory_affinity)
155 156
156 num_memory_chunks++; 157 num_memory_chunks++;
157 158
158 printk("Memory range 0x%lX to 0x%lX (type 0x%X) in proximity domain 0x%02X %s\n", 159 printk(KERN_DEBUG "Memory range %08lx to %08lx (type %x)"
160 " in proximity domain %02x %s\n",
159 start_pfn, end_pfn, 161 start_pfn, end_pfn,
160 memory_affinity->memory_type, 162 memory_affinity->memory_type,
161 pxm, 163 pxm,
@@ -186,7 +188,7 @@ static __init void node_read_chunk(int nid, struct node_memory_chunk_s *memory_c
186 * *possible* memory hotplug areas the same as normal RAM. 188 * *possible* memory hotplug areas the same as normal RAM.
187 */ 189 */
188 if (memory_chunk->start_pfn >= max_pfn) { 190 if (memory_chunk->start_pfn >= max_pfn) {
189 printk (KERN_INFO "Ignoring SRAT pfns: 0x%08lx -> %08lx\n", 191 printk(KERN_INFO "Ignoring SRAT pfns: %08lx - %08lx\n",
190 memory_chunk->start_pfn, memory_chunk->end_pfn); 192 memory_chunk->start_pfn, memory_chunk->end_pfn);
191 return; 193 return;
192 } 194 }
@@ -212,7 +214,8 @@ int __init get_memcfg_from_srat(void)
212 goto out_fail; 214 goto out_fail;
213 215
214 if (num_memory_chunks == 0) { 216 if (num_memory_chunks == 0) {
215 printk("could not finy any ACPI SRAT memory areas.\n"); 217 printk(KERN_WARNING
218 "could not finy any ACPI SRAT memory areas.\n");
216 goto out_fail; 219 goto out_fail;
217 } 220 }
218 221
@@ -239,20 +242,23 @@ int __init get_memcfg_from_srat(void)
239 for (i = 0; i < num_memory_chunks; i++) 242 for (i = 0; i < num_memory_chunks; i++)
240 node_memory_chunk[i].nid = pxm_to_node(node_memory_chunk[i].pxm); 243 node_memory_chunk[i].nid = pxm_to_node(node_memory_chunk[i].pxm);
241 244
242 printk("pxm bitmap: "); 245 printk(KERN_DEBUG "pxm bitmap: ");
243 for (i = 0; i < sizeof(pxm_bitmap); i++) { 246 for (i = 0; i < sizeof(pxm_bitmap); i++) {
244 printk("%02X ", pxm_bitmap[i]); 247 printk(KERN_CONT "%02x ", pxm_bitmap[i]);
245 } 248 }
246 printk("\n"); 249 printk(KERN_CONT "\n");
247 printk("Number of logical nodes in system = %d\n", num_online_nodes()); 250 printk(KERN_DEBUG "Number of logical nodes in system = %d\n",
248 printk("Number of memory chunks in system = %d\n", num_memory_chunks); 251 num_online_nodes());
252 printk(KERN_DEBUG "Number of memory chunks in system = %d\n",
253 num_memory_chunks);
249 254
250 for (i = 0; i < MAX_APICID; i++) 255 for (i = 0; i < MAX_APICID; i++)
251 apicid_2_node[i] = pxm_to_node(apicid_to_pxm[i]); 256 apicid_2_node[i] = pxm_to_node(apicid_to_pxm[i]);
252 257
253 for (j = 0; j < num_memory_chunks; j++){ 258 for (j = 0; j < num_memory_chunks; j++){
254 struct node_memory_chunk_s * chunk = &node_memory_chunk[j]; 259 struct node_memory_chunk_s * chunk = &node_memory_chunk[j];
255 printk("chunk %d nid %d start_pfn %08lx end_pfn %08lx\n", 260 printk(KERN_DEBUG
261 "chunk %d nid %d start_pfn %08lx end_pfn %08lx\n",
256 j, chunk->nid, chunk->start_pfn, chunk->end_pfn); 262 j, chunk->nid, chunk->start_pfn, chunk->end_pfn);
257 node_read_chunk(chunk->nid, chunk); 263 node_read_chunk(chunk->nid, chunk);
258 e820_register_active_regions(chunk->nid, chunk->start_pfn, 264 e820_register_active_regions(chunk->nid, chunk->start_pfn,
@@ -268,6 +274,7 @@ int __init get_memcfg_from_srat(void)
268 } 274 }
269 return 1; 275 return 1;
270out_fail: 276out_fail:
271 printk("failed to get NUMA memory information from SRAT table\n"); 277 printk(KERN_ERR "failed to get NUMA memory information from SRAT"
278 " table\n");
272 return 0; 279 return 0;
273} 280}
diff --git a/arch/x86/mm/discontig_32.c b/arch/x86/mm/discontig_32.c
index 1dfff700264c..f5ae31935ca8 100644
--- a/arch/x86/mm/discontig_32.c
+++ b/arch/x86/mm/discontig_32.c
@@ -76,13 +76,13 @@ void memory_present(int nid, unsigned long start, unsigned long end)
76{ 76{
77 unsigned long pfn; 77 unsigned long pfn;
78 78
79 printk(KERN_INFO "Node: %d, start_pfn: %ld, end_pfn: %ld\n", 79 printk(KERN_INFO "Node: %d, start_pfn: %lx, end_pfn: %lx\n",
80 nid, start, end); 80 nid, start, end);
81 printk(KERN_DEBUG " Setting physnode_map array to node %d for pfns:\n", nid); 81 printk(KERN_DEBUG " Setting physnode_map array to node %d for pfns:\n", nid);
82 printk(KERN_DEBUG " "); 82 printk(KERN_DEBUG " ");
83 for (pfn = start; pfn < end; pfn += PAGES_PER_ELEMENT) { 83 for (pfn = start; pfn < end; pfn += PAGES_PER_ELEMENT) {
84 physnode_map[pfn / PAGES_PER_ELEMENT] = nid; 84 physnode_map[pfn / PAGES_PER_ELEMENT] = nid;
85 printk(KERN_CONT "%ld ", pfn); 85 printk(KERN_CONT "%lx ", pfn);
86 } 86 }
87 printk(KERN_CONT "\n"); 87 printk(KERN_CONT "\n");
88} 88}
@@ -117,7 +117,7 @@ static unsigned long kva_pages;
117 */ 117 */
118int __init get_memcfg_numa_flat(void) 118int __init get_memcfg_numa_flat(void)
119{ 119{
120 printk("NUMA - single node, flat memory mode\n"); 120 printk(KERN_DEBUG "NUMA - single node, flat memory mode\n");
121 121
122 node_start_pfn[0] = 0; 122 node_start_pfn[0] = 0;
123 node_end_pfn[0] = max_pfn; 123 node_end_pfn[0] = max_pfn;
@@ -233,7 +233,7 @@ static unsigned long calculate_numa_remap_pages(void)
233 * The acpi/srat node info can show hot-add memroy zones 233 * The acpi/srat node info can show hot-add memroy zones
234 * where memory could be added but not currently present. 234 * where memory could be added but not currently present.
235 */ 235 */
236 printk("node %d pfn: [%lx - %lx]\n", 236 printk(KERN_DEBUG "node %d pfn: [%lx - %lx]\n",
237 nid, node_start_pfn[nid], node_end_pfn[nid]); 237 nid, node_start_pfn[nid], node_end_pfn[nid]);
238 if (node_start_pfn[nid] > max_pfn) 238 if (node_start_pfn[nid] > max_pfn)
239 continue; 239 continue;
@@ -268,7 +268,8 @@ static unsigned long calculate_numa_remap_pages(void)
268 node_remap_size[nid] = size; 268 node_remap_size[nid] = size;
269 node_remap_offset[nid] = reserve_pages; 269 node_remap_offset[nid] = reserve_pages;
270 reserve_pages += size; 270 reserve_pages += size;
271 printk("Reserving %ld pages of KVA for lmem_map of node %d at %llx\n", 271 printk(KERN_DEBUG "Reserving %ld pages of KVA for lmem_map of"
272 " node %d at %llx\n",
272 size, nid, node_kva_final>>PAGE_SHIFT); 273 size, nid, node_kva_final>>PAGE_SHIFT);
273 274
274 /* 275 /*
@@ -290,7 +291,7 @@ static unsigned long calculate_numa_remap_pages(void)
290 remove_active_range(nid, node_remap_start_pfn[nid], 291 remove_active_range(nid, node_remap_start_pfn[nid],
291 node_remap_start_pfn[nid] + size); 292 node_remap_start_pfn[nid] + size);
292 } 293 }
293 printk("Reserving total of %ld pages for numa KVA remap\n", 294 printk(KERN_INFO "Reserving total of %lx pages for numa KVA remap\n",
294 reserve_pages); 295 reserve_pages);
295 return reserve_pages; 296 return reserve_pages;
296} 297}
@@ -304,7 +305,7 @@ static void init_remap_allocator(int nid)
304 node_remap_alloc_vaddr[nid] = node_remap_start_vaddr[nid] + 305 node_remap_alloc_vaddr[nid] = node_remap_start_vaddr[nid] +
305 ALIGN(sizeof(pg_data_t), PAGE_SIZE); 306 ALIGN(sizeof(pg_data_t), PAGE_SIZE);
306 307
307 printk ("node %d will remap to vaddr %08lx - %08lx\n", nid, 308 printk(KERN_DEBUG "node %d will remap to vaddr %08lx - %08lx\n", nid,
308 (ulong) node_remap_start_vaddr[nid], 309 (ulong) node_remap_start_vaddr[nid],
309 (ulong) node_remap_end_vaddr[nid]); 310 (ulong) node_remap_end_vaddr[nid]);
310} 311}
@@ -340,9 +341,9 @@ void __init initmem_init(unsigned long start_pfn,
340 if (kva_start_pfn == -1UL) 341 if (kva_start_pfn == -1UL)
341 panic("Can not get kva space\n"); 342 panic("Can not get kva space\n");
342 343
343 printk("kva_start_pfn ~ %ld find_max_low_pfn() ~ %ld\n", 344 printk(KERN_INFO "kva_start_pfn ~ %lx max_low_pfn ~ %lx\n",
344 kva_start_pfn, max_low_pfn); 345 kva_start_pfn, max_low_pfn);
345 printk("max_pfn = %ld\n", max_pfn); 346 printk(KERN_INFO "max_pfn = %lx\n", max_pfn);
346 347
347 /* avoid clash with initrd */ 348 /* avoid clash with initrd */
348 reserve_early(kva_start_pfn<<PAGE_SHIFT, 349 reserve_early(kva_start_pfn<<PAGE_SHIFT,
@@ -362,17 +363,17 @@ void __init initmem_init(unsigned long start_pfn,
362#endif 363#endif
363 printk(KERN_NOTICE "%ldMB LOWMEM available.\n", 364 printk(KERN_NOTICE "%ldMB LOWMEM available.\n",
364 pages_to_mb(max_low_pfn)); 365 pages_to_mb(max_low_pfn));
365 printk("min_low_pfn = %ld, max_low_pfn = %ld, highstart_pfn = %ld\n", 366 printk(KERN_DEBUG "max_low_pfn = %lx, highstart_pfn = %lx\n",
366 min_low_pfn, max_low_pfn, highstart_pfn); 367 max_low_pfn, highstart_pfn);
367 368
368 printk("Low memory ends at vaddr %08lx\n", 369 printk(KERN_DEBUG "Low memory ends at vaddr %08lx\n",
369 (ulong) pfn_to_kaddr(max_low_pfn)); 370 (ulong) pfn_to_kaddr(max_low_pfn));
370 for_each_online_node(nid) { 371 for_each_online_node(nid) {
371 init_remap_allocator(nid); 372 init_remap_allocator(nid);
372 373
373 allocate_pgdat(nid); 374 allocate_pgdat(nid);
374 } 375 }
375 printk("High memory starts at vaddr %08lx\n", 376 printk(KERN_DEBUG "High memory starts at vaddr %08lx\n",
376 (ulong) pfn_to_kaddr(highstart_pfn)); 377 (ulong) pfn_to_kaddr(highstart_pfn));
377 for_each_online_node(nid) 378 for_each_online_node(nid)
378 propagate_e820_map_node(nid); 379 propagate_e820_map_node(nid);
@@ -413,7 +414,7 @@ void __init set_highmem_pages_init(void)
413 zone_end_pfn = zone_start_pfn + zone->spanned_pages; 414 zone_end_pfn = zone_start_pfn + zone->spanned_pages;
414 415
415 nid = zone_to_nid(zone); 416 nid = zone_to_nid(zone);
416 printk("Initializing %s for node %d (%08lx:%08lx)\n", 417 printk(KERN_INFO "Initializing %s for node %d (%08lx:%08lx)\n",
417 zone->name, nid, zone_start_pfn, zone_end_pfn); 418 zone->name, nid, zone_start_pfn, zone_end_pfn);
418 419
419 add_highpages_with_active_regions(nid, zone_start_pfn, 420 add_highpages_with_active_regions(nid, zone_start_pfn,