diff options
author | Tejun Heo <tj@kernel.org> | 2011-07-12 04:46:30 -0400 |
---|---|---|
committer | H. Peter Anvin <hpa@linux.intel.com> | 2011-07-14 14:45:31 -0400 |
commit | c13291a536b835b2ab278ab201f2cb1ce22f2785 (patch) | |
tree | 6bb3a2fd47e22d75308314b14f3a0f0a4d338141 /mm/page_alloc.c | |
parent | 96e907d1360240d1958fe8ce3a3ac640733330d4 (diff) |
bootmem: Use for_each_mem_pfn_range() in page_alloc.c
The previous patch added for_each_mem_pfn_range() which is more
versatile than for_each_active_range_index_in_nid(). This patch
replaces for_each_active_range_index_in_nid() and open coded
early_node_map[] walks with for_each_mem_pfn_range().
All conversions in this patch are straight-forward and shouldn't cause
any functional difference. After the conversions,
for_each_active_range_index_in_nid() doesn't have any user left and is
removed.
Signed-off-by: Tejun Heo <tj@kernel.org>
Link: http://lkml.kernel.org/r/1310460395-30913-4-git-send-email-tj@kernel.org
Cc: Yinghai Lu <yinghai@kernel.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
Diffstat (limited to 'mm/page_alloc.c')
-rw-r--r-- | mm/page_alloc.c | 165 |
1 files changed, 56 insertions, 109 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 3092a97268d1..902f03a4fd6b 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
@@ -3711,34 +3711,6 @@ __meminit int init_currently_empty_zone(struct zone *zone, | |||
3711 | } | 3711 | } |
3712 | 3712 | ||
3713 | #ifdef CONFIG_ARCH_POPULATES_NODE_MAP | 3713 | #ifdef CONFIG_ARCH_POPULATES_NODE_MAP |
3714 | /* | ||
3715 | * Basic iterator support. Return the first range of PFNs for a node | ||
3716 | * Note: nid == MAX_NUMNODES returns first region regardless of node | ||
3717 | */ | ||
3718 | static int __meminit first_active_region_index_in_nid(int nid) | ||
3719 | { | ||
3720 | int i; | ||
3721 | |||
3722 | for (i = 0; i < nr_nodemap_entries; i++) | ||
3723 | if (nid == MAX_NUMNODES || early_node_map[i].nid == nid) | ||
3724 | return i; | ||
3725 | |||
3726 | return -1; | ||
3727 | } | ||
3728 | |||
3729 | /* | ||
3730 | * Basic iterator support. Return the next active range of PFNs for a node | ||
3731 | * Note: nid == MAX_NUMNODES returns next region regardless of node | ||
3732 | */ | ||
3733 | static int __meminit next_active_region_index_in_nid(int index, int nid) | ||
3734 | { | ||
3735 | for (index = index + 1; index < nr_nodemap_entries; index++) | ||
3736 | if (nid == MAX_NUMNODES || early_node_map[index].nid == nid) | ||
3737 | return index; | ||
3738 | |||
3739 | return -1; | ||
3740 | } | ||
3741 | |||
3742 | #ifndef CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID | 3714 | #ifndef CONFIG_HAVE_ARCH_EARLY_PFN_TO_NID |
3743 | /* | 3715 | /* |
3744 | * Required by SPARSEMEM. Given a PFN, return what node the PFN is on. | 3716 | * Required by SPARSEMEM. Given a PFN, return what node the PFN is on. |
@@ -3748,15 +3720,12 @@ static int __meminit next_active_region_index_in_nid(int index, int nid) | |||
3748 | */ | 3720 | */ |
3749 | int __meminit __early_pfn_to_nid(unsigned long pfn) | 3721 | int __meminit __early_pfn_to_nid(unsigned long pfn) |
3750 | { | 3722 | { |
3751 | int i; | 3723 | unsigned long start_pfn, end_pfn; |
3752 | 3724 | int i, nid; | |
3753 | for (i = 0; i < nr_nodemap_entries; i++) { | ||
3754 | unsigned long start_pfn = early_node_map[i].start_pfn; | ||
3755 | unsigned long end_pfn = early_node_map[i].end_pfn; | ||
3756 | 3725 | ||
3726 | for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) | ||
3757 | if (start_pfn <= pfn && pfn < end_pfn) | 3727 | if (start_pfn <= pfn && pfn < end_pfn) |
3758 | return early_node_map[i].nid; | 3728 | return nid; |
3759 | } | ||
3760 | /* This is a memory hole */ | 3729 | /* This is a memory hole */ |
3761 | return -1; | 3730 | return -1; |
3762 | } | 3731 | } |
@@ -3785,11 +3754,6 @@ bool __meminit early_pfn_in_nid(unsigned long pfn, int node) | |||
3785 | } | 3754 | } |
3786 | #endif | 3755 | #endif |
3787 | 3756 | ||
3788 | /* Basic iterator support to walk early_node_map[] */ | ||
3789 | #define for_each_active_range_index_in_nid(i, nid) \ | ||
3790 | for (i = first_active_region_index_in_nid(nid); i != -1; \ | ||
3791 | i = next_active_region_index_in_nid(i, nid)) | ||
3792 | |||
3793 | /** | 3757 | /** |
3794 | * free_bootmem_with_active_regions - Call free_bootmem_node for each active range | 3758 | * free_bootmem_with_active_regions - Call free_bootmem_node for each active range |
3795 | * @nid: The node to free memory on. If MAX_NUMNODES, all nodes are freed. | 3759 | * @nid: The node to free memory on. If MAX_NUMNODES, all nodes are freed. |
@@ -3799,25 +3763,19 @@ bool __meminit early_pfn_in_nid(unsigned long pfn, int node) | |||
3799 | * add_active_ranges() contain no holes and may be freed, this | 3763 | * add_active_ranges() contain no holes and may be freed, this |
3800 | * this function may be used instead of calling free_bootmem() manually. | 3764 | * this function may be used instead of calling free_bootmem() manually. |
3801 | */ | 3765 | */ |
3802 | void __init free_bootmem_with_active_regions(int nid, | 3766 | void __init free_bootmem_with_active_regions(int nid, unsigned long max_low_pfn) |
3803 | unsigned long max_low_pfn) | ||
3804 | { | 3767 | { |
3805 | int i; | 3768 | unsigned long start_pfn, end_pfn; |
3806 | 3769 | int i, this_nid; | |
3807 | for_each_active_range_index_in_nid(i, nid) { | ||
3808 | unsigned long size_pages = 0; | ||
3809 | unsigned long end_pfn = early_node_map[i].end_pfn; | ||
3810 | |||
3811 | if (early_node_map[i].start_pfn >= max_low_pfn) | ||
3812 | continue; | ||
3813 | 3770 | ||
3814 | if (end_pfn > max_low_pfn) | 3771 | for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) { |
3815 | end_pfn = max_low_pfn; | 3772 | start_pfn = min(start_pfn, max_low_pfn); |
3773 | end_pfn = min(end_pfn, max_low_pfn); | ||
3816 | 3774 | ||
3817 | size_pages = end_pfn - early_node_map[i].start_pfn; | 3775 | if (start_pfn < end_pfn) |
3818 | free_bootmem_node(NODE_DATA(early_node_map[i].nid), | 3776 | free_bootmem_node(NODE_DATA(this_nid), |
3819 | PFN_PHYS(early_node_map[i].start_pfn), | 3777 | PFN_PHYS(start_pfn), |
3820 | size_pages << PAGE_SHIFT); | 3778 | (end_pfn - start_pfn) << PAGE_SHIFT); |
3821 | } | 3779 | } |
3822 | } | 3780 | } |
3823 | 3781 | ||
@@ -3891,15 +3849,12 @@ u64 __init find_memory_core_early(int nid, u64 size, u64 align, | |||
3891 | int __init add_from_early_node_map(struct range *range, int az, | 3849 | int __init add_from_early_node_map(struct range *range, int az, |
3892 | int nr_range, int nid) | 3850 | int nr_range, int nid) |
3893 | { | 3851 | { |
3852 | unsigned long start_pfn, end_pfn; | ||
3894 | int i; | 3853 | int i; |
3895 | u64 start, end; | ||
3896 | 3854 | ||
3897 | /* need to go over early_node_map to find out good range for node */ | 3855 | /* need to go over early_node_map to find out good range for node */ |
3898 | for_each_active_range_index_in_nid(i, nid) { | 3856 | for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) |
3899 | start = early_node_map[i].start_pfn; | 3857 | nr_range = add_range(range, az, nr_range, start_pfn, end_pfn); |
3900 | end = early_node_map[i].end_pfn; | ||
3901 | nr_range = add_range(range, az, nr_range, start, end); | ||
3902 | } | ||
3903 | return nr_range; | 3858 | return nr_range; |
3904 | } | 3859 | } |
3905 | 3860 | ||
@@ -3913,12 +3868,11 @@ int __init add_from_early_node_map(struct range *range, int az, | |||
3913 | */ | 3868 | */ |
3914 | void __init sparse_memory_present_with_active_regions(int nid) | 3869 | void __init sparse_memory_present_with_active_regions(int nid) |
3915 | { | 3870 | { |
3916 | int i; | 3871 | unsigned long start_pfn, end_pfn; |
3872 | int i, this_nid; | ||
3917 | 3873 | ||
3918 | for_each_active_range_index_in_nid(i, nid) | 3874 | for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) |
3919 | memory_present(early_node_map[i].nid, | 3875 | memory_present(this_nid, start_pfn, end_pfn); |
3920 | early_node_map[i].start_pfn, | ||
3921 | early_node_map[i].end_pfn); | ||
3922 | } | 3876 | } |
3923 | 3877 | ||
3924 | /** | 3878 | /** |
@@ -3935,13 +3889,15 @@ void __init sparse_memory_present_with_active_regions(int nid) | |||
3935 | void __meminit get_pfn_range_for_nid(unsigned int nid, | 3889 | void __meminit get_pfn_range_for_nid(unsigned int nid, |
3936 | unsigned long *start_pfn, unsigned long *end_pfn) | 3890 | unsigned long *start_pfn, unsigned long *end_pfn) |
3937 | { | 3891 | { |
3892 | unsigned long this_start_pfn, this_end_pfn; | ||
3938 | int i; | 3893 | int i; |
3894 | |||
3939 | *start_pfn = -1UL; | 3895 | *start_pfn = -1UL; |
3940 | *end_pfn = 0; | 3896 | *end_pfn = 0; |
3941 | 3897 | ||
3942 | for_each_active_range_index_in_nid(i, nid) { | 3898 | for_each_mem_pfn_range(i, nid, &this_start_pfn, &this_end_pfn, NULL) { |
3943 | *start_pfn = min(*start_pfn, early_node_map[i].start_pfn); | 3899 | *start_pfn = min(*start_pfn, this_start_pfn); |
3944 | *end_pfn = max(*end_pfn, early_node_map[i].end_pfn); | 3900 | *end_pfn = max(*end_pfn, this_end_pfn); |
3945 | } | 3901 | } |
3946 | 3902 | ||
3947 | if (*start_pfn == -1UL) | 3903 | if (*start_pfn == -1UL) |
@@ -4484,6 +4440,7 @@ void __init add_active_range(unsigned int nid, unsigned long start_pfn, | |||
4484 | void __init remove_active_range(unsigned int nid, unsigned long start_pfn, | 4440 | void __init remove_active_range(unsigned int nid, unsigned long start_pfn, |
4485 | unsigned long end_pfn) | 4441 | unsigned long end_pfn) |
4486 | { | 4442 | { |
4443 | unsigned long this_start_pfn, this_end_pfn; | ||
4487 | int i, j; | 4444 | int i, j; |
4488 | int removed = 0; | 4445 | int removed = 0; |
4489 | 4446 | ||
@@ -4491,26 +4448,22 @@ void __init remove_active_range(unsigned int nid, unsigned long start_pfn, | |||
4491 | nid, start_pfn, end_pfn); | 4448 | nid, start_pfn, end_pfn); |
4492 | 4449 | ||
4493 | /* Find the old active region end and shrink */ | 4450 | /* Find the old active region end and shrink */ |
4494 | for_each_active_range_index_in_nid(i, nid) { | 4451 | for_each_mem_pfn_range(i, nid, &this_start_pfn, &this_end_pfn, NULL) { |
4495 | if (early_node_map[i].start_pfn >= start_pfn && | 4452 | if (this_start_pfn >= start_pfn && this_end_pfn <= end_pfn) { |
4496 | early_node_map[i].end_pfn <= end_pfn) { | ||
4497 | /* clear it */ | 4453 | /* clear it */ |
4498 | early_node_map[i].start_pfn = 0; | 4454 | early_node_map[i].start_pfn = 0; |
4499 | early_node_map[i].end_pfn = 0; | 4455 | early_node_map[i].end_pfn = 0; |
4500 | removed = 1; | 4456 | removed = 1; |
4501 | continue; | 4457 | continue; |
4502 | } | 4458 | } |
4503 | if (early_node_map[i].start_pfn < start_pfn && | 4459 | if (this_start_pfn < start_pfn && this_end_pfn > start_pfn) { |
4504 | early_node_map[i].end_pfn > start_pfn) { | ||
4505 | unsigned long temp_end_pfn = early_node_map[i].end_pfn; | ||
4506 | early_node_map[i].end_pfn = start_pfn; | 4460 | early_node_map[i].end_pfn = start_pfn; |
4507 | if (temp_end_pfn > end_pfn) | 4461 | if (this_end_pfn > end_pfn) |
4508 | add_active_range(nid, end_pfn, temp_end_pfn); | 4462 | add_active_range(nid, end_pfn, this_end_pfn); |
4509 | continue; | 4463 | continue; |
4510 | } | 4464 | } |
4511 | if (early_node_map[i].start_pfn >= start_pfn && | 4465 | if (this_start_pfn >= start_pfn && this_end_pfn > end_pfn && |
4512 | early_node_map[i].end_pfn > end_pfn && | 4466 | this_start_pfn < end_pfn) { |
4513 | early_node_map[i].start_pfn < end_pfn) { | ||
4514 | early_node_map[i].start_pfn = end_pfn; | 4467 | early_node_map[i].start_pfn = end_pfn; |
4515 | continue; | 4468 | continue; |
4516 | } | 4469 | } |
@@ -4593,15 +4546,11 @@ void __init sort_node_map(void) | |||
4593 | unsigned long __init node_map_pfn_alignment(void) | 4546 | unsigned long __init node_map_pfn_alignment(void) |
4594 | { | 4547 | { |
4595 | unsigned long accl_mask = 0, last_end = 0; | 4548 | unsigned long accl_mask = 0, last_end = 0; |
4549 | unsigned long start, end, mask; | ||
4596 | int last_nid = -1; | 4550 | int last_nid = -1; |
4597 | int i; | 4551 | int i, nid; |
4598 | |||
4599 | for_each_active_range_index_in_nid(i, MAX_NUMNODES) { | ||
4600 | int nid = early_node_map[i].nid; | ||
4601 | unsigned long start = early_node_map[i].start_pfn; | ||
4602 | unsigned long end = early_node_map[i].end_pfn; | ||
4603 | unsigned long mask; | ||
4604 | 4552 | ||
4553 | for_each_mem_pfn_range(i, MAX_NUMNODES, &start, &end, &nid) { | ||
4605 | if (!start || last_nid < 0 || last_nid == nid) { | 4554 | if (!start || last_nid < 0 || last_nid == nid) { |
4606 | last_nid = nid; | 4555 | last_nid = nid; |
4607 | last_end = end; | 4556 | last_end = end; |
@@ -4628,12 +4577,12 @@ unsigned long __init node_map_pfn_alignment(void) | |||
4628 | /* Find the lowest pfn for a node */ | 4577 | /* Find the lowest pfn for a node */ |
4629 | static unsigned long __init find_min_pfn_for_node(int nid) | 4578 | static unsigned long __init find_min_pfn_for_node(int nid) |
4630 | { | 4579 | { |
4631 | int i; | ||
4632 | unsigned long min_pfn = ULONG_MAX; | 4580 | unsigned long min_pfn = ULONG_MAX; |
4581 | unsigned long start_pfn; | ||
4582 | int i; | ||
4633 | 4583 | ||
4634 | /* Assuming a sorted map, the first range found has the starting pfn */ | 4584 | for_each_mem_pfn_range(i, nid, &start_pfn, NULL, NULL) |
4635 | for_each_active_range_index_in_nid(i, nid) | 4585 | min_pfn = min(min_pfn, start_pfn); |
4636 | min_pfn = min(min_pfn, early_node_map[i].start_pfn); | ||
4637 | 4586 | ||
4638 | if (min_pfn == ULONG_MAX) { | 4587 | if (min_pfn == ULONG_MAX) { |
4639 | printk(KERN_WARNING | 4588 | printk(KERN_WARNING |
@@ -4662,15 +4611,16 @@ unsigned long __init find_min_pfn_with_active_regions(void) | |||
4662 | */ | 4611 | */ |
4663 | static unsigned long __init early_calculate_totalpages(void) | 4612 | static unsigned long __init early_calculate_totalpages(void) |
4664 | { | 4613 | { |
4665 | int i; | ||
4666 | unsigned long totalpages = 0; | 4614 | unsigned long totalpages = 0; |
4615 | unsigned long start_pfn, end_pfn; | ||
4616 | int i, nid; | ||
4617 | |||
4618 | for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { | ||
4619 | unsigned long pages = end_pfn - start_pfn; | ||
4667 | 4620 | ||
4668 | for (i = 0; i < nr_nodemap_entries; i++) { | ||
4669 | unsigned long pages = early_node_map[i].end_pfn - | ||
4670 | early_node_map[i].start_pfn; | ||
4671 | totalpages += pages; | 4621 | totalpages += pages; |
4672 | if (pages) | 4622 | if (pages) |
4673 | node_set_state(early_node_map[i].nid, N_HIGH_MEMORY); | 4623 | node_set_state(nid, N_HIGH_MEMORY); |
4674 | } | 4624 | } |
4675 | return totalpages; | 4625 | return totalpages; |
4676 | } | 4626 | } |
@@ -4725,6 +4675,8 @@ restart: | |||
4725 | /* Spread kernelcore memory as evenly as possible throughout nodes */ | 4675 | /* Spread kernelcore memory as evenly as possible throughout nodes */ |
4726 | kernelcore_node = required_kernelcore / usable_nodes; | 4676 | kernelcore_node = required_kernelcore / usable_nodes; |
4727 | for_each_node_state(nid, N_HIGH_MEMORY) { | 4677 | for_each_node_state(nid, N_HIGH_MEMORY) { |
4678 | unsigned long start_pfn, end_pfn; | ||
4679 | |||
4728 | /* | 4680 | /* |
4729 | * Recalculate kernelcore_node if the division per node | 4681 | * Recalculate kernelcore_node if the division per node |
4730 | * now exceeds what is necessary to satisfy the requested | 4682 | * now exceeds what is necessary to satisfy the requested |
@@ -4741,13 +4693,10 @@ restart: | |||
4741 | kernelcore_remaining = kernelcore_node; | 4693 | kernelcore_remaining = kernelcore_node; |
4742 | 4694 | ||
4743 | /* Go through each range of PFNs within this node */ | 4695 | /* Go through each range of PFNs within this node */ |
4744 | for_each_active_range_index_in_nid(i, nid) { | 4696 | for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { |
4745 | unsigned long start_pfn, end_pfn; | ||
4746 | unsigned long size_pages; | 4697 | unsigned long size_pages; |
4747 | 4698 | ||
4748 | start_pfn = max(early_node_map[i].start_pfn, | 4699 | start_pfn = max(start_pfn, zone_movable_pfn[nid]); |
4749 | zone_movable_pfn[nid]); | ||
4750 | end_pfn = early_node_map[i].end_pfn; | ||
4751 | if (start_pfn >= end_pfn) | 4700 | if (start_pfn >= end_pfn) |
4752 | continue; | 4701 | continue; |
4753 | 4702 | ||
@@ -4849,8 +4798,8 @@ static void check_for_regular_memory(pg_data_t *pgdat) | |||
4849 | */ | 4798 | */ |
4850 | void __init free_area_init_nodes(unsigned long *max_zone_pfn) | 4799 | void __init free_area_init_nodes(unsigned long *max_zone_pfn) |
4851 | { | 4800 | { |
4852 | unsigned long nid; | 4801 | unsigned long start_pfn, end_pfn; |
4853 | int i; | 4802 | int i, nid; |
4854 | 4803 | ||
4855 | /* Sort early_node_map as initialisation assumes it is sorted */ | 4804 | /* Sort early_node_map as initialisation assumes it is sorted */ |
4856 | sort_node_map(); | 4805 | sort_node_map(); |
@@ -4900,11 +4849,9 @@ void __init free_area_init_nodes(unsigned long *max_zone_pfn) | |||
4900 | } | 4849 | } |
4901 | 4850 | ||
4902 | /* Print out the early_node_map[] */ | 4851 | /* Print out the early_node_map[] */ |
4903 | printk("early_node_map[%d] active PFN ranges\n", nr_nodemap_entries); | 4852 | printk("Early memory PFN ranges\n"); |
4904 | for (i = 0; i < nr_nodemap_entries; i++) | 4853 | for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) |
4905 | printk(" %3d: %0#10lx -> %0#10lx\n", early_node_map[i].nid, | 4854 | printk(" %3d: %0#10lx -> %0#10lx\n", nid, start_pfn, end_pfn); |
4906 | early_node_map[i].start_pfn, | ||
4907 | early_node_map[i].end_pfn); | ||
4908 | 4855 | ||
4909 | /* Initialise every node */ | 4856 | /* Initialise every node */ |
4910 | mminit_verify_pageflags_layout(); | 4857 | mminit_verify_pageflags_layout(); |