aboutsummaryrefslogtreecommitdiffstats
path: root/include/linux/memblock.h
diff options
context:
space:
mode:
authorRobin Holt <holt@sgi.com>2015-06-30 17:56:41 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2015-06-30 22:44:55 -0400
commit8e7a7f8619f1f93736d9bb7e31caf4721bdc739d (patch)
treeb79b8131a17456108e029a7675770fa0a1f2952d /include/linux/memblock.h
parent6aaf0da8728c55ff627619f933ed161cc89057c6 (diff)
memblock: introduce a for_each_reserved_mem_region iterator
Struct page initialisation had been identified as one of the reasons why large machines take a long time to boot. Patches were posted a long time ago to defer initialisation until they were first used. This was rejected on the grounds it should not be necessary to hurt the fast paths. This series reuses much of the work from that time but defers the initialisation of memory to kswapd so that one thread per node initialises memory local to that node. After applying the series and setting the appropriate Kconfig variable I see this in the boot log on a 64G machine [ 7.383764] kswapd 0 initialised deferred memory in 188ms [ 7.404253] kswapd 1 initialised deferred memory in 208ms [ 7.411044] kswapd 3 initialised deferred memory in 216ms [ 7.411551] kswapd 2 initialised deferred memory in 216ms On a 1TB machine, I see [ 8.406511] kswapd 3 initialised deferred memory in 1116ms [ 8.428518] kswapd 1 initialised deferred memory in 1140ms [ 8.435977] kswapd 0 initialised deferred memory in 1148ms [ 8.437416] kswapd 2 initialised deferred memory in 1148ms Once booted the machine appears to work as normal. Boot times were measured from the time shutdown was called until ssh was available again. In the 64G case, the boot time savings are negligible. On the 1TB machine, the savings were 16 seconds. Nate Zimmer said: : On an older 8 TB box with lots and lots of cpus the boot time, as : measure from grub to login prompt, the boot time improved from 1484 : seconds to exactly 1000 seconds. Waiman Long said: : I ran a bootup timing test on a 12-TB 16-socket IvyBridge-EX system. From : grub menu to ssh login, the bootup time was 453s before the patch and 265s : after the patch - a saving of 188s (42%). Daniel Blueman said: : On a 7TB, 1728-core NumaConnect system with 108 NUMA nodes, we're seeing : stock 4.0 boot in 7136s. This drops to 2159s, or a 70% reduction with : this patchset. Non-temporal PMD init (https://lkml.org/lkml/2015/4/23/350) : drops this to 1045s. This patch (of 13): As part of initializing struct page's in 2MiB chunks, we noticed that at the end of free_all_bootmem(), there was nothing which had forced the reserved/allocated 4KiB pages to be initialized. This helper function will be used for that expansion. Signed-off-by: Robin Holt <holt@sgi.com> Signed-off-by: Nate Zimmer <nzimmer@sgi.com> Signed-off-by: Mel Gorman <mgorman@suse.de> Tested-by: Nate Zimmer <nzimmer@sgi.com> Tested-by: Waiman Long <waiman.long@hp.com> Tested-by: Daniel J Blueman <daniel@numascale.com> Acked-by: Pekka Enberg <penberg@kernel.org> Cc: Robin Holt <robinmholt@gmail.com> Cc: Dave Hansen <dave.hansen@intel.com> Cc: Waiman Long <waiman.long@hp.com> Cc: Scott Norton <scott.norton@hp.com> Cc: "Luck, Tony" <tony.luck@intel.com> Cc: Ingo Molnar <mingo@elte.hu> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'include/linux/memblock.h')
-rw-r--r--include/linux/memblock.h18
1 files changed, 18 insertions, 0 deletions
diff --git a/include/linux/memblock.h b/include/linux/memblock.h
index 0215ffd63069..cc4b01972060 100644
--- a/include/linux/memblock.h
+++ b/include/linux/memblock.h
@@ -101,6 +101,9 @@ void __next_mem_range_rev(u64 *idx, int nid, ulong flags,
101 struct memblock_type *type_b, phys_addr_t *out_start, 101 struct memblock_type *type_b, phys_addr_t *out_start,
102 phys_addr_t *out_end, int *out_nid); 102 phys_addr_t *out_end, int *out_nid);
103 103
104void __next_reserved_mem_region(u64 *idx, phys_addr_t *out_start,
105 phys_addr_t *out_end);
106
104/** 107/**
105 * for_each_mem_range - iterate through memblock areas from type_a and not 108 * for_each_mem_range - iterate through memblock areas from type_a and not
106 * included in type_b. Or just type_a if type_b is NULL. 109 * included in type_b. Or just type_a if type_b is NULL.
@@ -142,6 +145,21 @@ void __next_mem_range_rev(u64 *idx, int nid, ulong flags,
142 __next_mem_range_rev(&i, nid, flags, type_a, type_b, \ 145 __next_mem_range_rev(&i, nid, flags, type_a, type_b, \
143 p_start, p_end, p_nid)) 146 p_start, p_end, p_nid))
144 147
148/**
149 * for_each_reserved_mem_region - iterate over all reserved memblock areas
150 * @i: u64 used as loop variable
151 * @p_start: ptr to phys_addr_t for start address of the range, can be %NULL
152 * @p_end: ptr to phys_addr_t for end address of the range, can be %NULL
153 *
154 * Walks over reserved areas of memblock. Available as soon as memblock
155 * is initialized.
156 */
157#define for_each_reserved_mem_region(i, p_start, p_end) \
158 for (i = 0UL, \
159 __next_reserved_mem_region(&i, p_start, p_end); \
160 i != (u64)ULLONG_MAX; \
161 __next_reserved_mem_region(&i, p_start, p_end))
162
145#ifdef CONFIG_MOVABLE_NODE 163#ifdef CONFIG_MOVABLE_NODE
146static inline bool memblock_is_hotpluggable(struct memblock_region *m) 164static inline bool memblock_is_hotpluggable(struct memblock_region *m)
147{ 165{