diff options
author | Tejun Heo <tj@kernel.org> | 2009-03-06 00:33:59 -0500 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2009-03-06 00:33:59 -0500 |
commit | edcb463997ed7b2ffa3bac76e3e75957318f2e01 (patch) | |
tree | d1eefb78736404993b0de4a049392f9af578e29e /include/linux/percpu.h | |
parent | 3e24aa58907c62bc79d1094e941a374568f62522 (diff) |
percpu, module: implement reserved allocation and use it for module percpu variables
Impact: add reserved allocation functionality and use it for module
percpu variables
This patch implements reserved allocation from the first chunk. When
setting up the first chunk, arch can ask to set aside certain number
of bytes right after the core static area which is available only
through a separate reserved allocator. This will be used primarily
for module static percpu variables on architectures with limited
relocation range to ensure that the module perpcu symbols are inside
the relocatable range.
If reserved area is requested, the first chunk becomes reserved and
isn't available for regular allocation. If the first chunk also
includes piggy-back dynamic allocation area, a separate chunk mapping
the same region is created to serve dynamic allocation. The first one
is called static first chunk and the second dynamic first chunk.
Although they share the page map, their different area map
initializations guarantee they serve disjoint areas according to their
purposes.
If arch doesn't setup reserved area, reserved allocation is handled
like any other allocation.
Signed-off-by: Tejun Heo <tj@kernel.org>
Diffstat (limited to 'include/linux/percpu.h')
-rw-r--r-- | include/linux/percpu.h | 10 |
1 files changed, 6 insertions, 4 deletions
diff --git a/include/linux/percpu.h b/include/linux/percpu.h index a96fc53bbd62..8ff15153ae20 100644 --- a/include/linux/percpu.h +++ b/include/linux/percpu.h | |||
@@ -117,10 +117,10 @@ typedef struct page * (*pcpu_get_page_fn_t)(unsigned int cpu, int pageno); | |||
117 | typedef void (*pcpu_populate_pte_fn_t)(unsigned long addr); | 117 | typedef void (*pcpu_populate_pte_fn_t)(unsigned long addr); |
118 | 118 | ||
119 | extern size_t __init pcpu_setup_first_chunk(pcpu_get_page_fn_t get_page_fn, | 119 | extern size_t __init pcpu_setup_first_chunk(pcpu_get_page_fn_t get_page_fn, |
120 | size_t static_size, | 120 | size_t static_size, size_t reserved_size, |
121 | ssize_t unit_size, ssize_t dyn_size, | 121 | ssize_t unit_size, ssize_t dyn_size, |
122 | void *base_addr, | 122 | void *base_addr, |
123 | pcpu_populate_pte_fn_t populate_pte_fn); | 123 | pcpu_populate_pte_fn_t populate_pte_fn); |
124 | 124 | ||
125 | /* | 125 | /* |
126 | * Use this to get to a cpu's version of the per-cpu object | 126 | * Use this to get to a cpu's version of the per-cpu object |
@@ -129,6 +129,8 @@ extern size_t __init pcpu_setup_first_chunk(pcpu_get_page_fn_t get_page_fn, | |||
129 | */ | 129 | */ |
130 | #define per_cpu_ptr(ptr, cpu) SHIFT_PERCPU_PTR((ptr), per_cpu_offset((cpu))) | 130 | #define per_cpu_ptr(ptr, cpu) SHIFT_PERCPU_PTR((ptr), per_cpu_offset((cpu))) |
131 | 131 | ||
132 | extern void *__alloc_reserved_percpu(size_t size, size_t align); | ||
133 | |||
132 | #else /* CONFIG_HAVE_DYNAMIC_PER_CPU_AREA */ | 134 | #else /* CONFIG_HAVE_DYNAMIC_PER_CPU_AREA */ |
133 | 135 | ||
134 | struct percpu_data { | 136 | struct percpu_data { |