diff options
author | Catalin Marinas <catalin.marinas@arm.com> | 2014-10-10 10:37:28 -0400 |
---|---|---|
committer | Catalin Marinas <catalin.marinas@arm.com> | 2014-10-20 12:47:02 -0400 |
commit | 2a0b5c0d19298cad54573682321b5fe015fa5a0d (patch) | |
tree | d1b64ac1f1ce289bb1184fa4563930a2be9519e8 /arch/arm64/mm | |
parent | 04f905a9569ca2b6964a35563f135fabbb2470bc (diff) |
arm64: Align less than PAGE_SIZE pgds naturally
When the pgd size is smaller than PAGE_SIZE, pgd_alloc() uses kzalloc()
to save space. However, this is not always naturally aligned as required
by the architecture. This patch creates a kmem_cache for pgd allocations
with the correct alignment.
The current kernel configurations with 4K pages + 39-bit VA and 64K
pages + 42-bit VA use a full page for the pgd and are not affected. The
patch is required for 48-bit VA with 64K pages where the pgd is 512
bytes.
Reported-by: Christoffer Dall <christoffer.dall@linaro.org>
Reviewed-by: Christoffer Dall <christoffer.dall@linaro.org>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
Diffstat (limited to 'arch/arm64/mm')
-rw-r--r-- | arch/arm64/mm/pgd.c | 18 |
1 files changed, 16 insertions, 2 deletions
diff --git a/arch/arm64/mm/pgd.c b/arch/arm64/mm/pgd.c index 62c6101df260..6682b361d3ac 100644 --- a/arch/arm64/mm/pgd.c +++ b/arch/arm64/mm/pgd.c | |||
@@ -30,12 +30,14 @@ | |||
30 | 30 | ||
31 | #define PGD_SIZE (PTRS_PER_PGD * sizeof(pgd_t)) | 31 | #define PGD_SIZE (PTRS_PER_PGD * sizeof(pgd_t)) |
32 | 32 | ||
33 | static struct kmem_cache *pgd_cache; | ||
34 | |||
33 | pgd_t *pgd_alloc(struct mm_struct *mm) | 35 | pgd_t *pgd_alloc(struct mm_struct *mm) |
34 | { | 36 | { |
35 | if (PGD_SIZE == PAGE_SIZE) | 37 | if (PGD_SIZE == PAGE_SIZE) |
36 | return (pgd_t *)get_zeroed_page(GFP_KERNEL); | 38 | return (pgd_t *)get_zeroed_page(GFP_KERNEL); |
37 | else | 39 | else |
38 | return kzalloc(PGD_SIZE, GFP_KERNEL); | 40 | return kmem_cache_zalloc(pgd_cache, GFP_KERNEL); |
39 | } | 41 | } |
40 | 42 | ||
41 | void pgd_free(struct mm_struct *mm, pgd_t *pgd) | 43 | void pgd_free(struct mm_struct *mm, pgd_t *pgd) |
@@ -43,5 +45,17 @@ void pgd_free(struct mm_struct *mm, pgd_t *pgd) | |||
43 | if (PGD_SIZE == PAGE_SIZE) | 45 | if (PGD_SIZE == PAGE_SIZE) |
44 | free_page((unsigned long)pgd); | 46 | free_page((unsigned long)pgd); |
45 | else | 47 | else |
46 | kfree(pgd); | 48 | kmem_cache_free(pgd_cache, pgd); |
49 | } | ||
50 | |||
51 | static int __init pgd_cache_init(void) | ||
52 | { | ||
53 | /* | ||
54 | * Naturally aligned pgds required by the architecture. | ||
55 | */ | ||
56 | if (PGD_SIZE != PAGE_SIZE) | ||
57 | pgd_cache = kmem_cache_create("pgd_cache", PGD_SIZE, PGD_SIZE, | ||
58 | SLAB_PANIC, NULL); | ||
59 | return 0; | ||
47 | } | 60 | } |
61 | core_initcall(pgd_cache_init); | ||