aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--include/linux/ksm.h20
-rw-r--r--mm/ksm.c14
-rw-r--r--mm/mmap.c1
3 files changed, 13 insertions, 22 deletions
diff --git a/include/linux/ksm.h b/include/linux/ksm.h
index 0e26de6adb51..a485c14ecd5d 100644
--- a/include/linux/ksm.h
+++ b/include/linux/ksm.h
@@ -12,8 +12,6 @@
12#include <linux/sched.h> 12#include <linux/sched.h>
13#include <linux/vmstat.h> 13#include <linux/vmstat.h>
14 14
15struct mmu_gather;
16
17#ifdef CONFIG_KSM 15#ifdef CONFIG_KSM
18int ksm_madvise(struct vm_area_struct *vma, unsigned long start, 16int ksm_madvise(struct vm_area_struct *vma, unsigned long start,
19 unsigned long end, int advice, unsigned long *vm_flags); 17 unsigned long end, int advice, unsigned long *vm_flags);
@@ -27,19 +25,6 @@ static inline int ksm_fork(struct mm_struct *mm, struct mm_struct *oldmm)
27 return 0; 25 return 0;
28} 26}
29 27
30/*
31 * For KSM to handle OOM without deadlock when it's breaking COW in a
32 * likely victim of the OOM killer, exit_mmap() has to serialize with
33 * ksm_exit() after freeing mm's pages but before freeing its page tables.
34 * That leaves a window in which KSM might refault pages which have just
35 * been finally unmapped: guard against that with ksm_test_exit(), and
36 * use it after getting mmap_sem in ksm.c, to check if mm is exiting.
37 */
38static inline bool ksm_test_exit(struct mm_struct *mm)
39{
40 return atomic_read(&mm->mm_users) == 0;
41}
42
43static inline void ksm_exit(struct mm_struct *mm) 28static inline void ksm_exit(struct mm_struct *mm)
44{ 29{
45 if (test_bit(MMF_VM_MERGEABLE, &mm->flags)) 30 if (test_bit(MMF_VM_MERGEABLE, &mm->flags))
@@ -79,11 +64,6 @@ static inline int ksm_fork(struct mm_struct *mm, struct mm_struct *oldmm)
79 return 0; 64 return 0;
80} 65}
81 66
82static inline bool ksm_test_exit(struct mm_struct *mm)
83{
84 return 0;
85}
86
87static inline void ksm_exit(struct mm_struct *mm) 67static inline void ksm_exit(struct mm_struct *mm)
88{ 68{
89} 69}
diff --git a/mm/ksm.c b/mm/ksm.c
index 3bd54ce9eb38..e11e7a5ac84f 100644
--- a/mm/ksm.c
+++ b/mm/ksm.c
@@ -32,7 +32,6 @@
32#include <linux/mmu_notifier.h> 32#include <linux/mmu_notifier.h>
33#include <linux/ksm.h> 33#include <linux/ksm.h>
34 34
35#include <asm/tlb.h>
36#include <asm/tlbflush.h> 35#include <asm/tlbflush.h>
37 36
38/* 37/*
@@ -285,6 +284,19 @@ static inline int in_stable_tree(struct rmap_item *rmap_item)
285} 284}
286 285
287/* 286/*
287 * ksmd, and unmerge_and_remove_all_rmap_items(), must not touch an mm's
288 * page tables after it has passed through ksm_exit() - which, if necessary,
289 * takes mmap_sem briefly to serialize against them. ksm_exit() does not set
290 * a special flag: they can just back out as soon as mm_users goes to zero.
291 * ksm_test_exit() is used throughout to make this test for exit: in some
292 * places for correctness, in some places just to avoid unnecessary work.
293 */
294static inline bool ksm_test_exit(struct mm_struct *mm)
295{
296 return atomic_read(&mm->mm_users) == 0;
297}
298
299/*
288 * We use break_ksm to break COW on a ksm page: it's a stripped down 300 * We use break_ksm to break COW on a ksm page: it's a stripped down
289 * 301 *
290 * if (get_user_pages(current, mm, addr, 1, 1, 1, &page, NULL) == 1) 302 * if (get_user_pages(current, mm, addr, 1, 1, 1, &page, NULL) == 1)
diff --git a/mm/mmap.c b/mm/mmap.c
index 22dff49d579e..6eed98c00543 100644
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -27,7 +27,6 @@
27#include <linux/mount.h> 27#include <linux/mount.h>
28#include <linux/mempolicy.h> 28#include <linux/mempolicy.h>
29#include <linux/rmap.h> 29#include <linux/rmap.h>
30#include <linux/ksm.h>
31#include <linux/mmu_notifier.h> 30#include <linux/mmu_notifier.h>
32#include <linux/perf_event.h> 31#include <linux/perf_event.h>
33 32