diff options
Diffstat (limited to 'mm')
-rw-r--r-- | mm/huge_memory.c | 4 | ||||
-rw-r--r-- | mm/mmap.c | 10 | ||||
-rw-r--r-- | mm/rmap.c | 8 |
3 files changed, 11 insertions, 11 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 83326ad66d9b..90eef404ec2e 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c | |||
@@ -1139,7 +1139,7 @@ static int __split_huge_page_splitting(struct page *page, | |||
1139 | * We can't temporarily set the pmd to null in order | 1139 | * We can't temporarily set the pmd to null in order |
1140 | * to split it, the pmd must remain marked huge at all | 1140 | * to split it, the pmd must remain marked huge at all |
1141 | * times or the VM won't take the pmd_trans_huge paths | 1141 | * times or the VM won't take the pmd_trans_huge paths |
1142 | * and it won't wait on the anon_vma->root->lock to | 1142 | * and it won't wait on the anon_vma->root->mutex to |
1143 | * serialize against split_huge_page*. | 1143 | * serialize against split_huge_page*. |
1144 | */ | 1144 | */ |
1145 | pmdp_splitting_flush_notify(vma, address, pmd); | 1145 | pmdp_splitting_flush_notify(vma, address, pmd); |
@@ -1333,7 +1333,7 @@ static int __split_huge_page_map(struct page *page, | |||
1333 | return ret; | 1333 | return ret; |
1334 | } | 1334 | } |
1335 | 1335 | ||
1336 | /* must be called with anon_vma->root->lock hold */ | 1336 | /* must be called with anon_vma->root->mutex hold */ |
1337 | static void __split_huge_page(struct page *page, | 1337 | static void __split_huge_page(struct page *page, |
1338 | struct anon_vma *anon_vma) | 1338 | struct anon_vma *anon_vma) |
1339 | { | 1339 | { |
@@ -2502,15 +2502,15 @@ static void vm_lock_anon_vma(struct mm_struct *mm, struct anon_vma *anon_vma) | |||
2502 | * The LSB of head.next can't change from under us | 2502 | * The LSB of head.next can't change from under us |
2503 | * because we hold the mm_all_locks_mutex. | 2503 | * because we hold the mm_all_locks_mutex. |
2504 | */ | 2504 | */ |
2505 | spin_lock_nest_lock(&anon_vma->root->lock, &mm->mmap_sem); | 2505 | mutex_lock_nest_lock(&anon_vma->root->mutex, &mm->mmap_sem); |
2506 | /* | 2506 | /* |
2507 | * We can safely modify head.next after taking the | 2507 | * We can safely modify head.next after taking the |
2508 | * anon_vma->root->lock. If some other vma in this mm shares | 2508 | * anon_vma->root->mutex. If some other vma in this mm shares |
2509 | * the same anon_vma we won't take it again. | 2509 | * the same anon_vma we won't take it again. |
2510 | * | 2510 | * |
2511 | * No need of atomic instructions here, head.next | 2511 | * No need of atomic instructions here, head.next |
2512 | * can't change from under us thanks to the | 2512 | * can't change from under us thanks to the |
2513 | * anon_vma->root->lock. | 2513 | * anon_vma->root->mutex. |
2514 | */ | 2514 | */ |
2515 | if (__test_and_set_bit(0, (unsigned long *) | 2515 | if (__test_and_set_bit(0, (unsigned long *) |
2516 | &anon_vma->root->head.next)) | 2516 | &anon_vma->root->head.next)) |
@@ -2559,7 +2559,7 @@ static void vm_lock_mapping(struct mm_struct *mm, struct address_space *mapping) | |||
2559 | * vma in this mm is backed by the same anon_vma or address_space. | 2559 | * vma in this mm is backed by the same anon_vma or address_space. |
2560 | * | 2560 | * |
2561 | * We can take all the locks in random order because the VM code | 2561 | * We can take all the locks in random order because the VM code |
2562 | * taking i_mmap_mutex or anon_vma->lock outside the mmap_sem never | 2562 | * taking i_mmap_mutex or anon_vma->mutex outside the mmap_sem never |
2563 | * takes more than one of them in a row. Secondly we're protected | 2563 | * takes more than one of them in a row. Secondly we're protected |
2564 | * against a concurrent mm_take_all_locks() by the mm_all_locks_mutex. | 2564 | * against a concurrent mm_take_all_locks() by the mm_all_locks_mutex. |
2565 | * | 2565 | * |
@@ -2615,7 +2615,7 @@ static void vm_unlock_anon_vma(struct anon_vma *anon_vma) | |||
2615 | * | 2615 | * |
2616 | * No need of atomic instructions here, head.next | 2616 | * No need of atomic instructions here, head.next |
2617 | * can't change from under us until we release the | 2617 | * can't change from under us until we release the |
2618 | * anon_vma->root->lock. | 2618 | * anon_vma->root->mutex. |
2619 | */ | 2619 | */ |
2620 | if (!__test_and_clear_bit(0, (unsigned long *) | 2620 | if (!__test_and_clear_bit(0, (unsigned long *) |
2621 | &anon_vma->root->head.next)) | 2621 | &anon_vma->root->head.next)) |
@@ -25,7 +25,7 @@ | |||
25 | * mm->mmap_sem | 25 | * mm->mmap_sem |
26 | * page->flags PG_locked (lock_page) | 26 | * page->flags PG_locked (lock_page) |
27 | * mapping->i_mmap_mutex | 27 | * mapping->i_mmap_mutex |
28 | * anon_vma->lock | 28 | * anon_vma->mutex |
29 | * mm->page_table_lock or pte_lock | 29 | * mm->page_table_lock or pte_lock |
30 | * zone->lru_lock (in mark_page_accessed, isolate_lru_page) | 30 | * zone->lru_lock (in mark_page_accessed, isolate_lru_page) |
31 | * swap_lock (in swap_duplicate, swap_info_get) | 31 | * swap_lock (in swap_duplicate, swap_info_get) |
@@ -40,7 +40,7 @@ | |||
40 | * | 40 | * |
41 | * (code doesn't rely on that order so it could be switched around) | 41 | * (code doesn't rely on that order so it could be switched around) |
42 | * ->tasklist_lock | 42 | * ->tasklist_lock |
43 | * anon_vma->lock (memory_failure, collect_procs_anon) | 43 | * anon_vma->mutex (memory_failure, collect_procs_anon) |
44 | * pte map lock | 44 | * pte map lock |
45 | */ | 45 | */ |
46 | 46 | ||
@@ -307,7 +307,7 @@ static void anon_vma_ctor(void *data) | |||
307 | { | 307 | { |
308 | struct anon_vma *anon_vma = data; | 308 | struct anon_vma *anon_vma = data; |
309 | 309 | ||
310 | spin_lock_init(&anon_vma->lock); | 310 | mutex_init(&anon_vma->mutex); |
311 | atomic_set(&anon_vma->refcount, 0); | 311 | atomic_set(&anon_vma->refcount, 0); |
312 | INIT_LIST_HEAD(&anon_vma->head); | 312 | INIT_LIST_HEAD(&anon_vma->head); |
313 | } | 313 | } |
@@ -1143,7 +1143,7 @@ out_mlock: | |||
1143 | /* | 1143 | /* |
1144 | * We need mmap_sem locking, Otherwise VM_LOCKED check makes | 1144 | * We need mmap_sem locking, Otherwise VM_LOCKED check makes |
1145 | * unstable result and race. Plus, We can't wait here because | 1145 | * unstable result and race. Plus, We can't wait here because |
1146 | * we now hold anon_vma->lock or mapping->i_mmap_mutex. | 1146 | * we now hold anon_vma->mutex or mapping->i_mmap_mutex. |
1147 | * if trylock failed, the page remain in evictable lru and later | 1147 | * if trylock failed, the page remain in evictable lru and later |
1148 | * vmscan could retry to move the page to unevictable lru if the | 1148 | * vmscan could retry to move the page to unevictable lru if the |
1149 | * page is actually mlocked. | 1149 | * page is actually mlocked. |