diff options
Diffstat (limited to 'mm/mmap.c')
-rw-r--r-- | mm/mmap.c | 8 |
1 files changed, 4 insertions, 4 deletions
@@ -438,7 +438,7 @@ static void vma_gap_update(struct vm_area_struct *vma) | |||
438 | { | 438 | { |
439 | /* | 439 | /* |
440 | * As it turns out, RB_DECLARE_CALLBACKS() already created a callback | 440 | * As it turns out, RB_DECLARE_CALLBACKS() already created a callback |
441 | * function that does exacltly what we want. | 441 | * function that does exactly what we want. |
442 | */ | 442 | */ |
443 | vma_gap_callbacks_propagate(&vma->vm_rb, NULL); | 443 | vma_gap_callbacks_propagate(&vma->vm_rb, NULL); |
444 | } | 444 | } |
@@ -1012,7 +1012,7 @@ static inline int is_mergeable_vma(struct vm_area_struct *vma, | |||
1012 | * VM_SOFTDIRTY should not prevent from VMA merging, if we | 1012 | * VM_SOFTDIRTY should not prevent from VMA merging, if we |
1013 | * match the flags but dirty bit -- the caller should mark | 1013 | * match the flags but dirty bit -- the caller should mark |
1014 | * merged VMA as dirty. If dirty bit won't be excluded from | 1014 | * merged VMA as dirty. If dirty bit won't be excluded from |
1015 | * comparison, we increase pressue on the memory system forcing | 1015 | * comparison, we increase pressure on the memory system forcing |
1016 | * the kernel to generate new VMAs when old one could be | 1016 | * the kernel to generate new VMAs when old one could be |
1017 | * extended instead. | 1017 | * extended instead. |
1018 | */ | 1018 | */ |
@@ -1115,7 +1115,7 @@ can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, | |||
1115 | * PPPP NNNN PPPPPPPPPPPP PPPPPPPPNNNN PPPPNNNNNNNN | 1115 | * PPPP NNNN PPPPPPPPPPPP PPPPPPPPNNNN PPPPNNNNNNNN |
1116 | * might become case 1 below case 2 below case 3 below | 1116 | * might become case 1 below case 2 below case 3 below |
1117 | * | 1117 | * |
1118 | * It is important for case 8 that the the vma NNNN overlapping the | 1118 | * It is important for case 8 that the vma NNNN overlapping the |
1119 | * region AAAA is never going to extended over XXXX. Instead XXXX must | 1119 | * region AAAA is never going to extended over XXXX. Instead XXXX must |
1120 | * be extended in region AAAA and NNNN must be removed. This way in | 1120 | * be extended in region AAAA and NNNN must be removed. This way in |
1121 | * all cases where vma_merge succeeds, the moment vma_adjust drops the | 1121 | * all cases where vma_merge succeeds, the moment vma_adjust drops the |
@@ -1645,7 +1645,7 @@ SYSCALL_DEFINE1(old_mmap, struct mmap_arg_struct __user *, arg) | |||
1645 | #endif /* __ARCH_WANT_SYS_OLD_MMAP */ | 1645 | #endif /* __ARCH_WANT_SYS_OLD_MMAP */ |
1646 | 1646 | ||
1647 | /* | 1647 | /* |
1648 | * Some shared mappigns will want the pages marked read-only | 1648 | * Some shared mappings will want the pages marked read-only |
1649 | * to track write events. If so, we'll downgrade vm_page_prot | 1649 | * to track write events. If so, we'll downgrade vm_page_prot |
1650 | * to the private version (using protection_map[] without the | 1650 | * to the private version (using protection_map[] without the |
1651 | * VM_SHARED bit). | 1651 | * VM_SHARED bit). |