aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/filemap.c4
-rw-r--r--mm/memory.c4
-rw-r--r--mm/mempool.c6
-rw-r--r--mm/page-writeback.c5
-rw-r--r--mm/slab.c2
-rw-r--r--mm/vmalloc.c2
6 files changed, 9 insertions, 14 deletions
diff --git a/mm/filemap.c b/mm/filemap.c
index f30ef28405d3..00414849a867 100644
--- a/mm/filemap.c
+++ b/mm/filemap.c
@@ -327,7 +327,7 @@ EXPORT_SYMBOL(sync_page_range);
327 * @pos: beginning offset in pages to write 327 * @pos: beginning offset in pages to write
328 * @count: number of bytes to write 328 * @count: number of bytes to write
329 * 329 *
330 * Note: Holding i_mutex across sync_page_range_nolock is not a good idea 330 * Note: Holding i_mutex across sync_page_range_nolock() is not a good idea
331 * as it forces O_SYNC writers to different parts of the same file 331 * as it forces O_SYNC writers to different parts of the same file
332 * to be serialised right until io completion. 332 * to be serialised right until io completion.
333 */ 333 */
@@ -784,7 +784,7 @@ unsigned find_get_pages_tag(struct address_space *mapping, pgoff_t *index,
784 * @mapping: target address_space 784 * @mapping: target address_space
785 * @index: the page index 785 * @index: the page index
786 * 786 *
787 * Same as grab_cache_page, but do not wait if the page is unavailable. 787 * Same as grab_cache_page(), but do not wait if the page is unavailable.
788 * This is intended for speculative data generators, where the data can 788 * This is intended for speculative data generators, where the data can
789 * be regenerated if the page couldn't be grabbed. This routine should 789 * be regenerated if the page couldn't be grabbed. This routine should
790 * be safe to call while holding the lock for another page. 790 * be safe to call while holding the lock for another page.
diff --git a/mm/memory.c b/mm/memory.c
index 0e6a402d86be..072c1135ad37 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -1775,9 +1775,7 @@ restart:
1775} 1775}
1776 1776
1777/** 1777/**
1778 * unmap_mapping_range - unmap the portion of all mmaps 1778 * unmap_mapping_range - unmap the portion of all mmaps in the specified address_space corresponding to the specified page range in the underlying file.
1779 * in the specified address_space corresponding to the specified
1780 * page range in the underlying file.
1781 * @mapping: the address space containing mmaps to be unmapped. 1779 * @mapping: the address space containing mmaps to be unmapped.
1782 * @holebegin: byte in first page to unmap, relative to the start of 1780 * @holebegin: byte in first page to unmap, relative to the start of
1783 * the underlying file. This will be rounded down to a PAGE_SIZE 1781 * the underlying file. This will be rounded down to a PAGE_SIZE
diff --git a/mm/mempool.c b/mm/mempool.c
index ccd8cb8cd41f..cc1ca86dfc24 100644
--- a/mm/mempool.c
+++ b/mm/mempool.c
@@ -46,9 +46,9 @@ static void free_pool(mempool_t *pool)
46 * @pool_data: optional private data available to the user-defined functions. 46 * @pool_data: optional private data available to the user-defined functions.
47 * 47 *
48 * this function creates and allocates a guaranteed size, preallocated 48 * this function creates and allocates a guaranteed size, preallocated
49 * memory pool. The pool can be used from the mempool_alloc and mempool_free 49 * memory pool. The pool can be used from the mempool_alloc() and mempool_free()
50 * functions. This function might sleep. Both the alloc_fn() and the free_fn() 50 * functions. This function might sleep. Both the alloc_fn() and the free_fn()
51 * functions might sleep - as long as the mempool_alloc function is not called 51 * functions might sleep - as long as the mempool_alloc() function is not called
52 * from IRQ contexts. 52 * from IRQ contexts.
53 */ 53 */
54mempool_t *mempool_create(int min_nr, mempool_alloc_t *alloc_fn, 54mempool_t *mempool_create(int min_nr, mempool_alloc_t *alloc_fn,
@@ -195,7 +195,7 @@ EXPORT_SYMBOL(mempool_destroy);
195 * mempool_create(). 195 * mempool_create().
196 * @gfp_mask: the usual allocation bitmask. 196 * @gfp_mask: the usual allocation bitmask.
197 * 197 *
198 * this function only sleeps if the alloc_fn function sleeps or 198 * this function only sleeps if the alloc_fn() function sleeps or
199 * returns NULL. Note that due to preallocation, this function 199 * returns NULL. Note that due to preallocation, this function
200 * *never* fails when called from process contexts. (it might 200 * *never* fails when called from process contexts. (it might
201 * fail if called from an IRQ context.) 201 * fail if called from an IRQ context.)
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 438833cbbca4..fd96a555e500 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -549,9 +549,7 @@ void __init page_writeback_init(void)
549} 549}
550 550
551/** 551/**
552 * generic_writepages - walk the list of dirty pages of the given 552 * generic_writepages - walk the list of dirty pages of the given address space and writepage() all of them.
553 * address space and writepage() all of them.
554 *
555 * @mapping: address space structure to write 553 * @mapping: address space structure to write
556 * @wbc: subtract the number of written pages from *@wbc->nr_to_write 554 * @wbc: subtract the number of written pages from *@wbc->nr_to_write
557 * 555 *
@@ -698,7 +696,6 @@ int do_writepages(struct address_space *mapping, struct writeback_control *wbc)
698 696
699/** 697/**
700 * write_one_page - write out a single page and optionally wait on I/O 698 * write_one_page - write out a single page and optionally wait on I/O
701 *
702 * @page: the page to write 699 * @page: the page to write
703 * @wait: if true, wait on writeout 700 * @wait: if true, wait on writeout
704 * 701 *
diff --git a/mm/slab.c b/mm/slab.c
index 196df70eb8cb..70784b848b69 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -2520,7 +2520,7 @@ EXPORT_SYMBOL(kmem_cache_shrink);
2520 * kmem_cache_destroy - delete a cache 2520 * kmem_cache_destroy - delete a cache
2521 * @cachep: the cache to destroy 2521 * @cachep: the cache to destroy
2522 * 2522 *
2523 * Remove a struct kmem_cache object from the slab cache. 2523 * Remove a &struct kmem_cache object from the slab cache.
2524 * 2524 *
2525 * It is expected this function will be called by a module when it is 2525 * It is expected this function will be called by a module when it is
2526 * unloaded. This will remove the cache completely, and avoid a duplicate 2526 * unloaded. This will remove the cache completely, and avoid a duplicate
diff --git a/mm/vmalloc.c b/mm/vmalloc.c
index 86897ee792d6..9eef486da909 100644
--- a/mm/vmalloc.c
+++ b/mm/vmalloc.c
@@ -699,7 +699,7 @@ finished:
699 * that it is big enough to cover the vma. Will return failure if 699 * that it is big enough to cover the vma. Will return failure if
700 * that criteria isn't met. 700 * that criteria isn't met.
701 * 701 *
702 * Similar to remap_pfn_range (see mm/memory.c) 702 * Similar to remap_pfn_range() (see mm/memory.c)
703 */ 703 */
704int remap_vmalloc_range(struct vm_area_struct *vma, void *addr, 704int remap_vmalloc_range(struct vm_area_struct *vma, void *addr,
705 unsigned long pgoff) 705 unsigned long pgoff)