diff options
author | Jeremy Fitzhardinge <jeremy@goop.org> | 2007-05-06 17:48:54 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-05-07 15:12:51 -0400 |
commit | aee16b3cee2746880e40945a9b5bff4f309cfbc4 (patch) | |
tree | 887faaebf5562dc1fac5e090140da04dc7e2a174 | |
parent | eb3a1e1145ca8f12372c7c96aa0702d86a9002a9 (diff) |
Add apply_to_page_range() which applies a function to a pte range
Add a new mm function apply_to_page_range() which applies a given function to
every pte in a given virtual address range in a given mm structure. This is a
generic alternative to cut-and-pasting the Linux idiomatic pagetable walking
code in every place that a sequence of PTEs must be accessed.
Although this interface is intended to be useful in a wide range of
situations, it is currently used specifically by several Xen subsystems, for
example: to ensure that pagetables have been allocated for a virtual address
range, and to construct batched special pagetable update requests to map I/O
memory (in ioremap()).
[akpm@linux-foundation.org: fix warning, unpleasantly]
Signed-off-by: Ian Pratt <ian.pratt@xensource.com>
Signed-off-by: Christian Limpach <Christian.Limpach@cl.cam.ac.uk>
Signed-off-by: Chris Wright <chrisw@sous-sol.org>
Signed-off-by: Jeremy Fitzhardinge <jeremy@xensource.com>
Cc: Christoph Lameter <clameter@sgi.com>
Cc: Matt Mackall <mpm@waste.org>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | include/linux/mm.h | 5 | ||||
-rw-r--r-- | mm/memory.c | 94 |
2 files changed, 99 insertions, 0 deletions
diff --git a/include/linux/mm.h b/include/linux/mm.h index 60e0e4a592d2..7bf0bd882fc3 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h | |||
@@ -1130,6 +1130,11 @@ struct page *follow_page(struct vm_area_struct *, unsigned long address, | |||
1130 | #define FOLL_GET 0x04 /* do get_page on page */ | 1130 | #define FOLL_GET 0x04 /* do get_page on page */ |
1131 | #define FOLL_ANON 0x08 /* give ZERO_PAGE if no pgtable */ | 1131 | #define FOLL_ANON 0x08 /* give ZERO_PAGE if no pgtable */ |
1132 | 1132 | ||
1133 | typedef int (*pte_fn_t)(pte_t *pte, struct page *pmd_page, unsigned long addr, | ||
1134 | void *data); | ||
1135 | extern int apply_to_page_range(struct mm_struct *mm, unsigned long address, | ||
1136 | unsigned long size, pte_fn_t fn, void *data); | ||
1137 | |||
1133 | #ifdef CONFIG_PROC_FS | 1138 | #ifdef CONFIG_PROC_FS |
1134 | void vm_stat_account(struct mm_struct *, unsigned long, struct file *, long); | 1139 | void vm_stat_account(struct mm_struct *, unsigned long, struct file *, long); |
1135 | #else | 1140 | #else |
diff --git a/mm/memory.c b/mm/memory.c index e7066e71dfa3..044feb7e7134 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
@@ -1448,6 +1448,100 @@ int remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, | |||
1448 | } | 1448 | } |
1449 | EXPORT_SYMBOL(remap_pfn_range); | 1449 | EXPORT_SYMBOL(remap_pfn_range); |
1450 | 1450 | ||
1451 | static int apply_to_pte_range(struct mm_struct *mm, pmd_t *pmd, | ||
1452 | unsigned long addr, unsigned long end, | ||
1453 | pte_fn_t fn, void *data) | ||
1454 | { | ||
1455 | pte_t *pte; | ||
1456 | int err; | ||
1457 | struct page *pmd_page; | ||
1458 | spinlock_t *ptl = ptl; /* Suppress gcc warning */ | ||
1459 | |||
1460 | pte = (mm == &init_mm) ? | ||
1461 | pte_alloc_kernel(pmd, addr) : | ||
1462 | pte_alloc_map_lock(mm, pmd, addr, &ptl); | ||
1463 | if (!pte) | ||
1464 | return -ENOMEM; | ||
1465 | |||
1466 | BUG_ON(pmd_huge(*pmd)); | ||
1467 | |||
1468 | pmd_page = pmd_page(*pmd); | ||
1469 | |||
1470 | do { | ||
1471 | err = fn(pte, pmd_page, addr, data); | ||
1472 | if (err) | ||
1473 | break; | ||
1474 | } while (pte++, addr += PAGE_SIZE, addr != end); | ||
1475 | |||
1476 | if (mm != &init_mm) | ||
1477 | pte_unmap_unlock(pte-1, ptl); | ||
1478 | return err; | ||
1479 | } | ||
1480 | |||
1481 | static int apply_to_pmd_range(struct mm_struct *mm, pud_t *pud, | ||
1482 | unsigned long addr, unsigned long end, | ||
1483 | pte_fn_t fn, void *data) | ||
1484 | { | ||
1485 | pmd_t *pmd; | ||
1486 | unsigned long next; | ||
1487 | int err; | ||
1488 | |||
1489 | pmd = pmd_alloc(mm, pud, addr); | ||
1490 | if (!pmd) | ||
1491 | return -ENOMEM; | ||
1492 | do { | ||
1493 | next = pmd_addr_end(addr, end); | ||
1494 | err = apply_to_pte_range(mm, pmd, addr, next, fn, data); | ||
1495 | if (err) | ||
1496 | break; | ||
1497 | } while (pmd++, addr = next, addr != end); | ||
1498 | return err; | ||
1499 | } | ||
1500 | |||
1501 | static int apply_to_pud_range(struct mm_struct *mm, pgd_t *pgd, | ||
1502 | unsigned long addr, unsigned long end, | ||
1503 | pte_fn_t fn, void *data) | ||
1504 | { | ||
1505 | pud_t *pud; | ||
1506 | unsigned long next; | ||
1507 | int err; | ||
1508 | |||
1509 | pud = pud_alloc(mm, pgd, addr); | ||
1510 | if (!pud) | ||
1511 | return -ENOMEM; | ||
1512 | do { | ||
1513 | next = pud_addr_end(addr, end); | ||
1514 | err = apply_to_pmd_range(mm, pud, addr, next, fn, data); | ||
1515 | if (err) | ||
1516 | break; | ||
1517 | } while (pud++, addr = next, addr != end); | ||
1518 | return err; | ||
1519 | } | ||
1520 | |||
1521 | /* | ||
1522 | * Scan a region of virtual memory, filling in page tables as necessary | ||
1523 | * and calling a provided function on each leaf page table. | ||
1524 | */ | ||
1525 | int apply_to_page_range(struct mm_struct *mm, unsigned long addr, | ||
1526 | unsigned long size, pte_fn_t fn, void *data) | ||
1527 | { | ||
1528 | pgd_t *pgd; | ||
1529 | unsigned long next; | ||
1530 | unsigned long end = addr + size; | ||
1531 | int err; | ||
1532 | |||
1533 | BUG_ON(addr >= end); | ||
1534 | pgd = pgd_offset(mm, addr); | ||
1535 | do { | ||
1536 | next = pgd_addr_end(addr, end); | ||
1537 | err = apply_to_pud_range(mm, pgd, addr, next, fn, data); | ||
1538 | if (err) | ||
1539 | break; | ||
1540 | } while (pgd++, addr = next, addr != end); | ||
1541 | return err; | ||
1542 | } | ||
1543 | EXPORT_SYMBOL_GPL(apply_to_page_range); | ||
1544 | |||
1451 | /* | 1545 | /* |
1452 | * handle_pte_fault chooses page fault handler according to an entry | 1546 | * handle_pte_fault chooses page fault handler according to an entry |
1453 | * which was read non-atomically. Before making any commitment, on | 1547 | * which was read non-atomically. Before making any commitment, on |