aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorMagnus Damm <magnus.damm@gmail.com>2008-01-24 04:35:10 -0500
committerPaul Mundt <lethal@linux-sh.org>2008-01-27 23:19:04 -0500
commitf93e97eaead5c50af35d73cca7301ebbfdff116c (patch)
tree108a0b968123bb00a9453189fe93ad467d265cbf /arch
parent4862ec073975e28f432f164405e60fa6f5c9d071 (diff)
sh: declared coherent memory support V2
This patch adds declared coherent memory support to the sh architecture. All functions are based on the x86 implementation. Header files are adjusted to use the new functions instead of the former consistent_alloc() code. This version includes the few changes what were included in the fix patch together with modifications based on feedback from Paul. Signed-off-by: Magnus Damm <damm@igel.co.jp> Signed-off-by: Paul Mundt <lethal@linux-sh.org>
Diffstat (limited to 'arch')
-rw-r--r--arch/sh/mm/consistent.c174
1 files changed, 128 insertions, 46 deletions
diff --git a/arch/sh/mm/consistent.c b/arch/sh/mm/consistent.c
index 65ad30031ad7..7b2131c9eeda 100644
--- a/arch/sh/mm/consistent.c
+++ b/arch/sh/mm/consistent.c
@@ -3,6 +3,8 @@
3 * 3 *
4 * Copyright (C) 2004 - 2007 Paul Mundt 4 * Copyright (C) 2004 - 2007 Paul Mundt
5 * 5 *
6 * Declared coherent memory functions based on arch/x86/kernel/pci-dma_32.c
7 *
6 * This file is subject to the terms and conditions of the GNU General Public 8 * This file is subject to the terms and conditions of the GNU General Public
7 * License. See the file "COPYING" in the main directory of this archive 9 * License. See the file "COPYING" in the main directory of this archive
8 * for more details. 10 * for more details.
@@ -13,66 +15,146 @@
13#include <asm/addrspace.h> 15#include <asm/addrspace.h>
14#include <asm/io.h> 16#include <asm/io.h>
15 17
16void *consistent_alloc(gfp_t gfp, size_t size, dma_addr_t *handle) 18struct dma_coherent_mem {
17{ 19 void *virt_base;
18 struct page *page, *end, *free; 20 u32 device_base;
19 void *ret, *vp; 21 int size;
20 int order; 22 int flags;
21 23 unsigned long *bitmap;
22 size = PAGE_ALIGN(size); 24};
23 order = get_order(size);
24 25
25 page = alloc_pages(gfp, order); 26void *dma_alloc_coherent(struct device *dev, size_t size,
26 if (!page) 27 dma_addr_t *dma_handle, gfp_t gfp)
27 return NULL; 28{
28 split_page(page, order); 29 void *ret;
30 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
31 int order = get_order(size);
32
33 if (mem) {
34 int page = bitmap_find_free_region(mem->bitmap, mem->size,
35 order);
36 if (page >= 0) {
37 *dma_handle = mem->device_base + (page << PAGE_SHIFT);
38 ret = mem->virt_base + (page << PAGE_SHIFT);
39 memset(ret, 0, size);
40 return ret;
41 }
42 if (mem->flags & DMA_MEMORY_EXCLUSIVE)
43 return NULL;
44 }
29 45
30 ret = page_address(page); 46 ret = (void *)__get_free_pages(gfp, order);
31 *handle = virt_to_phys(ret);
32 47
33 vp = ioremap_nocache(*handle, size); 48 if (ret != NULL) {
34 if (!vp) { 49 memset(ret, 0, size);
35 free_pages((unsigned long)ret, order); 50 /*
36 return NULL; 51 * Pages from the page allocator may have data present in
52 * cache. So flush the cache before using uncached memory.
53 */
54 dma_cache_sync(NULL, ret, size, DMA_BIDIRECTIONAL);
55 *dma_handle = virt_to_phys(ret);
37 } 56 }
57 return ret;
58}
59EXPORT_SYMBOL(dma_alloc_coherent);
38 60
39 memset(vp, 0, size); 61void dma_free_coherent(struct device *dev, size_t size,
40 62 void *vaddr, dma_addr_t dma_handle)
41 /* 63{
42 * We must flush the cache before we pass it on to the device 64 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL;
43 */ 65 int order = get_order(size);
44 dma_cache_sync(NULL, ret, size, DMA_BIDIRECTIONAL);
45 66
46 page = virt_to_page(ret); 67 if (mem && vaddr >= mem->virt_base && vaddr < (mem->virt_base + (mem->size << PAGE_SHIFT))) {
47 free = page + (size >> PAGE_SHIFT); 68 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
48 end = page + (1 << order);
49 69
50 while (++page < end) { 70 bitmap_release_region(mem->bitmap, page, order);
51 /* Free any unused pages */ 71 } else {
52 if (page >= free) { 72 WARN_ON(irqs_disabled()); /* for portability */
53 __free_page(page); 73 BUG_ON(mem && mem->flags & DMA_MEMORY_EXCLUSIVE);
54 } 74 free_pages((unsigned long)vaddr, order);
55 } 75 }
56
57 return vp;
58} 76}
59EXPORT_SYMBOL(consistent_alloc); 77EXPORT_SYMBOL(dma_free_coherent);
60 78
61void consistent_free(void *vaddr, size_t size, dma_addr_t dma_handle) 79int dma_declare_coherent_memory(struct device *dev, dma_addr_t bus_addr,
80 dma_addr_t device_addr, size_t size, int flags)
62{ 81{
63 struct page *page; 82 void __iomem *mem_base = NULL;
64 unsigned long addr; 83 int pages = size >> PAGE_SHIFT;
65 84 int bitmap_size = BITS_TO_LONGS(pages) * sizeof(long);
66 addr = (unsigned long)phys_to_virt((unsigned long)dma_handle); 85
67 page = virt_to_page(addr); 86 if ((flags & (DMA_MEMORY_MAP | DMA_MEMORY_IO)) == 0)
87 goto out;
88 if (!size)
89 goto out;
90 if (dev->dma_mem)
91 goto out;
92
93 /* FIXME: this routine just ignores DMA_MEMORY_INCLUDES_CHILDREN */
94
95 mem_base = ioremap_nocache(bus_addr, size);
96 if (!mem_base)
97 goto out;
98
99 dev->dma_mem = kmalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL);
100 if (!dev->dma_mem)
101 goto out;
102 dev->dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
103 if (!dev->dma_mem->bitmap)
104 goto free1_out;
105
106 dev->dma_mem->virt_base = mem_base;
107 dev->dma_mem->device_base = device_addr;
108 dev->dma_mem->size = pages;
109 dev->dma_mem->flags = flags;
110
111 if (flags & DMA_MEMORY_MAP)
112 return DMA_MEMORY_MAP;
113
114 return DMA_MEMORY_IO;
115
116 free1_out:
117 kfree(dev->dma_mem);
118 out:
119 if (mem_base)
120 iounmap(mem_base);
121 return 0;
122}
123EXPORT_SYMBOL(dma_declare_coherent_memory);
68 124
69 free_pages(addr, get_order(size)); 125void dma_release_declared_memory(struct device *dev)
126{
127 struct dma_coherent_mem *mem = dev->dma_mem;
128
129 if (!mem)
130 return;
131 dev->dma_mem = NULL;
132 iounmap(mem->virt_base);
133 kfree(mem->bitmap);
134 kfree(mem);
135}
136EXPORT_SYMBOL(dma_release_declared_memory);
70 137
71 iounmap(vaddr); 138void *dma_mark_declared_memory_occupied(struct device *dev,
139 dma_addr_t device_addr, size_t size)
140{
141 struct dma_coherent_mem *mem = dev->dma_mem;
142 int pages = (size + (device_addr & ~PAGE_MASK) + PAGE_SIZE - 1) >> PAGE_SHIFT;
143 int pos, err;
144
145 if (!mem)
146 return ERR_PTR(-EINVAL);
147
148 pos = (device_addr - mem->device_base) >> PAGE_SHIFT;
149 err = bitmap_allocate_region(mem->bitmap, pos, get_order(pages));
150 if (err != 0)
151 return ERR_PTR(err);
152 return mem->virt_base + (pos << PAGE_SHIFT);
72} 153}
73EXPORT_SYMBOL(consistent_free); 154EXPORT_SYMBOL(dma_mark_declared_memory_occupied);
74 155
75void consistent_sync(void *vaddr, size_t size, int direction) 156void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
157 enum dma_data_direction direction)
76{ 158{
77#ifdef CONFIG_CPU_SH5 159#ifdef CONFIG_CPU_SH5
78 void *p1addr = vaddr; 160 void *p1addr = vaddr;
@@ -94,4 +176,4 @@ void consistent_sync(void *vaddr, size_t size, int direction)
94 BUG(); 176 BUG();
95 } 177 }
96} 178}
97EXPORT_SYMBOL(consistent_sync); 179EXPORT_SYMBOL(dma_cache_sync);