aboutsummaryrefslogtreecommitdiffstats
path: root/arch/arm/common/dmabounce.c
diff options
context:
space:
mode:
authorRussell King <rmk@dyn-67.arm.linux.org.uk>2007-02-06 12:39:31 -0500
committerRussell King <rmk+kernel@arm.linux.org.uk>2007-02-08 09:49:44 -0500
commit7ae5a761d2ffc4cf7d3248e09f4d3da234434f30 (patch)
tree8b936e37b9750c27e8f2c86318591912dc7e1cc3 /arch/arm/common/dmabounce.c
parent953233dc9958ba2b29753d0f24e37a33a076a5f6 (diff)
[ARM] Convert DMA cache handling to take const void * args
The DMA cache handling functions take virtual addresses, but in the form of unsigned long arguments. This leads to a little confusion about what exactly they take. So, convert them to take const void * instead. Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
Diffstat (limited to 'arch/arm/common/dmabounce.c')
-rw-r--r--arch/arm/common/dmabounce.c7
1 files changed, 3 insertions, 4 deletions
diff --git a/arch/arm/common/dmabounce.c b/arch/arm/common/dmabounce.c
index b4748e3171c6..2362c498f52e 100644
--- a/arch/arm/common/dmabounce.c
+++ b/arch/arm/common/dmabounce.c
@@ -321,12 +321,12 @@ unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size,
321 DO_STATS ( device_info->bounce_count++ ); 321 DO_STATS ( device_info->bounce_count++ );
322 322
323 if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL) { 323 if (dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL) {
324 unsigned long ptr; 324 void *ptr = buf->ptr;
325 325
326 dev_dbg(dev, 326 dev_dbg(dev,
327 "%s: copy back safe %p to unsafe %p size %d\n", 327 "%s: copy back safe %p to unsafe %p size %d\n",
328 __func__, buf->safe, buf->ptr, size); 328 __func__, buf->safe, ptr, size);
329 memcpy(buf->ptr, buf->safe, size); 329 memcpy(ptr, buf->safe, size);
330 330
331 /* 331 /*
332 * DMA buffers must have the same cache properties 332 * DMA buffers must have the same cache properties
@@ -336,7 +336,6 @@ unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size,
336 * bidirectional case because we know the cache 336 * bidirectional case because we know the cache
337 * lines will be coherent with the data written. 337 * lines will be coherent with the data written.
338 */ 338 */
339 ptr = (unsigned long)buf->ptr;
340 dmac_clean_range(ptr, ptr + size); 339 dmac_clean_range(ptr, ptr + size);
341 outer_clean_range(__pa(ptr), __pa(ptr) + size); 340 outer_clean_range(__pa(ptr), __pa(ptr) + size);
342 } 341 }