aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorKonrad Rzeszutek Wilk <konrad.wilk@oracle.com>2011-02-22 13:24:32 -0500
committerKonrad Rzeszutek Wilk <konrad.wilk@oracle.com>2011-02-22 13:26:23 -0500
commit5a893fc28f0393adb7c885a871b8c59e623fd528 (patch)
tree8d119f02b7d0ba7e60f0f554fa5000c2cce2fb87
parente0138c26cdeee8c033256ccd9e07d66db3c998be (diff)
ttm: Include the 'struct dev' when using the DMA API.
This makes the accounting when using 'debug_dma_dump_mappings()' and CONFIG_DMA_API_DEBUG=y be assigned to the correct device instead of 'fallback'. No functional change - just cosmetic. Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
-rw-r--r--drivers/gpu/drm/nouveau/nouveau_mem.c1
-rw-r--r--drivers/gpu/drm/radeon/radeon_ttm.c1
-rw-r--r--drivers/gpu/drm/ttm/ttm_page_alloc.c11
-rw-r--r--drivers/gpu/drm/ttm/ttm_tt.c4
-rw-r--r--drivers/gpu/drm/vmwgfx/vmwgfx_drv.c2
-rw-r--r--include/drm/ttm/ttm_bo_driver.h1
-rw-r--r--include/drm/ttm/ttm_page_alloc.h8
7 files changed, 18 insertions, 10 deletions
diff --git a/drivers/gpu/drm/nouveau/nouveau_mem.c b/drivers/gpu/drm/nouveau/nouveau_mem.c
index a163c7c612e7..931b22142ed2 100644
--- a/drivers/gpu/drm/nouveau/nouveau_mem.c
+++ b/drivers/gpu/drm/nouveau/nouveau_mem.c
@@ -559,6 +559,7 @@ nouveau_mem_vram_init(struct drm_device *dev)
559 if (ret) 559 if (ret)
560 return ret; 560 return ret;
561 561
562 dev_priv->ttm.bdev.dev = dev->dev;
562 ret = ttm_bo_device_init(&dev_priv->ttm.bdev, 563 ret = ttm_bo_device_init(&dev_priv->ttm.bdev,
563 dev_priv->ttm.bo_global_ref.ref.object, 564 dev_priv->ttm.bo_global_ref.ref.object,
564 &nouveau_bo_driver, DRM_FILE_PAGE_OFFSET, 565 &nouveau_bo_driver, DRM_FILE_PAGE_OFFSET,
diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c
index ca045058e498..cfe223f22394 100644
--- a/drivers/gpu/drm/radeon/radeon_ttm.c
+++ b/drivers/gpu/drm/radeon/radeon_ttm.c
@@ -513,6 +513,7 @@ int radeon_ttm_init(struct radeon_device *rdev)
513 if (r) { 513 if (r) {
514 return r; 514 return r;
515 } 515 }
516 rdev->mman.bdev.dev = rdev->dev;
516 /* No others user of address space so set it to 0 */ 517 /* No others user of address space so set it to 0 */
517 r = ttm_bo_device_init(&rdev->mman.bdev, 518 r = ttm_bo_device_init(&rdev->mman.bdev,
518 rdev->mman.bo_global_ref.ref.object, 519 rdev->mman.bo_global_ref.ref.object,
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c
index 737a2a2e46a5..35849dbf3ab5 100644
--- a/drivers/gpu/drm/ttm/ttm_page_alloc.c
+++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c
@@ -664,7 +664,7 @@ out:
664 */ 664 */
665int ttm_get_pages(struct list_head *pages, int flags, 665int ttm_get_pages(struct list_head *pages, int flags,
666 enum ttm_caching_state cstate, unsigned count, 666 enum ttm_caching_state cstate, unsigned count,
667 dma_addr_t *dma_address) 667 dma_addr_t *dma_address, struct device *dev)
668{ 668{
669 struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); 669 struct ttm_page_pool *pool = ttm_get_pool(flags, cstate);
670 struct page *p = NULL; 670 struct page *p = NULL;
@@ -685,7 +685,7 @@ int ttm_get_pages(struct list_head *pages, int flags,
685 for (r = 0; r < count; ++r) { 685 for (r = 0; r < count; ++r) {
686 if ((flags & TTM_PAGE_FLAG_DMA32) && dma_address) { 686 if ((flags & TTM_PAGE_FLAG_DMA32) && dma_address) {
687 void *addr; 687 void *addr;
688 addr = dma_alloc_coherent(NULL, PAGE_SIZE, 688 addr = dma_alloc_coherent(dev, PAGE_SIZE,
689 &dma_address[r], 689 &dma_address[r],
690 gfp_flags); 690 gfp_flags);
691 if (addr == NULL) 691 if (addr == NULL)
@@ -730,7 +730,7 @@ int ttm_get_pages(struct list_head *pages, int flags,
730 printk(KERN_ERR TTM_PFX 730 printk(KERN_ERR TTM_PFX
731 "Failed to allocate extra pages " 731 "Failed to allocate extra pages "
732 "for large request."); 732 "for large request.");
733 ttm_put_pages(pages, 0, flags, cstate, NULL); 733 ttm_put_pages(pages, 0, flags, cstate, NULL, NULL);
734 return r; 734 return r;
735 } 735 }
736 } 736 }
@@ -741,7 +741,8 @@ int ttm_get_pages(struct list_head *pages, int flags,
741 741
742/* Put all pages in pages list to correct pool to wait for reuse */ 742/* Put all pages in pages list to correct pool to wait for reuse */
743void ttm_put_pages(struct list_head *pages, unsigned page_count, int flags, 743void ttm_put_pages(struct list_head *pages, unsigned page_count, int flags,
744 enum ttm_caching_state cstate, dma_addr_t *dma_address) 744 enum ttm_caching_state cstate, dma_addr_t *dma_address,
745 struct device *dev)
745{ 746{
746 unsigned long irq_flags; 747 unsigned long irq_flags;
747 struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); 748 struct ttm_page_pool *pool = ttm_get_pool(flags, cstate);
@@ -757,7 +758,7 @@ void ttm_put_pages(struct list_head *pages, unsigned page_count, int flags,
757 void *addr = page_address(p); 758 void *addr = page_address(p);
758 WARN_ON(!addr || !dma_address[r]); 759 WARN_ON(!addr || !dma_address[r]);
759 if (addr) 760 if (addr)
760 dma_free_coherent(NULL, PAGE_SIZE, 761 dma_free_coherent(dev, PAGE_SIZE,
761 addr, 762 addr,
762 dma_address[r]); 763 dma_address[r]);
763 dma_address[r] = 0; 764 dma_address[r] = 0;
diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c
index 86d5b1745a45..0f8fc9ff0c53 100644
--- a/drivers/gpu/drm/ttm/ttm_tt.c
+++ b/drivers/gpu/drm/ttm/ttm_tt.c
@@ -110,7 +110,7 @@ static struct page *__ttm_tt_get_page(struct ttm_tt *ttm, int index)
110 INIT_LIST_HEAD(&h); 110 INIT_LIST_HEAD(&h);
111 111
112 ret = ttm_get_pages(&h, ttm->page_flags, ttm->caching_state, 1, 112 ret = ttm_get_pages(&h, ttm->page_flags, ttm->caching_state, 1,
113 &ttm->dma_address[index]); 113 &ttm->dma_address[index], ttm->be->bdev->dev);
114 114
115 if (ret != 0) 115 if (ret != 0)
116 return NULL; 116 return NULL;
@@ -304,7 +304,7 @@ static void ttm_tt_free_alloced_pages(struct ttm_tt *ttm)
304 } 304 }
305 } 305 }
306 ttm_put_pages(&h, count, ttm->page_flags, ttm->caching_state, 306 ttm_put_pages(&h, count, ttm->page_flags, ttm->caching_state,
307 ttm->dma_address); 307 ttm->dma_address, ttm->be->bdev->dev);
308 ttm->state = tt_unpopulated; 308 ttm->state = tt_unpopulated;
309 ttm->first_himem_page = ttm->num_pages; 309 ttm->first_himem_page = ttm->num_pages;
310 ttm->last_lomem_page = -1; 310 ttm->last_lomem_page = -1;
diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
index 10ca97ee0206..4a8c7893e8ff 100644
--- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
+++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
@@ -322,7 +322,7 @@ static int vmw_driver_load(struct drm_device *dev, unsigned long chipset)
322 ttm_lock_set_kill(&dev_priv->fbdev_master.lock, false, SIGTERM); 322 ttm_lock_set_kill(&dev_priv->fbdev_master.lock, false, SIGTERM);
323 dev_priv->active_master = &dev_priv->fbdev_master; 323 dev_priv->active_master = &dev_priv->fbdev_master;
324 324
325 325 dev_priv->bdev.dev = dev->dev;
326 ret = ttm_bo_device_init(&dev_priv->bdev, 326 ret = ttm_bo_device_init(&dev_priv->bdev,
327 dev_priv->bo_global_ref.ref.object, 327 dev_priv->bo_global_ref.ref.object,
328 &vmw_bo_driver, VMWGFX_FILE_PAGE_OFFSET, 328 &vmw_bo_driver, VMWGFX_FILE_PAGE_OFFSET,
diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
index ebcd3dd7203b..4d97014e8c8d 100644
--- a/include/drm/ttm/ttm_bo_driver.h
+++ b/include/drm/ttm/ttm_bo_driver.h
@@ -533,6 +533,7 @@ struct ttm_bo_device {
533 struct list_head device_list; 533 struct list_head device_list;
534 struct ttm_bo_global *glob; 534 struct ttm_bo_global *glob;
535 struct ttm_bo_driver *driver; 535 struct ttm_bo_driver *driver;
536 struct device *dev;
536 rwlock_t vm_lock; 537 rwlock_t vm_lock;
537 struct ttm_mem_type_manager man[TTM_NUM_MEM_TYPES]; 538 struct ttm_mem_type_manager man[TTM_NUM_MEM_TYPES];
538 /* 539 /*
diff --git a/include/drm/ttm/ttm_page_alloc.h b/include/drm/ttm/ttm_page_alloc.h
index 8062890f725e..ccb6b7a240e2 100644
--- a/include/drm/ttm/ttm_page_alloc.h
+++ b/include/drm/ttm/ttm_page_alloc.h
@@ -37,12 +37,14 @@
37 * @cstate: ttm caching state for the page. 37 * @cstate: ttm caching state for the page.
38 * @count: number of pages to allocate. 38 * @count: number of pages to allocate.
39 * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set). 39 * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set).
40 * @dev: struct device for appropiate DMA accounting.
40 */ 41 */
41int ttm_get_pages(struct list_head *pages, 42int ttm_get_pages(struct list_head *pages,
42 int flags, 43 int flags,
43 enum ttm_caching_state cstate, 44 enum ttm_caching_state cstate,
44 unsigned count, 45 unsigned count,
45 dma_addr_t *dma_address); 46 dma_addr_t *dma_address,
47 struct device *dev);
46/** 48/**
47 * Put linked list of pages to pool. 49 * Put linked list of pages to pool.
48 * 50 *
@@ -52,12 +54,14 @@ int ttm_get_pages(struct list_head *pages,
52 * @flags: ttm flags for page allocation. 54 * @flags: ttm flags for page allocation.
53 * @cstate: ttm caching state. 55 * @cstate: ttm caching state.
54 * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set). 56 * @dma_address: The DMA (bus) address of pages (if TTM_PAGE_FLAG_DMA32 set).
57 * @dev: struct device for appropiate DMA accounting.
55 */ 58 */
56void ttm_put_pages(struct list_head *pages, 59void ttm_put_pages(struct list_head *pages,
57 unsigned page_count, 60 unsigned page_count,
58 int flags, 61 int flags,
59 enum ttm_caching_state cstate, 62 enum ttm_caching_state cstate,
60 dma_addr_t *dma_address); 63 dma_addr_t *dma_address,
64 struct device *dev);
61/** 65/**
62 * Initialize pool allocator. 66 * Initialize pool allocator.
63 */ 67 */