diff options
author | Jerome Glisse <jglisse@redhat.com> | 2011-11-09 17:15:26 -0500 |
---|---|---|
committer | Dave Airlie <airlied@redhat.com> | 2011-12-06 05:40:02 -0500 |
commit | 8e7e70522d760c4ccd4cd370ebfa0ba69e006c6e (patch) | |
tree | a2b0f931e513f3aeba174b974bd5e869685fe288 /drivers/gpu/drm/ttm | |
parent | 3230cfc34fca9d17c1628cf0e4ac25199592a69a (diff) |
drm/ttm: isolate dma data from ttm_tt V4
Move dma data to a superset ttm_dma_tt structure which herit
from ttm_tt. This allow driver that don't use dma functionalities
to not have to waste memory for it.
V2 Rebase on top of no memory account changes (where/when is my
delorean when i need it ?)
V3 Make sure page list is initialized empty
V4 typo/syntax fixes
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Reviewed-by: Thomas Hellstrom <thellstrom@vmware.com>
Diffstat (limited to 'drivers/gpu/drm/ttm')
-rw-r--r-- | drivers/gpu/drm/ttm/ttm_page_alloc.c | 114 | ||||
-rw-r--r-- | drivers/gpu/drm/ttm/ttm_page_alloc_dma.c | 35 | ||||
-rw-r--r-- | drivers/gpu/drm/ttm/ttm_tt.c | 60 |
3 files changed, 124 insertions, 85 deletions
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c index 8d6267e434ab..499debda791e 100644 --- a/drivers/gpu/drm/ttm/ttm_page_alloc.c +++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c | |||
@@ -662,13 +662,61 @@ out: | |||
662 | return count; | 662 | return count; |
663 | } | 663 | } |
664 | 664 | ||
665 | /* Put all pages in pages list to correct pool to wait for reuse */ | ||
666 | static void ttm_put_pages(struct page **pages, unsigned npages, int flags, | ||
667 | enum ttm_caching_state cstate) | ||
668 | { | ||
669 | unsigned long irq_flags; | ||
670 | struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); | ||
671 | unsigned i; | ||
672 | |||
673 | if (pool == NULL) { | ||
674 | /* No pool for this memory type so free the pages */ | ||
675 | for (i = 0; i < npages; i++) { | ||
676 | if (pages[i]) { | ||
677 | if (page_count(pages[i]) != 1) | ||
678 | printk(KERN_ERR TTM_PFX | ||
679 | "Erroneous page count. " | ||
680 | "Leaking pages.\n"); | ||
681 | __free_page(pages[i]); | ||
682 | pages[i] = NULL; | ||
683 | } | ||
684 | } | ||
685 | return; | ||
686 | } | ||
687 | |||
688 | spin_lock_irqsave(&pool->lock, irq_flags); | ||
689 | for (i = 0; i < npages; i++) { | ||
690 | if (pages[i]) { | ||
691 | if (page_count(pages[i]) != 1) | ||
692 | printk(KERN_ERR TTM_PFX | ||
693 | "Erroneous page count. " | ||
694 | "Leaking pages.\n"); | ||
695 | list_add_tail(&pages[i]->lru, &pool->list); | ||
696 | pages[i] = NULL; | ||
697 | pool->npages++; | ||
698 | } | ||
699 | } | ||
700 | /* Check that we don't go over the pool limit */ | ||
701 | npages = 0; | ||
702 | if (pool->npages > _manager->options.max_size) { | ||
703 | npages = pool->npages - _manager->options.max_size; | ||
704 | /* free at least NUM_PAGES_TO_ALLOC number of pages | ||
705 | * to reduce calls to set_memory_wb */ | ||
706 | if (npages < NUM_PAGES_TO_ALLOC) | ||
707 | npages = NUM_PAGES_TO_ALLOC; | ||
708 | } | ||
709 | spin_unlock_irqrestore(&pool->lock, irq_flags); | ||
710 | if (npages) | ||
711 | ttm_page_pool_free(pool, npages); | ||
712 | } | ||
713 | |||
665 | /* | 714 | /* |
666 | * On success pages list will hold count number of correctly | 715 | * On success pages list will hold count number of correctly |
667 | * cached pages. | 716 | * cached pages. |
668 | */ | 717 | */ |
669 | int ttm_get_pages(struct page **pages, int flags, | 718 | static int ttm_get_pages(struct page **pages, unsigned npages, int flags, |
670 | enum ttm_caching_state cstate, unsigned npages, | 719 | enum ttm_caching_state cstate) |
671 | dma_addr_t *dma_address) | ||
672 | { | 720 | { |
673 | struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); | 721 | struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); |
674 | struct list_head plist; | 722 | struct list_head plist; |
@@ -736,7 +784,7 @@ int ttm_get_pages(struct page **pages, int flags, | |||
736 | printk(KERN_ERR TTM_PFX | 784 | printk(KERN_ERR TTM_PFX |
737 | "Failed to allocate extra pages " | 785 | "Failed to allocate extra pages " |
738 | "for large request."); | 786 | "for large request."); |
739 | ttm_put_pages(pages, count, flags, cstate, NULL); | 787 | ttm_put_pages(pages, count, flags, cstate); |
740 | return r; | 788 | return r; |
741 | } | 789 | } |
742 | } | 790 | } |
@@ -744,55 +792,6 @@ int ttm_get_pages(struct page **pages, int flags, | |||
744 | return 0; | 792 | return 0; |
745 | } | 793 | } |
746 | 794 | ||
747 | /* Put all pages in pages list to correct pool to wait for reuse */ | ||
748 | void ttm_put_pages(struct page **pages, unsigned npages, int flags, | ||
749 | enum ttm_caching_state cstate, dma_addr_t *dma_address) | ||
750 | { | ||
751 | unsigned long irq_flags; | ||
752 | struct ttm_page_pool *pool = ttm_get_pool(flags, cstate); | ||
753 | unsigned i; | ||
754 | |||
755 | if (pool == NULL) { | ||
756 | /* No pool for this memory type so free the pages */ | ||
757 | for (i = 0; i < npages; i++) { | ||
758 | if (pages[i]) { | ||
759 | if (page_count(pages[i]) != 1) | ||
760 | printk(KERN_ERR TTM_PFX | ||
761 | "Erroneous page count. " | ||
762 | "Leaking pages.\n"); | ||
763 | __free_page(pages[i]); | ||
764 | pages[i] = NULL; | ||
765 | } | ||
766 | } | ||
767 | return; | ||
768 | } | ||
769 | |||
770 | spin_lock_irqsave(&pool->lock, irq_flags); | ||
771 | for (i = 0; i < npages; i++) { | ||
772 | if (pages[i]) { | ||
773 | if (page_count(pages[i]) != 1) | ||
774 | printk(KERN_ERR TTM_PFX | ||
775 | "Erroneous page count. " | ||
776 | "Leaking pages.\n"); | ||
777 | list_add_tail(&pages[i]->lru, &pool->list); | ||
778 | pages[i] = NULL; | ||
779 | pool->npages++; | ||
780 | } | ||
781 | } | ||
782 | /* Check that we don't go over the pool limit */ | ||
783 | npages = 0; | ||
784 | if (pool->npages > _manager->options.max_size) { | ||
785 | npages = pool->npages - _manager->options.max_size; | ||
786 | /* free at least NUM_PAGES_TO_ALLOC number of pages | ||
787 | * to reduce calls to set_memory_wb */ | ||
788 | if (npages < NUM_PAGES_TO_ALLOC) | ||
789 | npages = NUM_PAGES_TO_ALLOC; | ||
790 | } | ||
791 | spin_unlock_irqrestore(&pool->lock, irq_flags); | ||
792 | if (npages) | ||
793 | ttm_page_pool_free(pool, npages); | ||
794 | } | ||
795 | |||
796 | static void ttm_page_pool_init_locked(struct ttm_page_pool *pool, int flags, | 795 | static void ttm_page_pool_init_locked(struct ttm_page_pool *pool, int flags, |
797 | char *name) | 796 | char *name) |
798 | { | 797 | { |
@@ -865,9 +864,9 @@ int ttm_pool_populate(struct ttm_tt *ttm) | |||
865 | return 0; | 864 | return 0; |
866 | 865 | ||
867 | for (i = 0; i < ttm->num_pages; ++i) { | 866 | for (i = 0; i < ttm->num_pages; ++i) { |
868 | ret = ttm_get_pages(&ttm->pages[i], ttm->page_flags, | 867 | ret = ttm_get_pages(&ttm->pages[i], 1, |
869 | ttm->caching_state, 1, | 868 | ttm->page_flags, |
870 | &ttm->dma_address[i]); | 869 | ttm->caching_state); |
871 | if (ret != 0) { | 870 | if (ret != 0) { |
872 | ttm_pool_unpopulate(ttm); | 871 | ttm_pool_unpopulate(ttm); |
873 | return -ENOMEM; | 872 | return -ENOMEM; |
@@ -904,8 +903,7 @@ void ttm_pool_unpopulate(struct ttm_tt *ttm) | |||
904 | ttm->pages[i]); | 903 | ttm->pages[i]); |
905 | ttm_put_pages(&ttm->pages[i], 1, | 904 | ttm_put_pages(&ttm->pages[i], 1, |
906 | ttm->page_flags, | 905 | ttm->page_flags, |
907 | ttm->caching_state, | 906 | ttm->caching_state); |
908 | ttm->dma_address); | ||
909 | } | 907 | } |
910 | } | 908 | } |
911 | ttm->state = tt_unpopulated; | 909 | ttm->state = tt_unpopulated; |
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c b/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c index 7a4779304877..6678abca0d98 100644 --- a/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c +++ b/drivers/gpu/drm/ttm/ttm_page_alloc_dma.c | |||
@@ -789,7 +789,7 @@ out: | |||
789 | 789 | ||
790 | /* | 790 | /* |
791 | * @return count of pages still required to fulfill the request. | 791 | * @return count of pages still required to fulfill the request. |
792 | */ | 792 | */ |
793 | static int ttm_dma_page_pool_fill_locked(struct dma_pool *pool, | 793 | static int ttm_dma_page_pool_fill_locked(struct dma_pool *pool, |
794 | unsigned long *irq_flags) | 794 | unsigned long *irq_flags) |
795 | { | 795 | { |
@@ -838,10 +838,11 @@ static int ttm_dma_page_pool_fill_locked(struct dma_pool *pool, | |||
838 | * allocates one page at a time. | 838 | * allocates one page at a time. |
839 | */ | 839 | */ |
840 | static int ttm_dma_pool_get_pages(struct dma_pool *pool, | 840 | static int ttm_dma_pool_get_pages(struct dma_pool *pool, |
841 | struct ttm_tt *ttm, | 841 | struct ttm_dma_tt *ttm_dma, |
842 | unsigned index) | 842 | unsigned index) |
843 | { | 843 | { |
844 | struct dma_page *d_page; | 844 | struct dma_page *d_page; |
845 | struct ttm_tt *ttm = &ttm_dma->ttm; | ||
845 | unsigned long irq_flags; | 846 | unsigned long irq_flags; |
846 | int count, r = -ENOMEM; | 847 | int count, r = -ENOMEM; |
847 | 848 | ||
@@ -850,8 +851,8 @@ static int ttm_dma_pool_get_pages(struct dma_pool *pool, | |||
850 | if (count) { | 851 | if (count) { |
851 | d_page = list_first_entry(&pool->free_list, struct dma_page, page_list); | 852 | d_page = list_first_entry(&pool->free_list, struct dma_page, page_list); |
852 | ttm->pages[index] = d_page->p; | 853 | ttm->pages[index] = d_page->p; |
853 | ttm->dma_address[index] = d_page->dma; | 854 | ttm_dma->dma_address[index] = d_page->dma; |
854 | list_move_tail(&d_page->page_list, &ttm->alloc_list); | 855 | list_move_tail(&d_page->page_list, &ttm_dma->pages_list); |
855 | r = 0; | 856 | r = 0; |
856 | pool->npages_in_use += 1; | 857 | pool->npages_in_use += 1; |
857 | pool->npages_free -= 1; | 858 | pool->npages_free -= 1; |
@@ -864,8 +865,9 @@ static int ttm_dma_pool_get_pages(struct dma_pool *pool, | |||
864 | * On success pages list will hold count number of correctly | 865 | * On success pages list will hold count number of correctly |
865 | * cached pages. On failure will hold the negative return value (-ENOMEM, etc). | 866 | * cached pages. On failure will hold the negative return value (-ENOMEM, etc). |
866 | */ | 867 | */ |
867 | int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) | 868 | int ttm_dma_populate(struct ttm_dma_tt *ttm_dma, struct device *dev) |
868 | { | 869 | { |
870 | struct ttm_tt *ttm = &ttm_dma->ttm; | ||
869 | struct ttm_mem_global *mem_glob = ttm->glob->mem_glob; | 871 | struct ttm_mem_global *mem_glob = ttm->glob->mem_glob; |
870 | struct dma_pool *pool; | 872 | struct dma_pool *pool; |
871 | enum pool_type type; | 873 | enum pool_type type; |
@@ -892,18 +894,18 @@ int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) | |||
892 | } | 894 | } |
893 | } | 895 | } |
894 | 896 | ||
895 | INIT_LIST_HEAD(&ttm->alloc_list); | 897 | INIT_LIST_HEAD(&ttm_dma->pages_list); |
896 | for (i = 0; i < ttm->num_pages; ++i) { | 898 | for (i = 0; i < ttm->num_pages; ++i) { |
897 | ret = ttm_dma_pool_get_pages(pool, ttm, i); | 899 | ret = ttm_dma_pool_get_pages(pool, ttm_dma, i); |
898 | if (ret != 0) { | 900 | if (ret != 0) { |
899 | ttm_dma_unpopulate(ttm, dev); | 901 | ttm_dma_unpopulate(ttm_dma, dev); |
900 | return -ENOMEM; | 902 | return -ENOMEM; |
901 | } | 903 | } |
902 | 904 | ||
903 | ret = ttm_mem_global_alloc_page(mem_glob, ttm->pages[i], | 905 | ret = ttm_mem_global_alloc_page(mem_glob, ttm->pages[i], |
904 | false, false); | 906 | false, false); |
905 | if (unlikely(ret != 0)) { | 907 | if (unlikely(ret != 0)) { |
906 | ttm_dma_unpopulate(ttm, dev); | 908 | ttm_dma_unpopulate(ttm_dma, dev); |
907 | return -ENOMEM; | 909 | return -ENOMEM; |
908 | } | 910 | } |
909 | } | 911 | } |
@@ -911,7 +913,7 @@ int ttm_dma_populate(struct ttm_tt *ttm, struct device *dev) | |||
911 | if (unlikely(ttm->page_flags & TTM_PAGE_FLAG_SWAPPED)) { | 913 | if (unlikely(ttm->page_flags & TTM_PAGE_FLAG_SWAPPED)) { |
912 | ret = ttm_tt_swapin(ttm); | 914 | ret = ttm_tt_swapin(ttm); |
913 | if (unlikely(ret != 0)) { | 915 | if (unlikely(ret != 0)) { |
914 | ttm_dma_unpopulate(ttm, dev); | 916 | ttm_dma_unpopulate(ttm_dma, dev); |
915 | return ret; | 917 | return ret; |
916 | } | 918 | } |
917 | } | 919 | } |
@@ -937,8 +939,9 @@ static int ttm_dma_pool_get_num_unused_pages(void) | |||
937 | } | 939 | } |
938 | 940 | ||
939 | /* Put all pages in pages list to correct pool to wait for reuse */ | 941 | /* Put all pages in pages list to correct pool to wait for reuse */ |
940 | void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) | 942 | void ttm_dma_unpopulate(struct ttm_dma_tt *ttm_dma, struct device *dev) |
941 | { | 943 | { |
944 | struct ttm_tt *ttm = &ttm_dma->ttm; | ||
942 | struct dma_pool *pool; | 945 | struct dma_pool *pool; |
943 | struct dma_page *d_page, *next; | 946 | struct dma_page *d_page, *next; |
944 | enum pool_type type; | 947 | enum pool_type type; |
@@ -956,7 +959,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) | |||
956 | ttm_to_type(ttm->page_flags, tt_cached)) == pool); | 959 | ttm_to_type(ttm->page_flags, tt_cached)) == pool); |
957 | 960 | ||
958 | /* make sure pages array match list and count number of pages */ | 961 | /* make sure pages array match list and count number of pages */ |
959 | list_for_each_entry(d_page, &ttm->alloc_list, page_list) { | 962 | list_for_each_entry(d_page, &ttm_dma->pages_list, page_list) { |
960 | ttm->pages[count] = d_page->p; | 963 | ttm->pages[count] = d_page->p; |
961 | count++; | 964 | count++; |
962 | } | 965 | } |
@@ -967,7 +970,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) | |||
967 | pool->nfrees += count; | 970 | pool->nfrees += count; |
968 | } else { | 971 | } else { |
969 | pool->npages_free += count; | 972 | pool->npages_free += count; |
970 | list_splice(&ttm->alloc_list, &pool->free_list); | 973 | list_splice(&ttm_dma->pages_list, &pool->free_list); |
971 | if (pool->npages_free > _manager->options.max_size) { | 974 | if (pool->npages_free > _manager->options.max_size) { |
972 | count = pool->npages_free - _manager->options.max_size; | 975 | count = pool->npages_free - _manager->options.max_size; |
973 | } | 976 | } |
@@ -975,7 +978,7 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) | |||
975 | spin_unlock_irqrestore(&pool->lock, irq_flags); | 978 | spin_unlock_irqrestore(&pool->lock, irq_flags); |
976 | 979 | ||
977 | if (is_cached) { | 980 | if (is_cached) { |
978 | list_for_each_entry_safe(d_page, next, &ttm->alloc_list, page_list) { | 981 | list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, page_list) { |
979 | ttm_mem_global_free_page(ttm->glob->mem_glob, | 982 | ttm_mem_global_free_page(ttm->glob->mem_glob, |
980 | d_page->p); | 983 | d_page->p); |
981 | ttm_dma_page_put(pool, d_page); | 984 | ttm_dma_page_put(pool, d_page); |
@@ -987,10 +990,10 @@ void ttm_dma_unpopulate(struct ttm_tt *ttm, struct device *dev) | |||
987 | } | 990 | } |
988 | } | 991 | } |
989 | 992 | ||
990 | INIT_LIST_HEAD(&ttm->alloc_list); | 993 | INIT_LIST_HEAD(&ttm_dma->pages_list); |
991 | for (i = 0; i < ttm->num_pages; i++) { | 994 | for (i = 0; i < ttm->num_pages; i++) { |
992 | ttm->pages[i] = NULL; | 995 | ttm->pages[i] = NULL; |
993 | ttm->dma_address[i] = 0; | 996 | ttm_dma->dma_address[i] = 0; |
994 | } | 997 | } |
995 | 998 | ||
996 | /* shrink pool if necessary */ | 999 | /* shrink pool if necessary */ |
diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c index 1625739b434b..58e1fa14fe3a 100644 --- a/drivers/gpu/drm/ttm/ttm_tt.c +++ b/drivers/gpu/drm/ttm/ttm_tt.c | |||
@@ -48,17 +48,14 @@ | |||
48 | */ | 48 | */ |
49 | static void ttm_tt_alloc_page_directory(struct ttm_tt *ttm) | 49 | static void ttm_tt_alloc_page_directory(struct ttm_tt *ttm) |
50 | { | 50 | { |
51 | ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(*ttm->pages)); | 51 | ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(void*)); |
52 | ttm->dma_address = drm_calloc_large(ttm->num_pages, | ||
53 | sizeof(*ttm->dma_address)); | ||
54 | } | 52 | } |
55 | 53 | ||
56 | static void ttm_tt_free_page_directory(struct ttm_tt *ttm) | 54 | static void ttm_dma_tt_alloc_page_directory(struct ttm_dma_tt *ttm) |
57 | { | 55 | { |
58 | drm_free_large(ttm->pages); | 56 | ttm->ttm.pages = drm_calloc_large(ttm->ttm.num_pages, sizeof(void*)); |
59 | ttm->pages = NULL; | 57 | ttm->dma_address = drm_calloc_large(ttm->ttm.num_pages, |
60 | drm_free_large(ttm->dma_address); | 58 | sizeof(*ttm->dma_address)); |
61 | ttm->dma_address = NULL; | ||
62 | } | 59 | } |
63 | 60 | ||
64 | #ifdef CONFIG_X86 | 61 | #ifdef CONFIG_X86 |
@@ -173,7 +170,6 @@ void ttm_tt_destroy(struct ttm_tt *ttm) | |||
173 | 170 | ||
174 | if (likely(ttm->pages != NULL)) { | 171 | if (likely(ttm->pages != NULL)) { |
175 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); | 172 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); |
176 | ttm_tt_free_page_directory(ttm); | ||
177 | } | 173 | } |
178 | 174 | ||
179 | if (!(ttm->page_flags & TTM_PAGE_FLAG_PERSISTENT_SWAP) && | 175 | if (!(ttm->page_flags & TTM_PAGE_FLAG_PERSISTENT_SWAP) && |
@@ -196,9 +192,8 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev, | |||
196 | ttm->dummy_read_page = dummy_read_page; | 192 | ttm->dummy_read_page = dummy_read_page; |
197 | ttm->state = tt_unpopulated; | 193 | ttm->state = tt_unpopulated; |
198 | 194 | ||
199 | INIT_LIST_HEAD(&ttm->alloc_list); | ||
200 | ttm_tt_alloc_page_directory(ttm); | 195 | ttm_tt_alloc_page_directory(ttm); |
201 | if (!ttm->pages || !ttm->dma_address) { | 196 | if (!ttm->pages) { |
202 | ttm_tt_destroy(ttm); | 197 | ttm_tt_destroy(ttm); |
203 | printk(KERN_ERR TTM_PFX "Failed allocating page table\n"); | 198 | printk(KERN_ERR TTM_PFX "Failed allocating page table\n"); |
204 | return -ENOMEM; | 199 | return -ENOMEM; |
@@ -207,6 +202,49 @@ int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev, | |||
207 | } | 202 | } |
208 | EXPORT_SYMBOL(ttm_tt_init); | 203 | EXPORT_SYMBOL(ttm_tt_init); |
209 | 204 | ||
205 | void ttm_tt_fini(struct ttm_tt *ttm) | ||
206 | { | ||
207 | drm_free_large(ttm->pages); | ||
208 | ttm->pages = NULL; | ||
209 | } | ||
210 | EXPORT_SYMBOL(ttm_tt_fini); | ||
211 | |||
212 | int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev, | ||
213 | unsigned long size, uint32_t page_flags, | ||
214 | struct page *dummy_read_page) | ||
215 | { | ||
216 | struct ttm_tt *ttm = &ttm_dma->ttm; | ||
217 | |||
218 | ttm->bdev = bdev; | ||
219 | ttm->glob = bdev->glob; | ||
220 | ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; | ||
221 | ttm->caching_state = tt_cached; | ||
222 | ttm->page_flags = page_flags; | ||
223 | ttm->dummy_read_page = dummy_read_page; | ||
224 | ttm->state = tt_unpopulated; | ||
225 | |||
226 | INIT_LIST_HEAD(&ttm_dma->pages_list); | ||
227 | ttm_dma_tt_alloc_page_directory(ttm_dma); | ||
228 | if (!ttm->pages || !ttm_dma->dma_address) { | ||
229 | ttm_tt_destroy(ttm); | ||
230 | printk(KERN_ERR TTM_PFX "Failed allocating page table\n"); | ||
231 | return -ENOMEM; | ||
232 | } | ||
233 | return 0; | ||
234 | } | ||
235 | EXPORT_SYMBOL(ttm_dma_tt_init); | ||
236 | |||
237 | void ttm_dma_tt_fini(struct ttm_dma_tt *ttm_dma) | ||
238 | { | ||
239 | struct ttm_tt *ttm = &ttm_dma->ttm; | ||
240 | |||
241 | drm_free_large(ttm->pages); | ||
242 | ttm->pages = NULL; | ||
243 | drm_free_large(ttm_dma->dma_address); | ||
244 | ttm_dma->dma_address = NULL; | ||
245 | } | ||
246 | EXPORT_SYMBOL(ttm_dma_tt_fini); | ||
247 | |||
210 | void ttm_tt_unbind(struct ttm_tt *ttm) | 248 | void ttm_tt_unbind(struct ttm_tt *ttm) |
211 | { | 249 | { |
212 | int ret; | 250 | int ret; |