diff options
Diffstat (limited to 'drivers/gpu/drm/ttm/ttm_page_alloc.c')
-rw-r--r-- | drivers/gpu/drm/ttm/ttm_page_alloc.c | 13 |
1 files changed, 6 insertions, 7 deletions
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c index 4d688c8d7853..316f831ad5f0 100644 --- a/drivers/gpu/drm/ttm/ttm_page_alloc.c +++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c | |||
@@ -329,7 +329,7 @@ static int ttm_page_pool_free(struct ttm_page_pool *pool, unsigned nr_free, | |||
329 | pages_to_free = kmalloc(npages_to_free * sizeof(struct page *), | 329 | pages_to_free = kmalloc(npages_to_free * sizeof(struct page *), |
330 | GFP_KERNEL); | 330 | GFP_KERNEL); |
331 | if (!pages_to_free) { | 331 | if (!pages_to_free) { |
332 | pr_err("Failed to allocate memory for pool free operation\n"); | 332 | pr_debug("Failed to allocate memory for pool free operation\n"); |
333 | return 0; | 333 | return 0; |
334 | } | 334 | } |
335 | 335 | ||
@@ -517,7 +517,7 @@ static int ttm_alloc_new_pages(struct list_head *pages, gfp_t gfp_flags, | |||
517 | caching_array = kmalloc(max_cpages*sizeof(struct page *), GFP_KERNEL); | 517 | caching_array = kmalloc(max_cpages*sizeof(struct page *), GFP_KERNEL); |
518 | 518 | ||
519 | if (!caching_array) { | 519 | if (!caching_array) { |
520 | pr_err("Unable to allocate table for new pages\n"); | 520 | pr_debug("Unable to allocate table for new pages\n"); |
521 | return -ENOMEM; | 521 | return -ENOMEM; |
522 | } | 522 | } |
523 | 523 | ||
@@ -525,7 +525,7 @@ static int ttm_alloc_new_pages(struct list_head *pages, gfp_t gfp_flags, | |||
525 | p = alloc_pages(gfp_flags, order); | 525 | p = alloc_pages(gfp_flags, order); |
526 | 526 | ||
527 | if (!p) { | 527 | if (!p) { |
528 | pr_err("Unable to get page %u\n", i); | 528 | pr_debug("Unable to get page %u\n", i); |
529 | 529 | ||
530 | /* store already allocated pages in the pool after | 530 | /* store already allocated pages in the pool after |
531 | * setting the caching state */ | 531 | * setting the caching state */ |
@@ -625,7 +625,7 @@ static void ttm_page_pool_fill_locked(struct ttm_page_pool *pool, int ttm_flags, | |||
625 | ++pool->nrefills; | 625 | ++pool->nrefills; |
626 | pool->npages += alloc_size; | 626 | pool->npages += alloc_size; |
627 | } else { | 627 | } else { |
628 | pr_err("Failed to fill pool (%p)\n", pool); | 628 | pr_debug("Failed to fill pool (%p)\n", pool); |
629 | /* If we have any pages left put them to the pool. */ | 629 | /* If we have any pages left put them to the pool. */ |
630 | list_for_each_entry(p, &new_pages, lru) { | 630 | list_for_each_entry(p, &new_pages, lru) { |
631 | ++cpages; | 631 | ++cpages; |
@@ -885,8 +885,7 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags, | |||
885 | while (npages) { | 885 | while (npages) { |
886 | p = alloc_page(gfp_flags); | 886 | p = alloc_page(gfp_flags); |
887 | if (!p) { | 887 | if (!p) { |
888 | 888 | pr_debug("Unable to allocate page\n"); | |
889 | pr_err("Unable to allocate page\n"); | ||
890 | return -ENOMEM; | 889 | return -ENOMEM; |
891 | } | 890 | } |
892 | 891 | ||
@@ -925,7 +924,7 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags, | |||
925 | /* If there is any pages in the list put them back to | 924 | /* If there is any pages in the list put them back to |
926 | * the pool. | 925 | * the pool. |
927 | */ | 926 | */ |
928 | pr_err("Failed to allocate extra pages for large request\n"); | 927 | pr_debug("Failed to allocate extra pages for large request\n"); |
929 | ttm_put_pages(pages, count, flags, cstate); | 928 | ttm_put_pages(pages, count, flags, cstate); |
930 | return r; | 929 | return r; |
931 | } | 930 | } |