diff options
| author | Dave Airlie <airlied@redhat.com> | 2017-11-06 01:18:59 -0500 |
|---|---|---|
| committer | Dave Airlie <airlied@redhat.com> | 2017-11-06 01:18:59 -0500 |
| commit | 8a6fb5b5823d863b07f670dc9e791d4622d5b7e9 (patch) | |
| tree | dc853580b18fedbdaa6f79d1d9cdd9b7495fed08 /drivers/gpu/drm/ttm/ttm_page_alloc.c | |
| parent | 36a5fdf76d3281345e000e115f33817570a76420 (diff) | |
| parent | 767601d100a53e653233aebca7c262ce0addfa99 (diff) | |
Merge branch 'drm-next-4.15' of git://people.freedesktop.org/~agd5f/linux into drm-next
some more amd/ttm fixes.
* 'drm-next-4.15' of git://people.freedesktop.org/~agd5f/linux:
drm/ttm: Downgrade pr_err to pr_debug for memory allocation failures
drm/ttm: Always and only destroy bo->ttm_resv in ttm_bo_release_list
drm/amd/amdgpu: Enabling ACP clock in hw_init (v2)
drm/amdgpu/virt: don't dereference undefined 'module' struct
Diffstat (limited to 'drivers/gpu/drm/ttm/ttm_page_alloc.c')
| -rw-r--r-- | drivers/gpu/drm/ttm/ttm_page_alloc.c | 13 |
1 files changed, 6 insertions, 7 deletions
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c index 4d688c8d7853..316f831ad5f0 100644 --- a/drivers/gpu/drm/ttm/ttm_page_alloc.c +++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c | |||
| @@ -329,7 +329,7 @@ static int ttm_page_pool_free(struct ttm_page_pool *pool, unsigned nr_free, | |||
| 329 | pages_to_free = kmalloc(npages_to_free * sizeof(struct page *), | 329 | pages_to_free = kmalloc(npages_to_free * sizeof(struct page *), |
| 330 | GFP_KERNEL); | 330 | GFP_KERNEL); |
| 331 | if (!pages_to_free) { | 331 | if (!pages_to_free) { |
| 332 | pr_err("Failed to allocate memory for pool free operation\n"); | 332 | pr_debug("Failed to allocate memory for pool free operation\n"); |
| 333 | return 0; | 333 | return 0; |
| 334 | } | 334 | } |
| 335 | 335 | ||
| @@ -517,7 +517,7 @@ static int ttm_alloc_new_pages(struct list_head *pages, gfp_t gfp_flags, | |||
| 517 | caching_array = kmalloc(max_cpages*sizeof(struct page *), GFP_KERNEL); | 517 | caching_array = kmalloc(max_cpages*sizeof(struct page *), GFP_KERNEL); |
| 518 | 518 | ||
| 519 | if (!caching_array) { | 519 | if (!caching_array) { |
| 520 | pr_err("Unable to allocate table for new pages\n"); | 520 | pr_debug("Unable to allocate table for new pages\n"); |
| 521 | return -ENOMEM; | 521 | return -ENOMEM; |
| 522 | } | 522 | } |
| 523 | 523 | ||
| @@ -525,7 +525,7 @@ static int ttm_alloc_new_pages(struct list_head *pages, gfp_t gfp_flags, | |||
| 525 | p = alloc_pages(gfp_flags, order); | 525 | p = alloc_pages(gfp_flags, order); |
| 526 | 526 | ||
| 527 | if (!p) { | 527 | if (!p) { |
| 528 | pr_err("Unable to get page %u\n", i); | 528 | pr_debug("Unable to get page %u\n", i); |
| 529 | 529 | ||
| 530 | /* store already allocated pages in the pool after | 530 | /* store already allocated pages in the pool after |
| 531 | * setting the caching state */ | 531 | * setting the caching state */ |
| @@ -625,7 +625,7 @@ static void ttm_page_pool_fill_locked(struct ttm_page_pool *pool, int ttm_flags, | |||
| 625 | ++pool->nrefills; | 625 | ++pool->nrefills; |
| 626 | pool->npages += alloc_size; | 626 | pool->npages += alloc_size; |
| 627 | } else { | 627 | } else { |
| 628 | pr_err("Failed to fill pool (%p)\n", pool); | 628 | pr_debug("Failed to fill pool (%p)\n", pool); |
| 629 | /* If we have any pages left put them to the pool. */ | 629 | /* If we have any pages left put them to the pool. */ |
| 630 | list_for_each_entry(p, &new_pages, lru) { | 630 | list_for_each_entry(p, &new_pages, lru) { |
| 631 | ++cpages; | 631 | ++cpages; |
| @@ -885,8 +885,7 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags, | |||
| 885 | while (npages) { | 885 | while (npages) { |
| 886 | p = alloc_page(gfp_flags); | 886 | p = alloc_page(gfp_flags); |
| 887 | if (!p) { | 887 | if (!p) { |
| 888 | 888 | pr_debug("Unable to allocate page\n"); | |
| 889 | pr_err("Unable to allocate page\n"); | ||
| 890 | return -ENOMEM; | 889 | return -ENOMEM; |
| 891 | } | 890 | } |
| 892 | 891 | ||
| @@ -925,7 +924,7 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags, | |||
| 925 | /* If there is any pages in the list put them back to | 924 | /* If there is any pages in the list put them back to |
| 926 | * the pool. | 925 | * the pool. |
| 927 | */ | 926 | */ |
| 928 | pr_err("Failed to allocate extra pages for large request\n"); | 927 | pr_debug("Failed to allocate extra pages for large request\n"); |
| 929 | ttm_put_pages(pages, count, flags, cstate); | 928 | ttm_put_pages(pages, count, flags, cstate); |
| 930 | return r; | 929 | return r; |
| 931 | } | 930 | } |
