From 0d8830394adf94fee8f69bc034293d1880e9d8db Mon Sep 17 00:00:00 2001 From: Deepak Nibade Date: Fri, 10 Feb 2017 17:35:58 +0530 Subject: gpu: nvgpu: use nvgpu list for page chunks Use nvgpu list APIs instead of linux list APIs to store chunks of page allocator Jira NVGPU-13 Change-Id: I63375fc2df683e018c48a90b76eca368438cc32f Signed-off-by: Deepak Nibade Reviewed-on: http://git-master/r/1326814 Reviewed-by: Konsta Holtta Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: svccoveritychecker GVS: Gerrit_Virtual_Submit Reviewed-by: Terje Bergstrom --- drivers/gpu/nvgpu/common/mm/page_allocator.c | 42 +++++++++++++----------- drivers/gpu/nvgpu/common/pramin.c | 6 ++-- drivers/gpu/nvgpu/gk20a/mm_gk20a.c | 11 ++++--- drivers/gpu/nvgpu/include/nvgpu/page_allocator.h | 12 +++++-- 4 files changed, 43 insertions(+), 28 deletions(-) diff --git a/drivers/gpu/nvgpu/common/mm/page_allocator.c b/drivers/gpu/nvgpu/common/mm/page_allocator.c index 7d2cedc9..d74db3e3 100644 --- a/drivers/gpu/nvgpu/common/mm/page_allocator.c +++ b/drivers/gpu/nvgpu/common/mm/page_allocator.c @@ -135,11 +135,11 @@ static void __nvgpu_free_pages(struct nvgpu_page_allocator *a, { struct page_alloc_chunk *chunk; - while (!list_empty(&alloc->alloc_chunks)) { - chunk = list_first_entry(&alloc->alloc_chunks, - struct page_alloc_chunk, + while (!nvgpu_list_empty(&alloc->alloc_chunks)) { + chunk = nvgpu_list_first_entry(&alloc->alloc_chunks, + page_alloc_chunk, list_entry); - list_del(&chunk->list_entry); + nvgpu_list_del(&chunk->list_entry); if (free_buddy_alloc) nvgpu_free(&a->source_allocator, chunk->base); @@ -322,8 +322,8 @@ static int __do_slab_alloc(struct nvgpu_page_allocator *a, alloc->length = slab_page->slab_size; alloc->base = slab_page->page_addr + (offs * slab_page->slab_size); - chunk = list_first_entry(&alloc->alloc_chunks, - struct page_alloc_chunk, list_entry); + chunk = nvgpu_list_first_entry(&alloc->alloc_chunks, + page_alloc_chunk, list_entry); chunk->base = alloc->base; chunk->length = alloc->length; @@ -359,8 +359,8 @@ static struct nvgpu_page_alloc *__nvgpu_alloc_slab( goto fail; } - INIT_LIST_HEAD(&alloc->alloc_chunks); - list_add(&chunk->list_entry, &alloc->alloc_chunks); + nvgpu_init_list_node(&alloc->alloc_chunks); + nvgpu_list_add(&chunk->list_entry, &alloc->alloc_chunks); err = __do_slab_alloc(a, slab, alloc); if (err) @@ -448,7 +448,7 @@ static struct nvgpu_page_alloc *__do_nvgpu_alloc_pages( memset(alloc, 0, sizeof(*alloc)); - INIT_LIST_HEAD(&alloc->alloc_chunks); + nvgpu_init_list_node(&alloc->alloc_chunks); alloc->length = pages << a->page_shift; while (pages) { @@ -504,23 +504,23 @@ static struct nvgpu_page_alloc *__do_nvgpu_alloc_pages( c->base = chunk_addr; c->length = chunk_len; - list_add(&c->list_entry, &alloc->alloc_chunks); + nvgpu_list_add(&c->list_entry, &alloc->alloc_chunks); i++; } alloc->nr_chunks = i; - c = list_first_entry(&alloc->alloc_chunks, - struct page_alloc_chunk, list_entry); + c = nvgpu_list_first_entry(&alloc->alloc_chunks, + page_alloc_chunk, list_entry); alloc->base = c->base; return alloc; fail_cleanup: - while (!list_empty(&alloc->alloc_chunks)) { - c = list_first_entry(&alloc->alloc_chunks, - struct page_alloc_chunk, list_entry); - list_del(&c->list_entry); + while (!nvgpu_list_empty(&alloc->alloc_chunks)) { + c = nvgpu_list_first_entry(&alloc->alloc_chunks, + page_alloc_chunk, list_entry); + nvgpu_list_del(&c->list_entry); nvgpu_free(&a->source_allocator, c->base); nvgpu_kmem_cache_free(a->chunk_cache, c); } @@ -548,7 +548,8 @@ static struct nvgpu_page_alloc *__nvgpu_alloc_pages( palloc_dbg(a, "Alloc 0x%llx (%llu) id=0x%010llx\n", pages << a->page_shift, pages, alloc->base); - list_for_each_entry(c, &alloc->alloc_chunks, list_entry) { + nvgpu_list_for_each_entry(c, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { palloc_dbg(a, " Chunk %2d: 0x%010llx + 0x%llx\n", i++, c->base, c->length); } @@ -664,11 +665,11 @@ static struct nvgpu_page_alloc *__nvgpu_alloc_pages_fixed( alloc->nr_chunks = 1; alloc->length = length; - INIT_LIST_HEAD(&alloc->alloc_chunks); + nvgpu_init_list_node(&alloc->alloc_chunks); c->base = alloc->base; c->length = length; - list_add(&c->list_entry, &alloc->alloc_chunks); + nvgpu_list_add(&c->list_entry, &alloc->alloc_chunks); return alloc; @@ -708,7 +709,8 @@ static u64 nvgpu_page_alloc_fixed(struct nvgpu_allocator *__a, palloc_dbg(a, "Alloc [fixed] @ 0x%010llx + 0x%llx (%llu)\n", alloc->base, aligned_len, pages); - list_for_each_entry(c, &alloc->alloc_chunks, list_entry) { + nvgpu_list_for_each_entry(c, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { palloc_dbg(a, " Chunk %2d: 0x%010llx + 0x%llx\n", i++, c->base, c->length); } diff --git a/drivers/gpu/nvgpu/common/pramin.c b/drivers/gpu/nvgpu/common/pramin.c index b9216309..aa732368 100644 --- a/drivers/gpu/nvgpu/common/pramin.c +++ b/drivers/gpu/nvgpu/common/pramin.c @@ -88,7 +88,8 @@ void nvgpu_pramin_access_batched(struct gk20a *g, struct mem_desc *mem, u32 byteoff, start_reg, until_end, n; alloc = get_vidmem_page_alloc(mem->sgt->sgl); - list_for_each_entry(chunk, &alloc->alloc_chunks, list_entry) { + nvgpu_list_for_each_entry(chunk, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { if (offset >= chunk->length) offset -= chunk->length; else @@ -113,7 +114,8 @@ void nvgpu_pramin_access_batched(struct gk20a *g, struct mem_desc *mem, size -= n; if (n == (chunk->length - offset)) { - chunk = list_next_entry(chunk, list_entry); + chunk = nvgpu_list_next_entry(chunk, page_alloc_chunk, + list_entry); offset = 0; } else { offset += n / sizeof(u32); diff --git a/drivers/gpu/nvgpu/gk20a/mm_gk20a.c b/drivers/gpu/nvgpu/gk20a/mm_gk20a.c index 94d31273..3cda3034 100644 --- a/drivers/gpu/nvgpu/gk20a/mm_gk20a.c +++ b/drivers/gpu/nvgpu/gk20a/mm_gk20a.c @@ -33,6 +33,7 @@ #include #include #include +#include #include #include #include @@ -2174,7 +2175,8 @@ static u64 gk20a_mm_get_align(struct gk20a *g, struct scatterlist *sgl, struct nvgpu_page_alloc *alloc = get_vidmem_page_alloc(sgl); struct page_alloc_chunk *chunk = NULL; - list_for_each_entry(chunk, &alloc->alloc_chunks, list_entry) { + nvgpu_list_for_each_entry(chunk, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { chunk_align = 1ULL << __ffs(chunk->base | chunk->length); if (align) @@ -2875,7 +2877,8 @@ static int gk20a_gmmu_clear_vidmem_mem(struct gk20a *g, struct mem_desc *mem) alloc = get_vidmem_page_alloc(mem->sgt->sgl); - list_for_each_entry(chunk, &alloc->alloc_chunks, list_entry) { + nvgpu_list_for_each_entry(chunk, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { if (gk20a_last_fence) gk20a_fence_put(gk20a_last_fence); @@ -3732,8 +3735,8 @@ static int update_gmmu_ptes_locked(struct vm_gk20a *vm, if (sgt) { alloc = get_vidmem_page_alloc(sgt->sgl); - list_for_each_entry(chunk, &alloc->alloc_chunks, - list_entry) { + nvgpu_list_for_each_entry(chunk, &alloc->alloc_chunks, + page_alloc_chunk, list_entry) { if (space_to_skip && space_to_skip > chunk->length) { space_to_skip -= chunk->length; diff --git a/drivers/gpu/nvgpu/include/nvgpu/page_allocator.h b/drivers/gpu/nvgpu/include/nvgpu/page_allocator.h index fa586dba..92f48ac5 100644 --- a/drivers/gpu/nvgpu/include/nvgpu/page_allocator.h +++ b/drivers/gpu/nvgpu/include/nvgpu/page_allocator.h @@ -22,6 +22,7 @@ #include #include +#include struct nvgpu_allocator; @@ -78,19 +79,26 @@ struct page_alloc_slab_page { }; struct page_alloc_chunk { - struct list_head list_entry; + struct nvgpu_list_node list_entry; u64 base; u64 length; }; +static inline struct page_alloc_chunk * +page_alloc_chunk_from_list_entry(struct nvgpu_list_node *node) +{ + return (struct page_alloc_chunk *) + ((uintptr_t)node - offsetof(struct page_alloc_chunk, list_entry)); +}; + /* * Struct to handle internal management of page allocation. It holds a list * of the chunks of pages that make up the overall allocation - much like a * scatter gather table. */ struct nvgpu_page_alloc { - struct list_head alloc_chunks; + struct nvgpu_list_node alloc_chunks; int nr_chunks; u64 length; -- cgit v1.2.2