aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c')
-rw-r--r--drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c34
1 files changed, 17 insertions, 17 deletions
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
index 917ac5e074a0..81127ffcefb2 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
@@ -59,7 +59,7 @@ int amdgpu_gem_object_create(struct amdgpu_device *adev, unsigned long size,
59 59
60retry: 60retry:
61 r = amdgpu_bo_create(adev, size, alignment, kernel, initial_domain, 61 r = amdgpu_bo_create(adev, size, alignment, kernel, initial_domain,
62 flags, NULL, NULL, &robj); 62 flags, NULL, NULL, 0, &robj);
63 if (r) { 63 if (r) {
64 if (r != -ERESTARTSYS) { 64 if (r != -ERESTARTSYS) {
65 if (initial_domain == AMDGPU_GEM_DOMAIN_VRAM) { 65 if (initial_domain == AMDGPU_GEM_DOMAIN_VRAM) {
@@ -91,7 +91,7 @@ void amdgpu_gem_force_release(struct amdgpu_device *adev)
91 spin_lock(&file->table_lock); 91 spin_lock(&file->table_lock);
92 idr_for_each_entry(&file->object_idr, gobj, handle) { 92 idr_for_each_entry(&file->object_idr, gobj, handle) {
93 WARN_ONCE(1, "And also active allocations!\n"); 93 WARN_ONCE(1, "And also active allocations!\n");
94 drm_gem_object_unreference_unlocked(gobj); 94 drm_gem_object_put_unlocked(gobj);
95 } 95 }
96 idr_destroy(&file->object_idr); 96 idr_destroy(&file->object_idr);
97 spin_unlock(&file->table_lock); 97 spin_unlock(&file->table_lock);
@@ -263,7 +263,7 @@ int amdgpu_gem_create_ioctl(struct drm_device *dev, void *data,
263 263
264 r = drm_gem_handle_create(filp, gobj, &handle); 264 r = drm_gem_handle_create(filp, gobj, &handle);
265 /* drop reference from allocate - handle holds it now */ 265 /* drop reference from allocate - handle holds it now */
266 drm_gem_object_unreference_unlocked(gobj); 266 drm_gem_object_put_unlocked(gobj);
267 if (r) 267 if (r)
268 return r; 268 return r;
269 269
@@ -306,7 +306,7 @@ int amdgpu_gem_userptr_ioctl(struct drm_device *dev, void *data,
306 return r; 306 return r;
307 307
308 bo = gem_to_amdgpu_bo(gobj); 308 bo = gem_to_amdgpu_bo(gobj);
309 bo->prefered_domains = AMDGPU_GEM_DOMAIN_GTT; 309 bo->preferred_domains = AMDGPU_GEM_DOMAIN_GTT;
310 bo->allowed_domains = AMDGPU_GEM_DOMAIN_GTT; 310 bo->allowed_domains = AMDGPU_GEM_DOMAIN_GTT;
311 r = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags); 311 r = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags);
312 if (r) 312 if (r)
@@ -341,7 +341,7 @@ int amdgpu_gem_userptr_ioctl(struct drm_device *dev, void *data,
341 341
342 r = drm_gem_handle_create(filp, gobj, &handle); 342 r = drm_gem_handle_create(filp, gobj, &handle);
343 /* drop reference from allocate - handle holds it now */ 343 /* drop reference from allocate - handle holds it now */
344 drm_gem_object_unreference_unlocked(gobj); 344 drm_gem_object_put_unlocked(gobj);
345 if (r) 345 if (r)
346 return r; 346 return r;
347 347
@@ -355,7 +355,7 @@ unlock_mmap_sem:
355 up_read(&current->mm->mmap_sem); 355 up_read(&current->mm->mmap_sem);
356 356
357release_object: 357release_object:
358 drm_gem_object_unreference_unlocked(gobj); 358 drm_gem_object_put_unlocked(gobj);
359 359
360 return r; 360 return r;
361} 361}
@@ -374,11 +374,11 @@ int amdgpu_mode_dumb_mmap(struct drm_file *filp,
374 robj = gem_to_amdgpu_bo(gobj); 374 robj = gem_to_amdgpu_bo(gobj);
375 if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) || 375 if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) ||
376 (robj->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS)) { 376 (robj->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS)) {
377 drm_gem_object_unreference_unlocked(gobj); 377 drm_gem_object_put_unlocked(gobj);
378 return -EPERM; 378 return -EPERM;
379 } 379 }
380 *offset_p = amdgpu_bo_mmap_offset(robj); 380 *offset_p = amdgpu_bo_mmap_offset(robj);
381 drm_gem_object_unreference_unlocked(gobj); 381 drm_gem_object_put_unlocked(gobj);
382 return 0; 382 return 0;
383} 383}
384 384
@@ -448,7 +448,7 @@ int amdgpu_gem_wait_idle_ioctl(struct drm_device *dev, void *data,
448 } else 448 } else
449 r = ret; 449 r = ret;
450 450
451 drm_gem_object_unreference_unlocked(gobj); 451 drm_gem_object_put_unlocked(gobj);
452 return r; 452 return r;
453} 453}
454 454
@@ -491,7 +491,7 @@ int amdgpu_gem_metadata_ioctl(struct drm_device *dev, void *data,
491unreserve: 491unreserve:
492 amdgpu_bo_unreserve(robj); 492 amdgpu_bo_unreserve(robj);
493out: 493out:
494 drm_gem_object_unreference_unlocked(gobj); 494 drm_gem_object_put_unlocked(gobj);
495 return r; 495 return r;
496} 496}
497 497
@@ -664,7 +664,7 @@ error_backoff:
664 ttm_eu_backoff_reservation(&ticket, &list); 664 ttm_eu_backoff_reservation(&ticket, &list);
665 665
666error_unref: 666error_unref:
667 drm_gem_object_unreference_unlocked(gobj); 667 drm_gem_object_put_unlocked(gobj);
668 return r; 668 return r;
669} 669}
670 670
@@ -689,11 +689,11 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev, void *data,
689 switch (args->op) { 689 switch (args->op) {
690 case AMDGPU_GEM_OP_GET_GEM_CREATE_INFO: { 690 case AMDGPU_GEM_OP_GET_GEM_CREATE_INFO: {
691 struct drm_amdgpu_gem_create_in info; 691 struct drm_amdgpu_gem_create_in info;
692 void __user *out = (void __user *)(uintptr_t)args->value; 692 void __user *out = u64_to_user_ptr(args->value);
693 693
694 info.bo_size = robj->gem_base.size; 694 info.bo_size = robj->gem_base.size;
695 info.alignment = robj->tbo.mem.page_alignment << PAGE_SHIFT; 695 info.alignment = robj->tbo.mem.page_alignment << PAGE_SHIFT;
696 info.domains = robj->prefered_domains; 696 info.domains = robj->preferred_domains;
697 info.domain_flags = robj->flags; 697 info.domain_flags = robj->flags;
698 amdgpu_bo_unreserve(robj); 698 amdgpu_bo_unreserve(robj);
699 if (copy_to_user(out, &info, sizeof(info))) 699 if (copy_to_user(out, &info, sizeof(info)))
@@ -711,10 +711,10 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev, void *data,
711 amdgpu_bo_unreserve(robj); 711 amdgpu_bo_unreserve(robj);
712 break; 712 break;
713 } 713 }
714 robj->prefered_domains = args->value & (AMDGPU_GEM_DOMAIN_VRAM | 714 robj->preferred_domains = args->value & (AMDGPU_GEM_DOMAIN_VRAM |
715 AMDGPU_GEM_DOMAIN_GTT | 715 AMDGPU_GEM_DOMAIN_GTT |
716 AMDGPU_GEM_DOMAIN_CPU); 716 AMDGPU_GEM_DOMAIN_CPU);
717 robj->allowed_domains = robj->prefered_domains; 717 robj->allowed_domains = robj->preferred_domains;
718 if (robj->allowed_domains == AMDGPU_GEM_DOMAIN_VRAM) 718 if (robj->allowed_domains == AMDGPU_GEM_DOMAIN_VRAM)
719 robj->allowed_domains |= AMDGPU_GEM_DOMAIN_GTT; 719 robj->allowed_domains |= AMDGPU_GEM_DOMAIN_GTT;
720 720
@@ -726,7 +726,7 @@ int amdgpu_gem_op_ioctl(struct drm_device *dev, void *data,
726 } 726 }
727 727
728out: 728out:
729 drm_gem_object_unreference_unlocked(gobj); 729 drm_gem_object_put_unlocked(gobj);
730 return r; 730 return r;
731} 731}
732 732
@@ -754,7 +754,7 @@ int amdgpu_mode_dumb_create(struct drm_file *file_priv,
754 754
755 r = drm_gem_handle_create(file_priv, gobj, &handle); 755 r = drm_gem_handle_create(file_priv, gobj, &handle);
756 /* drop reference from allocate - handle holds it now */ 756 /* drop reference from allocate - handle holds it now */
757 drm_gem_object_unreference_unlocked(gobj); 757 drm_gem_object_put_unlocked(gobj);
758 if (r) { 758 if (r) {
759 return r; 759 return r;
760 } 760 }