diff options
-rw-r--r-- | drivers/gpu/drm/i915/i915_debugfs.c | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_drv.c | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_drv.h | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 132 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_gem_debug.c | 4 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_gem_tiling.c | 10 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_irq.c | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_display.c | 28 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_fb.c | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_overlay.c | 6 |
10 files changed, 96 insertions, 94 deletions
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index 1376dfe44c95..bb3a4a8aba08 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c | |||
@@ -225,7 +225,7 @@ static int i915_gem_fence_regs_info(struct seq_file *m, void *data) | |||
225 | } else { | 225 | } else { |
226 | struct drm_i915_gem_object *obj_priv; | 226 | struct drm_i915_gem_object *obj_priv; |
227 | 227 | ||
228 | obj_priv = obj->driver_private; | 228 | obj_priv = to_intel_bo(obj); |
229 | seq_printf(m, "Fenced object[%2d] = %p: %s " | 229 | seq_printf(m, "Fenced object[%2d] = %p: %s " |
230 | "%08x %08zx %08x %s %08x %08x %d", | 230 | "%08x %08zx %08x %s %08x %08x %d", |
231 | i, obj, get_pin_flag(obj_priv), | 231 | i, obj, get_pin_flag(obj_priv), |
diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c index 1a39ec75d76b..0af3dcc85ce9 100644 --- a/drivers/gpu/drm/i915/i915_drv.c +++ b/drivers/gpu/drm/i915/i915_drv.c | |||
@@ -361,7 +361,7 @@ int i965_reset(struct drm_device *dev, u8 flags) | |||
361 | !dev_priv->mm.suspended) { | 361 | !dev_priv->mm.suspended) { |
362 | drm_i915_ring_buffer_t *ring = &dev_priv->ring; | 362 | drm_i915_ring_buffer_t *ring = &dev_priv->ring; |
363 | struct drm_gem_object *obj = ring->ring_obj; | 363 | struct drm_gem_object *obj = ring->ring_obj; |
364 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 364 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
365 | dev_priv->mm.suspended = 0; | 365 | dev_priv->mm.suspended = 0; |
366 | 366 | ||
367 | /* Stop the ring if it's running. */ | 367 | /* Stop the ring if it's running. */ |
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index aba8260fbc5e..b7cb4aadd059 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
@@ -731,6 +731,8 @@ struct drm_i915_gem_object { | |||
731 | atomic_t pending_flip; | 731 | atomic_t pending_flip; |
732 | }; | 732 | }; |
733 | 733 | ||
734 | #define to_intel_bo(x) ((struct drm_i915_gem_object *) (x)->driver_private) | ||
735 | |||
734 | /** | 736 | /** |
735 | * Request queue structure. | 737 | * Request queue structure. |
736 | * | 738 | * |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 933e865a8929..b85727ce308e 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
@@ -162,7 +162,7 @@ fast_shmem_read(struct page **pages, | |||
162 | static int i915_gem_object_needs_bit17_swizzle(struct drm_gem_object *obj) | 162 | static int i915_gem_object_needs_bit17_swizzle(struct drm_gem_object *obj) |
163 | { | 163 | { |
164 | drm_i915_private_t *dev_priv = obj->dev->dev_private; | 164 | drm_i915_private_t *dev_priv = obj->dev->dev_private; |
165 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 165 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
166 | 166 | ||
167 | return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && | 167 | return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && |
168 | obj_priv->tiling_mode != I915_TILING_NONE; | 168 | obj_priv->tiling_mode != I915_TILING_NONE; |
@@ -263,7 +263,7 @@ i915_gem_shmem_pread_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
263 | struct drm_i915_gem_pread *args, | 263 | struct drm_i915_gem_pread *args, |
264 | struct drm_file *file_priv) | 264 | struct drm_file *file_priv) |
265 | { | 265 | { |
266 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 266 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
267 | ssize_t remain; | 267 | ssize_t remain; |
268 | loff_t offset, page_base; | 268 | loff_t offset, page_base; |
269 | char __user *user_data; | 269 | char __user *user_data; |
@@ -284,7 +284,7 @@ i915_gem_shmem_pread_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
284 | if (ret != 0) | 284 | if (ret != 0) |
285 | goto fail_put_pages; | 285 | goto fail_put_pages; |
286 | 286 | ||
287 | obj_priv = obj->driver_private; | 287 | obj_priv = to_intel_bo(obj); |
288 | offset = args->offset; | 288 | offset = args->offset; |
289 | 289 | ||
290 | while (remain > 0) { | 290 | while (remain > 0) { |
@@ -353,7 +353,7 @@ i915_gem_shmem_pread_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
353 | struct drm_i915_gem_pread *args, | 353 | struct drm_i915_gem_pread *args, |
354 | struct drm_file *file_priv) | 354 | struct drm_file *file_priv) |
355 | { | 355 | { |
356 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 356 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
357 | struct mm_struct *mm = current->mm; | 357 | struct mm_struct *mm = current->mm; |
358 | struct page **user_pages; | 358 | struct page **user_pages; |
359 | ssize_t remain; | 359 | ssize_t remain; |
@@ -402,7 +402,7 @@ i915_gem_shmem_pread_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
402 | if (ret != 0) | 402 | if (ret != 0) |
403 | goto fail_put_pages; | 403 | goto fail_put_pages; |
404 | 404 | ||
405 | obj_priv = obj->driver_private; | 405 | obj_priv = to_intel_bo(obj); |
406 | offset = args->offset; | 406 | offset = args->offset; |
407 | 407 | ||
408 | while (remain > 0) { | 408 | while (remain > 0) { |
@@ -478,7 +478,7 @@ i915_gem_pread_ioctl(struct drm_device *dev, void *data, | |||
478 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 478 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
479 | if (obj == NULL) | 479 | if (obj == NULL) |
480 | return -EBADF; | 480 | return -EBADF; |
481 | obj_priv = obj->driver_private; | 481 | obj_priv = to_intel_bo(obj); |
482 | 482 | ||
483 | /* Bounds check source. | 483 | /* Bounds check source. |
484 | * | 484 | * |
@@ -580,7 +580,7 @@ i915_gem_gtt_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
580 | struct drm_i915_gem_pwrite *args, | 580 | struct drm_i915_gem_pwrite *args, |
581 | struct drm_file *file_priv) | 581 | struct drm_file *file_priv) |
582 | { | 582 | { |
583 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 583 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
584 | drm_i915_private_t *dev_priv = dev->dev_private; | 584 | drm_i915_private_t *dev_priv = dev->dev_private; |
585 | ssize_t remain; | 585 | ssize_t remain; |
586 | loff_t offset, page_base; | 586 | loff_t offset, page_base; |
@@ -604,7 +604,7 @@ i915_gem_gtt_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
604 | if (ret) | 604 | if (ret) |
605 | goto fail; | 605 | goto fail; |
606 | 606 | ||
607 | obj_priv = obj->driver_private; | 607 | obj_priv = to_intel_bo(obj); |
608 | offset = obj_priv->gtt_offset + args->offset; | 608 | offset = obj_priv->gtt_offset + args->offset; |
609 | 609 | ||
610 | while (remain > 0) { | 610 | while (remain > 0) { |
@@ -654,7 +654,7 @@ i915_gem_gtt_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
654 | struct drm_i915_gem_pwrite *args, | 654 | struct drm_i915_gem_pwrite *args, |
655 | struct drm_file *file_priv) | 655 | struct drm_file *file_priv) |
656 | { | 656 | { |
657 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 657 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
658 | drm_i915_private_t *dev_priv = dev->dev_private; | 658 | drm_i915_private_t *dev_priv = dev->dev_private; |
659 | ssize_t remain; | 659 | ssize_t remain; |
660 | loff_t gtt_page_base, offset; | 660 | loff_t gtt_page_base, offset; |
@@ -698,7 +698,7 @@ i915_gem_gtt_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
698 | if (ret) | 698 | if (ret) |
699 | goto out_unpin_object; | 699 | goto out_unpin_object; |
700 | 700 | ||
701 | obj_priv = obj->driver_private; | 701 | obj_priv = to_intel_bo(obj); |
702 | offset = obj_priv->gtt_offset + args->offset; | 702 | offset = obj_priv->gtt_offset + args->offset; |
703 | 703 | ||
704 | while (remain > 0) { | 704 | while (remain > 0) { |
@@ -760,7 +760,7 @@ i915_gem_shmem_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
760 | struct drm_i915_gem_pwrite *args, | 760 | struct drm_i915_gem_pwrite *args, |
761 | struct drm_file *file_priv) | 761 | struct drm_file *file_priv) |
762 | { | 762 | { |
763 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 763 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
764 | ssize_t remain; | 764 | ssize_t remain; |
765 | loff_t offset, page_base; | 765 | loff_t offset, page_base; |
766 | char __user *user_data; | 766 | char __user *user_data; |
@@ -780,7 +780,7 @@ i915_gem_shmem_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
780 | if (ret != 0) | 780 | if (ret != 0) |
781 | goto fail_put_pages; | 781 | goto fail_put_pages; |
782 | 782 | ||
783 | obj_priv = obj->driver_private; | 783 | obj_priv = to_intel_bo(obj); |
784 | offset = args->offset; | 784 | offset = args->offset; |
785 | obj_priv->dirty = 1; | 785 | obj_priv->dirty = 1; |
786 | 786 | ||
@@ -828,7 +828,7 @@ i915_gem_shmem_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
828 | struct drm_i915_gem_pwrite *args, | 828 | struct drm_i915_gem_pwrite *args, |
829 | struct drm_file *file_priv) | 829 | struct drm_file *file_priv) |
830 | { | 830 | { |
831 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 831 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
832 | struct mm_struct *mm = current->mm; | 832 | struct mm_struct *mm = current->mm; |
833 | struct page **user_pages; | 833 | struct page **user_pages; |
834 | ssize_t remain; | 834 | ssize_t remain; |
@@ -876,7 +876,7 @@ i915_gem_shmem_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
876 | if (ret != 0) | 876 | if (ret != 0) |
877 | goto fail_put_pages; | 877 | goto fail_put_pages; |
878 | 878 | ||
879 | obj_priv = obj->driver_private; | 879 | obj_priv = to_intel_bo(obj); |
880 | offset = args->offset; | 880 | offset = args->offset; |
881 | obj_priv->dirty = 1; | 881 | obj_priv->dirty = 1; |
882 | 882 | ||
@@ -951,7 +951,7 @@ i915_gem_pwrite_ioctl(struct drm_device *dev, void *data, | |||
951 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 951 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
952 | if (obj == NULL) | 952 | if (obj == NULL) |
953 | return -EBADF; | 953 | return -EBADF; |
954 | obj_priv = obj->driver_private; | 954 | obj_priv = to_intel_bo(obj); |
955 | 955 | ||
956 | /* Bounds check destination. | 956 | /* Bounds check destination. |
957 | * | 957 | * |
@@ -1033,7 +1033,7 @@ i915_gem_set_domain_ioctl(struct drm_device *dev, void *data, | |||
1033 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 1033 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
1034 | if (obj == NULL) | 1034 | if (obj == NULL) |
1035 | return -EBADF; | 1035 | return -EBADF; |
1036 | obj_priv = obj->driver_private; | 1036 | obj_priv = to_intel_bo(obj); |
1037 | 1037 | ||
1038 | mutex_lock(&dev->struct_mutex); | 1038 | mutex_lock(&dev->struct_mutex); |
1039 | 1039 | ||
@@ -1095,7 +1095,7 @@ i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data, | |||
1095 | DRM_INFO("%s: sw_finish %d (%p %zd)\n", | 1095 | DRM_INFO("%s: sw_finish %d (%p %zd)\n", |
1096 | __func__, args->handle, obj, obj->size); | 1096 | __func__, args->handle, obj, obj->size); |
1097 | #endif | 1097 | #endif |
1098 | obj_priv = obj->driver_private; | 1098 | obj_priv = to_intel_bo(obj); |
1099 | 1099 | ||
1100 | /* Pinned buffers may be scanout, so flush the cache */ | 1100 | /* Pinned buffers may be scanout, so flush the cache */ |
1101 | if (obj_priv->pin_count) | 1101 | if (obj_priv->pin_count) |
@@ -1166,7 +1166,7 @@ int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
1166 | struct drm_gem_object *obj = vma->vm_private_data; | 1166 | struct drm_gem_object *obj = vma->vm_private_data; |
1167 | struct drm_device *dev = obj->dev; | 1167 | struct drm_device *dev = obj->dev; |
1168 | struct drm_i915_private *dev_priv = dev->dev_private; | 1168 | struct drm_i915_private *dev_priv = dev->dev_private; |
1169 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1169 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1170 | pgoff_t page_offset; | 1170 | pgoff_t page_offset; |
1171 | unsigned long pfn; | 1171 | unsigned long pfn; |
1172 | int ret = 0; | 1172 | int ret = 0; |
@@ -1233,7 +1233,7 @@ i915_gem_create_mmap_offset(struct drm_gem_object *obj) | |||
1233 | { | 1233 | { |
1234 | struct drm_device *dev = obj->dev; | 1234 | struct drm_device *dev = obj->dev; |
1235 | struct drm_gem_mm *mm = dev->mm_private; | 1235 | struct drm_gem_mm *mm = dev->mm_private; |
1236 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1236 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1237 | struct drm_map_list *list; | 1237 | struct drm_map_list *list; |
1238 | struct drm_local_map *map; | 1238 | struct drm_local_map *map; |
1239 | int ret = 0; | 1239 | int ret = 0; |
@@ -1304,7 +1304,7 @@ void | |||
1304 | i915_gem_release_mmap(struct drm_gem_object *obj) | 1304 | i915_gem_release_mmap(struct drm_gem_object *obj) |
1305 | { | 1305 | { |
1306 | struct drm_device *dev = obj->dev; | 1306 | struct drm_device *dev = obj->dev; |
1307 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1307 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1308 | 1308 | ||
1309 | if (dev->dev_mapping) | 1309 | if (dev->dev_mapping) |
1310 | unmap_mapping_range(dev->dev_mapping, | 1310 | unmap_mapping_range(dev->dev_mapping, |
@@ -1315,7 +1315,7 @@ static void | |||
1315 | i915_gem_free_mmap_offset(struct drm_gem_object *obj) | 1315 | i915_gem_free_mmap_offset(struct drm_gem_object *obj) |
1316 | { | 1316 | { |
1317 | struct drm_device *dev = obj->dev; | 1317 | struct drm_device *dev = obj->dev; |
1318 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1318 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1319 | struct drm_gem_mm *mm = dev->mm_private; | 1319 | struct drm_gem_mm *mm = dev->mm_private; |
1320 | struct drm_map_list *list; | 1320 | struct drm_map_list *list; |
1321 | 1321 | ||
@@ -1346,7 +1346,7 @@ static uint32_t | |||
1346 | i915_gem_get_gtt_alignment(struct drm_gem_object *obj) | 1346 | i915_gem_get_gtt_alignment(struct drm_gem_object *obj) |
1347 | { | 1347 | { |
1348 | struct drm_device *dev = obj->dev; | 1348 | struct drm_device *dev = obj->dev; |
1349 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1349 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1350 | int start, i; | 1350 | int start, i; |
1351 | 1351 | ||
1352 | /* | 1352 | /* |
@@ -1405,7 +1405,7 @@ i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, | |||
1405 | 1405 | ||
1406 | mutex_lock(&dev->struct_mutex); | 1406 | mutex_lock(&dev->struct_mutex); |
1407 | 1407 | ||
1408 | obj_priv = obj->driver_private; | 1408 | obj_priv = to_intel_bo(obj); |
1409 | 1409 | ||
1410 | if (obj_priv->madv != I915_MADV_WILLNEED) { | 1410 | if (obj_priv->madv != I915_MADV_WILLNEED) { |
1411 | DRM_ERROR("Attempting to mmap a purgeable buffer\n"); | 1411 | DRM_ERROR("Attempting to mmap a purgeable buffer\n"); |
@@ -1449,7 +1449,7 @@ i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, | |||
1449 | void | 1449 | void |
1450 | i915_gem_object_put_pages(struct drm_gem_object *obj) | 1450 | i915_gem_object_put_pages(struct drm_gem_object *obj) |
1451 | { | 1451 | { |
1452 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1452 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1453 | int page_count = obj->size / PAGE_SIZE; | 1453 | int page_count = obj->size / PAGE_SIZE; |
1454 | int i; | 1454 | int i; |
1455 | 1455 | ||
@@ -1485,7 +1485,7 @@ i915_gem_object_move_to_active(struct drm_gem_object *obj, uint32_t seqno) | |||
1485 | { | 1485 | { |
1486 | struct drm_device *dev = obj->dev; | 1486 | struct drm_device *dev = obj->dev; |
1487 | drm_i915_private_t *dev_priv = dev->dev_private; | 1487 | drm_i915_private_t *dev_priv = dev->dev_private; |
1488 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1488 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1489 | 1489 | ||
1490 | /* Add a reference if we're newly entering the active list. */ | 1490 | /* Add a reference if we're newly entering the active list. */ |
1491 | if (!obj_priv->active) { | 1491 | if (!obj_priv->active) { |
@@ -1505,7 +1505,7 @@ i915_gem_object_move_to_flushing(struct drm_gem_object *obj) | |||
1505 | { | 1505 | { |
1506 | struct drm_device *dev = obj->dev; | 1506 | struct drm_device *dev = obj->dev; |
1507 | drm_i915_private_t *dev_priv = dev->dev_private; | 1507 | drm_i915_private_t *dev_priv = dev->dev_private; |
1508 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1508 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1509 | 1509 | ||
1510 | BUG_ON(!obj_priv->active); | 1510 | BUG_ON(!obj_priv->active); |
1511 | list_move_tail(&obj_priv->list, &dev_priv->mm.flushing_list); | 1511 | list_move_tail(&obj_priv->list, &dev_priv->mm.flushing_list); |
@@ -1516,7 +1516,7 @@ i915_gem_object_move_to_flushing(struct drm_gem_object *obj) | |||
1516 | static void | 1516 | static void |
1517 | i915_gem_object_truncate(struct drm_gem_object *obj) | 1517 | i915_gem_object_truncate(struct drm_gem_object *obj) |
1518 | { | 1518 | { |
1519 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1519 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1520 | struct inode *inode; | 1520 | struct inode *inode; |
1521 | 1521 | ||
1522 | inode = obj->filp->f_path.dentry->d_inode; | 1522 | inode = obj->filp->f_path.dentry->d_inode; |
@@ -1537,7 +1537,7 @@ i915_gem_object_move_to_inactive(struct drm_gem_object *obj) | |||
1537 | { | 1537 | { |
1538 | struct drm_device *dev = obj->dev; | 1538 | struct drm_device *dev = obj->dev; |
1539 | drm_i915_private_t *dev_priv = dev->dev_private; | 1539 | drm_i915_private_t *dev_priv = dev->dev_private; |
1540 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1540 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1541 | 1541 | ||
1542 | i915_verify_inactive(dev, __FILE__, __LINE__); | 1542 | i915_verify_inactive(dev, __FILE__, __LINE__); |
1543 | if (obj_priv->pin_count != 0) | 1543 | if (obj_priv->pin_count != 0) |
@@ -1964,7 +1964,7 @@ static int | |||
1964 | i915_gem_object_wait_rendering(struct drm_gem_object *obj) | 1964 | i915_gem_object_wait_rendering(struct drm_gem_object *obj) |
1965 | { | 1965 | { |
1966 | struct drm_device *dev = obj->dev; | 1966 | struct drm_device *dev = obj->dev; |
1967 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1967 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1968 | int ret; | 1968 | int ret; |
1969 | 1969 | ||
1970 | /* This function only exists to support waiting for existing rendering, | 1970 | /* This function only exists to support waiting for existing rendering, |
@@ -1996,7 +1996,7 @@ i915_gem_object_unbind(struct drm_gem_object *obj) | |||
1996 | { | 1996 | { |
1997 | struct drm_device *dev = obj->dev; | 1997 | struct drm_device *dev = obj->dev; |
1998 | drm_i915_private_t *dev_priv = dev->dev_private; | 1998 | drm_i915_private_t *dev_priv = dev->dev_private; |
1999 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1999 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2000 | int ret = 0; | 2000 | int ret = 0; |
2001 | 2001 | ||
2002 | #if WATCH_BUF | 2002 | #if WATCH_BUF |
@@ -2172,7 +2172,7 @@ i915_gem_evict_something(struct drm_device *dev, int min_size) | |||
2172 | #if WATCH_LRU | 2172 | #if WATCH_LRU |
2173 | DRM_INFO("%s: evicting %p\n", __func__, obj); | 2173 | DRM_INFO("%s: evicting %p\n", __func__, obj); |
2174 | #endif | 2174 | #endif |
2175 | obj_priv = obj->driver_private; | 2175 | obj_priv = to_intel_bo(obj); |
2176 | BUG_ON(obj_priv->pin_count != 0); | 2176 | BUG_ON(obj_priv->pin_count != 0); |
2177 | BUG_ON(obj_priv->active); | 2177 | BUG_ON(obj_priv->active); |
2178 | 2178 | ||
@@ -2243,7 +2243,7 @@ int | |||
2243 | i915_gem_object_get_pages(struct drm_gem_object *obj, | 2243 | i915_gem_object_get_pages(struct drm_gem_object *obj, |
2244 | gfp_t gfpmask) | 2244 | gfp_t gfpmask) |
2245 | { | 2245 | { |
2246 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2246 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2247 | int page_count, i; | 2247 | int page_count, i; |
2248 | struct address_space *mapping; | 2248 | struct address_space *mapping; |
2249 | struct inode *inode; | 2249 | struct inode *inode; |
@@ -2296,7 +2296,7 @@ static void sandybridge_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
2296 | struct drm_gem_object *obj = reg->obj; | 2296 | struct drm_gem_object *obj = reg->obj; |
2297 | struct drm_device *dev = obj->dev; | 2297 | struct drm_device *dev = obj->dev; |
2298 | drm_i915_private_t *dev_priv = dev->dev_private; | 2298 | drm_i915_private_t *dev_priv = dev->dev_private; |
2299 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2299 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2300 | int regnum = obj_priv->fence_reg; | 2300 | int regnum = obj_priv->fence_reg; |
2301 | uint64_t val; | 2301 | uint64_t val; |
2302 | 2302 | ||
@@ -2318,7 +2318,7 @@ static void i965_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
2318 | struct drm_gem_object *obj = reg->obj; | 2318 | struct drm_gem_object *obj = reg->obj; |
2319 | struct drm_device *dev = obj->dev; | 2319 | struct drm_device *dev = obj->dev; |
2320 | drm_i915_private_t *dev_priv = dev->dev_private; | 2320 | drm_i915_private_t *dev_priv = dev->dev_private; |
2321 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2321 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2322 | int regnum = obj_priv->fence_reg; | 2322 | int regnum = obj_priv->fence_reg; |
2323 | uint64_t val; | 2323 | uint64_t val; |
2324 | 2324 | ||
@@ -2338,7 +2338,7 @@ static void i915_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
2338 | struct drm_gem_object *obj = reg->obj; | 2338 | struct drm_gem_object *obj = reg->obj; |
2339 | struct drm_device *dev = obj->dev; | 2339 | struct drm_device *dev = obj->dev; |
2340 | drm_i915_private_t *dev_priv = dev->dev_private; | 2340 | drm_i915_private_t *dev_priv = dev->dev_private; |
2341 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2341 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2342 | int regnum = obj_priv->fence_reg; | 2342 | int regnum = obj_priv->fence_reg; |
2343 | int tile_width; | 2343 | int tile_width; |
2344 | uint32_t fence_reg, val; | 2344 | uint32_t fence_reg, val; |
@@ -2380,7 +2380,7 @@ static void i830_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
2380 | struct drm_gem_object *obj = reg->obj; | 2380 | struct drm_gem_object *obj = reg->obj; |
2381 | struct drm_device *dev = obj->dev; | 2381 | struct drm_device *dev = obj->dev; |
2382 | drm_i915_private_t *dev_priv = dev->dev_private; | 2382 | drm_i915_private_t *dev_priv = dev->dev_private; |
2383 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2383 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2384 | int regnum = obj_priv->fence_reg; | 2384 | int regnum = obj_priv->fence_reg; |
2385 | uint32_t val; | 2385 | uint32_t val; |
2386 | uint32_t pitch_val; | 2386 | uint32_t pitch_val; |
@@ -2424,7 +2424,7 @@ static int i915_find_fence_reg(struct drm_device *dev) | |||
2424 | if (!reg->obj) | 2424 | if (!reg->obj) |
2425 | return i; | 2425 | return i; |
2426 | 2426 | ||
2427 | obj_priv = reg->obj->driver_private; | 2427 | obj_priv = to_intel_bo(reg->obj); |
2428 | if (!obj_priv->pin_count) | 2428 | if (!obj_priv->pin_count) |
2429 | avail++; | 2429 | avail++; |
2430 | } | 2430 | } |
@@ -2479,7 +2479,7 @@ i915_gem_object_get_fence_reg(struct drm_gem_object *obj) | |||
2479 | { | 2479 | { |
2480 | struct drm_device *dev = obj->dev; | 2480 | struct drm_device *dev = obj->dev; |
2481 | struct drm_i915_private *dev_priv = dev->dev_private; | 2481 | struct drm_i915_private *dev_priv = dev->dev_private; |
2482 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2482 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2483 | struct drm_i915_fence_reg *reg = NULL; | 2483 | struct drm_i915_fence_reg *reg = NULL; |
2484 | int ret; | 2484 | int ret; |
2485 | 2485 | ||
@@ -2546,7 +2546,7 @@ i915_gem_clear_fence_reg(struct drm_gem_object *obj) | |||
2546 | { | 2546 | { |
2547 | struct drm_device *dev = obj->dev; | 2547 | struct drm_device *dev = obj->dev; |
2548 | drm_i915_private_t *dev_priv = dev->dev_private; | 2548 | drm_i915_private_t *dev_priv = dev->dev_private; |
2549 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2549 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2550 | 2550 | ||
2551 | if (IS_GEN6(dev)) { | 2551 | if (IS_GEN6(dev)) { |
2552 | I915_WRITE64(FENCE_REG_SANDYBRIDGE_0 + | 2552 | I915_WRITE64(FENCE_REG_SANDYBRIDGE_0 + |
@@ -2582,7 +2582,7 @@ int | |||
2582 | i915_gem_object_put_fence_reg(struct drm_gem_object *obj) | 2582 | i915_gem_object_put_fence_reg(struct drm_gem_object *obj) |
2583 | { | 2583 | { |
2584 | struct drm_device *dev = obj->dev; | 2584 | struct drm_device *dev = obj->dev; |
2585 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2585 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2586 | 2586 | ||
2587 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE) | 2587 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE) |
2588 | return 0; | 2588 | return 0; |
@@ -2620,7 +2620,7 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment) | |||
2620 | { | 2620 | { |
2621 | struct drm_device *dev = obj->dev; | 2621 | struct drm_device *dev = obj->dev; |
2622 | drm_i915_private_t *dev_priv = dev->dev_private; | 2622 | drm_i915_private_t *dev_priv = dev->dev_private; |
2623 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2623 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2624 | struct drm_mm_node *free_space; | 2624 | struct drm_mm_node *free_space; |
2625 | gfp_t gfpmask = __GFP_NORETRY | __GFP_NOWARN; | 2625 | gfp_t gfpmask = __GFP_NORETRY | __GFP_NOWARN; |
2626 | int ret; | 2626 | int ret; |
@@ -2727,7 +2727,7 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment) | |||
2727 | void | 2727 | void |
2728 | i915_gem_clflush_object(struct drm_gem_object *obj) | 2728 | i915_gem_clflush_object(struct drm_gem_object *obj) |
2729 | { | 2729 | { |
2730 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2730 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2731 | 2731 | ||
2732 | /* If we don't have a page list set up, then we're not pinned | 2732 | /* If we don't have a page list set up, then we're not pinned |
2733 | * to GPU, and we can ignore the cache flush because it'll happen | 2733 | * to GPU, and we can ignore the cache flush because it'll happen |
@@ -2828,7 +2828,7 @@ i915_gem_object_flush_write_domain(struct drm_gem_object *obj) | |||
2828 | int | 2828 | int |
2829 | i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, int write) | 2829 | i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, int write) |
2830 | { | 2830 | { |
2831 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2831 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2832 | uint32_t old_write_domain, old_read_domains; | 2832 | uint32_t old_write_domain, old_read_domains; |
2833 | int ret; | 2833 | int ret; |
2834 | 2834 | ||
@@ -2878,7 +2878,7 @@ int | |||
2878 | i915_gem_object_set_to_display_plane(struct drm_gem_object *obj) | 2878 | i915_gem_object_set_to_display_plane(struct drm_gem_object *obj) |
2879 | { | 2879 | { |
2880 | struct drm_device *dev = obj->dev; | 2880 | struct drm_device *dev = obj->dev; |
2881 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2881 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
2882 | uint32_t old_write_domain, old_read_domains; | 2882 | uint32_t old_write_domain, old_read_domains; |
2883 | int ret; | 2883 | int ret; |
2884 | 2884 | ||
@@ -3091,7 +3091,7 @@ static void | |||
3091 | i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) | 3091 | i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) |
3092 | { | 3092 | { |
3093 | struct drm_device *dev = obj->dev; | 3093 | struct drm_device *dev = obj->dev; |
3094 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3094 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
3095 | uint32_t invalidate_domains = 0; | 3095 | uint32_t invalidate_domains = 0; |
3096 | uint32_t flush_domains = 0; | 3096 | uint32_t flush_domains = 0; |
3097 | uint32_t old_read_domains; | 3097 | uint32_t old_read_domains; |
@@ -3176,7 +3176,7 @@ i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) | |||
3176 | static void | 3176 | static void |
3177 | i915_gem_object_set_to_full_cpu_read_domain(struct drm_gem_object *obj) | 3177 | i915_gem_object_set_to_full_cpu_read_domain(struct drm_gem_object *obj) |
3178 | { | 3178 | { |
3179 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3179 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
3180 | 3180 | ||
3181 | if (!obj_priv->page_cpu_valid) | 3181 | if (!obj_priv->page_cpu_valid) |
3182 | return; | 3182 | return; |
@@ -3216,7 +3216,7 @@ static int | |||
3216 | i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, | 3216 | i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, |
3217 | uint64_t offset, uint64_t size) | 3217 | uint64_t offset, uint64_t size) |
3218 | { | 3218 | { |
3219 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3219 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
3220 | uint32_t old_read_domains; | 3220 | uint32_t old_read_domains; |
3221 | int i, ret; | 3221 | int i, ret; |
3222 | 3222 | ||
@@ -3285,7 +3285,7 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
3285 | { | 3285 | { |
3286 | struct drm_device *dev = obj->dev; | 3286 | struct drm_device *dev = obj->dev; |
3287 | drm_i915_private_t *dev_priv = dev->dev_private; | 3287 | drm_i915_private_t *dev_priv = dev->dev_private; |
3288 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3288 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
3289 | int i, ret; | 3289 | int i, ret; |
3290 | void __iomem *reloc_page; | 3290 | void __iomem *reloc_page; |
3291 | bool need_fence; | 3291 | bool need_fence; |
@@ -3336,7 +3336,7 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
3336 | i915_gem_object_unpin(obj); | 3336 | i915_gem_object_unpin(obj); |
3337 | return -EBADF; | 3337 | return -EBADF; |
3338 | } | 3338 | } |
3339 | target_obj_priv = target_obj->driver_private; | 3339 | target_obj_priv = to_intel_bo(target_obj); |
3340 | 3340 | ||
3341 | #if WATCH_RELOC | 3341 | #if WATCH_RELOC |
3342 | DRM_INFO("%s: obj %p offset %08x target %d " | 3342 | DRM_INFO("%s: obj %p offset %08x target %d " |
@@ -3688,7 +3688,7 @@ i915_gem_wait_for_pending_flip(struct drm_device *dev, | |||
3688 | prepare_to_wait(&dev_priv->pending_flip_queue, | 3688 | prepare_to_wait(&dev_priv->pending_flip_queue, |
3689 | &wait, TASK_INTERRUPTIBLE); | 3689 | &wait, TASK_INTERRUPTIBLE); |
3690 | for (i = 0; i < count; i++) { | 3690 | for (i = 0; i < count; i++) { |
3691 | obj_priv = object_list[i]->driver_private; | 3691 | obj_priv = to_intel_bo(object_list[i]); |
3692 | if (atomic_read(&obj_priv->pending_flip) > 0) | 3692 | if (atomic_read(&obj_priv->pending_flip) > 0) |
3693 | break; | 3693 | break; |
3694 | } | 3694 | } |
@@ -3797,7 +3797,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
3797 | goto err; | 3797 | goto err; |
3798 | } | 3798 | } |
3799 | 3799 | ||
3800 | obj_priv = object_list[i]->driver_private; | 3800 | obj_priv = to_intel_bo(object_list[i]); |
3801 | if (obj_priv->in_execbuffer) { | 3801 | if (obj_priv->in_execbuffer) { |
3802 | DRM_ERROR("Object %p appears more than once in object list\n", | 3802 | DRM_ERROR("Object %p appears more than once in object list\n", |
3803 | object_list[i]); | 3803 | object_list[i]); |
@@ -3923,7 +3923,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
3923 | 3923 | ||
3924 | for (i = 0; i < args->buffer_count; i++) { | 3924 | for (i = 0; i < args->buffer_count; i++) { |
3925 | struct drm_gem_object *obj = object_list[i]; | 3925 | struct drm_gem_object *obj = object_list[i]; |
3926 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3926 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
3927 | uint32_t old_write_domain = obj->write_domain; | 3927 | uint32_t old_write_domain = obj->write_domain; |
3928 | 3928 | ||
3929 | obj->write_domain = obj->pending_write_domain; | 3929 | obj->write_domain = obj->pending_write_domain; |
@@ -3998,7 +3998,7 @@ err: | |||
3998 | 3998 | ||
3999 | for (i = 0; i < args->buffer_count; i++) { | 3999 | for (i = 0; i < args->buffer_count; i++) { |
4000 | if (object_list[i]) { | 4000 | if (object_list[i]) { |
4001 | obj_priv = object_list[i]->driver_private; | 4001 | obj_priv = to_intel_bo(object_list[i]); |
4002 | obj_priv->in_execbuffer = false; | 4002 | obj_priv->in_execbuffer = false; |
4003 | } | 4003 | } |
4004 | drm_gem_object_unreference(object_list[i]); | 4004 | drm_gem_object_unreference(object_list[i]); |
@@ -4176,7 +4176,7 @@ int | |||
4176 | i915_gem_object_pin(struct drm_gem_object *obj, uint32_t alignment) | 4176 | i915_gem_object_pin(struct drm_gem_object *obj, uint32_t alignment) |
4177 | { | 4177 | { |
4178 | struct drm_device *dev = obj->dev; | 4178 | struct drm_device *dev = obj->dev; |
4179 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4179 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
4180 | int ret; | 4180 | int ret; |
4181 | 4181 | ||
4182 | i915_verify_inactive(dev, __FILE__, __LINE__); | 4182 | i915_verify_inactive(dev, __FILE__, __LINE__); |
@@ -4209,7 +4209,7 @@ i915_gem_object_unpin(struct drm_gem_object *obj) | |||
4209 | { | 4209 | { |
4210 | struct drm_device *dev = obj->dev; | 4210 | struct drm_device *dev = obj->dev; |
4211 | drm_i915_private_t *dev_priv = dev->dev_private; | 4211 | drm_i915_private_t *dev_priv = dev->dev_private; |
4212 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4212 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
4213 | 4213 | ||
4214 | i915_verify_inactive(dev, __FILE__, __LINE__); | 4214 | i915_verify_inactive(dev, __FILE__, __LINE__); |
4215 | obj_priv->pin_count--; | 4215 | obj_priv->pin_count--; |
@@ -4249,7 +4249,7 @@ i915_gem_pin_ioctl(struct drm_device *dev, void *data, | |||
4249 | mutex_unlock(&dev->struct_mutex); | 4249 | mutex_unlock(&dev->struct_mutex); |
4250 | return -EBADF; | 4250 | return -EBADF; |
4251 | } | 4251 | } |
4252 | obj_priv = obj->driver_private; | 4252 | obj_priv = to_intel_bo(obj); |
4253 | 4253 | ||
4254 | if (obj_priv->madv != I915_MADV_WILLNEED) { | 4254 | if (obj_priv->madv != I915_MADV_WILLNEED) { |
4255 | DRM_ERROR("Attempting to pin a purgeable buffer\n"); | 4255 | DRM_ERROR("Attempting to pin a purgeable buffer\n"); |
@@ -4306,7 +4306,7 @@ i915_gem_unpin_ioctl(struct drm_device *dev, void *data, | |||
4306 | return -EBADF; | 4306 | return -EBADF; |
4307 | } | 4307 | } |
4308 | 4308 | ||
4309 | obj_priv = obj->driver_private; | 4309 | obj_priv = to_intel_bo(obj); |
4310 | if (obj_priv->pin_filp != file_priv) { | 4310 | if (obj_priv->pin_filp != file_priv) { |
4311 | DRM_ERROR("Not pinned by caller in i915_gem_pin_ioctl(): %d\n", | 4311 | DRM_ERROR("Not pinned by caller in i915_gem_pin_ioctl(): %d\n", |
4312 | args->handle); | 4312 | args->handle); |
@@ -4348,7 +4348,7 @@ i915_gem_busy_ioctl(struct drm_device *dev, void *data, | |||
4348 | */ | 4348 | */ |
4349 | i915_gem_retire_requests(dev); | 4349 | i915_gem_retire_requests(dev); |
4350 | 4350 | ||
4351 | obj_priv = obj->driver_private; | 4351 | obj_priv = to_intel_bo(obj); |
4352 | /* Don't count being on the flushing list against the object being | 4352 | /* Don't count being on the flushing list against the object being |
4353 | * done. Otherwise, a buffer left on the flushing list but not getting | 4353 | * done. Otherwise, a buffer left on the flushing list but not getting |
4354 | * flushed (because nobody's flushing that domain) won't ever return | 4354 | * flushed (because nobody's flushing that domain) won't ever return |
@@ -4394,7 +4394,7 @@ i915_gem_madvise_ioctl(struct drm_device *dev, void *data, | |||
4394 | } | 4394 | } |
4395 | 4395 | ||
4396 | mutex_lock(&dev->struct_mutex); | 4396 | mutex_lock(&dev->struct_mutex); |
4397 | obj_priv = obj->driver_private; | 4397 | obj_priv = to_intel_bo(obj); |
4398 | 4398 | ||
4399 | if (obj_priv->pin_count) { | 4399 | if (obj_priv->pin_count) { |
4400 | drm_gem_object_unreference(obj); | 4400 | drm_gem_object_unreference(obj); |
@@ -4455,7 +4455,7 @@ int i915_gem_init_object(struct drm_gem_object *obj) | |||
4455 | void i915_gem_free_object(struct drm_gem_object *obj) | 4455 | void i915_gem_free_object(struct drm_gem_object *obj) |
4456 | { | 4456 | { |
4457 | struct drm_device *dev = obj->dev; | 4457 | struct drm_device *dev = obj->dev; |
4458 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4458 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
4459 | 4459 | ||
4460 | trace_i915_gem_object_destroy(obj); | 4460 | trace_i915_gem_object_destroy(obj); |
4461 | 4461 | ||
@@ -4564,7 +4564,7 @@ i915_gem_init_hws(struct drm_device *dev) | |||
4564 | DRM_ERROR("Failed to allocate status page\n"); | 4564 | DRM_ERROR("Failed to allocate status page\n"); |
4565 | return -ENOMEM; | 4565 | return -ENOMEM; |
4566 | } | 4566 | } |
4567 | obj_priv = obj->driver_private; | 4567 | obj_priv = to_intel_bo(obj); |
4568 | obj_priv->agp_type = AGP_USER_CACHED_MEMORY; | 4568 | obj_priv->agp_type = AGP_USER_CACHED_MEMORY; |
4569 | 4569 | ||
4570 | ret = i915_gem_object_pin(obj, 4096); | 4570 | ret = i915_gem_object_pin(obj, 4096); |
@@ -4608,7 +4608,7 @@ i915_gem_cleanup_hws(struct drm_device *dev) | |||
4608 | return; | 4608 | return; |
4609 | 4609 | ||
4610 | obj = dev_priv->hws_obj; | 4610 | obj = dev_priv->hws_obj; |
4611 | obj_priv = obj->driver_private; | 4611 | obj_priv = to_intel_bo(obj); |
4612 | 4612 | ||
4613 | kunmap(obj_priv->pages[0]); | 4613 | kunmap(obj_priv->pages[0]); |
4614 | i915_gem_object_unpin(obj); | 4614 | i915_gem_object_unpin(obj); |
@@ -4642,7 +4642,7 @@ i915_gem_init_ringbuffer(struct drm_device *dev) | |||
4642 | i915_gem_cleanup_hws(dev); | 4642 | i915_gem_cleanup_hws(dev); |
4643 | return -ENOMEM; | 4643 | return -ENOMEM; |
4644 | } | 4644 | } |
4645 | obj_priv = obj->driver_private; | 4645 | obj_priv = to_intel_bo(obj); |
4646 | 4646 | ||
4647 | ret = i915_gem_object_pin(obj, 4096); | 4647 | ret = i915_gem_object_pin(obj, 4096); |
4648 | if (ret != 0) { | 4648 | if (ret != 0) { |
@@ -4935,7 +4935,7 @@ void i915_gem_detach_phys_object(struct drm_device *dev, | |||
4935 | int ret; | 4935 | int ret; |
4936 | int page_count; | 4936 | int page_count; |
4937 | 4937 | ||
4938 | obj_priv = obj->driver_private; | 4938 | obj_priv = to_intel_bo(obj); |
4939 | if (!obj_priv->phys_obj) | 4939 | if (!obj_priv->phys_obj) |
4940 | return; | 4940 | return; |
4941 | 4941 | ||
@@ -4974,7 +4974,7 @@ i915_gem_attach_phys_object(struct drm_device *dev, | |||
4974 | if (id > I915_MAX_PHYS_OBJECT) | 4974 | if (id > I915_MAX_PHYS_OBJECT) |
4975 | return -EINVAL; | 4975 | return -EINVAL; |
4976 | 4976 | ||
4977 | obj_priv = obj->driver_private; | 4977 | obj_priv = to_intel_bo(obj); |
4978 | 4978 | ||
4979 | if (obj_priv->phys_obj) { | 4979 | if (obj_priv->phys_obj) { |
4980 | if (obj_priv->phys_obj->id == id) | 4980 | if (obj_priv->phys_obj->id == id) |
@@ -5025,7 +5025,7 @@ i915_gem_phys_pwrite(struct drm_device *dev, struct drm_gem_object *obj, | |||
5025 | struct drm_i915_gem_pwrite *args, | 5025 | struct drm_i915_gem_pwrite *args, |
5026 | struct drm_file *file_priv) | 5026 | struct drm_file *file_priv) |
5027 | { | 5027 | { |
5028 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 5028 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
5029 | void *obj_addr; | 5029 | void *obj_addr; |
5030 | int ret; | 5030 | int ret; |
5031 | char __user *user_data; | 5031 | char __user *user_data; |
diff --git a/drivers/gpu/drm/i915/i915_gem_debug.c b/drivers/gpu/drm/i915/i915_gem_debug.c index e602614bd3f8..35507cf53fa3 100644 --- a/drivers/gpu/drm/i915/i915_gem_debug.c +++ b/drivers/gpu/drm/i915/i915_gem_debug.c | |||
@@ -72,7 +72,7 @@ void | |||
72 | i915_gem_dump_object(struct drm_gem_object *obj, int len, | 72 | i915_gem_dump_object(struct drm_gem_object *obj, int len, |
73 | const char *where, uint32_t mark) | 73 | const char *where, uint32_t mark) |
74 | { | 74 | { |
75 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 75 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
76 | int page; | 76 | int page; |
77 | 77 | ||
78 | DRM_INFO("%s: object at offset %08x\n", where, obj_priv->gtt_offset); | 78 | DRM_INFO("%s: object at offset %08x\n", where, obj_priv->gtt_offset); |
@@ -137,7 +137,7 @@ void | |||
137 | i915_gem_object_check_coherency(struct drm_gem_object *obj, int handle) | 137 | i915_gem_object_check_coherency(struct drm_gem_object *obj, int handle) |
138 | { | 138 | { |
139 | struct drm_device *dev = obj->dev; | 139 | struct drm_device *dev = obj->dev; |
140 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 140 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
141 | int page; | 141 | int page; |
142 | uint32_t *gtt_mapping; | 142 | uint32_t *gtt_mapping; |
143 | uint32_t *backing_map = NULL; | 143 | uint32_t *backing_map = NULL; |
diff --git a/drivers/gpu/drm/i915/i915_gem_tiling.c b/drivers/gpu/drm/i915/i915_gem_tiling.c index c01c878e51ba..449157f71610 100644 --- a/drivers/gpu/drm/i915/i915_gem_tiling.c +++ b/drivers/gpu/drm/i915/i915_gem_tiling.c | |||
@@ -240,7 +240,7 @@ bool | |||
240 | i915_gem_object_fence_offset_ok(struct drm_gem_object *obj, int tiling_mode) | 240 | i915_gem_object_fence_offset_ok(struct drm_gem_object *obj, int tiling_mode) |
241 | { | 241 | { |
242 | struct drm_device *dev = obj->dev; | 242 | struct drm_device *dev = obj->dev; |
243 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 243 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
244 | 244 | ||
245 | if (obj_priv->gtt_space == NULL) | 245 | if (obj_priv->gtt_space == NULL) |
246 | return true; | 246 | return true; |
@@ -280,7 +280,7 @@ i915_gem_set_tiling(struct drm_device *dev, void *data, | |||
280 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 280 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
281 | if (obj == NULL) | 281 | if (obj == NULL) |
282 | return -EINVAL; | 282 | return -EINVAL; |
283 | obj_priv = obj->driver_private; | 283 | obj_priv = to_intel_bo(obj); |
284 | 284 | ||
285 | if (!i915_tiling_ok(dev, args->stride, obj->size, args->tiling_mode)) { | 285 | if (!i915_tiling_ok(dev, args->stride, obj->size, args->tiling_mode)) { |
286 | drm_gem_object_unreference_unlocked(obj); | 286 | drm_gem_object_unreference_unlocked(obj); |
@@ -364,7 +364,7 @@ i915_gem_get_tiling(struct drm_device *dev, void *data, | |||
364 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 364 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
365 | if (obj == NULL) | 365 | if (obj == NULL) |
366 | return -EINVAL; | 366 | return -EINVAL; |
367 | obj_priv = obj->driver_private; | 367 | obj_priv = to_intel_bo(obj); |
368 | 368 | ||
369 | mutex_lock(&dev->struct_mutex); | 369 | mutex_lock(&dev->struct_mutex); |
370 | 370 | ||
@@ -427,7 +427,7 @@ i915_gem_object_do_bit_17_swizzle(struct drm_gem_object *obj) | |||
427 | { | 427 | { |
428 | struct drm_device *dev = obj->dev; | 428 | struct drm_device *dev = obj->dev; |
429 | drm_i915_private_t *dev_priv = dev->dev_private; | 429 | drm_i915_private_t *dev_priv = dev->dev_private; |
430 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 430 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
431 | int page_count = obj->size >> PAGE_SHIFT; | 431 | int page_count = obj->size >> PAGE_SHIFT; |
432 | int i; | 432 | int i; |
433 | 433 | ||
@@ -456,7 +456,7 @@ i915_gem_object_save_bit_17_swizzle(struct drm_gem_object *obj) | |||
456 | { | 456 | { |
457 | struct drm_device *dev = obj->dev; | 457 | struct drm_device *dev = obj->dev; |
458 | drm_i915_private_t *dev_priv = dev->dev_private; | 458 | drm_i915_private_t *dev_priv = dev->dev_private; |
459 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 459 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
460 | int page_count = obj->size >> PAGE_SHIFT; | 460 | int page_count = obj->size >> PAGE_SHIFT; |
461 | int i; | 461 | int i; |
462 | 462 | ||
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index 5388354da0d1..bfbdad92d73d 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
@@ -443,7 +443,7 @@ i915_error_object_create(struct drm_device *dev, | |||
443 | if (src == NULL) | 443 | if (src == NULL) |
444 | return NULL; | 444 | return NULL; |
445 | 445 | ||
446 | src_priv = src->driver_private; | 446 | src_priv = to_intel_bo(src); |
447 | if (src_priv->pages == NULL) | 447 | if (src_priv->pages == NULL) |
448 | return NULL; | 448 | return NULL; |
449 | 449 | ||
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index f0214908a935..7adb3a54aac6 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
@@ -1002,7 +1002,7 @@ static void i8xx_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
1002 | struct drm_i915_private *dev_priv = dev->dev_private; | 1002 | struct drm_i915_private *dev_priv = dev->dev_private; |
1003 | struct drm_framebuffer *fb = crtc->fb; | 1003 | struct drm_framebuffer *fb = crtc->fb; |
1004 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); | 1004 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); |
1005 | struct drm_i915_gem_object *obj_priv = intel_fb->obj->driver_private; | 1005 | struct drm_i915_gem_object *obj_priv = to_intel_bo(intel_fb->obj); |
1006 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 1006 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
1007 | int plane, i; | 1007 | int plane, i; |
1008 | u32 fbc_ctl, fbc_ctl2; | 1008 | u32 fbc_ctl, fbc_ctl2; |
@@ -1079,7 +1079,7 @@ static void g4x_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
1079 | struct drm_i915_private *dev_priv = dev->dev_private; | 1079 | struct drm_i915_private *dev_priv = dev->dev_private; |
1080 | struct drm_framebuffer *fb = crtc->fb; | 1080 | struct drm_framebuffer *fb = crtc->fb; |
1081 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); | 1081 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); |
1082 | struct drm_i915_gem_object *obj_priv = intel_fb->obj->driver_private; | 1082 | struct drm_i915_gem_object *obj_priv = to_intel_bo(intel_fb->obj); |
1083 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 1083 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
1084 | int plane = (intel_crtc->plane == 0 ? DPFC_CTL_PLANEA : | 1084 | int plane = (intel_crtc->plane == 0 ? DPFC_CTL_PLANEA : |
1085 | DPFC_CTL_PLANEB); | 1085 | DPFC_CTL_PLANEB); |
@@ -1175,7 +1175,7 @@ static void intel_update_fbc(struct drm_crtc *crtc, | |||
1175 | return; | 1175 | return; |
1176 | 1176 | ||
1177 | intel_fb = to_intel_framebuffer(fb); | 1177 | intel_fb = to_intel_framebuffer(fb); |
1178 | obj_priv = intel_fb->obj->driver_private; | 1178 | obj_priv = to_intel_bo(intel_fb->obj); |
1179 | 1179 | ||
1180 | /* | 1180 | /* |
1181 | * If FBC is already on, we just have to verify that we can | 1181 | * If FBC is already on, we just have to verify that we can |
@@ -1242,7 +1242,7 @@ out_disable: | |||
1242 | static int | 1242 | static int |
1243 | intel_pin_and_fence_fb_obj(struct drm_device *dev, struct drm_gem_object *obj) | 1243 | intel_pin_and_fence_fb_obj(struct drm_device *dev, struct drm_gem_object *obj) |
1244 | { | 1244 | { |
1245 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1245 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
1246 | u32 alignment; | 1246 | u32 alignment; |
1247 | int ret; | 1247 | int ret; |
1248 | 1248 | ||
@@ -1322,7 +1322,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
1322 | 1322 | ||
1323 | intel_fb = to_intel_framebuffer(crtc->fb); | 1323 | intel_fb = to_intel_framebuffer(crtc->fb); |
1324 | obj = intel_fb->obj; | 1324 | obj = intel_fb->obj; |
1325 | obj_priv = obj->driver_private; | 1325 | obj_priv = to_intel_bo(obj); |
1326 | 1326 | ||
1327 | mutex_lock(&dev->struct_mutex); | 1327 | mutex_lock(&dev->struct_mutex); |
1328 | ret = intel_pin_and_fence_fb_obj(dev, obj); | 1328 | ret = intel_pin_and_fence_fb_obj(dev, obj); |
@@ -1400,7 +1400,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
1400 | 1400 | ||
1401 | if (old_fb) { | 1401 | if (old_fb) { |
1402 | intel_fb = to_intel_framebuffer(old_fb); | 1402 | intel_fb = to_intel_framebuffer(old_fb); |
1403 | obj_priv = intel_fb->obj->driver_private; | 1403 | obj_priv = to_intel_bo(intel_fb->obj); |
1404 | i915_gem_object_unpin(intel_fb->obj); | 1404 | i915_gem_object_unpin(intel_fb->obj); |
1405 | } | 1405 | } |
1406 | intel_increase_pllclock(crtc, true); | 1406 | intel_increase_pllclock(crtc, true); |
@@ -3510,7 +3510,7 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc, | |||
3510 | if (!bo) | 3510 | if (!bo) |
3511 | return -ENOENT; | 3511 | return -ENOENT; |
3512 | 3512 | ||
3513 | obj_priv = bo->driver_private; | 3513 | obj_priv = to_intel_bo(bo); |
3514 | 3514 | ||
3515 | if (bo->size < width * height * 4) { | 3515 | if (bo->size < width * height * 4) { |
3516 | DRM_ERROR("buffer is to small\n"); | 3516 | DRM_ERROR("buffer is to small\n"); |
@@ -4155,7 +4155,7 @@ void intel_finish_page_flip(struct drm_device *dev, int pipe) | |||
4155 | work = intel_crtc->unpin_work; | 4155 | work = intel_crtc->unpin_work; |
4156 | if (work == NULL || !work->pending) { | 4156 | if (work == NULL || !work->pending) { |
4157 | if (work && !work->pending) { | 4157 | if (work && !work->pending) { |
4158 | obj_priv = work->pending_flip_obj->driver_private; | 4158 | obj_priv = to_intel_bo(work->pending_flip_obj); |
4159 | DRM_DEBUG_DRIVER("flip finish: %p (%d) not pending?\n", | 4159 | DRM_DEBUG_DRIVER("flip finish: %p (%d) not pending?\n", |
4160 | obj_priv, | 4160 | obj_priv, |
4161 | atomic_read(&obj_priv->pending_flip)); | 4161 | atomic_read(&obj_priv->pending_flip)); |
@@ -4180,7 +4180,7 @@ void intel_finish_page_flip(struct drm_device *dev, int pipe) | |||
4180 | 4180 | ||
4181 | spin_unlock_irqrestore(&dev->event_lock, flags); | 4181 | spin_unlock_irqrestore(&dev->event_lock, flags); |
4182 | 4182 | ||
4183 | obj_priv = work->pending_flip_obj->driver_private; | 4183 | obj_priv = to_intel_bo(work->pending_flip_obj); |
4184 | 4184 | ||
4185 | /* Initial scanout buffer will have a 0 pending flip count */ | 4185 | /* Initial scanout buffer will have a 0 pending flip count */ |
4186 | if ((atomic_read(&obj_priv->pending_flip) == 0) || | 4186 | if ((atomic_read(&obj_priv->pending_flip) == 0) || |
@@ -4251,7 +4251,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc, | |||
4251 | ret = intel_pin_and_fence_fb_obj(dev, obj); | 4251 | ret = intel_pin_and_fence_fb_obj(dev, obj); |
4252 | if (ret != 0) { | 4252 | if (ret != 0) { |
4253 | DRM_DEBUG_DRIVER("flip queue: %p pin & fence failed\n", | 4253 | DRM_DEBUG_DRIVER("flip queue: %p pin & fence failed\n", |
4254 | obj->driver_private); | 4254 | to_intel_bo(obj)); |
4255 | kfree(work); | 4255 | kfree(work); |
4256 | intel_crtc->unpin_work = NULL; | 4256 | intel_crtc->unpin_work = NULL; |
4257 | mutex_unlock(&dev->struct_mutex); | 4257 | mutex_unlock(&dev->struct_mutex); |
@@ -4265,7 +4265,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc, | |||
4265 | crtc->fb = fb; | 4265 | crtc->fb = fb; |
4266 | i915_gem_object_flush_write_domain(obj); | 4266 | i915_gem_object_flush_write_domain(obj); |
4267 | drm_vblank_get(dev, intel_crtc->pipe); | 4267 | drm_vblank_get(dev, intel_crtc->pipe); |
4268 | obj_priv = obj->driver_private; | 4268 | obj_priv = to_intel_bo(obj); |
4269 | atomic_inc(&obj_priv->pending_flip); | 4269 | atomic_inc(&obj_priv->pending_flip); |
4270 | work->pending_flip_obj = obj; | 4270 | work->pending_flip_obj = obj; |
4271 | 4271 | ||
@@ -4778,14 +4778,14 @@ void intel_init_clock_gating(struct drm_device *dev) | |||
4778 | struct drm_i915_gem_object *obj_priv = NULL; | 4778 | struct drm_i915_gem_object *obj_priv = NULL; |
4779 | 4779 | ||
4780 | if (dev_priv->pwrctx) { | 4780 | if (dev_priv->pwrctx) { |
4781 | obj_priv = dev_priv->pwrctx->driver_private; | 4781 | obj_priv = to_intel_bo(dev_priv->pwrctx); |
4782 | } else { | 4782 | } else { |
4783 | struct drm_gem_object *pwrctx; | 4783 | struct drm_gem_object *pwrctx; |
4784 | 4784 | ||
4785 | pwrctx = intel_alloc_power_context(dev); | 4785 | pwrctx = intel_alloc_power_context(dev); |
4786 | if (pwrctx) { | 4786 | if (pwrctx) { |
4787 | dev_priv->pwrctx = pwrctx; | 4787 | dev_priv->pwrctx = pwrctx; |
4788 | obj_priv = pwrctx->driver_private; | 4788 | obj_priv = to_intel_bo(pwrctx); |
4789 | } | 4789 | } |
4790 | } | 4790 | } |
4791 | 4791 | ||
@@ -4956,7 +4956,7 @@ void intel_modeset_cleanup(struct drm_device *dev) | |||
4956 | if (dev_priv->pwrctx) { | 4956 | if (dev_priv->pwrctx) { |
4957 | struct drm_i915_gem_object *obj_priv; | 4957 | struct drm_i915_gem_object *obj_priv; |
4958 | 4958 | ||
4959 | obj_priv = dev_priv->pwrctx->driver_private; | 4959 | obj_priv = to_intel_bo(dev_priv->pwrctx); |
4960 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset &~ PWRCTX_EN); | 4960 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset &~ PWRCTX_EN); |
4961 | I915_READ(PWRCTXA); | 4961 | I915_READ(PWRCTXA); |
4962 | i915_gem_object_unpin(dev_priv->pwrctx); | 4962 | i915_gem_object_unpin(dev_priv->pwrctx); |
diff --git a/drivers/gpu/drm/i915/intel_fb.c b/drivers/gpu/drm/i915/intel_fb.c index 8cd791dc5b29..c9fbdfa9c575 100644 --- a/drivers/gpu/drm/i915/intel_fb.c +++ b/drivers/gpu/drm/i915/intel_fb.c | |||
@@ -145,7 +145,7 @@ static int intelfb_create(struct drm_device *dev, uint32_t fb_width, | |||
145 | ret = -ENOMEM; | 145 | ret = -ENOMEM; |
146 | goto out; | 146 | goto out; |
147 | } | 147 | } |
148 | obj_priv = fbo->driver_private; | 148 | obj_priv = to_intel_bo(fbo); |
149 | 149 | ||
150 | mutex_lock(&dev->struct_mutex); | 150 | mutex_lock(&dev->struct_mutex); |
151 | 151 | ||
diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c index 60595fc26fdd..6d524a1fc271 100644 --- a/drivers/gpu/drm/i915/intel_overlay.c +++ b/drivers/gpu/drm/i915/intel_overlay.c | |||
@@ -724,7 +724,7 @@ int intel_overlay_do_put_image(struct intel_overlay *overlay, | |||
724 | int ret, tmp_width; | 724 | int ret, tmp_width; |
725 | struct overlay_registers *regs; | 725 | struct overlay_registers *regs; |
726 | bool scale_changed = false; | 726 | bool scale_changed = false; |
727 | struct drm_i915_gem_object *bo_priv = new_bo->driver_private; | 727 | struct drm_i915_gem_object *bo_priv = to_intel_bo(new_bo); |
728 | struct drm_device *dev = overlay->dev; | 728 | struct drm_device *dev = overlay->dev; |
729 | 729 | ||
730 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); | 730 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); |
@@ -809,7 +809,7 @@ int intel_overlay_do_put_image(struct intel_overlay *overlay, | |||
809 | intel_overlay_continue(overlay, scale_changed); | 809 | intel_overlay_continue(overlay, scale_changed); |
810 | 810 | ||
811 | overlay->old_vid_bo = overlay->vid_bo; | 811 | overlay->old_vid_bo = overlay->vid_bo; |
812 | overlay->vid_bo = new_bo->driver_private; | 812 | overlay->vid_bo = to_intel_bo(new_bo); |
813 | 813 | ||
814 | return 0; | 814 | return 0; |
815 | 815 | ||
@@ -1344,7 +1344,7 @@ void intel_setup_overlay(struct drm_device *dev) | |||
1344 | reg_bo = drm_gem_object_alloc(dev, PAGE_SIZE); | 1344 | reg_bo = drm_gem_object_alloc(dev, PAGE_SIZE); |
1345 | if (!reg_bo) | 1345 | if (!reg_bo) |
1346 | goto out_free; | 1346 | goto out_free; |
1347 | overlay->reg_bo = reg_bo->driver_private; | 1347 | overlay->reg_bo = to_intel_bo(reg_bo); |
1348 | 1348 | ||
1349 | if (OVERLAY_NONPHYSICAL(dev)) { | 1349 | if (OVERLAY_NONPHYSICAL(dev)) { |
1350 | ret = i915_gem_object_pin(reg_bo, PAGE_SIZE); | 1350 | ret = i915_gem_object_pin(reg_bo, PAGE_SIZE); |