diff options
| -rw-r--r-- | drivers/gpu/drm/i915/i915_debugfs.c | 2 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_drv.c | 2 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_drv.h | 2 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 132 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem_debug.c | 4 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem_tiling.c | 10 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_irq.c | 2 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_display.c | 28 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_fb.c | 2 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_overlay.c | 6 |
10 files changed, 96 insertions, 94 deletions
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index 1376dfe44c95..bb3a4a8aba08 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c | |||
| @@ -225,7 +225,7 @@ static int i915_gem_fence_regs_info(struct seq_file *m, void *data) | |||
| 225 | } else { | 225 | } else { |
| 226 | struct drm_i915_gem_object *obj_priv; | 226 | struct drm_i915_gem_object *obj_priv; |
| 227 | 227 | ||
| 228 | obj_priv = obj->driver_private; | 228 | obj_priv = to_intel_bo(obj); |
| 229 | seq_printf(m, "Fenced object[%2d] = %p: %s " | 229 | seq_printf(m, "Fenced object[%2d] = %p: %s " |
| 230 | "%08x %08zx %08x %s %08x %08x %d", | 230 | "%08x %08zx %08x %s %08x %08x %d", |
| 231 | i, obj, get_pin_flag(obj_priv), | 231 | i, obj, get_pin_flag(obj_priv), |
diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c index 1a39ec75d76b..0af3dcc85ce9 100644 --- a/drivers/gpu/drm/i915/i915_drv.c +++ b/drivers/gpu/drm/i915/i915_drv.c | |||
| @@ -361,7 +361,7 @@ int i965_reset(struct drm_device *dev, u8 flags) | |||
| 361 | !dev_priv->mm.suspended) { | 361 | !dev_priv->mm.suspended) { |
| 362 | drm_i915_ring_buffer_t *ring = &dev_priv->ring; | 362 | drm_i915_ring_buffer_t *ring = &dev_priv->ring; |
| 363 | struct drm_gem_object *obj = ring->ring_obj; | 363 | struct drm_gem_object *obj = ring->ring_obj; |
| 364 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 364 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 365 | dev_priv->mm.suspended = 0; | 365 | dev_priv->mm.suspended = 0; |
| 366 | 366 | ||
| 367 | /* Stop the ring if it's running. */ | 367 | /* Stop the ring if it's running. */ |
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index aba8260fbc5e..b7cb4aadd059 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
| @@ -731,6 +731,8 @@ struct drm_i915_gem_object { | |||
| 731 | atomic_t pending_flip; | 731 | atomic_t pending_flip; |
| 732 | }; | 732 | }; |
| 733 | 733 | ||
| 734 | #define to_intel_bo(x) ((struct drm_i915_gem_object *) (x)->driver_private) | ||
| 735 | |||
| 734 | /** | 736 | /** |
| 735 | * Request queue structure. | 737 | * Request queue structure. |
| 736 | * | 738 | * |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 933e865a8929..b85727ce308e 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
| @@ -162,7 +162,7 @@ fast_shmem_read(struct page **pages, | |||
| 162 | static int i915_gem_object_needs_bit17_swizzle(struct drm_gem_object *obj) | 162 | static int i915_gem_object_needs_bit17_swizzle(struct drm_gem_object *obj) |
| 163 | { | 163 | { |
| 164 | drm_i915_private_t *dev_priv = obj->dev->dev_private; | 164 | drm_i915_private_t *dev_priv = obj->dev->dev_private; |
| 165 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 165 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 166 | 166 | ||
| 167 | return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && | 167 | return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && |
| 168 | obj_priv->tiling_mode != I915_TILING_NONE; | 168 | obj_priv->tiling_mode != I915_TILING_NONE; |
| @@ -263,7 +263,7 @@ i915_gem_shmem_pread_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 263 | struct drm_i915_gem_pread *args, | 263 | struct drm_i915_gem_pread *args, |
| 264 | struct drm_file *file_priv) | 264 | struct drm_file *file_priv) |
| 265 | { | 265 | { |
| 266 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 266 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 267 | ssize_t remain; | 267 | ssize_t remain; |
| 268 | loff_t offset, page_base; | 268 | loff_t offset, page_base; |
| 269 | char __user *user_data; | 269 | char __user *user_data; |
| @@ -284,7 +284,7 @@ i915_gem_shmem_pread_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 284 | if (ret != 0) | 284 | if (ret != 0) |
| 285 | goto fail_put_pages; | 285 | goto fail_put_pages; |
| 286 | 286 | ||
| 287 | obj_priv = obj->driver_private; | 287 | obj_priv = to_intel_bo(obj); |
| 288 | offset = args->offset; | 288 | offset = args->offset; |
| 289 | 289 | ||
| 290 | while (remain > 0) { | 290 | while (remain > 0) { |
| @@ -353,7 +353,7 @@ i915_gem_shmem_pread_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 353 | struct drm_i915_gem_pread *args, | 353 | struct drm_i915_gem_pread *args, |
| 354 | struct drm_file *file_priv) | 354 | struct drm_file *file_priv) |
| 355 | { | 355 | { |
| 356 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 356 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 357 | struct mm_struct *mm = current->mm; | 357 | struct mm_struct *mm = current->mm; |
| 358 | struct page **user_pages; | 358 | struct page **user_pages; |
| 359 | ssize_t remain; | 359 | ssize_t remain; |
| @@ -402,7 +402,7 @@ i915_gem_shmem_pread_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 402 | if (ret != 0) | 402 | if (ret != 0) |
| 403 | goto fail_put_pages; | 403 | goto fail_put_pages; |
| 404 | 404 | ||
| 405 | obj_priv = obj->driver_private; | 405 | obj_priv = to_intel_bo(obj); |
| 406 | offset = args->offset; | 406 | offset = args->offset; |
| 407 | 407 | ||
| 408 | while (remain > 0) { | 408 | while (remain > 0) { |
| @@ -478,7 +478,7 @@ i915_gem_pread_ioctl(struct drm_device *dev, void *data, | |||
| 478 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 478 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
| 479 | if (obj == NULL) | 479 | if (obj == NULL) |
| 480 | return -EBADF; | 480 | return -EBADF; |
| 481 | obj_priv = obj->driver_private; | 481 | obj_priv = to_intel_bo(obj); |
| 482 | 482 | ||
| 483 | /* Bounds check source. | 483 | /* Bounds check source. |
| 484 | * | 484 | * |
| @@ -580,7 +580,7 @@ i915_gem_gtt_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 580 | struct drm_i915_gem_pwrite *args, | 580 | struct drm_i915_gem_pwrite *args, |
| 581 | struct drm_file *file_priv) | 581 | struct drm_file *file_priv) |
| 582 | { | 582 | { |
| 583 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 583 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 584 | drm_i915_private_t *dev_priv = dev->dev_private; | 584 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 585 | ssize_t remain; | 585 | ssize_t remain; |
| 586 | loff_t offset, page_base; | 586 | loff_t offset, page_base; |
| @@ -604,7 +604,7 @@ i915_gem_gtt_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 604 | if (ret) | 604 | if (ret) |
| 605 | goto fail; | 605 | goto fail; |
| 606 | 606 | ||
| 607 | obj_priv = obj->driver_private; | 607 | obj_priv = to_intel_bo(obj); |
| 608 | offset = obj_priv->gtt_offset + args->offset; | 608 | offset = obj_priv->gtt_offset + args->offset; |
| 609 | 609 | ||
| 610 | while (remain > 0) { | 610 | while (remain > 0) { |
| @@ -654,7 +654,7 @@ i915_gem_gtt_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 654 | struct drm_i915_gem_pwrite *args, | 654 | struct drm_i915_gem_pwrite *args, |
| 655 | struct drm_file *file_priv) | 655 | struct drm_file *file_priv) |
| 656 | { | 656 | { |
| 657 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 657 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 658 | drm_i915_private_t *dev_priv = dev->dev_private; | 658 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 659 | ssize_t remain; | 659 | ssize_t remain; |
| 660 | loff_t gtt_page_base, offset; | 660 | loff_t gtt_page_base, offset; |
| @@ -698,7 +698,7 @@ i915_gem_gtt_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 698 | if (ret) | 698 | if (ret) |
| 699 | goto out_unpin_object; | 699 | goto out_unpin_object; |
| 700 | 700 | ||
| 701 | obj_priv = obj->driver_private; | 701 | obj_priv = to_intel_bo(obj); |
| 702 | offset = obj_priv->gtt_offset + args->offset; | 702 | offset = obj_priv->gtt_offset + args->offset; |
| 703 | 703 | ||
| 704 | while (remain > 0) { | 704 | while (remain > 0) { |
| @@ -760,7 +760,7 @@ i915_gem_shmem_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 760 | struct drm_i915_gem_pwrite *args, | 760 | struct drm_i915_gem_pwrite *args, |
| 761 | struct drm_file *file_priv) | 761 | struct drm_file *file_priv) |
| 762 | { | 762 | { |
| 763 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 763 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 764 | ssize_t remain; | 764 | ssize_t remain; |
| 765 | loff_t offset, page_base; | 765 | loff_t offset, page_base; |
| 766 | char __user *user_data; | 766 | char __user *user_data; |
| @@ -780,7 +780,7 @@ i915_gem_shmem_pwrite_fast(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 780 | if (ret != 0) | 780 | if (ret != 0) |
| 781 | goto fail_put_pages; | 781 | goto fail_put_pages; |
| 782 | 782 | ||
| 783 | obj_priv = obj->driver_private; | 783 | obj_priv = to_intel_bo(obj); |
| 784 | offset = args->offset; | 784 | offset = args->offset; |
| 785 | obj_priv->dirty = 1; | 785 | obj_priv->dirty = 1; |
| 786 | 786 | ||
| @@ -828,7 +828,7 @@ i915_gem_shmem_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 828 | struct drm_i915_gem_pwrite *args, | 828 | struct drm_i915_gem_pwrite *args, |
| 829 | struct drm_file *file_priv) | 829 | struct drm_file *file_priv) |
| 830 | { | 830 | { |
| 831 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 831 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 832 | struct mm_struct *mm = current->mm; | 832 | struct mm_struct *mm = current->mm; |
| 833 | struct page **user_pages; | 833 | struct page **user_pages; |
| 834 | ssize_t remain; | 834 | ssize_t remain; |
| @@ -876,7 +876,7 @@ i915_gem_shmem_pwrite_slow(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 876 | if (ret != 0) | 876 | if (ret != 0) |
| 877 | goto fail_put_pages; | 877 | goto fail_put_pages; |
| 878 | 878 | ||
| 879 | obj_priv = obj->driver_private; | 879 | obj_priv = to_intel_bo(obj); |
| 880 | offset = args->offset; | 880 | offset = args->offset; |
| 881 | obj_priv->dirty = 1; | 881 | obj_priv->dirty = 1; |
| 882 | 882 | ||
| @@ -951,7 +951,7 @@ i915_gem_pwrite_ioctl(struct drm_device *dev, void *data, | |||
| 951 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 951 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
| 952 | if (obj == NULL) | 952 | if (obj == NULL) |
| 953 | return -EBADF; | 953 | return -EBADF; |
| 954 | obj_priv = obj->driver_private; | 954 | obj_priv = to_intel_bo(obj); |
| 955 | 955 | ||
| 956 | /* Bounds check destination. | 956 | /* Bounds check destination. |
| 957 | * | 957 | * |
| @@ -1033,7 +1033,7 @@ i915_gem_set_domain_ioctl(struct drm_device *dev, void *data, | |||
| 1033 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 1033 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
| 1034 | if (obj == NULL) | 1034 | if (obj == NULL) |
| 1035 | return -EBADF; | 1035 | return -EBADF; |
| 1036 | obj_priv = obj->driver_private; | 1036 | obj_priv = to_intel_bo(obj); |
| 1037 | 1037 | ||
| 1038 | mutex_lock(&dev->struct_mutex); | 1038 | mutex_lock(&dev->struct_mutex); |
| 1039 | 1039 | ||
| @@ -1095,7 +1095,7 @@ i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data, | |||
| 1095 | DRM_INFO("%s: sw_finish %d (%p %zd)\n", | 1095 | DRM_INFO("%s: sw_finish %d (%p %zd)\n", |
| 1096 | __func__, args->handle, obj, obj->size); | 1096 | __func__, args->handle, obj, obj->size); |
| 1097 | #endif | 1097 | #endif |
| 1098 | obj_priv = obj->driver_private; | 1098 | obj_priv = to_intel_bo(obj); |
| 1099 | 1099 | ||
| 1100 | /* Pinned buffers may be scanout, so flush the cache */ | 1100 | /* Pinned buffers may be scanout, so flush the cache */ |
| 1101 | if (obj_priv->pin_count) | 1101 | if (obj_priv->pin_count) |
| @@ -1166,7 +1166,7 @@ int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
| 1166 | struct drm_gem_object *obj = vma->vm_private_data; | 1166 | struct drm_gem_object *obj = vma->vm_private_data; |
| 1167 | struct drm_device *dev = obj->dev; | 1167 | struct drm_device *dev = obj->dev; |
| 1168 | struct drm_i915_private *dev_priv = dev->dev_private; | 1168 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1169 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1169 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1170 | pgoff_t page_offset; | 1170 | pgoff_t page_offset; |
| 1171 | unsigned long pfn; | 1171 | unsigned long pfn; |
| 1172 | int ret = 0; | 1172 | int ret = 0; |
| @@ -1233,7 +1233,7 @@ i915_gem_create_mmap_offset(struct drm_gem_object *obj) | |||
| 1233 | { | 1233 | { |
| 1234 | struct drm_device *dev = obj->dev; | 1234 | struct drm_device *dev = obj->dev; |
| 1235 | struct drm_gem_mm *mm = dev->mm_private; | 1235 | struct drm_gem_mm *mm = dev->mm_private; |
| 1236 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1236 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1237 | struct drm_map_list *list; | 1237 | struct drm_map_list *list; |
| 1238 | struct drm_local_map *map; | 1238 | struct drm_local_map *map; |
| 1239 | int ret = 0; | 1239 | int ret = 0; |
| @@ -1304,7 +1304,7 @@ void | |||
| 1304 | i915_gem_release_mmap(struct drm_gem_object *obj) | 1304 | i915_gem_release_mmap(struct drm_gem_object *obj) |
| 1305 | { | 1305 | { |
| 1306 | struct drm_device *dev = obj->dev; | 1306 | struct drm_device *dev = obj->dev; |
| 1307 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1307 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1308 | 1308 | ||
| 1309 | if (dev->dev_mapping) | 1309 | if (dev->dev_mapping) |
| 1310 | unmap_mapping_range(dev->dev_mapping, | 1310 | unmap_mapping_range(dev->dev_mapping, |
| @@ -1315,7 +1315,7 @@ static void | |||
| 1315 | i915_gem_free_mmap_offset(struct drm_gem_object *obj) | 1315 | i915_gem_free_mmap_offset(struct drm_gem_object *obj) |
| 1316 | { | 1316 | { |
| 1317 | struct drm_device *dev = obj->dev; | 1317 | struct drm_device *dev = obj->dev; |
| 1318 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1318 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1319 | struct drm_gem_mm *mm = dev->mm_private; | 1319 | struct drm_gem_mm *mm = dev->mm_private; |
| 1320 | struct drm_map_list *list; | 1320 | struct drm_map_list *list; |
| 1321 | 1321 | ||
| @@ -1346,7 +1346,7 @@ static uint32_t | |||
| 1346 | i915_gem_get_gtt_alignment(struct drm_gem_object *obj) | 1346 | i915_gem_get_gtt_alignment(struct drm_gem_object *obj) |
| 1347 | { | 1347 | { |
| 1348 | struct drm_device *dev = obj->dev; | 1348 | struct drm_device *dev = obj->dev; |
| 1349 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1349 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1350 | int start, i; | 1350 | int start, i; |
| 1351 | 1351 | ||
| 1352 | /* | 1352 | /* |
| @@ -1405,7 +1405,7 @@ i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, | |||
| 1405 | 1405 | ||
| 1406 | mutex_lock(&dev->struct_mutex); | 1406 | mutex_lock(&dev->struct_mutex); |
| 1407 | 1407 | ||
| 1408 | obj_priv = obj->driver_private; | 1408 | obj_priv = to_intel_bo(obj); |
| 1409 | 1409 | ||
| 1410 | if (obj_priv->madv != I915_MADV_WILLNEED) { | 1410 | if (obj_priv->madv != I915_MADV_WILLNEED) { |
| 1411 | DRM_ERROR("Attempting to mmap a purgeable buffer\n"); | 1411 | DRM_ERROR("Attempting to mmap a purgeable buffer\n"); |
| @@ -1449,7 +1449,7 @@ i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, | |||
| 1449 | void | 1449 | void |
| 1450 | i915_gem_object_put_pages(struct drm_gem_object *obj) | 1450 | i915_gem_object_put_pages(struct drm_gem_object *obj) |
| 1451 | { | 1451 | { |
| 1452 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1452 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1453 | int page_count = obj->size / PAGE_SIZE; | 1453 | int page_count = obj->size / PAGE_SIZE; |
| 1454 | int i; | 1454 | int i; |
| 1455 | 1455 | ||
| @@ -1485,7 +1485,7 @@ i915_gem_object_move_to_active(struct drm_gem_object *obj, uint32_t seqno) | |||
| 1485 | { | 1485 | { |
| 1486 | struct drm_device *dev = obj->dev; | 1486 | struct drm_device *dev = obj->dev; |
| 1487 | drm_i915_private_t *dev_priv = dev->dev_private; | 1487 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1488 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1488 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1489 | 1489 | ||
| 1490 | /* Add a reference if we're newly entering the active list. */ | 1490 | /* Add a reference if we're newly entering the active list. */ |
| 1491 | if (!obj_priv->active) { | 1491 | if (!obj_priv->active) { |
| @@ -1505,7 +1505,7 @@ i915_gem_object_move_to_flushing(struct drm_gem_object *obj) | |||
| 1505 | { | 1505 | { |
| 1506 | struct drm_device *dev = obj->dev; | 1506 | struct drm_device *dev = obj->dev; |
| 1507 | drm_i915_private_t *dev_priv = dev->dev_private; | 1507 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1508 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1508 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1509 | 1509 | ||
| 1510 | BUG_ON(!obj_priv->active); | 1510 | BUG_ON(!obj_priv->active); |
| 1511 | list_move_tail(&obj_priv->list, &dev_priv->mm.flushing_list); | 1511 | list_move_tail(&obj_priv->list, &dev_priv->mm.flushing_list); |
| @@ -1516,7 +1516,7 @@ i915_gem_object_move_to_flushing(struct drm_gem_object *obj) | |||
| 1516 | static void | 1516 | static void |
| 1517 | i915_gem_object_truncate(struct drm_gem_object *obj) | 1517 | i915_gem_object_truncate(struct drm_gem_object *obj) |
| 1518 | { | 1518 | { |
| 1519 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1519 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1520 | struct inode *inode; | 1520 | struct inode *inode; |
| 1521 | 1521 | ||
| 1522 | inode = obj->filp->f_path.dentry->d_inode; | 1522 | inode = obj->filp->f_path.dentry->d_inode; |
| @@ -1537,7 +1537,7 @@ i915_gem_object_move_to_inactive(struct drm_gem_object *obj) | |||
| 1537 | { | 1537 | { |
| 1538 | struct drm_device *dev = obj->dev; | 1538 | struct drm_device *dev = obj->dev; |
| 1539 | drm_i915_private_t *dev_priv = dev->dev_private; | 1539 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1540 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1540 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1541 | 1541 | ||
| 1542 | i915_verify_inactive(dev, __FILE__, __LINE__); | 1542 | i915_verify_inactive(dev, __FILE__, __LINE__); |
| 1543 | if (obj_priv->pin_count != 0) | 1543 | if (obj_priv->pin_count != 0) |
| @@ -1964,7 +1964,7 @@ static int | |||
| 1964 | i915_gem_object_wait_rendering(struct drm_gem_object *obj) | 1964 | i915_gem_object_wait_rendering(struct drm_gem_object *obj) |
| 1965 | { | 1965 | { |
| 1966 | struct drm_device *dev = obj->dev; | 1966 | struct drm_device *dev = obj->dev; |
| 1967 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1967 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1968 | int ret; | 1968 | int ret; |
| 1969 | 1969 | ||
| 1970 | /* This function only exists to support waiting for existing rendering, | 1970 | /* This function only exists to support waiting for existing rendering, |
| @@ -1996,7 +1996,7 @@ i915_gem_object_unbind(struct drm_gem_object *obj) | |||
| 1996 | { | 1996 | { |
| 1997 | struct drm_device *dev = obj->dev; | 1997 | struct drm_device *dev = obj->dev; |
| 1998 | drm_i915_private_t *dev_priv = dev->dev_private; | 1998 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1999 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1999 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2000 | int ret = 0; | 2000 | int ret = 0; |
| 2001 | 2001 | ||
| 2002 | #if WATCH_BUF | 2002 | #if WATCH_BUF |
| @@ -2172,7 +2172,7 @@ i915_gem_evict_something(struct drm_device *dev, int min_size) | |||
| 2172 | #if WATCH_LRU | 2172 | #if WATCH_LRU |
| 2173 | DRM_INFO("%s: evicting %p\n", __func__, obj); | 2173 | DRM_INFO("%s: evicting %p\n", __func__, obj); |
| 2174 | #endif | 2174 | #endif |
| 2175 | obj_priv = obj->driver_private; | 2175 | obj_priv = to_intel_bo(obj); |
| 2176 | BUG_ON(obj_priv->pin_count != 0); | 2176 | BUG_ON(obj_priv->pin_count != 0); |
| 2177 | BUG_ON(obj_priv->active); | 2177 | BUG_ON(obj_priv->active); |
| 2178 | 2178 | ||
| @@ -2243,7 +2243,7 @@ int | |||
| 2243 | i915_gem_object_get_pages(struct drm_gem_object *obj, | 2243 | i915_gem_object_get_pages(struct drm_gem_object *obj, |
| 2244 | gfp_t gfpmask) | 2244 | gfp_t gfpmask) |
| 2245 | { | 2245 | { |
| 2246 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2246 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2247 | int page_count, i; | 2247 | int page_count, i; |
| 2248 | struct address_space *mapping; | 2248 | struct address_space *mapping; |
| 2249 | struct inode *inode; | 2249 | struct inode *inode; |
| @@ -2296,7 +2296,7 @@ static void sandybridge_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
| 2296 | struct drm_gem_object *obj = reg->obj; | 2296 | struct drm_gem_object *obj = reg->obj; |
| 2297 | struct drm_device *dev = obj->dev; | 2297 | struct drm_device *dev = obj->dev; |
| 2298 | drm_i915_private_t *dev_priv = dev->dev_private; | 2298 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2299 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2299 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2300 | int regnum = obj_priv->fence_reg; | 2300 | int regnum = obj_priv->fence_reg; |
| 2301 | uint64_t val; | 2301 | uint64_t val; |
| 2302 | 2302 | ||
| @@ -2318,7 +2318,7 @@ static void i965_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
| 2318 | struct drm_gem_object *obj = reg->obj; | 2318 | struct drm_gem_object *obj = reg->obj; |
| 2319 | struct drm_device *dev = obj->dev; | 2319 | struct drm_device *dev = obj->dev; |
| 2320 | drm_i915_private_t *dev_priv = dev->dev_private; | 2320 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2321 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2321 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2322 | int regnum = obj_priv->fence_reg; | 2322 | int regnum = obj_priv->fence_reg; |
| 2323 | uint64_t val; | 2323 | uint64_t val; |
| 2324 | 2324 | ||
| @@ -2338,7 +2338,7 @@ static void i915_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
| 2338 | struct drm_gem_object *obj = reg->obj; | 2338 | struct drm_gem_object *obj = reg->obj; |
| 2339 | struct drm_device *dev = obj->dev; | 2339 | struct drm_device *dev = obj->dev; |
| 2340 | drm_i915_private_t *dev_priv = dev->dev_private; | 2340 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2341 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2341 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2342 | int regnum = obj_priv->fence_reg; | 2342 | int regnum = obj_priv->fence_reg; |
| 2343 | int tile_width; | 2343 | int tile_width; |
| 2344 | uint32_t fence_reg, val; | 2344 | uint32_t fence_reg, val; |
| @@ -2380,7 +2380,7 @@ static void i830_write_fence_reg(struct drm_i915_fence_reg *reg) | |||
| 2380 | struct drm_gem_object *obj = reg->obj; | 2380 | struct drm_gem_object *obj = reg->obj; |
| 2381 | struct drm_device *dev = obj->dev; | 2381 | struct drm_device *dev = obj->dev; |
| 2382 | drm_i915_private_t *dev_priv = dev->dev_private; | 2382 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2383 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2383 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2384 | int regnum = obj_priv->fence_reg; | 2384 | int regnum = obj_priv->fence_reg; |
| 2385 | uint32_t val; | 2385 | uint32_t val; |
| 2386 | uint32_t pitch_val; | 2386 | uint32_t pitch_val; |
| @@ -2424,7 +2424,7 @@ static int i915_find_fence_reg(struct drm_device *dev) | |||
| 2424 | if (!reg->obj) | 2424 | if (!reg->obj) |
| 2425 | return i; | 2425 | return i; |
| 2426 | 2426 | ||
| 2427 | obj_priv = reg->obj->driver_private; | 2427 | obj_priv = to_intel_bo(reg->obj); |
| 2428 | if (!obj_priv->pin_count) | 2428 | if (!obj_priv->pin_count) |
| 2429 | avail++; | 2429 | avail++; |
| 2430 | } | 2430 | } |
| @@ -2479,7 +2479,7 @@ i915_gem_object_get_fence_reg(struct drm_gem_object *obj) | |||
| 2479 | { | 2479 | { |
| 2480 | struct drm_device *dev = obj->dev; | 2480 | struct drm_device *dev = obj->dev; |
| 2481 | struct drm_i915_private *dev_priv = dev->dev_private; | 2481 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 2482 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2482 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2483 | struct drm_i915_fence_reg *reg = NULL; | 2483 | struct drm_i915_fence_reg *reg = NULL; |
| 2484 | int ret; | 2484 | int ret; |
| 2485 | 2485 | ||
| @@ -2546,7 +2546,7 @@ i915_gem_clear_fence_reg(struct drm_gem_object *obj) | |||
| 2546 | { | 2546 | { |
| 2547 | struct drm_device *dev = obj->dev; | 2547 | struct drm_device *dev = obj->dev; |
| 2548 | drm_i915_private_t *dev_priv = dev->dev_private; | 2548 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2549 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2549 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2550 | 2550 | ||
| 2551 | if (IS_GEN6(dev)) { | 2551 | if (IS_GEN6(dev)) { |
| 2552 | I915_WRITE64(FENCE_REG_SANDYBRIDGE_0 + | 2552 | I915_WRITE64(FENCE_REG_SANDYBRIDGE_0 + |
| @@ -2582,7 +2582,7 @@ int | |||
| 2582 | i915_gem_object_put_fence_reg(struct drm_gem_object *obj) | 2582 | i915_gem_object_put_fence_reg(struct drm_gem_object *obj) |
| 2583 | { | 2583 | { |
| 2584 | struct drm_device *dev = obj->dev; | 2584 | struct drm_device *dev = obj->dev; |
| 2585 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2585 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2586 | 2586 | ||
| 2587 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE) | 2587 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE) |
| 2588 | return 0; | 2588 | return 0; |
| @@ -2620,7 +2620,7 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment) | |||
| 2620 | { | 2620 | { |
| 2621 | struct drm_device *dev = obj->dev; | 2621 | struct drm_device *dev = obj->dev; |
| 2622 | drm_i915_private_t *dev_priv = dev->dev_private; | 2622 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 2623 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2623 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2624 | struct drm_mm_node *free_space; | 2624 | struct drm_mm_node *free_space; |
| 2625 | gfp_t gfpmask = __GFP_NORETRY | __GFP_NOWARN; | 2625 | gfp_t gfpmask = __GFP_NORETRY | __GFP_NOWARN; |
| 2626 | int ret; | 2626 | int ret; |
| @@ -2727,7 +2727,7 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment) | |||
| 2727 | void | 2727 | void |
| 2728 | i915_gem_clflush_object(struct drm_gem_object *obj) | 2728 | i915_gem_clflush_object(struct drm_gem_object *obj) |
| 2729 | { | 2729 | { |
| 2730 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2730 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2731 | 2731 | ||
| 2732 | /* If we don't have a page list set up, then we're not pinned | 2732 | /* If we don't have a page list set up, then we're not pinned |
| 2733 | * to GPU, and we can ignore the cache flush because it'll happen | 2733 | * to GPU, and we can ignore the cache flush because it'll happen |
| @@ -2828,7 +2828,7 @@ i915_gem_object_flush_write_domain(struct drm_gem_object *obj) | |||
| 2828 | int | 2828 | int |
| 2829 | i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, int write) | 2829 | i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, int write) |
| 2830 | { | 2830 | { |
| 2831 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2831 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2832 | uint32_t old_write_domain, old_read_domains; | 2832 | uint32_t old_write_domain, old_read_domains; |
| 2833 | int ret; | 2833 | int ret; |
| 2834 | 2834 | ||
| @@ -2878,7 +2878,7 @@ int | |||
| 2878 | i915_gem_object_set_to_display_plane(struct drm_gem_object *obj) | 2878 | i915_gem_object_set_to_display_plane(struct drm_gem_object *obj) |
| 2879 | { | 2879 | { |
| 2880 | struct drm_device *dev = obj->dev; | 2880 | struct drm_device *dev = obj->dev; |
| 2881 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 2881 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 2882 | uint32_t old_write_domain, old_read_domains; | 2882 | uint32_t old_write_domain, old_read_domains; |
| 2883 | int ret; | 2883 | int ret; |
| 2884 | 2884 | ||
| @@ -3091,7 +3091,7 @@ static void | |||
| 3091 | i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) | 3091 | i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) |
| 3092 | { | 3092 | { |
| 3093 | struct drm_device *dev = obj->dev; | 3093 | struct drm_device *dev = obj->dev; |
| 3094 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3094 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 3095 | uint32_t invalidate_domains = 0; | 3095 | uint32_t invalidate_domains = 0; |
| 3096 | uint32_t flush_domains = 0; | 3096 | uint32_t flush_domains = 0; |
| 3097 | uint32_t old_read_domains; | 3097 | uint32_t old_read_domains; |
| @@ -3176,7 +3176,7 @@ i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj) | |||
| 3176 | static void | 3176 | static void |
| 3177 | i915_gem_object_set_to_full_cpu_read_domain(struct drm_gem_object *obj) | 3177 | i915_gem_object_set_to_full_cpu_read_domain(struct drm_gem_object *obj) |
| 3178 | { | 3178 | { |
| 3179 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3179 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 3180 | 3180 | ||
| 3181 | if (!obj_priv->page_cpu_valid) | 3181 | if (!obj_priv->page_cpu_valid) |
| 3182 | return; | 3182 | return; |
| @@ -3216,7 +3216,7 @@ static int | |||
| 3216 | i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, | 3216 | i915_gem_object_set_cpu_read_domain_range(struct drm_gem_object *obj, |
| 3217 | uint64_t offset, uint64_t size) | 3217 | uint64_t offset, uint64_t size) |
| 3218 | { | 3218 | { |
| 3219 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3219 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 3220 | uint32_t old_read_domains; | 3220 | uint32_t old_read_domains; |
| 3221 | int i, ret; | 3221 | int i, ret; |
| 3222 | 3222 | ||
| @@ -3285,7 +3285,7 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
| 3285 | { | 3285 | { |
| 3286 | struct drm_device *dev = obj->dev; | 3286 | struct drm_device *dev = obj->dev; |
| 3287 | drm_i915_private_t *dev_priv = dev->dev_private; | 3287 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 3288 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3288 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 3289 | int i, ret; | 3289 | int i, ret; |
| 3290 | void __iomem *reloc_page; | 3290 | void __iomem *reloc_page; |
| 3291 | bool need_fence; | 3291 | bool need_fence; |
| @@ -3336,7 +3336,7 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
| 3336 | i915_gem_object_unpin(obj); | 3336 | i915_gem_object_unpin(obj); |
| 3337 | return -EBADF; | 3337 | return -EBADF; |
| 3338 | } | 3338 | } |
| 3339 | target_obj_priv = target_obj->driver_private; | 3339 | target_obj_priv = to_intel_bo(target_obj); |
| 3340 | 3340 | ||
| 3341 | #if WATCH_RELOC | 3341 | #if WATCH_RELOC |
| 3342 | DRM_INFO("%s: obj %p offset %08x target %d " | 3342 | DRM_INFO("%s: obj %p offset %08x target %d " |
| @@ -3688,7 +3688,7 @@ i915_gem_wait_for_pending_flip(struct drm_device *dev, | |||
| 3688 | prepare_to_wait(&dev_priv->pending_flip_queue, | 3688 | prepare_to_wait(&dev_priv->pending_flip_queue, |
| 3689 | &wait, TASK_INTERRUPTIBLE); | 3689 | &wait, TASK_INTERRUPTIBLE); |
| 3690 | for (i = 0; i < count; i++) { | 3690 | for (i = 0; i < count; i++) { |
| 3691 | obj_priv = object_list[i]->driver_private; | 3691 | obj_priv = to_intel_bo(object_list[i]); |
| 3692 | if (atomic_read(&obj_priv->pending_flip) > 0) | 3692 | if (atomic_read(&obj_priv->pending_flip) > 0) |
| 3693 | break; | 3693 | break; |
| 3694 | } | 3694 | } |
| @@ -3797,7 +3797,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
| 3797 | goto err; | 3797 | goto err; |
| 3798 | } | 3798 | } |
| 3799 | 3799 | ||
| 3800 | obj_priv = object_list[i]->driver_private; | 3800 | obj_priv = to_intel_bo(object_list[i]); |
| 3801 | if (obj_priv->in_execbuffer) { | 3801 | if (obj_priv->in_execbuffer) { |
| 3802 | DRM_ERROR("Object %p appears more than once in object list\n", | 3802 | DRM_ERROR("Object %p appears more than once in object list\n", |
| 3803 | object_list[i]); | 3803 | object_list[i]); |
| @@ -3923,7 +3923,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
| 3923 | 3923 | ||
| 3924 | for (i = 0; i < args->buffer_count; i++) { | 3924 | for (i = 0; i < args->buffer_count; i++) { |
| 3925 | struct drm_gem_object *obj = object_list[i]; | 3925 | struct drm_gem_object *obj = object_list[i]; |
| 3926 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 3926 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 3927 | uint32_t old_write_domain = obj->write_domain; | 3927 | uint32_t old_write_domain = obj->write_domain; |
| 3928 | 3928 | ||
| 3929 | obj->write_domain = obj->pending_write_domain; | 3929 | obj->write_domain = obj->pending_write_domain; |
| @@ -3998,7 +3998,7 @@ err: | |||
| 3998 | 3998 | ||
| 3999 | for (i = 0; i < args->buffer_count; i++) { | 3999 | for (i = 0; i < args->buffer_count; i++) { |
| 4000 | if (object_list[i]) { | 4000 | if (object_list[i]) { |
| 4001 | obj_priv = object_list[i]->driver_private; | 4001 | obj_priv = to_intel_bo(object_list[i]); |
| 4002 | obj_priv->in_execbuffer = false; | 4002 | obj_priv->in_execbuffer = false; |
| 4003 | } | 4003 | } |
| 4004 | drm_gem_object_unreference(object_list[i]); | 4004 | drm_gem_object_unreference(object_list[i]); |
| @@ -4176,7 +4176,7 @@ int | |||
| 4176 | i915_gem_object_pin(struct drm_gem_object *obj, uint32_t alignment) | 4176 | i915_gem_object_pin(struct drm_gem_object *obj, uint32_t alignment) |
| 4177 | { | 4177 | { |
| 4178 | struct drm_device *dev = obj->dev; | 4178 | struct drm_device *dev = obj->dev; |
| 4179 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4179 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 4180 | int ret; | 4180 | int ret; |
| 4181 | 4181 | ||
| 4182 | i915_verify_inactive(dev, __FILE__, __LINE__); | 4182 | i915_verify_inactive(dev, __FILE__, __LINE__); |
| @@ -4209,7 +4209,7 @@ i915_gem_object_unpin(struct drm_gem_object *obj) | |||
| 4209 | { | 4209 | { |
| 4210 | struct drm_device *dev = obj->dev; | 4210 | struct drm_device *dev = obj->dev; |
| 4211 | drm_i915_private_t *dev_priv = dev->dev_private; | 4211 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 4212 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4212 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 4213 | 4213 | ||
| 4214 | i915_verify_inactive(dev, __FILE__, __LINE__); | 4214 | i915_verify_inactive(dev, __FILE__, __LINE__); |
| 4215 | obj_priv->pin_count--; | 4215 | obj_priv->pin_count--; |
| @@ -4249,7 +4249,7 @@ i915_gem_pin_ioctl(struct drm_device *dev, void *data, | |||
| 4249 | mutex_unlock(&dev->struct_mutex); | 4249 | mutex_unlock(&dev->struct_mutex); |
| 4250 | return -EBADF; | 4250 | return -EBADF; |
| 4251 | } | 4251 | } |
| 4252 | obj_priv = obj->driver_private; | 4252 | obj_priv = to_intel_bo(obj); |
| 4253 | 4253 | ||
| 4254 | if (obj_priv->madv != I915_MADV_WILLNEED) { | 4254 | if (obj_priv->madv != I915_MADV_WILLNEED) { |
| 4255 | DRM_ERROR("Attempting to pin a purgeable buffer\n"); | 4255 | DRM_ERROR("Attempting to pin a purgeable buffer\n"); |
| @@ -4306,7 +4306,7 @@ i915_gem_unpin_ioctl(struct drm_device *dev, void *data, | |||
| 4306 | return -EBADF; | 4306 | return -EBADF; |
| 4307 | } | 4307 | } |
| 4308 | 4308 | ||
| 4309 | obj_priv = obj->driver_private; | 4309 | obj_priv = to_intel_bo(obj); |
| 4310 | if (obj_priv->pin_filp != file_priv) { | 4310 | if (obj_priv->pin_filp != file_priv) { |
| 4311 | DRM_ERROR("Not pinned by caller in i915_gem_pin_ioctl(): %d\n", | 4311 | DRM_ERROR("Not pinned by caller in i915_gem_pin_ioctl(): %d\n", |
| 4312 | args->handle); | 4312 | args->handle); |
| @@ -4348,7 +4348,7 @@ i915_gem_busy_ioctl(struct drm_device *dev, void *data, | |||
| 4348 | */ | 4348 | */ |
| 4349 | i915_gem_retire_requests(dev); | 4349 | i915_gem_retire_requests(dev); |
| 4350 | 4350 | ||
| 4351 | obj_priv = obj->driver_private; | 4351 | obj_priv = to_intel_bo(obj); |
| 4352 | /* Don't count being on the flushing list against the object being | 4352 | /* Don't count being on the flushing list against the object being |
| 4353 | * done. Otherwise, a buffer left on the flushing list but not getting | 4353 | * done. Otherwise, a buffer left on the flushing list but not getting |
| 4354 | * flushed (because nobody's flushing that domain) won't ever return | 4354 | * flushed (because nobody's flushing that domain) won't ever return |
| @@ -4394,7 +4394,7 @@ i915_gem_madvise_ioctl(struct drm_device *dev, void *data, | |||
| 4394 | } | 4394 | } |
| 4395 | 4395 | ||
| 4396 | mutex_lock(&dev->struct_mutex); | 4396 | mutex_lock(&dev->struct_mutex); |
| 4397 | obj_priv = obj->driver_private; | 4397 | obj_priv = to_intel_bo(obj); |
| 4398 | 4398 | ||
| 4399 | if (obj_priv->pin_count) { | 4399 | if (obj_priv->pin_count) { |
| 4400 | drm_gem_object_unreference(obj); | 4400 | drm_gem_object_unreference(obj); |
| @@ -4455,7 +4455,7 @@ int i915_gem_init_object(struct drm_gem_object *obj) | |||
| 4455 | void i915_gem_free_object(struct drm_gem_object *obj) | 4455 | void i915_gem_free_object(struct drm_gem_object *obj) |
| 4456 | { | 4456 | { |
| 4457 | struct drm_device *dev = obj->dev; | 4457 | struct drm_device *dev = obj->dev; |
| 4458 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 4458 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 4459 | 4459 | ||
| 4460 | trace_i915_gem_object_destroy(obj); | 4460 | trace_i915_gem_object_destroy(obj); |
| 4461 | 4461 | ||
| @@ -4564,7 +4564,7 @@ i915_gem_init_hws(struct drm_device *dev) | |||
| 4564 | DRM_ERROR("Failed to allocate status page\n"); | 4564 | DRM_ERROR("Failed to allocate status page\n"); |
| 4565 | return -ENOMEM; | 4565 | return -ENOMEM; |
| 4566 | } | 4566 | } |
| 4567 | obj_priv = obj->driver_private; | 4567 | obj_priv = to_intel_bo(obj); |
| 4568 | obj_priv->agp_type = AGP_USER_CACHED_MEMORY; | 4568 | obj_priv->agp_type = AGP_USER_CACHED_MEMORY; |
| 4569 | 4569 | ||
| 4570 | ret = i915_gem_object_pin(obj, 4096); | 4570 | ret = i915_gem_object_pin(obj, 4096); |
| @@ -4608,7 +4608,7 @@ i915_gem_cleanup_hws(struct drm_device *dev) | |||
| 4608 | return; | 4608 | return; |
| 4609 | 4609 | ||
| 4610 | obj = dev_priv->hws_obj; | 4610 | obj = dev_priv->hws_obj; |
| 4611 | obj_priv = obj->driver_private; | 4611 | obj_priv = to_intel_bo(obj); |
| 4612 | 4612 | ||
| 4613 | kunmap(obj_priv->pages[0]); | 4613 | kunmap(obj_priv->pages[0]); |
| 4614 | i915_gem_object_unpin(obj); | 4614 | i915_gem_object_unpin(obj); |
| @@ -4642,7 +4642,7 @@ i915_gem_init_ringbuffer(struct drm_device *dev) | |||
| 4642 | i915_gem_cleanup_hws(dev); | 4642 | i915_gem_cleanup_hws(dev); |
| 4643 | return -ENOMEM; | 4643 | return -ENOMEM; |
| 4644 | } | 4644 | } |
| 4645 | obj_priv = obj->driver_private; | 4645 | obj_priv = to_intel_bo(obj); |
| 4646 | 4646 | ||
| 4647 | ret = i915_gem_object_pin(obj, 4096); | 4647 | ret = i915_gem_object_pin(obj, 4096); |
| 4648 | if (ret != 0) { | 4648 | if (ret != 0) { |
| @@ -4935,7 +4935,7 @@ void i915_gem_detach_phys_object(struct drm_device *dev, | |||
| 4935 | int ret; | 4935 | int ret; |
| 4936 | int page_count; | 4936 | int page_count; |
| 4937 | 4937 | ||
| 4938 | obj_priv = obj->driver_private; | 4938 | obj_priv = to_intel_bo(obj); |
| 4939 | if (!obj_priv->phys_obj) | 4939 | if (!obj_priv->phys_obj) |
| 4940 | return; | 4940 | return; |
| 4941 | 4941 | ||
| @@ -4974,7 +4974,7 @@ i915_gem_attach_phys_object(struct drm_device *dev, | |||
| 4974 | if (id > I915_MAX_PHYS_OBJECT) | 4974 | if (id > I915_MAX_PHYS_OBJECT) |
| 4975 | return -EINVAL; | 4975 | return -EINVAL; |
| 4976 | 4976 | ||
| 4977 | obj_priv = obj->driver_private; | 4977 | obj_priv = to_intel_bo(obj); |
| 4978 | 4978 | ||
| 4979 | if (obj_priv->phys_obj) { | 4979 | if (obj_priv->phys_obj) { |
| 4980 | if (obj_priv->phys_obj->id == id) | 4980 | if (obj_priv->phys_obj->id == id) |
| @@ -5025,7 +5025,7 @@ i915_gem_phys_pwrite(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 5025 | struct drm_i915_gem_pwrite *args, | 5025 | struct drm_i915_gem_pwrite *args, |
| 5026 | struct drm_file *file_priv) | 5026 | struct drm_file *file_priv) |
| 5027 | { | 5027 | { |
| 5028 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 5028 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 5029 | void *obj_addr; | 5029 | void *obj_addr; |
| 5030 | int ret; | 5030 | int ret; |
| 5031 | char __user *user_data; | 5031 | char __user *user_data; |
diff --git a/drivers/gpu/drm/i915/i915_gem_debug.c b/drivers/gpu/drm/i915/i915_gem_debug.c index e602614bd3f8..35507cf53fa3 100644 --- a/drivers/gpu/drm/i915/i915_gem_debug.c +++ b/drivers/gpu/drm/i915/i915_gem_debug.c | |||
| @@ -72,7 +72,7 @@ void | |||
| 72 | i915_gem_dump_object(struct drm_gem_object *obj, int len, | 72 | i915_gem_dump_object(struct drm_gem_object *obj, int len, |
| 73 | const char *where, uint32_t mark) | 73 | const char *where, uint32_t mark) |
| 74 | { | 74 | { |
| 75 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 75 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 76 | int page; | 76 | int page; |
| 77 | 77 | ||
| 78 | DRM_INFO("%s: object at offset %08x\n", where, obj_priv->gtt_offset); | 78 | DRM_INFO("%s: object at offset %08x\n", where, obj_priv->gtt_offset); |
| @@ -137,7 +137,7 @@ void | |||
| 137 | i915_gem_object_check_coherency(struct drm_gem_object *obj, int handle) | 137 | i915_gem_object_check_coherency(struct drm_gem_object *obj, int handle) |
| 138 | { | 138 | { |
| 139 | struct drm_device *dev = obj->dev; | 139 | struct drm_device *dev = obj->dev; |
| 140 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 140 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 141 | int page; | 141 | int page; |
| 142 | uint32_t *gtt_mapping; | 142 | uint32_t *gtt_mapping; |
| 143 | uint32_t *backing_map = NULL; | 143 | uint32_t *backing_map = NULL; |
diff --git a/drivers/gpu/drm/i915/i915_gem_tiling.c b/drivers/gpu/drm/i915/i915_gem_tiling.c index c01c878e51ba..449157f71610 100644 --- a/drivers/gpu/drm/i915/i915_gem_tiling.c +++ b/drivers/gpu/drm/i915/i915_gem_tiling.c | |||
| @@ -240,7 +240,7 @@ bool | |||
| 240 | i915_gem_object_fence_offset_ok(struct drm_gem_object *obj, int tiling_mode) | 240 | i915_gem_object_fence_offset_ok(struct drm_gem_object *obj, int tiling_mode) |
| 241 | { | 241 | { |
| 242 | struct drm_device *dev = obj->dev; | 242 | struct drm_device *dev = obj->dev; |
| 243 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 243 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 244 | 244 | ||
| 245 | if (obj_priv->gtt_space == NULL) | 245 | if (obj_priv->gtt_space == NULL) |
| 246 | return true; | 246 | return true; |
| @@ -280,7 +280,7 @@ i915_gem_set_tiling(struct drm_device *dev, void *data, | |||
| 280 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 280 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
| 281 | if (obj == NULL) | 281 | if (obj == NULL) |
| 282 | return -EINVAL; | 282 | return -EINVAL; |
| 283 | obj_priv = obj->driver_private; | 283 | obj_priv = to_intel_bo(obj); |
| 284 | 284 | ||
| 285 | if (!i915_tiling_ok(dev, args->stride, obj->size, args->tiling_mode)) { | 285 | if (!i915_tiling_ok(dev, args->stride, obj->size, args->tiling_mode)) { |
| 286 | drm_gem_object_unreference_unlocked(obj); | 286 | drm_gem_object_unreference_unlocked(obj); |
| @@ -364,7 +364,7 @@ i915_gem_get_tiling(struct drm_device *dev, void *data, | |||
| 364 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); | 364 | obj = drm_gem_object_lookup(dev, file_priv, args->handle); |
| 365 | if (obj == NULL) | 365 | if (obj == NULL) |
| 366 | return -EINVAL; | 366 | return -EINVAL; |
| 367 | obj_priv = obj->driver_private; | 367 | obj_priv = to_intel_bo(obj); |
| 368 | 368 | ||
| 369 | mutex_lock(&dev->struct_mutex); | 369 | mutex_lock(&dev->struct_mutex); |
| 370 | 370 | ||
| @@ -427,7 +427,7 @@ i915_gem_object_do_bit_17_swizzle(struct drm_gem_object *obj) | |||
| 427 | { | 427 | { |
| 428 | struct drm_device *dev = obj->dev; | 428 | struct drm_device *dev = obj->dev; |
| 429 | drm_i915_private_t *dev_priv = dev->dev_private; | 429 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 430 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 430 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 431 | int page_count = obj->size >> PAGE_SHIFT; | 431 | int page_count = obj->size >> PAGE_SHIFT; |
| 432 | int i; | 432 | int i; |
| 433 | 433 | ||
| @@ -456,7 +456,7 @@ i915_gem_object_save_bit_17_swizzle(struct drm_gem_object *obj) | |||
| 456 | { | 456 | { |
| 457 | struct drm_device *dev = obj->dev; | 457 | struct drm_device *dev = obj->dev; |
| 458 | drm_i915_private_t *dev_priv = dev->dev_private; | 458 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 459 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 459 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 460 | int page_count = obj->size >> PAGE_SHIFT; | 460 | int page_count = obj->size >> PAGE_SHIFT; |
| 461 | int i; | 461 | int i; |
| 462 | 462 | ||
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index 5388354da0d1..bfbdad92d73d 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
| @@ -443,7 +443,7 @@ i915_error_object_create(struct drm_device *dev, | |||
| 443 | if (src == NULL) | 443 | if (src == NULL) |
| 444 | return NULL; | 444 | return NULL; |
| 445 | 445 | ||
| 446 | src_priv = src->driver_private; | 446 | src_priv = to_intel_bo(src); |
| 447 | if (src_priv->pages == NULL) | 447 | if (src_priv->pages == NULL) |
| 448 | return NULL; | 448 | return NULL; |
| 449 | 449 | ||
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index f0214908a935..7adb3a54aac6 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
| @@ -1002,7 +1002,7 @@ static void i8xx_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
| 1002 | struct drm_i915_private *dev_priv = dev->dev_private; | 1002 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1003 | struct drm_framebuffer *fb = crtc->fb; | 1003 | struct drm_framebuffer *fb = crtc->fb; |
| 1004 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); | 1004 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); |
| 1005 | struct drm_i915_gem_object *obj_priv = intel_fb->obj->driver_private; | 1005 | struct drm_i915_gem_object *obj_priv = to_intel_bo(intel_fb->obj); |
| 1006 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 1006 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
| 1007 | int plane, i; | 1007 | int plane, i; |
| 1008 | u32 fbc_ctl, fbc_ctl2; | 1008 | u32 fbc_ctl, fbc_ctl2; |
| @@ -1079,7 +1079,7 @@ static void g4x_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
| 1079 | struct drm_i915_private *dev_priv = dev->dev_private; | 1079 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1080 | struct drm_framebuffer *fb = crtc->fb; | 1080 | struct drm_framebuffer *fb = crtc->fb; |
| 1081 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); | 1081 | struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); |
| 1082 | struct drm_i915_gem_object *obj_priv = intel_fb->obj->driver_private; | 1082 | struct drm_i915_gem_object *obj_priv = to_intel_bo(intel_fb->obj); |
| 1083 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 1083 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
| 1084 | int plane = (intel_crtc->plane == 0 ? DPFC_CTL_PLANEA : | 1084 | int plane = (intel_crtc->plane == 0 ? DPFC_CTL_PLANEA : |
| 1085 | DPFC_CTL_PLANEB); | 1085 | DPFC_CTL_PLANEB); |
| @@ -1175,7 +1175,7 @@ static void intel_update_fbc(struct drm_crtc *crtc, | |||
| 1175 | return; | 1175 | return; |
| 1176 | 1176 | ||
| 1177 | intel_fb = to_intel_framebuffer(fb); | 1177 | intel_fb = to_intel_framebuffer(fb); |
| 1178 | obj_priv = intel_fb->obj->driver_private; | 1178 | obj_priv = to_intel_bo(intel_fb->obj); |
| 1179 | 1179 | ||
| 1180 | /* | 1180 | /* |
| 1181 | * If FBC is already on, we just have to verify that we can | 1181 | * If FBC is already on, we just have to verify that we can |
| @@ -1242,7 +1242,7 @@ out_disable: | |||
| 1242 | static int | 1242 | static int |
| 1243 | intel_pin_and_fence_fb_obj(struct drm_device *dev, struct drm_gem_object *obj) | 1243 | intel_pin_and_fence_fb_obj(struct drm_device *dev, struct drm_gem_object *obj) |
| 1244 | { | 1244 | { |
| 1245 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1245 | struct drm_i915_gem_object *obj_priv = to_intel_bo(obj); |
| 1246 | u32 alignment; | 1246 | u32 alignment; |
| 1247 | int ret; | 1247 | int ret; |
| 1248 | 1248 | ||
| @@ -1322,7 +1322,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1322 | 1322 | ||
| 1323 | intel_fb = to_intel_framebuffer(crtc->fb); | 1323 | intel_fb = to_intel_framebuffer(crtc->fb); |
| 1324 | obj = intel_fb->obj; | 1324 | obj = intel_fb->obj; |
| 1325 | obj_priv = obj->driver_private; | 1325 | obj_priv = to_intel_bo(obj); |
| 1326 | 1326 | ||
| 1327 | mutex_lock(&dev->struct_mutex); | 1327 | mutex_lock(&dev->struct_mutex); |
| 1328 | ret = intel_pin_and_fence_fb_obj(dev, obj); | 1328 | ret = intel_pin_and_fence_fb_obj(dev, obj); |
| @@ -1400,7 +1400,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1400 | 1400 | ||
| 1401 | if (old_fb) { | 1401 | if (old_fb) { |
| 1402 | intel_fb = to_intel_framebuffer(old_fb); | 1402 | intel_fb = to_intel_framebuffer(old_fb); |
| 1403 | obj_priv = intel_fb->obj->driver_private; | 1403 | obj_priv = to_intel_bo(intel_fb->obj); |
| 1404 | i915_gem_object_unpin(intel_fb->obj); | 1404 | i915_gem_object_unpin(intel_fb->obj); |
| 1405 | } | 1405 | } |
| 1406 | intel_increase_pllclock(crtc, true); | 1406 | intel_increase_pllclock(crtc, true); |
| @@ -3510,7 +3510,7 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc, | |||
| 3510 | if (!bo) | 3510 | if (!bo) |
| 3511 | return -ENOENT; | 3511 | return -ENOENT; |
| 3512 | 3512 | ||
| 3513 | obj_priv = bo->driver_private; | 3513 | obj_priv = to_intel_bo(bo); |
| 3514 | 3514 | ||
| 3515 | if (bo->size < width * height * 4) { | 3515 | if (bo->size < width * height * 4) { |
| 3516 | DRM_ERROR("buffer is to small\n"); | 3516 | DRM_ERROR("buffer is to small\n"); |
| @@ -4155,7 +4155,7 @@ void intel_finish_page_flip(struct drm_device *dev, int pipe) | |||
| 4155 | work = intel_crtc->unpin_work; | 4155 | work = intel_crtc->unpin_work; |
| 4156 | if (work == NULL || !work->pending) { | 4156 | if (work == NULL || !work->pending) { |
| 4157 | if (work && !work->pending) { | 4157 | if (work && !work->pending) { |
| 4158 | obj_priv = work->pending_flip_obj->driver_private; | 4158 | obj_priv = to_intel_bo(work->pending_flip_obj); |
| 4159 | DRM_DEBUG_DRIVER("flip finish: %p (%d) not pending?\n", | 4159 | DRM_DEBUG_DRIVER("flip finish: %p (%d) not pending?\n", |
| 4160 | obj_priv, | 4160 | obj_priv, |
| 4161 | atomic_read(&obj_priv->pending_flip)); | 4161 | atomic_read(&obj_priv->pending_flip)); |
| @@ -4180,7 +4180,7 @@ void intel_finish_page_flip(struct drm_device *dev, int pipe) | |||
| 4180 | 4180 | ||
| 4181 | spin_unlock_irqrestore(&dev->event_lock, flags); | 4181 | spin_unlock_irqrestore(&dev->event_lock, flags); |
| 4182 | 4182 | ||
| 4183 | obj_priv = work->pending_flip_obj->driver_private; | 4183 | obj_priv = to_intel_bo(work->pending_flip_obj); |
| 4184 | 4184 | ||
| 4185 | /* Initial scanout buffer will have a 0 pending flip count */ | 4185 | /* Initial scanout buffer will have a 0 pending flip count */ |
| 4186 | if ((atomic_read(&obj_priv->pending_flip) == 0) || | 4186 | if ((atomic_read(&obj_priv->pending_flip) == 0) || |
| @@ -4251,7 +4251,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc, | |||
| 4251 | ret = intel_pin_and_fence_fb_obj(dev, obj); | 4251 | ret = intel_pin_and_fence_fb_obj(dev, obj); |
| 4252 | if (ret != 0) { | 4252 | if (ret != 0) { |
| 4253 | DRM_DEBUG_DRIVER("flip queue: %p pin & fence failed\n", | 4253 | DRM_DEBUG_DRIVER("flip queue: %p pin & fence failed\n", |
| 4254 | obj->driver_private); | 4254 | to_intel_bo(obj)); |
| 4255 | kfree(work); | 4255 | kfree(work); |
| 4256 | intel_crtc->unpin_work = NULL; | 4256 | intel_crtc->unpin_work = NULL; |
| 4257 | mutex_unlock(&dev->struct_mutex); | 4257 | mutex_unlock(&dev->struct_mutex); |
| @@ -4265,7 +4265,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc, | |||
| 4265 | crtc->fb = fb; | 4265 | crtc->fb = fb; |
| 4266 | i915_gem_object_flush_write_domain(obj); | 4266 | i915_gem_object_flush_write_domain(obj); |
| 4267 | drm_vblank_get(dev, intel_crtc->pipe); | 4267 | drm_vblank_get(dev, intel_crtc->pipe); |
| 4268 | obj_priv = obj->driver_private; | 4268 | obj_priv = to_intel_bo(obj); |
| 4269 | atomic_inc(&obj_priv->pending_flip); | 4269 | atomic_inc(&obj_priv->pending_flip); |
| 4270 | work->pending_flip_obj = obj; | 4270 | work->pending_flip_obj = obj; |
| 4271 | 4271 | ||
| @@ -4778,14 +4778,14 @@ void intel_init_clock_gating(struct drm_device *dev) | |||
| 4778 | struct drm_i915_gem_object *obj_priv = NULL; | 4778 | struct drm_i915_gem_object *obj_priv = NULL; |
| 4779 | 4779 | ||
| 4780 | if (dev_priv->pwrctx) { | 4780 | if (dev_priv->pwrctx) { |
| 4781 | obj_priv = dev_priv->pwrctx->driver_private; | 4781 | obj_priv = to_intel_bo(dev_priv->pwrctx); |
| 4782 | } else { | 4782 | } else { |
| 4783 | struct drm_gem_object *pwrctx; | 4783 | struct drm_gem_object *pwrctx; |
| 4784 | 4784 | ||
| 4785 | pwrctx = intel_alloc_power_context(dev); | 4785 | pwrctx = intel_alloc_power_context(dev); |
| 4786 | if (pwrctx) { | 4786 | if (pwrctx) { |
| 4787 | dev_priv->pwrctx = pwrctx; | 4787 | dev_priv->pwrctx = pwrctx; |
| 4788 | obj_priv = pwrctx->driver_private; | 4788 | obj_priv = to_intel_bo(pwrctx); |
| 4789 | } | 4789 | } |
| 4790 | } | 4790 | } |
| 4791 | 4791 | ||
| @@ -4956,7 +4956,7 @@ void intel_modeset_cleanup(struct drm_device *dev) | |||
| 4956 | if (dev_priv->pwrctx) { | 4956 | if (dev_priv->pwrctx) { |
| 4957 | struct drm_i915_gem_object *obj_priv; | 4957 | struct drm_i915_gem_object *obj_priv; |
| 4958 | 4958 | ||
| 4959 | obj_priv = dev_priv->pwrctx->driver_private; | 4959 | obj_priv = to_intel_bo(dev_priv->pwrctx); |
| 4960 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset &~ PWRCTX_EN); | 4960 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset &~ PWRCTX_EN); |
| 4961 | I915_READ(PWRCTXA); | 4961 | I915_READ(PWRCTXA); |
| 4962 | i915_gem_object_unpin(dev_priv->pwrctx); | 4962 | i915_gem_object_unpin(dev_priv->pwrctx); |
diff --git a/drivers/gpu/drm/i915/intel_fb.c b/drivers/gpu/drm/i915/intel_fb.c index 8cd791dc5b29..c9fbdfa9c575 100644 --- a/drivers/gpu/drm/i915/intel_fb.c +++ b/drivers/gpu/drm/i915/intel_fb.c | |||
| @@ -145,7 +145,7 @@ static int intelfb_create(struct drm_device *dev, uint32_t fb_width, | |||
| 145 | ret = -ENOMEM; | 145 | ret = -ENOMEM; |
| 146 | goto out; | 146 | goto out; |
| 147 | } | 147 | } |
| 148 | obj_priv = fbo->driver_private; | 148 | obj_priv = to_intel_bo(fbo); |
| 149 | 149 | ||
| 150 | mutex_lock(&dev->struct_mutex); | 150 | mutex_lock(&dev->struct_mutex); |
| 151 | 151 | ||
diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c index 60595fc26fdd..6d524a1fc271 100644 --- a/drivers/gpu/drm/i915/intel_overlay.c +++ b/drivers/gpu/drm/i915/intel_overlay.c | |||
| @@ -724,7 +724,7 @@ int intel_overlay_do_put_image(struct intel_overlay *overlay, | |||
| 724 | int ret, tmp_width; | 724 | int ret, tmp_width; |
| 725 | struct overlay_registers *regs; | 725 | struct overlay_registers *regs; |
| 726 | bool scale_changed = false; | 726 | bool scale_changed = false; |
| 727 | struct drm_i915_gem_object *bo_priv = new_bo->driver_private; | 727 | struct drm_i915_gem_object *bo_priv = to_intel_bo(new_bo); |
| 728 | struct drm_device *dev = overlay->dev; | 728 | struct drm_device *dev = overlay->dev; |
| 729 | 729 | ||
| 730 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); | 730 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); |
| @@ -809,7 +809,7 @@ int intel_overlay_do_put_image(struct intel_overlay *overlay, | |||
| 809 | intel_overlay_continue(overlay, scale_changed); | 809 | intel_overlay_continue(overlay, scale_changed); |
| 810 | 810 | ||
| 811 | overlay->old_vid_bo = overlay->vid_bo; | 811 | overlay->old_vid_bo = overlay->vid_bo; |
| 812 | overlay->vid_bo = new_bo->driver_private; | 812 | overlay->vid_bo = to_intel_bo(new_bo); |
| 813 | 813 | ||
| 814 | return 0; | 814 | return 0; |
| 815 | 815 | ||
| @@ -1344,7 +1344,7 @@ void intel_setup_overlay(struct drm_device *dev) | |||
| 1344 | reg_bo = drm_gem_object_alloc(dev, PAGE_SIZE); | 1344 | reg_bo = drm_gem_object_alloc(dev, PAGE_SIZE); |
| 1345 | if (!reg_bo) | 1345 | if (!reg_bo) |
| 1346 | goto out_free; | 1346 | goto out_free; |
| 1347 | overlay->reg_bo = reg_bo->driver_private; | 1347 | overlay->reg_bo = to_intel_bo(reg_bo); |
| 1348 | 1348 | ||
| 1349 | if (OVERLAY_NONPHYSICAL(dev)) { | 1349 | if (OVERLAY_NONPHYSICAL(dev)) { |
| 1350 | ret = i915_gem_object_pin(reg_bo, PAGE_SIZE); | 1350 | ret = i915_gem_object_pin(reg_bo, PAGE_SIZE); |
