diff options
| author | Dave Airlie <airlied@gmail.com> | 2011-01-11 19:49:03 -0500 |
|---|---|---|
| committer | Dave Airlie <airlied@gmail.com> | 2011-01-11 19:49:03 -0500 |
| commit | 784fe39fa80d557847baeca2695915d17b09827f (patch) | |
| tree | a258083fb42459b6f5415e397c0ce551713fc1ed /drivers/gpu/drm/i915 | |
| parent | 4162cf64973df51fc885825bc9ca4d055891c49f (diff) | |
| parent | 6fe4f14044f181e146cdc15485428f95fa541ce8 (diff) | |
Merge branch 'drm-intel-fixes' of ssh://master.kernel.org/pub/scm/linux/kernel/git/ickle/drm-intel
* 'drm-intel-fixes' of ssh://master.kernel.org/pub/scm/linux/kernel/git/ickle/drm-intel: (37 commits)
drm/i915/execbuffer: Reorder binding of objects to favour restrictions
drm/i915: If we hit OOM when allocating GTT pages, clear the aperture
drm/i915/evict: Ensure we completely cleanup on failure
drm/i915/execbuffer: Correctly clear the current object list upon EFAULT
drm/i915/debugfs: Show all objects in the gtt
drm/i915: Record AGP memory type upon error
drm/i915: Periodically flush the active lists and requests
drm/i915/gtt: Unmap the PCI pages after unbinding them from the GTT
drm/i915: Record the error batchbuffer on each ring
drm/i915: Include TLB miss overhead for computing WM
drm/i915: Propagate error from flushing the ring
drm/i915: detect & report PCH display error interrupts
drm/i915: cleanup rc6 code
drm/i915: fix rc6 enabling around suspend/resume
drm/i915: re-enable rc6 support for Ironlake+
drm/i915: Make the ring IMR handling private
drm/i915/ringbuffer: Simplify the ring irq refcounting
drm/i915/debugfs: Show the per-ring IMR
drm/i915: Mask USER interrupts on gen6 (until required)
drm/i915: Handle ringbuffer stalls when flushing
...
Diffstat (limited to 'drivers/gpu/drm/i915')
| -rw-r--r-- | drivers/gpu/drm/i915/i915_debugfs.c | 87 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_dma.c | 8 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_drv.c | 9 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_drv.h | 24 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 156 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem_evict.c | 9 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem_execbuffer.c | 119 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem_gtt.c | 10 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_irq.c | 269 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_reg.h | 95 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_suspend.c | 8 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_crt.c | 30 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_display.c | 434 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_dp.c | 50 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_drv.h | 3 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_fb.c | 20 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_lvds.c | 14 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_panel.c | 31 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_ringbuffer.c | 255 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_ringbuffer.h | 36 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/intel_sdvo.c | 33 |
21 files changed, 1067 insertions, 633 deletions
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index 92f75782c332..19a3d58044dd 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c | |||
| @@ -106,10 +106,19 @@ static const char *get_tiling_flag(struct drm_i915_gem_object *obj) | |||
| 106 | } | 106 | } |
| 107 | } | 107 | } |
| 108 | 108 | ||
| 109 | static const char *agp_type_str(int type) | ||
| 110 | { | ||
| 111 | switch (type) { | ||
| 112 | case 0: return " uncached"; | ||
| 113 | case 1: return " snooped"; | ||
| 114 | default: return ""; | ||
| 115 | } | ||
| 116 | } | ||
| 117 | |||
| 109 | static void | 118 | static void |
| 110 | describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) | 119 | describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) |
| 111 | { | 120 | { |
| 112 | seq_printf(m, "%p: %s%s %8zd %04x %04x %d %d%s%s", | 121 | seq_printf(m, "%p: %s%s %8zd %04x %04x %d %d%s%s%s", |
| 113 | &obj->base, | 122 | &obj->base, |
| 114 | get_pin_flag(obj), | 123 | get_pin_flag(obj), |
| 115 | get_tiling_flag(obj), | 124 | get_tiling_flag(obj), |
| @@ -118,6 +127,7 @@ describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) | |||
| 118 | obj->base.write_domain, | 127 | obj->base.write_domain, |
| 119 | obj->last_rendering_seqno, | 128 | obj->last_rendering_seqno, |
| 120 | obj->last_fenced_seqno, | 129 | obj->last_fenced_seqno, |
| 130 | agp_type_str(obj->agp_type == AGP_USER_CACHED_MEMORY), | ||
| 121 | obj->dirty ? " dirty" : "", | 131 | obj->dirty ? " dirty" : "", |
| 122 | obj->madv == I915_MADV_DONTNEED ? " purgeable" : ""); | 132 | obj->madv == I915_MADV_DONTNEED ? " purgeable" : ""); |
| 123 | if (obj->base.name) | 133 | if (obj->base.name) |
| @@ -276,6 +286,37 @@ static int i915_gem_object_info(struct seq_file *m, void* data) | |||
| 276 | return 0; | 286 | return 0; |
| 277 | } | 287 | } |
| 278 | 288 | ||
| 289 | static int i915_gem_gtt_info(struct seq_file *m, void* data) | ||
| 290 | { | ||
| 291 | struct drm_info_node *node = (struct drm_info_node *) m->private; | ||
| 292 | struct drm_device *dev = node->minor->dev; | ||
| 293 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 294 | struct drm_i915_gem_object *obj; | ||
| 295 | size_t total_obj_size, total_gtt_size; | ||
| 296 | int count, ret; | ||
| 297 | |||
| 298 | ret = mutex_lock_interruptible(&dev->struct_mutex); | ||
| 299 | if (ret) | ||
| 300 | return ret; | ||
| 301 | |||
| 302 | total_obj_size = total_gtt_size = count = 0; | ||
| 303 | list_for_each_entry(obj, &dev_priv->mm.gtt_list, gtt_list) { | ||
| 304 | seq_printf(m, " "); | ||
| 305 | describe_obj(m, obj); | ||
| 306 | seq_printf(m, "\n"); | ||
| 307 | total_obj_size += obj->base.size; | ||
| 308 | total_gtt_size += obj->gtt_space->size; | ||
| 309 | count++; | ||
| 310 | } | ||
| 311 | |||
| 312 | mutex_unlock(&dev->struct_mutex); | ||
| 313 | |||
| 314 | seq_printf(m, "Total %d objects, %zu bytes, %zu GTT size\n", | ||
| 315 | count, total_obj_size, total_gtt_size); | ||
| 316 | |||
| 317 | return 0; | ||
| 318 | } | ||
| 319 | |||
| 279 | 320 | ||
| 280 | static int i915_gem_pageflip_info(struct seq_file *m, void *data) | 321 | static int i915_gem_pageflip_info(struct seq_file *m, void *data) |
| 281 | { | 322 | { |
| @@ -456,8 +497,14 @@ static int i915_interrupt_info(struct seq_file *m, void *data) | |||
| 456 | } | 497 | } |
| 457 | seq_printf(m, "Interrupts received: %d\n", | 498 | seq_printf(m, "Interrupts received: %d\n", |
| 458 | atomic_read(&dev_priv->irq_received)); | 499 | atomic_read(&dev_priv->irq_received)); |
| 459 | for (i = 0; i < I915_NUM_RINGS; i++) | 500 | for (i = 0; i < I915_NUM_RINGS; i++) { |
| 501 | if (IS_GEN6(dev)) { | ||
| 502 | seq_printf(m, "Graphics Interrupt mask (%s): %08x\n", | ||
| 503 | dev_priv->ring[i].name, | ||
| 504 | I915_READ_IMR(&dev_priv->ring[i])); | ||
| 505 | } | ||
| 460 | i915_ring_seqno_info(m, &dev_priv->ring[i]); | 506 | i915_ring_seqno_info(m, &dev_priv->ring[i]); |
| 507 | } | ||
| 461 | mutex_unlock(&dev->struct_mutex); | 508 | mutex_unlock(&dev->struct_mutex); |
| 462 | 509 | ||
| 463 | return 0; | 510 | return 0; |
| @@ -656,7 +703,7 @@ static void print_error_buffers(struct seq_file *m, | |||
| 656 | seq_printf(m, "%s [%d]:\n", name, count); | 703 | seq_printf(m, "%s [%d]:\n", name, count); |
| 657 | 704 | ||
| 658 | while (count--) { | 705 | while (count--) { |
| 659 | seq_printf(m, " %08x %8zd %04x %04x %08x%s%s%s%s%s", | 706 | seq_printf(m, " %08x %8zd %04x %04x %08x%s%s%s%s%s%s", |
| 660 | err->gtt_offset, | 707 | err->gtt_offset, |
| 661 | err->size, | 708 | err->size, |
| 662 | err->read_domains, | 709 | err->read_domains, |
| @@ -666,7 +713,8 @@ static void print_error_buffers(struct seq_file *m, | |||
| 666 | tiling_flag(err->tiling), | 713 | tiling_flag(err->tiling), |
| 667 | dirty_flag(err->dirty), | 714 | dirty_flag(err->dirty), |
| 668 | purgeable_flag(err->purgeable), | 715 | purgeable_flag(err->purgeable), |
| 669 | ring_str(err->ring)); | 716 | ring_str(err->ring), |
| 717 | agp_type_str(err->agp_type)); | ||
| 670 | 718 | ||
| 671 | if (err->name) | 719 | if (err->name) |
| 672 | seq_printf(m, " (name: %d)", err->name); | 720 | seq_printf(m, " (name: %d)", err->name); |
| @@ -744,7 +792,9 @@ static int i915_error_state(struct seq_file *m, void *unused) | |||
| 744 | if (error->batchbuffer[i]) { | 792 | if (error->batchbuffer[i]) { |
| 745 | struct drm_i915_error_object *obj = error->batchbuffer[i]; | 793 | struct drm_i915_error_object *obj = error->batchbuffer[i]; |
| 746 | 794 | ||
| 747 | seq_printf(m, "--- gtt_offset = 0x%08x\n", obj->gtt_offset); | 795 | seq_printf(m, "%s --- gtt_offset = 0x%08x\n", |
| 796 | dev_priv->ring[i].name, | ||
| 797 | obj->gtt_offset); | ||
| 748 | offset = 0; | 798 | offset = 0; |
| 749 | for (page = 0; page < obj->page_count; page++) { | 799 | for (page = 0; page < obj->page_count; page++) { |
| 750 | for (elt = 0; elt < PAGE_SIZE/4; elt++) { | 800 | for (elt = 0; elt < PAGE_SIZE/4; elt++) { |
| @@ -890,7 +940,7 @@ static int i915_drpc_info(struct seq_file *m, void *unused) | |||
| 890 | struct drm_device *dev = node->minor->dev; | 940 | struct drm_device *dev = node->minor->dev; |
| 891 | drm_i915_private_t *dev_priv = dev->dev_private; | 941 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 892 | u32 rgvmodectl = I915_READ(MEMMODECTL); | 942 | u32 rgvmodectl = I915_READ(MEMMODECTL); |
| 893 | u32 rstdbyctl = I915_READ(MCHBAR_RENDER_STANDBY); | 943 | u32 rstdbyctl = I915_READ(RSTDBYCTL); |
| 894 | u16 crstandvid = I915_READ16(CRSTANDVID); | 944 | u16 crstandvid = I915_READ16(CRSTANDVID); |
| 895 | 945 | ||
| 896 | seq_printf(m, "HD boost: %s\n", (rgvmodectl & MEMMODE_BOOST_EN) ? | 946 | seq_printf(m, "HD boost: %s\n", (rgvmodectl & MEMMODE_BOOST_EN) ? |
| @@ -913,6 +963,30 @@ static int i915_drpc_info(struct seq_file *m, void *unused) | |||
| 913 | seq_printf(m, "RS2 VID: %d\n", ((crstandvid >> 8) & 0x3f)); | 963 | seq_printf(m, "RS2 VID: %d\n", ((crstandvid >> 8) & 0x3f)); |
| 914 | seq_printf(m, "Render standby enabled: %s\n", | 964 | seq_printf(m, "Render standby enabled: %s\n", |
| 915 | (rstdbyctl & RCX_SW_EXIT) ? "no" : "yes"); | 965 | (rstdbyctl & RCX_SW_EXIT) ? "no" : "yes"); |
| 966 | seq_printf(m, "Current RS state: "); | ||
| 967 | switch (rstdbyctl & RSX_STATUS_MASK) { | ||
| 968 | case RSX_STATUS_ON: | ||
| 969 | seq_printf(m, "on\n"); | ||
| 970 | break; | ||
| 971 | case RSX_STATUS_RC1: | ||
| 972 | seq_printf(m, "RC1\n"); | ||
| 973 | break; | ||
| 974 | case RSX_STATUS_RC1E: | ||
| 975 | seq_printf(m, "RC1E\n"); | ||
| 976 | break; | ||
| 977 | case RSX_STATUS_RS1: | ||
| 978 | seq_printf(m, "RS1\n"); | ||
| 979 | break; | ||
| 980 | case RSX_STATUS_RS2: | ||
| 981 | seq_printf(m, "RS2 (RC6)\n"); | ||
| 982 | break; | ||
| 983 | case RSX_STATUS_RS3: | ||
| 984 | seq_printf(m, "RC3 (RC6+)\n"); | ||
| 985 | break; | ||
| 986 | default: | ||
| 987 | seq_printf(m, "unknown\n"); | ||
| 988 | break; | ||
| 989 | } | ||
| 916 | 990 | ||
| 917 | return 0; | 991 | return 0; |
| 918 | } | 992 | } |
| @@ -1187,6 +1261,7 @@ static int i915_wedged_create(struct dentry *root, struct drm_minor *minor) | |||
| 1187 | static struct drm_info_list i915_debugfs_list[] = { | 1261 | static struct drm_info_list i915_debugfs_list[] = { |
| 1188 | {"i915_capabilities", i915_capabilities, 0, 0}, | 1262 | {"i915_capabilities", i915_capabilities, 0, 0}, |
| 1189 | {"i915_gem_objects", i915_gem_object_info, 0}, | 1263 | {"i915_gem_objects", i915_gem_object_info, 0}, |
| 1264 | {"i915_gem_gtt", i915_gem_gtt_info, 0}, | ||
| 1190 | {"i915_gem_active", i915_gem_object_list_info, 0, (void *) ACTIVE_LIST}, | 1265 | {"i915_gem_active", i915_gem_object_list_info, 0, (void *) ACTIVE_LIST}, |
| 1191 | {"i915_gem_flushing", i915_gem_object_list_info, 0, (void *) FLUSHING_LIST}, | 1266 | {"i915_gem_flushing", i915_gem_object_list_info, 0, (void *) FLUSHING_LIST}, |
| 1192 | {"i915_gem_inactive", i915_gem_object_list_info, 0, (void *) INACTIVE_LIST}, | 1267 | {"i915_gem_inactive", i915_gem_object_list_info, 0, (void *) INACTIVE_LIST}, |
diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c index 0568dbdc10ef..844f3c972b04 100644 --- a/drivers/gpu/drm/i915/i915_dma.c +++ b/drivers/gpu/drm/i915/i915_dma.c | |||
| @@ -1962,13 +1962,6 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
| 1962 | /* enable GEM by default */ | 1962 | /* enable GEM by default */ |
| 1963 | dev_priv->has_gem = 1; | 1963 | dev_priv->has_gem = 1; |
| 1964 | 1964 | ||
| 1965 | if (dev_priv->has_gem == 0 && | ||
| 1966 | drm_core_check_feature(dev, DRIVER_MODESET)) { | ||
| 1967 | DRM_ERROR("kernel modesetting requires GEM, disabling driver.\n"); | ||
| 1968 | ret = -ENODEV; | ||
| 1969 | goto out_workqueue_free; | ||
| 1970 | } | ||
| 1971 | |||
| 1972 | dev->driver->get_vblank_counter = i915_get_vblank_counter; | 1965 | dev->driver->get_vblank_counter = i915_get_vblank_counter; |
| 1973 | dev->max_vblank_count = 0xffffff; /* only 24 bits of frame count */ | 1966 | dev->max_vblank_count = 0xffffff; /* only 24 bits of frame count */ |
| 1974 | if (IS_G4X(dev) || IS_GEN5(dev) || IS_GEN6(dev)) { | 1967 | if (IS_G4X(dev) || IS_GEN5(dev) || IS_GEN6(dev)) { |
| @@ -2055,7 +2048,6 @@ out_gem_unload: | |||
| 2055 | 2048 | ||
| 2056 | intel_teardown_gmbus(dev); | 2049 | intel_teardown_gmbus(dev); |
| 2057 | intel_teardown_mchbar(dev); | 2050 | intel_teardown_mchbar(dev); |
| 2058 | out_workqueue_free: | ||
| 2059 | destroy_workqueue(dev_priv->wq); | 2051 | destroy_workqueue(dev_priv->wq); |
| 2060 | out_iomapfree: | 2052 | out_iomapfree: |
| 2061 | io_mapping_free(dev_priv->mm.gtt_mapping); | 2053 | io_mapping_free(dev_priv->mm.gtt_mapping); |
diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c index 872493331988..0de75a23f8e7 100644 --- a/drivers/gpu/drm/i915/i915_drv.c +++ b/drivers/gpu/drm/i915/i915_drv.c | |||
| @@ -49,6 +49,9 @@ module_param_named(powersave, i915_powersave, int, 0600); | |||
| 49 | unsigned int i915_lvds_downclock = 0; | 49 | unsigned int i915_lvds_downclock = 0; |
| 50 | module_param_named(lvds_downclock, i915_lvds_downclock, int, 0400); | 50 | module_param_named(lvds_downclock, i915_lvds_downclock, int, 0400); |
| 51 | 51 | ||
| 52 | bool i915_try_reset = true; | ||
| 53 | module_param_named(reset, i915_try_reset, bool, 0600); | ||
| 54 | |||
| 52 | static struct drm_driver driver; | 55 | static struct drm_driver driver; |
| 53 | extern int intel_agp_enabled; | 56 | extern int intel_agp_enabled; |
| 54 | 57 | ||
| @@ -352,6 +355,9 @@ static int i915_drm_thaw(struct drm_device *dev) | |||
| 352 | 355 | ||
| 353 | /* Resume the modeset for every activated CRTC */ | 356 | /* Resume the modeset for every activated CRTC */ |
| 354 | drm_helper_resume_force_mode(dev); | 357 | drm_helper_resume_force_mode(dev); |
| 358 | |||
| 359 | if (dev_priv->renderctx && dev_priv->pwrctx) | ||
| 360 | ironlake_enable_rc6(dev); | ||
| 355 | } | 361 | } |
| 356 | 362 | ||
| 357 | intel_opregion_init(dev); | 363 | intel_opregion_init(dev); |
| @@ -475,6 +481,9 @@ int i915_reset(struct drm_device *dev, u8 flags) | |||
| 475 | bool need_display = true; | 481 | bool need_display = true; |
| 476 | int ret; | 482 | int ret; |
| 477 | 483 | ||
| 484 | if (!i915_try_reset) | ||
| 485 | return 0; | ||
| 486 | |||
| 478 | if (!mutex_trylock(&dev->struct_mutex)) | 487 | if (!mutex_trylock(&dev->struct_mutex)) |
| 479 | return -EBUSY; | 488 | return -EBUSY; |
| 480 | 489 | ||
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index aac1bf332f75..385fc7ec39d3 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
| @@ -172,20 +172,21 @@ struct drm_i915_error_state { | |||
| 172 | int page_count; | 172 | int page_count; |
| 173 | u32 gtt_offset; | 173 | u32 gtt_offset; |
| 174 | u32 *pages[0]; | 174 | u32 *pages[0]; |
| 175 | } *ringbuffer, *batchbuffer[2]; | 175 | } *ringbuffer, *batchbuffer[I915_NUM_RINGS]; |
| 176 | struct drm_i915_error_buffer { | 176 | struct drm_i915_error_buffer { |
| 177 | size_t size; | 177 | u32 size; |
| 178 | u32 name; | 178 | u32 name; |
| 179 | u32 seqno; | 179 | u32 seqno; |
| 180 | u32 gtt_offset; | 180 | u32 gtt_offset; |
| 181 | u32 read_domains; | 181 | u32 read_domains; |
| 182 | u32 write_domain; | 182 | u32 write_domain; |
| 183 | u32 fence_reg; | 183 | s32 fence_reg:5; |
| 184 | s32 pinned:2; | 184 | s32 pinned:2; |
| 185 | u32 tiling:2; | 185 | u32 tiling:2; |
| 186 | u32 dirty:1; | 186 | u32 dirty:1; |
| 187 | u32 purgeable:1; | 187 | u32 purgeable:1; |
| 188 | u32 ring:4; | 188 | u32 ring:4; |
| 189 | u32 agp_type:1; | ||
| 189 | } *active_bo, *pinned_bo; | 190 | } *active_bo, *pinned_bo; |
| 190 | u32 active_bo_count, pinned_bo_count; | 191 | u32 active_bo_count, pinned_bo_count; |
| 191 | struct intel_overlay_error_state *overlay; | 192 | struct intel_overlay_error_state *overlay; |
| @@ -332,6 +333,7 @@ typedef struct drm_i915_private { | |||
| 332 | 333 | ||
| 333 | /* LVDS info */ | 334 | /* LVDS info */ |
| 334 | int backlight_level; /* restore backlight to this value */ | 335 | int backlight_level; /* restore backlight to this value */ |
| 336 | bool backlight_enabled; | ||
| 335 | struct drm_display_mode *panel_fixed_mode; | 337 | struct drm_display_mode *panel_fixed_mode; |
| 336 | struct drm_display_mode *lfp_lvds_vbt_mode; /* if any */ | 338 | struct drm_display_mode *lfp_lvds_vbt_mode; /* if any */ |
| 337 | struct drm_display_mode *sdvo_lvds_vbt_mode; /* if any */ | 339 | struct drm_display_mode *sdvo_lvds_vbt_mode; /* if any */ |
| @@ -794,6 +796,7 @@ struct drm_i915_gem_object { | |||
| 794 | */ | 796 | */ |
| 795 | struct hlist_node exec_node; | 797 | struct hlist_node exec_node; |
| 796 | unsigned long exec_handle; | 798 | unsigned long exec_handle; |
| 799 | struct drm_i915_gem_exec_object2 *exec_entry; | ||
| 797 | 800 | ||
| 798 | /** | 801 | /** |
| 799 | * Current offset of the object in GTT space. | 802 | * Current offset of the object in GTT space. |
| @@ -1006,12 +1009,6 @@ extern u32 i915_get_vblank_counter(struct drm_device *dev, int crtc); | |||
| 1006 | extern u32 gm45_get_vblank_counter(struct drm_device *dev, int crtc); | 1009 | extern u32 gm45_get_vblank_counter(struct drm_device *dev, int crtc); |
| 1007 | extern int i915_vblank_swap(struct drm_device *dev, void *data, | 1010 | extern int i915_vblank_swap(struct drm_device *dev, void *data, |
| 1008 | struct drm_file *file_priv); | 1011 | struct drm_file *file_priv); |
| 1009 | extern void i915_enable_irq(drm_i915_private_t *dev_priv, u32 mask); | ||
| 1010 | extern void i915_disable_irq(drm_i915_private_t *dev_priv, u32 mask); | ||
| 1011 | extern void ironlake_enable_graphics_irq(drm_i915_private_t *dev_priv, | ||
| 1012 | u32 mask); | ||
| 1013 | extern void ironlake_disable_graphics_irq(drm_i915_private_t *dev_priv, | ||
| 1014 | u32 mask); | ||
| 1015 | 1012 | ||
| 1016 | void | 1013 | void |
| 1017 | i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); | 1014 | i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
| @@ -1091,10 +1088,10 @@ int i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, | |||
| 1091 | struct drm_file *file_priv); | 1088 | struct drm_file *file_priv); |
| 1092 | void i915_gem_load(struct drm_device *dev); | 1089 | void i915_gem_load(struct drm_device *dev); |
| 1093 | int i915_gem_init_object(struct drm_gem_object *obj); | 1090 | int i915_gem_init_object(struct drm_gem_object *obj); |
| 1094 | void i915_gem_flush_ring(struct drm_device *dev, | 1091 | int __must_check i915_gem_flush_ring(struct drm_device *dev, |
| 1095 | struct intel_ring_buffer *ring, | 1092 | struct intel_ring_buffer *ring, |
| 1096 | uint32_t invalidate_domains, | 1093 | uint32_t invalidate_domains, |
| 1097 | uint32_t flush_domains); | 1094 | uint32_t flush_domains); |
| 1098 | struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, | 1095 | struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, |
| 1099 | size_t size); | 1096 | size_t size); |
| 1100 | void i915_gem_free_object(struct drm_gem_object *obj); | 1097 | void i915_gem_free_object(struct drm_gem_object *obj); |
| @@ -1265,6 +1262,7 @@ extern void intel_disable_fbc(struct drm_device *dev); | |||
| 1265 | extern void intel_enable_fbc(struct drm_crtc *crtc, unsigned long interval); | 1262 | extern void intel_enable_fbc(struct drm_crtc *crtc, unsigned long interval); |
| 1266 | extern bool intel_fbc_enabled(struct drm_device *dev); | 1263 | extern bool intel_fbc_enabled(struct drm_device *dev); |
| 1267 | extern bool ironlake_set_drps(struct drm_device *dev, u8 val); | 1264 | extern bool ironlake_set_drps(struct drm_device *dev, u8 val); |
| 1265 | extern void ironlake_enable_rc6(struct drm_device *dev); | ||
| 1268 | extern void gen6_set_rps(struct drm_device *dev, u8 val); | 1266 | extern void gen6_set_rps(struct drm_device *dev, u8 val); |
| 1269 | extern void intel_detect_pch (struct drm_device *dev); | 1267 | extern void intel_detect_pch (struct drm_device *dev); |
| 1270 | extern int intel_trans_dp_port_sel (struct drm_crtc *crtc); | 1268 | extern int intel_trans_dp_port_sel (struct drm_crtc *crtc); |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index c79c0b62ef60..3dfc848ff755 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
| @@ -35,18 +35,18 @@ | |||
| 35 | #include <linux/swap.h> | 35 | #include <linux/swap.h> |
| 36 | #include <linux/pci.h> | 36 | #include <linux/pci.h> |
| 37 | 37 | ||
| 38 | static void i915_gem_object_flush_gpu_write_domain(struct drm_i915_gem_object *obj); | 38 | static __must_check int i915_gem_object_flush_gpu_write_domain(struct drm_i915_gem_object *obj); |
| 39 | static void i915_gem_object_flush_gtt_write_domain(struct drm_i915_gem_object *obj); | 39 | static void i915_gem_object_flush_gtt_write_domain(struct drm_i915_gem_object *obj); |
| 40 | static void i915_gem_object_flush_cpu_write_domain(struct drm_i915_gem_object *obj); | 40 | static void i915_gem_object_flush_cpu_write_domain(struct drm_i915_gem_object *obj); |
| 41 | static int i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, | 41 | static __must_check int i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, |
| 42 | bool write); | 42 | bool write); |
| 43 | static int i915_gem_object_set_cpu_read_domain_range(struct drm_i915_gem_object *obj, | 43 | static __must_check int i915_gem_object_set_cpu_read_domain_range(struct drm_i915_gem_object *obj, |
| 44 | uint64_t offset, | 44 | uint64_t offset, |
| 45 | uint64_t size); | 45 | uint64_t size); |
| 46 | static void i915_gem_object_set_to_full_cpu_read_domain(struct drm_i915_gem_object *obj); | 46 | static void i915_gem_object_set_to_full_cpu_read_domain(struct drm_i915_gem_object *obj); |
| 47 | static int i915_gem_object_bind_to_gtt(struct drm_i915_gem_object *obj, | 47 | static __must_check int i915_gem_object_bind_to_gtt(struct drm_i915_gem_object *obj, |
| 48 | unsigned alignment, | 48 | unsigned alignment, |
| 49 | bool map_and_fenceable); | 49 | bool map_and_fenceable); |
| 50 | static void i915_gem_clear_fence_reg(struct drm_device *dev, | 50 | static void i915_gem_clear_fence_reg(struct drm_device *dev, |
| 51 | struct drm_i915_fence_reg *reg); | 51 | struct drm_i915_fence_reg *reg); |
| 52 | static int i915_gem_phys_pwrite(struct drm_device *dev, | 52 | static int i915_gem_phys_pwrite(struct drm_device *dev, |
| @@ -1935,6 +1935,8 @@ i915_gem_retire_work_handler(struct work_struct *work) | |||
| 1935 | { | 1935 | { |
| 1936 | drm_i915_private_t *dev_priv; | 1936 | drm_i915_private_t *dev_priv; |
| 1937 | struct drm_device *dev; | 1937 | struct drm_device *dev; |
| 1938 | bool idle; | ||
| 1939 | int i; | ||
| 1938 | 1940 | ||
| 1939 | dev_priv = container_of(work, drm_i915_private_t, | 1941 | dev_priv = container_of(work, drm_i915_private_t, |
| 1940 | mm.retire_work.work); | 1942 | mm.retire_work.work); |
| @@ -1948,11 +1950,31 @@ i915_gem_retire_work_handler(struct work_struct *work) | |||
| 1948 | 1950 | ||
| 1949 | i915_gem_retire_requests(dev); | 1951 | i915_gem_retire_requests(dev); |
| 1950 | 1952 | ||
| 1951 | if (!dev_priv->mm.suspended && | 1953 | /* Send a periodic flush down the ring so we don't hold onto GEM |
| 1952 | (!list_empty(&dev_priv->ring[RCS].request_list) || | 1954 | * objects indefinitely. |
| 1953 | !list_empty(&dev_priv->ring[VCS].request_list) || | 1955 | */ |
| 1954 | !list_empty(&dev_priv->ring[BCS].request_list))) | 1956 | idle = true; |
| 1957 | for (i = 0; i < I915_NUM_RINGS; i++) { | ||
| 1958 | struct intel_ring_buffer *ring = &dev_priv->ring[i]; | ||
| 1959 | |||
| 1960 | if (!list_empty(&ring->gpu_write_list)) { | ||
| 1961 | struct drm_i915_gem_request *request; | ||
| 1962 | int ret; | ||
| 1963 | |||
| 1964 | ret = i915_gem_flush_ring(dev, ring, 0, | ||
| 1965 | I915_GEM_GPU_DOMAINS); | ||
| 1966 | request = kzalloc(sizeof(*request), GFP_KERNEL); | ||
| 1967 | if (ret || request == NULL || | ||
| 1968 | i915_add_request(dev, NULL, request, ring)) | ||
| 1969 | kfree(request); | ||
| 1970 | } | ||
| 1971 | |||
| 1972 | idle &= list_empty(&ring->request_list); | ||
| 1973 | } | ||
| 1974 | |||
| 1975 | if (!dev_priv->mm.suspended && !idle) | ||
| 1955 | queue_delayed_work(dev_priv->wq, &dev_priv->mm.retire_work, HZ); | 1976 | queue_delayed_work(dev_priv->wq, &dev_priv->mm.retire_work, HZ); |
| 1977 | |||
| 1956 | mutex_unlock(&dev->struct_mutex); | 1978 | mutex_unlock(&dev->struct_mutex); |
| 1957 | } | 1979 | } |
| 1958 | 1980 | ||
| @@ -2142,25 +2164,37 @@ i915_gem_object_unbind(struct drm_i915_gem_object *obj) | |||
| 2142 | return ret; | 2164 | return ret; |
| 2143 | } | 2165 | } |
| 2144 | 2166 | ||
| 2145 | void | 2167 | int |
| 2146 | i915_gem_flush_ring(struct drm_device *dev, | 2168 | i915_gem_flush_ring(struct drm_device *dev, |
| 2147 | struct intel_ring_buffer *ring, | 2169 | struct intel_ring_buffer *ring, |
| 2148 | uint32_t invalidate_domains, | 2170 | uint32_t invalidate_domains, |
| 2149 | uint32_t flush_domains) | 2171 | uint32_t flush_domains) |
| 2150 | { | 2172 | { |
| 2151 | ring->flush(ring, invalidate_domains, flush_domains); | 2173 | int ret; |
| 2174 | |||
| 2175 | ret = ring->flush(ring, invalidate_domains, flush_domains); | ||
| 2176 | if (ret) | ||
| 2177 | return ret; | ||
| 2178 | |||
| 2152 | i915_gem_process_flushing_list(dev, flush_domains, ring); | 2179 | i915_gem_process_flushing_list(dev, flush_domains, ring); |
| 2180 | return 0; | ||
| 2153 | } | 2181 | } |
| 2154 | 2182 | ||
| 2155 | static int i915_ring_idle(struct drm_device *dev, | 2183 | static int i915_ring_idle(struct drm_device *dev, |
| 2156 | struct intel_ring_buffer *ring) | 2184 | struct intel_ring_buffer *ring) |
| 2157 | { | 2185 | { |
| 2186 | int ret; | ||
| 2187 | |||
| 2158 | if (list_empty(&ring->gpu_write_list) && list_empty(&ring->active_list)) | 2188 | if (list_empty(&ring->gpu_write_list) && list_empty(&ring->active_list)) |
| 2159 | return 0; | 2189 | return 0; |
| 2160 | 2190 | ||
| 2161 | if (!list_empty(&ring->gpu_write_list)) | 2191 | if (!list_empty(&ring->gpu_write_list)) { |
| 2162 | i915_gem_flush_ring(dev, ring, | 2192 | ret = i915_gem_flush_ring(dev, ring, |
| 2163 | I915_GEM_GPU_DOMAINS, I915_GEM_GPU_DOMAINS); | 2193 | I915_GEM_GPU_DOMAINS, I915_GEM_GPU_DOMAINS); |
| 2194 | if (ret) | ||
| 2195 | return ret; | ||
| 2196 | } | ||
| 2197 | |||
| 2164 | return i915_wait_request(dev, | 2198 | return i915_wait_request(dev, |
| 2165 | i915_gem_next_request_seqno(dev, ring), | 2199 | i915_gem_next_request_seqno(dev, ring), |
| 2166 | ring); | 2200 | ring); |
| @@ -2370,10 +2404,13 @@ i915_gem_object_flush_fence(struct drm_i915_gem_object *obj, | |||
| 2370 | int ret; | 2404 | int ret; |
| 2371 | 2405 | ||
| 2372 | if (obj->fenced_gpu_access) { | 2406 | if (obj->fenced_gpu_access) { |
| 2373 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) | 2407 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) { |
| 2374 | i915_gem_flush_ring(obj->base.dev, | 2408 | ret = i915_gem_flush_ring(obj->base.dev, |
| 2375 | obj->last_fenced_ring, | 2409 | obj->last_fenced_ring, |
| 2376 | 0, obj->base.write_domain); | 2410 | 0, obj->base.write_domain); |
| 2411 | if (ret) | ||
| 2412 | return ret; | ||
| 2413 | } | ||
| 2377 | 2414 | ||
| 2378 | obj->fenced_gpu_access = false; | 2415 | obj->fenced_gpu_access = false; |
| 2379 | } | 2416 | } |
| @@ -2393,6 +2430,12 @@ i915_gem_object_flush_fence(struct drm_i915_gem_object *obj, | |||
| 2393 | obj->last_fenced_ring = NULL; | 2430 | obj->last_fenced_ring = NULL; |
| 2394 | } | 2431 | } |
| 2395 | 2432 | ||
| 2433 | /* Ensure that all CPU reads are completed before installing a fence | ||
| 2434 | * and all writes before removing the fence. | ||
| 2435 | */ | ||
| 2436 | if (obj->base.read_domains & I915_GEM_DOMAIN_GTT) | ||
| 2437 | mb(); | ||
| 2438 | |||
| 2396 | return 0; | 2439 | return 0; |
| 2397 | } | 2440 | } |
| 2398 | 2441 | ||
| @@ -2523,9 +2566,12 @@ i915_gem_object_get_fence(struct drm_i915_gem_object *obj, | |||
| 2523 | return ret; | 2566 | return ret; |
| 2524 | } else if (obj->tiling_changed) { | 2567 | } else if (obj->tiling_changed) { |
| 2525 | if (obj->fenced_gpu_access) { | 2568 | if (obj->fenced_gpu_access) { |
| 2526 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) | 2569 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) { |
| 2527 | i915_gem_flush_ring(obj->base.dev, obj->ring, | 2570 | ret = i915_gem_flush_ring(obj->base.dev, obj->ring, |
| 2528 | 0, obj->base.write_domain); | 2571 | 0, obj->base.write_domain); |
| 2572 | if (ret) | ||
| 2573 | return ret; | ||
| 2574 | } | ||
| 2529 | 2575 | ||
| 2530 | obj->fenced_gpu_access = false; | 2576 | obj->fenced_gpu_access = false; |
| 2531 | } | 2577 | } |
| @@ -2736,10 +2782,8 @@ i915_gem_object_bind_to_gtt(struct drm_i915_gem_object *obj, | |||
| 2736 | obj->gtt_space = NULL; | 2782 | obj->gtt_space = NULL; |
| 2737 | 2783 | ||
| 2738 | if (ret == -ENOMEM) { | 2784 | if (ret == -ENOMEM) { |
| 2739 | /* first try to clear up some space from the GTT */ | 2785 | /* first try to reclaim some memory by clearing the GTT */ |
| 2740 | ret = i915_gem_evict_something(dev, size, | 2786 | ret = i915_gem_evict_everything(dev, false); |
| 2741 | alignment, | ||
| 2742 | map_and_fenceable); | ||
| 2743 | if (ret) { | 2787 | if (ret) { |
| 2744 | /* now try to shrink everyone else */ | 2788 | /* now try to shrink everyone else */ |
| 2745 | if (gfpmask) { | 2789 | if (gfpmask) { |
| @@ -2747,7 +2791,7 @@ i915_gem_object_bind_to_gtt(struct drm_i915_gem_object *obj, | |||
| 2747 | goto search_free; | 2791 | goto search_free; |
| 2748 | } | 2792 | } |
| 2749 | 2793 | ||
| 2750 | return ret; | 2794 | return -ENOMEM; |
| 2751 | } | 2795 | } |
| 2752 | 2796 | ||
| 2753 | goto search_free; | 2797 | goto search_free; |
| @@ -2762,9 +2806,7 @@ i915_gem_object_bind_to_gtt(struct drm_i915_gem_object *obj, | |||
| 2762 | drm_mm_put_block(obj->gtt_space); | 2806 | drm_mm_put_block(obj->gtt_space); |
| 2763 | obj->gtt_space = NULL; | 2807 | obj->gtt_space = NULL; |
| 2764 | 2808 | ||
| 2765 | ret = i915_gem_evict_something(dev, size, | 2809 | if (i915_gem_evict_everything(dev, false)) |
| 2766 | alignment, map_and_fenceable); | ||
| 2767 | if (ret) | ||
| 2768 | return ret; | 2810 | return ret; |
| 2769 | 2811 | ||
| 2770 | goto search_free; | 2812 | goto search_free; |
| @@ -2811,17 +2853,16 @@ i915_gem_clflush_object(struct drm_i915_gem_object *obj) | |||
| 2811 | } | 2853 | } |
| 2812 | 2854 | ||
| 2813 | /** Flushes any GPU write domain for the object if it's dirty. */ | 2855 | /** Flushes any GPU write domain for the object if it's dirty. */ |
| 2814 | static void | 2856 | static int |
| 2815 | i915_gem_object_flush_gpu_write_domain(struct drm_i915_gem_object *obj) | 2857 | i915_gem_object_flush_gpu_write_domain(struct drm_i915_gem_object *obj) |
| 2816 | { | 2858 | { |
| 2817 | struct drm_device *dev = obj->base.dev; | 2859 | struct drm_device *dev = obj->base.dev; |
| 2818 | 2860 | ||
| 2819 | if ((obj->base.write_domain & I915_GEM_GPU_DOMAINS) == 0) | 2861 | if ((obj->base.write_domain & I915_GEM_GPU_DOMAINS) == 0) |
| 2820 | return; | 2862 | return 0; |
| 2821 | 2863 | ||
| 2822 | /* Queue the GPU write cache flushing we need. */ | 2864 | /* Queue the GPU write cache flushing we need. */ |
| 2823 | i915_gem_flush_ring(dev, obj->ring, 0, obj->base.write_domain); | 2865 | return i915_gem_flush_ring(dev, obj->ring, 0, obj->base.write_domain); |
| 2824 | BUG_ON(obj->base.write_domain); | ||
| 2825 | } | 2866 | } |
| 2826 | 2867 | ||
| 2827 | /** Flushes the GTT write domain for the object if it's dirty. */ | 2868 | /** Flushes the GTT write domain for the object if it's dirty. */ |
| @@ -2833,10 +2874,16 @@ i915_gem_object_flush_gtt_write_domain(struct drm_i915_gem_object *obj) | |||
| 2833 | if (obj->base.write_domain != I915_GEM_DOMAIN_GTT) | 2874 | if (obj->base.write_domain != I915_GEM_DOMAIN_GTT) |
| 2834 | return; | 2875 | return; |
| 2835 | 2876 | ||
| 2836 | /* No actual flushing is required for the GTT write domain. Writes | 2877 | /* No actual flushing is required for the GTT write domain. Writes |
| 2837 | * to it immediately go to main memory as far as we know, so there's | 2878 | * to it immediately go to main memory as far as we know, so there's |
| 2838 | * no chipset flush. It also doesn't land in render cache. | 2879 | * no chipset flush. It also doesn't land in render cache. |
| 2880 | * | ||
| 2881 | * However, we do have to enforce the order so that all writes through | ||
| 2882 | * the GTT land before any writes to the device, such as updates to | ||
| 2883 | * the GATT itself. | ||
| 2839 | */ | 2884 | */ |
| 2885 | wmb(); | ||
| 2886 | |||
| 2840 | i915_gem_release_mmap(obj); | 2887 | i915_gem_release_mmap(obj); |
| 2841 | 2888 | ||
| 2842 | old_write_domain = obj->base.write_domain; | 2889 | old_write_domain = obj->base.write_domain; |
| @@ -2882,7 +2929,10 @@ i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write) | |||
| 2882 | if (obj->gtt_space == NULL) | 2929 | if (obj->gtt_space == NULL) |
| 2883 | return -EINVAL; | 2930 | return -EINVAL; |
| 2884 | 2931 | ||
| 2885 | i915_gem_object_flush_gpu_write_domain(obj); | 2932 | ret = i915_gem_object_flush_gpu_write_domain(obj); |
| 2933 | if (ret) | ||
| 2934 | return ret; | ||
| 2935 | |||
| 2886 | if (obj->pending_gpu_write || write) { | 2936 | if (obj->pending_gpu_write || write) { |
| 2887 | ret = i915_gem_object_wait_rendering(obj, true); | 2937 | ret = i915_gem_object_wait_rendering(obj, true); |
| 2888 | if (ret) | 2938 | if (ret) |
| @@ -2927,7 +2977,10 @@ i915_gem_object_set_to_display_plane(struct drm_i915_gem_object *obj, | |||
| 2927 | if (obj->gtt_space == NULL) | 2977 | if (obj->gtt_space == NULL) |
| 2928 | return -EINVAL; | 2978 | return -EINVAL; |
| 2929 | 2979 | ||
| 2930 | i915_gem_object_flush_gpu_write_domain(obj); | 2980 | ret = i915_gem_object_flush_gpu_write_domain(obj); |
| 2981 | if (ret) | ||
| 2982 | return ret; | ||
| 2983 | |||
| 2931 | 2984 | ||
| 2932 | /* Currently, we are always called from an non-interruptible context. */ | 2985 | /* Currently, we are always called from an non-interruptible context. */ |
| 2933 | if (pipelined != obj->ring) { | 2986 | if (pipelined != obj->ring) { |
| @@ -2952,12 +3005,17 @@ int | |||
| 2952 | i915_gem_object_flush_gpu(struct drm_i915_gem_object *obj, | 3005 | i915_gem_object_flush_gpu(struct drm_i915_gem_object *obj, |
| 2953 | bool interruptible) | 3006 | bool interruptible) |
| 2954 | { | 3007 | { |
| 3008 | int ret; | ||
| 3009 | |||
| 2955 | if (!obj->active) | 3010 | if (!obj->active) |
| 2956 | return 0; | 3011 | return 0; |
| 2957 | 3012 | ||
| 2958 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) | 3013 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) { |
| 2959 | i915_gem_flush_ring(obj->base.dev, obj->ring, | 3014 | ret = i915_gem_flush_ring(obj->base.dev, obj->ring, |
| 2960 | 0, obj->base.write_domain); | 3015 | 0, obj->base.write_domain); |
| 3016 | if (ret) | ||
| 3017 | return ret; | ||
| 3018 | } | ||
| 2961 | 3019 | ||
| 2962 | return i915_gem_object_wait_rendering(obj, interruptible); | 3020 | return i915_gem_object_wait_rendering(obj, interruptible); |
| 2963 | } | 3021 | } |
| @@ -2974,7 +3032,10 @@ i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write) | |||
| 2974 | uint32_t old_write_domain, old_read_domains; | 3032 | uint32_t old_write_domain, old_read_domains; |
| 2975 | int ret; | 3033 | int ret; |
| 2976 | 3034 | ||
| 2977 | i915_gem_object_flush_gpu_write_domain(obj); | 3035 | ret = i915_gem_object_flush_gpu_write_domain(obj); |
| 3036 | if (ret) | ||
| 3037 | return ret; | ||
| 3038 | |||
| 2978 | ret = i915_gem_object_wait_rendering(obj, true); | 3039 | ret = i915_gem_object_wait_rendering(obj, true); |
| 2979 | if (ret) | 3040 | if (ret) |
| 2980 | return ret; | 3041 | return ret; |
| @@ -3069,7 +3130,10 @@ i915_gem_object_set_cpu_read_domain_range(struct drm_i915_gem_object *obj, | |||
| 3069 | if (offset == 0 && size == obj->base.size) | 3130 | if (offset == 0 && size == obj->base.size) |
| 3070 | return i915_gem_object_set_to_cpu_domain(obj, 0); | 3131 | return i915_gem_object_set_to_cpu_domain(obj, 0); |
| 3071 | 3132 | ||
| 3072 | i915_gem_object_flush_gpu_write_domain(obj); | 3133 | ret = i915_gem_object_flush_gpu_write_domain(obj); |
| 3134 | if (ret) | ||
| 3135 | return ret; | ||
| 3136 | |||
| 3073 | ret = i915_gem_object_wait_rendering(obj, true); | 3137 | ret = i915_gem_object_wait_rendering(obj, true); |
| 3074 | if (ret) | 3138 | if (ret) |
| 3075 | return ret; | 3139 | return ret; |
| @@ -3362,8 +3426,8 @@ i915_gem_busy_ioctl(struct drm_device *dev, void *data, | |||
| 3362 | * flush earlier is beneficial. | 3426 | * flush earlier is beneficial. |
| 3363 | */ | 3427 | */ |
| 3364 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) { | 3428 | if (obj->base.write_domain & I915_GEM_GPU_DOMAINS) { |
| 3365 | i915_gem_flush_ring(dev, obj->ring, | 3429 | ret = i915_gem_flush_ring(dev, obj->ring, |
| 3366 | 0, obj->base.write_domain); | 3430 | 0, obj->base.write_domain); |
| 3367 | } else if (obj->ring->outstanding_lazy_request == | 3431 | } else if (obj->ring->outstanding_lazy_request == |
| 3368 | obj->last_rendering_seqno) { | 3432 | obj->last_rendering_seqno) { |
| 3369 | struct drm_i915_gem_request *request; | 3433 | struct drm_i915_gem_request *request; |
diff --git a/drivers/gpu/drm/i915/i915_gem_evict.c b/drivers/gpu/drm/i915/i915_gem_evict.c index 78b8cf90c922..3d39005540aa 100644 --- a/drivers/gpu/drm/i915/i915_gem_evict.c +++ b/drivers/gpu/drm/i915/i915_gem_evict.c | |||
| @@ -127,9 +127,15 @@ i915_gem_evict_something(struct drm_device *dev, int min_size, | |||
| 127 | } | 127 | } |
| 128 | 128 | ||
| 129 | /* Nothing found, clean up and bail out! */ | 129 | /* Nothing found, clean up and bail out! */ |
| 130 | list_for_each_entry(obj, &unwind_list, exec_list) { | 130 | while (!list_empty(&unwind_list)) { |
| 131 | obj = list_first_entry(&unwind_list, | ||
| 132 | struct drm_i915_gem_object, | ||
| 133 | exec_list); | ||
| 134 | |||
| 131 | ret = drm_mm_scan_remove_block(obj->gtt_space); | 135 | ret = drm_mm_scan_remove_block(obj->gtt_space); |
| 132 | BUG_ON(ret); | 136 | BUG_ON(ret); |
| 137 | |||
| 138 | list_del_init(&obj->exec_list); | ||
| 133 | drm_gem_object_unreference(&obj->base); | 139 | drm_gem_object_unreference(&obj->base); |
| 134 | } | 140 | } |
| 135 | 141 | ||
| @@ -162,6 +168,7 @@ found: | |||
| 162 | exec_list); | 168 | exec_list); |
| 163 | if (ret == 0) | 169 | if (ret == 0) |
| 164 | ret = i915_gem_object_unbind(obj); | 170 | ret = i915_gem_object_unbind(obj); |
| 171 | |||
| 165 | list_del_init(&obj->exec_list); | 172 | list_del_init(&obj->exec_list); |
| 166 | drm_gem_object_unreference(&obj->base); | 173 | drm_gem_object_unreference(&obj->base); |
| 167 | } | 174 | } |
diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c index 61129e6759eb..e69834341ef0 100644 --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c | |||
| @@ -268,7 +268,6 @@ eb_destroy(struct eb_objects *eb) | |||
| 268 | static int | 268 | static int |
| 269 | i915_gem_execbuffer_relocate_entry(struct drm_i915_gem_object *obj, | 269 | i915_gem_execbuffer_relocate_entry(struct drm_i915_gem_object *obj, |
| 270 | struct eb_objects *eb, | 270 | struct eb_objects *eb, |
| 271 | struct drm_i915_gem_exec_object2 *entry, | ||
| 272 | struct drm_i915_gem_relocation_entry *reloc) | 271 | struct drm_i915_gem_relocation_entry *reloc) |
| 273 | { | 272 | { |
| 274 | struct drm_device *dev = obj->base.dev; | 273 | struct drm_device *dev = obj->base.dev; |
| @@ -411,10 +410,10 @@ i915_gem_execbuffer_relocate_entry(struct drm_i915_gem_object *obj, | |||
| 411 | 410 | ||
| 412 | static int | 411 | static int |
| 413 | i915_gem_execbuffer_relocate_object(struct drm_i915_gem_object *obj, | 412 | i915_gem_execbuffer_relocate_object(struct drm_i915_gem_object *obj, |
| 414 | struct eb_objects *eb, | 413 | struct eb_objects *eb) |
| 415 | struct drm_i915_gem_exec_object2 *entry) | ||
| 416 | { | 414 | { |
| 417 | struct drm_i915_gem_relocation_entry __user *user_relocs; | 415 | struct drm_i915_gem_relocation_entry __user *user_relocs; |
| 416 | struct drm_i915_gem_exec_object2 *entry = obj->exec_entry; | ||
| 418 | int i, ret; | 417 | int i, ret; |
| 419 | 418 | ||
| 420 | user_relocs = (void __user *)(uintptr_t)entry->relocs_ptr; | 419 | user_relocs = (void __user *)(uintptr_t)entry->relocs_ptr; |
| @@ -426,7 +425,7 @@ i915_gem_execbuffer_relocate_object(struct drm_i915_gem_object *obj, | |||
| 426 | sizeof(reloc))) | 425 | sizeof(reloc))) |
| 427 | return -EFAULT; | 426 | return -EFAULT; |
| 428 | 427 | ||
| 429 | ret = i915_gem_execbuffer_relocate_entry(obj, eb, entry, &reloc); | 428 | ret = i915_gem_execbuffer_relocate_entry(obj, eb, &reloc); |
| 430 | if (ret) | 429 | if (ret) |
| 431 | return ret; | 430 | return ret; |
| 432 | 431 | ||
| @@ -442,13 +441,13 @@ i915_gem_execbuffer_relocate_object(struct drm_i915_gem_object *obj, | |||
| 442 | static int | 441 | static int |
| 443 | i915_gem_execbuffer_relocate_object_slow(struct drm_i915_gem_object *obj, | 442 | i915_gem_execbuffer_relocate_object_slow(struct drm_i915_gem_object *obj, |
| 444 | struct eb_objects *eb, | 443 | struct eb_objects *eb, |
| 445 | struct drm_i915_gem_exec_object2 *entry, | ||
| 446 | struct drm_i915_gem_relocation_entry *relocs) | 444 | struct drm_i915_gem_relocation_entry *relocs) |
| 447 | { | 445 | { |
| 446 | const struct drm_i915_gem_exec_object2 *entry = obj->exec_entry; | ||
| 448 | int i, ret; | 447 | int i, ret; |
| 449 | 448 | ||
| 450 | for (i = 0; i < entry->relocation_count; i++) { | 449 | for (i = 0; i < entry->relocation_count; i++) { |
| 451 | ret = i915_gem_execbuffer_relocate_entry(obj, eb, entry, &relocs[i]); | 450 | ret = i915_gem_execbuffer_relocate_entry(obj, eb, &relocs[i]); |
| 452 | if (ret) | 451 | if (ret) |
| 453 | return ret; | 452 | return ret; |
| 454 | } | 453 | } |
| @@ -459,8 +458,7 @@ i915_gem_execbuffer_relocate_object_slow(struct drm_i915_gem_object *obj, | |||
| 459 | static int | 458 | static int |
| 460 | i915_gem_execbuffer_relocate(struct drm_device *dev, | 459 | i915_gem_execbuffer_relocate(struct drm_device *dev, |
| 461 | struct eb_objects *eb, | 460 | struct eb_objects *eb, |
| 462 | struct list_head *objects, | 461 | struct list_head *objects) |
| 463 | struct drm_i915_gem_exec_object2 *exec) | ||
| 464 | { | 462 | { |
| 465 | struct drm_i915_gem_object *obj; | 463 | struct drm_i915_gem_object *obj; |
| 466 | int ret; | 464 | int ret; |
| @@ -468,7 +466,7 @@ i915_gem_execbuffer_relocate(struct drm_device *dev, | |||
| 468 | list_for_each_entry(obj, objects, exec_list) { | 466 | list_for_each_entry(obj, objects, exec_list) { |
| 469 | obj->base.pending_read_domains = 0; | 467 | obj->base.pending_read_domains = 0; |
| 470 | obj->base.pending_write_domain = 0; | 468 | obj->base.pending_write_domain = 0; |
| 471 | ret = i915_gem_execbuffer_relocate_object(obj, eb, exec++); | 469 | ret = i915_gem_execbuffer_relocate_object(obj, eb); |
| 472 | if (ret) | 470 | if (ret) |
| 473 | return ret; | 471 | return ret; |
| 474 | } | 472 | } |
| @@ -479,13 +477,36 @@ i915_gem_execbuffer_relocate(struct drm_device *dev, | |||
| 479 | static int | 477 | static int |
| 480 | i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, | 478 | i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, |
| 481 | struct drm_file *file, | 479 | struct drm_file *file, |
| 482 | struct list_head *objects, | 480 | struct list_head *objects) |
| 483 | struct drm_i915_gem_exec_object2 *exec) | ||
| 484 | { | 481 | { |
| 485 | struct drm_i915_gem_object *obj; | 482 | struct drm_i915_gem_object *obj; |
| 486 | struct drm_i915_gem_exec_object2 *entry; | ||
| 487 | int ret, retry; | 483 | int ret, retry; |
| 488 | bool has_fenced_gpu_access = INTEL_INFO(ring->dev)->gen < 4; | 484 | bool has_fenced_gpu_access = INTEL_INFO(ring->dev)->gen < 4; |
| 485 | struct list_head ordered_objects; | ||
| 486 | |||
| 487 | INIT_LIST_HEAD(&ordered_objects); | ||
| 488 | while (!list_empty(objects)) { | ||
| 489 | struct drm_i915_gem_exec_object2 *entry; | ||
| 490 | bool need_fence, need_mappable; | ||
| 491 | |||
| 492 | obj = list_first_entry(objects, | ||
| 493 | struct drm_i915_gem_object, | ||
| 494 | exec_list); | ||
| 495 | entry = obj->exec_entry; | ||
| 496 | |||
| 497 | need_fence = | ||
| 498 | has_fenced_gpu_access && | ||
| 499 | entry->flags & EXEC_OBJECT_NEEDS_FENCE && | ||
| 500 | obj->tiling_mode != I915_TILING_NONE; | ||
| 501 | need_mappable = | ||
| 502 | entry->relocation_count ? true : need_fence; | ||
| 503 | |||
| 504 | if (need_mappable) | ||
| 505 | list_move(&obj->exec_list, &ordered_objects); | ||
| 506 | else | ||
| 507 | list_move_tail(&obj->exec_list, &ordered_objects); | ||
| 508 | } | ||
| 509 | list_splice(&ordered_objects, objects); | ||
| 489 | 510 | ||
| 490 | /* Attempt to pin all of the buffers into the GTT. | 511 | /* Attempt to pin all of the buffers into the GTT. |
| 491 | * This is done in 3 phases: | 512 | * This is done in 3 phases: |
| @@ -504,14 +525,11 @@ i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, | |||
| 504 | ret = 0; | 525 | ret = 0; |
| 505 | 526 | ||
| 506 | /* Unbind any ill-fitting objects or pin. */ | 527 | /* Unbind any ill-fitting objects or pin. */ |
| 507 | entry = exec; | ||
| 508 | list_for_each_entry(obj, objects, exec_list) { | 528 | list_for_each_entry(obj, objects, exec_list) { |
| 529 | struct drm_i915_gem_exec_object2 *entry = obj->exec_entry; | ||
| 509 | bool need_fence, need_mappable; | 530 | bool need_fence, need_mappable; |
| 510 | 531 | if (!obj->gtt_space) | |
| 511 | if (!obj->gtt_space) { | ||
| 512 | entry++; | ||
| 513 | continue; | 532 | continue; |
| 514 | } | ||
| 515 | 533 | ||
| 516 | need_fence = | 534 | need_fence = |
| 517 | has_fenced_gpu_access && | 535 | has_fenced_gpu_access && |
| @@ -534,8 +552,8 @@ i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, | |||
| 534 | } | 552 | } |
| 535 | 553 | ||
| 536 | /* Bind fresh objects */ | 554 | /* Bind fresh objects */ |
| 537 | entry = exec; | ||
| 538 | list_for_each_entry(obj, objects, exec_list) { | 555 | list_for_each_entry(obj, objects, exec_list) { |
| 556 | struct drm_i915_gem_exec_object2 *entry = obj->exec_entry; | ||
| 539 | bool need_fence; | 557 | bool need_fence; |
| 540 | 558 | ||
| 541 | need_fence = | 559 | need_fence = |
| @@ -570,7 +588,6 @@ i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, | |||
| 570 | } | 588 | } |
| 571 | 589 | ||
| 572 | entry->offset = obj->gtt_offset; | 590 | entry->offset = obj->gtt_offset; |
| 573 | entry++; | ||
| 574 | } | 591 | } |
| 575 | 592 | ||
| 576 | /* Decrement pin count for bound objects */ | 593 | /* Decrement pin count for bound objects */ |
| @@ -622,7 +639,7 @@ i915_gem_execbuffer_relocate_slow(struct drm_device *dev, | |||
| 622 | int i, total, ret; | 639 | int i, total, ret; |
| 623 | 640 | ||
| 624 | /* We may process another execbuffer during the unlock... */ | 641 | /* We may process another execbuffer during the unlock... */ |
| 625 | while (list_empty(objects)) { | 642 | while (!list_empty(objects)) { |
| 626 | obj = list_first_entry(objects, | 643 | obj = list_first_entry(objects, |
| 627 | struct drm_i915_gem_object, | 644 | struct drm_i915_gem_object, |
| 628 | exec_list); | 645 | exec_list); |
| @@ -665,7 +682,6 @@ i915_gem_execbuffer_relocate_slow(struct drm_device *dev, | |||
| 665 | } | 682 | } |
| 666 | 683 | ||
| 667 | /* reacquire the objects */ | 684 | /* reacquire the objects */ |
| 668 | INIT_LIST_HEAD(objects); | ||
| 669 | eb_reset(eb); | 685 | eb_reset(eb); |
| 670 | for (i = 0; i < count; i++) { | 686 | for (i = 0; i < count; i++) { |
| 671 | struct drm_i915_gem_object *obj; | 687 | struct drm_i915_gem_object *obj; |
| @@ -681,10 +697,11 @@ i915_gem_execbuffer_relocate_slow(struct drm_device *dev, | |||
| 681 | 697 | ||
| 682 | list_add_tail(&obj->exec_list, objects); | 698 | list_add_tail(&obj->exec_list, objects); |
| 683 | obj->exec_handle = exec[i].handle; | 699 | obj->exec_handle = exec[i].handle; |
| 700 | obj->exec_entry = &exec[i]; | ||
| 684 | eb_add_object(eb, obj); | 701 | eb_add_object(eb, obj); |
| 685 | } | 702 | } |
| 686 | 703 | ||
| 687 | ret = i915_gem_execbuffer_reserve(ring, file, objects, exec); | 704 | ret = i915_gem_execbuffer_reserve(ring, file, objects); |
| 688 | if (ret) | 705 | if (ret) |
| 689 | goto err; | 706 | goto err; |
| 690 | 707 | ||
| @@ -693,7 +710,6 @@ i915_gem_execbuffer_relocate_slow(struct drm_device *dev, | |||
| 693 | obj->base.pending_read_domains = 0; | 710 | obj->base.pending_read_domains = 0; |
| 694 | obj->base.pending_write_domain = 0; | 711 | obj->base.pending_write_domain = 0; |
| 695 | ret = i915_gem_execbuffer_relocate_object_slow(obj, eb, | 712 | ret = i915_gem_execbuffer_relocate_object_slow(obj, eb, |
| 696 | exec, | ||
| 697 | reloc + total); | 713 | reloc + total); |
| 698 | if (ret) | 714 | if (ret) |
| 699 | goto err; | 715 | goto err; |
| @@ -713,25 +729,34 @@ err: | |||
| 713 | return ret; | 729 | return ret; |
| 714 | } | 730 | } |
| 715 | 731 | ||
| 716 | static void | 732 | static int |
| 717 | i915_gem_execbuffer_flush(struct drm_device *dev, | 733 | i915_gem_execbuffer_flush(struct drm_device *dev, |
| 718 | uint32_t invalidate_domains, | 734 | uint32_t invalidate_domains, |
| 719 | uint32_t flush_domains, | 735 | uint32_t flush_domains, |
| 720 | uint32_t flush_rings) | 736 | uint32_t flush_rings) |
| 721 | { | 737 | { |
| 722 | drm_i915_private_t *dev_priv = dev->dev_private; | 738 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 723 | int i; | 739 | int i, ret; |
| 724 | 740 | ||
| 725 | if (flush_domains & I915_GEM_DOMAIN_CPU) | 741 | if (flush_domains & I915_GEM_DOMAIN_CPU) |
| 726 | intel_gtt_chipset_flush(); | 742 | intel_gtt_chipset_flush(); |
| 727 | 743 | ||
| 744 | if (flush_domains & I915_GEM_DOMAIN_GTT) | ||
| 745 | wmb(); | ||
| 746 | |||
| 728 | if ((flush_domains | invalidate_domains) & I915_GEM_GPU_DOMAINS) { | 747 | if ((flush_domains | invalidate_domains) & I915_GEM_GPU_DOMAINS) { |
| 729 | for (i = 0; i < I915_NUM_RINGS; i++) | 748 | for (i = 0; i < I915_NUM_RINGS; i++) |
| 730 | if (flush_rings & (1 << i)) | 749 | if (flush_rings & (1 << i)) { |
| 731 | i915_gem_flush_ring(dev, &dev_priv->ring[i], | 750 | ret = i915_gem_flush_ring(dev, |
| 732 | invalidate_domains, | 751 | &dev_priv->ring[i], |
| 733 | flush_domains); | 752 | invalidate_domains, |
| 753 | flush_domains); | ||
| 754 | if (ret) | ||
| 755 | return ret; | ||
| 756 | } | ||
| 734 | } | 757 | } |
| 758 | |||
| 759 | return 0; | ||
| 735 | } | 760 | } |
| 736 | 761 | ||
| 737 | static int | 762 | static int |
| @@ -795,10 +820,12 @@ i915_gem_execbuffer_move_to_gpu(struct intel_ring_buffer *ring, | |||
| 795 | cd.invalidate_domains, | 820 | cd.invalidate_domains, |
| 796 | cd.flush_domains); | 821 | cd.flush_domains); |
| 797 | #endif | 822 | #endif |
| 798 | i915_gem_execbuffer_flush(ring->dev, | 823 | ret = i915_gem_execbuffer_flush(ring->dev, |
| 799 | cd.invalidate_domains, | 824 | cd.invalidate_domains, |
| 800 | cd.flush_domains, | 825 | cd.flush_domains, |
| 801 | cd.flush_rings); | 826 | cd.flush_rings); |
| 827 | if (ret) | ||
| 828 | return ret; | ||
| 802 | } | 829 | } |
| 803 | 830 | ||
| 804 | list_for_each_entry(obj, objects, exec_list) { | 831 | list_for_each_entry(obj, objects, exec_list) { |
| @@ -921,7 +948,7 @@ i915_gem_execbuffer_retire_commands(struct drm_device *dev, | |||
| 921 | struct intel_ring_buffer *ring) | 948 | struct intel_ring_buffer *ring) |
| 922 | { | 949 | { |
| 923 | struct drm_i915_gem_request *request; | 950 | struct drm_i915_gem_request *request; |
| 924 | u32 flush_domains; | 951 | u32 invalidate; |
| 925 | 952 | ||
| 926 | /* | 953 | /* |
| 927 | * Ensure that the commands in the batch buffer are | 954 | * Ensure that the commands in the batch buffer are |
| @@ -929,11 +956,13 @@ i915_gem_execbuffer_retire_commands(struct drm_device *dev, | |||
| 929 | * | 956 | * |
| 930 | * The sampler always gets flushed on i965 (sigh). | 957 | * The sampler always gets flushed on i965 (sigh). |
| 931 | */ | 958 | */ |
| 932 | flush_domains = 0; | 959 | invalidate = I915_GEM_DOMAIN_COMMAND; |
| 933 | if (INTEL_INFO(dev)->gen >= 4) | 960 | if (INTEL_INFO(dev)->gen >= 4) |
| 934 | flush_domains |= I915_GEM_DOMAIN_SAMPLER; | 961 | invalidate |= I915_GEM_DOMAIN_SAMPLER; |
| 935 | 962 | if (ring->flush(ring, invalidate, 0)) { | |
| 936 | ring->flush(ring, I915_GEM_DOMAIN_COMMAND, flush_domains); | 963 | i915_gem_next_request_seqno(dev, ring); |
| 964 | return; | ||
| 965 | } | ||
| 937 | 966 | ||
| 938 | /* Add a breadcrumb for the completion of the batch buffer */ | 967 | /* Add a breadcrumb for the completion of the batch buffer */ |
| 939 | request = kzalloc(sizeof(*request), GFP_KERNEL); | 968 | request = kzalloc(sizeof(*request), GFP_KERNEL); |
| @@ -1098,16 +1127,22 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
| 1098 | 1127 | ||
| 1099 | list_add_tail(&obj->exec_list, &objects); | 1128 | list_add_tail(&obj->exec_list, &objects); |
| 1100 | obj->exec_handle = exec[i].handle; | 1129 | obj->exec_handle = exec[i].handle; |
| 1130 | obj->exec_entry = &exec[i]; | ||
| 1101 | eb_add_object(eb, obj); | 1131 | eb_add_object(eb, obj); |
| 1102 | } | 1132 | } |
| 1103 | 1133 | ||
| 1134 | /* take note of the batch buffer before we might reorder the lists */ | ||
| 1135 | batch_obj = list_entry(objects.prev, | ||
| 1136 | struct drm_i915_gem_object, | ||
| 1137 | exec_list); | ||
| 1138 | |||
| 1104 | /* Move the objects en-masse into the GTT, evicting if necessary. */ | 1139 | /* Move the objects en-masse into the GTT, evicting if necessary. */ |
| 1105 | ret = i915_gem_execbuffer_reserve(ring, file, &objects, exec); | 1140 | ret = i915_gem_execbuffer_reserve(ring, file, &objects); |
| 1106 | if (ret) | 1141 | if (ret) |
| 1107 | goto err; | 1142 | goto err; |
| 1108 | 1143 | ||
| 1109 | /* The objects are in their final locations, apply the relocations. */ | 1144 | /* The objects are in their final locations, apply the relocations. */ |
| 1110 | ret = i915_gem_execbuffer_relocate(dev, eb, &objects, exec); | 1145 | ret = i915_gem_execbuffer_relocate(dev, eb, &objects); |
| 1111 | if (ret) { | 1146 | if (ret) { |
| 1112 | if (ret == -EFAULT) { | 1147 | if (ret == -EFAULT) { |
| 1113 | ret = i915_gem_execbuffer_relocate_slow(dev, file, ring, | 1148 | ret = i915_gem_execbuffer_relocate_slow(dev, file, ring, |
| @@ -1121,9 +1156,6 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, | |||
| 1121 | } | 1156 | } |
| 1122 | 1157 | ||
| 1123 | /* Set the pending read domains for the batch buffer to COMMAND */ | 1158 | /* Set the pending read domains for the batch buffer to COMMAND */ |
| 1124 | batch_obj = list_entry(objects.prev, | ||
| 1125 | struct drm_i915_gem_object, | ||
| 1126 | exec_list); | ||
| 1127 | if (batch_obj->base.pending_write_domain) { | 1159 | if (batch_obj->base.pending_write_domain) { |
| 1128 | DRM_ERROR("Attempting to use self-modifying batch buffer\n"); | 1160 | DRM_ERROR("Attempting to use self-modifying batch buffer\n"); |
| 1129 | ret = -EINVAL; | 1161 | ret = -EINVAL; |
| @@ -1340,4 +1372,3 @@ i915_gem_execbuffer2(struct drm_device *dev, void *data, | |||
| 1340 | drm_free_large(exec2_list); | 1372 | drm_free_large(exec2_list); |
| 1341 | return ret; | 1373 | return ret; |
| 1342 | } | 1374 | } |
| 1343 | |||
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c index 86673e77d7cb..70433ae50ac8 100644 --- a/drivers/gpu/drm/i915/i915_gem_gtt.c +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c | |||
| @@ -85,15 +85,11 @@ int i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj) | |||
| 85 | 85 | ||
| 86 | void i915_gem_gtt_unbind_object(struct drm_i915_gem_object *obj) | 86 | void i915_gem_gtt_unbind_object(struct drm_i915_gem_object *obj) |
| 87 | { | 87 | { |
| 88 | struct drm_device *dev = obj->base.dev; | 88 | intel_gtt_clear_range(obj->gtt_space->start >> PAGE_SHIFT, |
| 89 | struct drm_i915_private *dev_priv = dev->dev_private; | 89 | obj->base.size >> PAGE_SHIFT); |
| 90 | 90 | ||
| 91 | if (dev_priv->mm.gtt->needs_dmar) { | 91 | if (obj->sg_list) { |
| 92 | intel_gtt_unmap_memory(obj->sg_list, obj->num_sg); | 92 | intel_gtt_unmap_memory(obj->sg_list, obj->num_sg); |
| 93 | obj->sg_list = NULL; | 93 | obj->sg_list = NULL; |
| 94 | obj->num_sg = 0; | ||
| 95 | } | 94 | } |
| 96 | |||
| 97 | intel_gtt_clear_range(obj->gtt_space->start >> PAGE_SHIFT, | ||
| 98 | obj->base.size >> PAGE_SHIFT); | ||
| 99 | } | 95 | } |
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index 0dadc025b77b..e418e8bb61e6 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
| @@ -64,26 +64,6 @@ | |||
| 64 | #define DRM_I915_VBLANK_PIPE_ALL (DRM_I915_VBLANK_PIPE_A | \ | 64 | #define DRM_I915_VBLANK_PIPE_ALL (DRM_I915_VBLANK_PIPE_A | \ |
| 65 | DRM_I915_VBLANK_PIPE_B) | 65 | DRM_I915_VBLANK_PIPE_B) |
| 66 | 66 | ||
| 67 | void | ||
| 68 | ironlake_enable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 69 | { | ||
| 70 | if ((dev_priv->gt_irq_mask & mask) != 0) { | ||
| 71 | dev_priv->gt_irq_mask &= ~mask; | ||
| 72 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); | ||
| 73 | POSTING_READ(GTIMR); | ||
| 74 | } | ||
| 75 | } | ||
| 76 | |||
| 77 | void | ||
| 78 | ironlake_disable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 79 | { | ||
| 80 | if ((dev_priv->gt_irq_mask & mask) != mask) { | ||
| 81 | dev_priv->gt_irq_mask |= mask; | ||
| 82 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); | ||
| 83 | POSTING_READ(GTIMR); | ||
| 84 | } | ||
| 85 | } | ||
| 86 | |||
| 87 | /* For display hotplug interrupt */ | 67 | /* For display hotplug interrupt */ |
| 88 | static void | 68 | static void |
| 89 | ironlake_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask) | 69 | ironlake_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask) |
| @@ -105,26 +85,6 @@ ironlake_disable_display_irq(drm_i915_private_t *dev_priv, u32 mask) | |||
| 105 | } | 85 | } |
| 106 | } | 86 | } |
| 107 | 87 | ||
| 108 | void | ||
| 109 | i915_enable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 110 | { | ||
| 111 | if ((dev_priv->irq_mask & mask) != 0) { | ||
| 112 | dev_priv->irq_mask &= ~mask; | ||
| 113 | I915_WRITE(IMR, dev_priv->irq_mask); | ||
| 114 | POSTING_READ(IMR); | ||
| 115 | } | ||
| 116 | } | ||
| 117 | |||
| 118 | void | ||
| 119 | i915_disable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 120 | { | ||
| 121 | if ((dev_priv->irq_mask & mask) != mask) { | ||
| 122 | dev_priv->irq_mask |= mask; | ||
| 123 | I915_WRITE(IMR, dev_priv->irq_mask); | ||
| 124 | POSTING_READ(IMR); | ||
| 125 | } | ||
| 126 | } | ||
| 127 | |||
| 128 | static inline u32 | 88 | static inline u32 |
| 129 | i915_pipestat(int pipe) | 89 | i915_pipestat(int pipe) |
| 130 | { | 90 | { |
| @@ -389,9 +349,12 @@ static void notify_ring(struct drm_device *dev, | |||
| 389 | { | 349 | { |
| 390 | struct drm_i915_private *dev_priv = dev->dev_private; | 350 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 391 | u32 seqno = ring->get_seqno(ring); | 351 | u32 seqno = ring->get_seqno(ring); |
| 392 | ring->irq_seqno = seqno; | 352 | |
| 393 | trace_i915_gem_request_complete(dev, seqno); | 353 | trace_i915_gem_request_complete(dev, seqno); |
| 354 | |||
| 355 | ring->irq_seqno = seqno; | ||
| 394 | wake_up_all(&ring->irq_queue); | 356 | wake_up_all(&ring->irq_queue); |
| 357 | |||
| 395 | dev_priv->hangcheck_count = 0; | 358 | dev_priv->hangcheck_count = 0; |
| 396 | mod_timer(&dev_priv->hangcheck_timer, | 359 | mod_timer(&dev_priv->hangcheck_timer, |
| 397 | jiffies + msecs_to_jiffies(DRM_I915_HANGCHECK_PERIOD)); | 360 | jiffies + msecs_to_jiffies(DRM_I915_HANGCHECK_PERIOD)); |
| @@ -435,6 +398,50 @@ static void gen6_pm_irq_handler(struct drm_device *dev) | |||
| 435 | I915_WRITE(GEN6_PMIIR, pm_iir); | 398 | I915_WRITE(GEN6_PMIIR, pm_iir); |
| 436 | } | 399 | } |
| 437 | 400 | ||
| 401 | static void pch_irq_handler(struct drm_device *dev) | ||
| 402 | { | ||
| 403 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | ||
| 404 | u32 pch_iir; | ||
| 405 | |||
| 406 | pch_iir = I915_READ(SDEIIR); | ||
| 407 | |||
| 408 | if (pch_iir & SDE_AUDIO_POWER_MASK) | ||
| 409 | DRM_DEBUG_DRIVER("PCH audio power change on port %d\n", | ||
| 410 | (pch_iir & SDE_AUDIO_POWER_MASK) >> | ||
| 411 | SDE_AUDIO_POWER_SHIFT); | ||
| 412 | |||
| 413 | if (pch_iir & SDE_GMBUS) | ||
| 414 | DRM_DEBUG_DRIVER("PCH GMBUS interrupt\n"); | ||
| 415 | |||
| 416 | if (pch_iir & SDE_AUDIO_HDCP_MASK) | ||
| 417 | DRM_DEBUG_DRIVER("PCH HDCP audio interrupt\n"); | ||
| 418 | |||
| 419 | if (pch_iir & SDE_AUDIO_TRANS_MASK) | ||
| 420 | DRM_DEBUG_DRIVER("PCH transcoder audio interrupt\n"); | ||
| 421 | |||
| 422 | if (pch_iir & SDE_POISON) | ||
| 423 | DRM_ERROR("PCH poison interrupt\n"); | ||
| 424 | |||
| 425 | if (pch_iir & SDE_FDI_MASK) { | ||
| 426 | u32 fdia, fdib; | ||
| 427 | |||
| 428 | fdia = I915_READ(FDI_RXA_IIR); | ||
| 429 | fdib = I915_READ(FDI_RXB_IIR); | ||
| 430 | DRM_DEBUG_DRIVER("PCH FDI RX interrupt; FDI RXA IIR: 0x%08x, FDI RXB IIR: 0x%08x\n", fdia, fdib); | ||
| 431 | } | ||
| 432 | |||
| 433 | if (pch_iir & (SDE_TRANSB_CRC_DONE | SDE_TRANSA_CRC_DONE)) | ||
| 434 | DRM_DEBUG_DRIVER("PCH transcoder CRC done interrupt\n"); | ||
| 435 | |||
| 436 | if (pch_iir & (SDE_TRANSB_CRC_ERR | SDE_TRANSA_CRC_ERR)) | ||
| 437 | DRM_DEBUG_DRIVER("PCH transcoder CRC error interrupt\n"); | ||
| 438 | |||
| 439 | if (pch_iir & SDE_TRANSB_FIFO_UNDER) | ||
| 440 | DRM_DEBUG_DRIVER("PCH transcoder B underrun interrupt\n"); | ||
| 441 | if (pch_iir & SDE_TRANSA_FIFO_UNDER) | ||
| 442 | DRM_DEBUG_DRIVER("PCH transcoder A underrun interrupt\n"); | ||
| 443 | } | ||
| 444 | |||
| 438 | static irqreturn_t ironlake_irq_handler(struct drm_device *dev) | 445 | static irqreturn_t ironlake_irq_handler(struct drm_device *dev) |
| 439 | { | 446 | { |
| 440 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 447 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| @@ -502,8 +509,11 @@ static irqreturn_t ironlake_irq_handler(struct drm_device *dev) | |||
| 502 | drm_handle_vblank(dev, 1); | 509 | drm_handle_vblank(dev, 1); |
| 503 | 510 | ||
| 504 | /* check event from PCH */ | 511 | /* check event from PCH */ |
| 505 | if ((de_iir & DE_PCH_EVENT) && (pch_iir & hotplug_mask)) | 512 | if (de_iir & DE_PCH_EVENT) { |
| 506 | queue_work(dev_priv->wq, &dev_priv->hotplug_work); | 513 | if (pch_iir & hotplug_mask) |
| 514 | queue_work(dev_priv->wq, &dev_priv->hotplug_work); | ||
| 515 | pch_irq_handler(dev); | ||
| 516 | } | ||
| 507 | 517 | ||
| 508 | if (de_iir & DE_PCU_EVENT) { | 518 | if (de_iir & DE_PCU_EVENT) { |
| 509 | I915_WRITE16(MEMINTRSTS, I915_READ(MEMINTRSTS)); | 519 | I915_WRITE16(MEMINTRSTS, I915_READ(MEMINTRSTS)); |
| @@ -556,10 +566,9 @@ static void i915_error_work_func(struct work_struct *work) | |||
| 556 | 566 | ||
| 557 | #ifdef CONFIG_DEBUG_FS | 567 | #ifdef CONFIG_DEBUG_FS |
| 558 | static struct drm_i915_error_object * | 568 | static struct drm_i915_error_object * |
| 559 | i915_error_object_create(struct drm_device *dev, | 569 | i915_error_object_create(struct drm_i915_private *dev_priv, |
| 560 | struct drm_i915_gem_object *src) | 570 | struct drm_i915_gem_object *src) |
| 561 | { | 571 | { |
| 562 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 563 | struct drm_i915_error_object *dst; | 572 | struct drm_i915_error_object *dst; |
| 564 | int page, page_count; | 573 | int page, page_count; |
| 565 | u32 reloc_offset; | 574 | u32 reloc_offset; |
| @@ -632,52 +641,6 @@ i915_error_state_free(struct drm_device *dev, | |||
| 632 | kfree(error); | 641 | kfree(error); |
| 633 | } | 642 | } |
| 634 | 643 | ||
| 635 | static u32 | ||
| 636 | i915_get_bbaddr(struct drm_device *dev, u32 *ring) | ||
| 637 | { | ||
| 638 | u32 cmd; | ||
| 639 | |||
| 640 | if (IS_I830(dev) || IS_845G(dev)) | ||
| 641 | cmd = MI_BATCH_BUFFER; | ||
| 642 | else if (INTEL_INFO(dev)->gen >= 4) | ||
| 643 | cmd = (MI_BATCH_BUFFER_START | (2 << 6) | | ||
| 644 | MI_BATCH_NON_SECURE_I965); | ||
| 645 | else | ||
| 646 | cmd = (MI_BATCH_BUFFER_START | (2 << 6)); | ||
| 647 | |||
| 648 | return ring[0] == cmd ? ring[1] : 0; | ||
| 649 | } | ||
| 650 | |||
| 651 | static u32 | ||
| 652 | i915_ringbuffer_last_batch(struct drm_device *dev, | ||
| 653 | struct intel_ring_buffer *ring) | ||
| 654 | { | ||
| 655 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 656 | u32 head, bbaddr; | ||
| 657 | u32 *val; | ||
| 658 | |||
| 659 | /* Locate the current position in the ringbuffer and walk back | ||
| 660 | * to find the most recently dispatched batch buffer. | ||
| 661 | */ | ||
| 662 | head = I915_READ_HEAD(ring) & HEAD_ADDR; | ||
| 663 | |||
| 664 | val = (u32 *)(ring->virtual_start + head); | ||
| 665 | while (--val >= (u32 *)ring->virtual_start) { | ||
| 666 | bbaddr = i915_get_bbaddr(dev, val); | ||
| 667 | if (bbaddr) | ||
| 668 | return bbaddr; | ||
| 669 | } | ||
| 670 | |||
| 671 | val = (u32 *)(ring->virtual_start + ring->size); | ||
| 672 | while (--val >= (u32 *)ring->virtual_start) { | ||
| 673 | bbaddr = i915_get_bbaddr(dev, val); | ||
| 674 | if (bbaddr) | ||
| 675 | return bbaddr; | ||
| 676 | } | ||
| 677 | |||
| 678 | return 0; | ||
| 679 | } | ||
| 680 | |||
| 681 | static u32 capture_bo_list(struct drm_i915_error_buffer *err, | 644 | static u32 capture_bo_list(struct drm_i915_error_buffer *err, |
| 682 | int count, | 645 | int count, |
| 683 | struct list_head *head) | 646 | struct list_head *head) |
| @@ -702,6 +665,7 @@ static u32 capture_bo_list(struct drm_i915_error_buffer *err, | |||
| 702 | err->dirty = obj->dirty; | 665 | err->dirty = obj->dirty; |
| 703 | err->purgeable = obj->madv != I915_MADV_WILLNEED; | 666 | err->purgeable = obj->madv != I915_MADV_WILLNEED; |
| 704 | err->ring = obj->ring ? obj->ring->id : 0; | 667 | err->ring = obj->ring ? obj->ring->id : 0; |
| 668 | err->agp_type = obj->agp_type == AGP_USER_CACHED_MEMORY; | ||
| 705 | 669 | ||
| 706 | if (++i == count) | 670 | if (++i == count) |
| 707 | break; | 671 | break; |
| @@ -741,6 +705,36 @@ static void i915_gem_record_fences(struct drm_device *dev, | |||
| 741 | } | 705 | } |
| 742 | } | 706 | } |
| 743 | 707 | ||
| 708 | static struct drm_i915_error_object * | ||
| 709 | i915_error_first_batchbuffer(struct drm_i915_private *dev_priv, | ||
| 710 | struct intel_ring_buffer *ring) | ||
| 711 | { | ||
| 712 | struct drm_i915_gem_object *obj; | ||
| 713 | u32 seqno; | ||
| 714 | |||
| 715 | if (!ring->get_seqno) | ||
| 716 | return NULL; | ||
| 717 | |||
| 718 | seqno = ring->get_seqno(ring); | ||
| 719 | list_for_each_entry(obj, &dev_priv->mm.active_list, mm_list) { | ||
| 720 | if (obj->ring != ring) | ||
| 721 | continue; | ||
| 722 | |||
| 723 | if (!i915_seqno_passed(obj->last_rendering_seqno, seqno)) | ||
| 724 | continue; | ||
| 725 | |||
| 726 | if ((obj->base.read_domains & I915_GEM_DOMAIN_COMMAND) == 0) | ||
| 727 | continue; | ||
| 728 | |||
| 729 | /* We need to copy these to an anonymous buffer as the simplest | ||
| 730 | * method to avoid being overwritten by userspace. | ||
| 731 | */ | ||
| 732 | return i915_error_object_create(dev_priv, obj); | ||
| 733 | } | ||
| 734 | |||
| 735 | return NULL; | ||
| 736 | } | ||
| 737 | |||
| 744 | /** | 738 | /** |
| 745 | * i915_capture_error_state - capture an error record for later analysis | 739 | * i915_capture_error_state - capture an error record for later analysis |
| 746 | * @dev: drm device | 740 | * @dev: drm device |
| @@ -755,10 +749,8 @@ static void i915_capture_error_state(struct drm_device *dev) | |||
| 755 | struct drm_i915_private *dev_priv = dev->dev_private; | 749 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 756 | struct drm_i915_gem_object *obj; | 750 | struct drm_i915_gem_object *obj; |
| 757 | struct drm_i915_error_state *error; | 751 | struct drm_i915_error_state *error; |
| 758 | struct drm_i915_gem_object *batchbuffer[2]; | ||
| 759 | unsigned long flags; | 752 | unsigned long flags; |
| 760 | u32 bbaddr; | 753 | int i; |
| 761 | int count; | ||
| 762 | 754 | ||
| 763 | spin_lock_irqsave(&dev_priv->error_lock, flags); | 755 | spin_lock_irqsave(&dev_priv->error_lock, flags); |
| 764 | error = dev_priv->first_error; | 756 | error = dev_priv->first_error; |
| @@ -817,83 +809,30 @@ static void i915_capture_error_state(struct drm_device *dev) | |||
| 817 | } | 809 | } |
| 818 | i915_gem_record_fences(dev, error); | 810 | i915_gem_record_fences(dev, error); |
| 819 | 811 | ||
| 820 | bbaddr = i915_ringbuffer_last_batch(dev, &dev_priv->ring[RCS]); | 812 | /* Record the active batchbuffers */ |
| 821 | 813 | for (i = 0; i < I915_NUM_RINGS; i++) | |
| 822 | /* Grab the current batchbuffer, most likely to have crashed. */ | 814 | error->batchbuffer[i] = |
| 823 | batchbuffer[0] = NULL; | 815 | i915_error_first_batchbuffer(dev_priv, |
| 824 | batchbuffer[1] = NULL; | 816 | &dev_priv->ring[i]); |
| 825 | count = 0; | ||
| 826 | list_for_each_entry(obj, &dev_priv->mm.active_list, mm_list) { | ||
| 827 | if (batchbuffer[0] == NULL && | ||
| 828 | bbaddr >= obj->gtt_offset && | ||
| 829 | bbaddr < obj->gtt_offset + obj->base.size) | ||
| 830 | batchbuffer[0] = obj; | ||
| 831 | |||
| 832 | if (batchbuffer[1] == NULL && | ||
| 833 | error->acthd >= obj->gtt_offset && | ||
| 834 | error->acthd < obj->gtt_offset + obj->base.size) | ||
| 835 | batchbuffer[1] = obj; | ||
| 836 | |||
| 837 | count++; | ||
| 838 | } | ||
| 839 | /* Scan the other lists for completeness for those bizarre errors. */ | ||
| 840 | if (batchbuffer[0] == NULL || batchbuffer[1] == NULL) { | ||
| 841 | list_for_each_entry(obj, &dev_priv->mm.flushing_list, mm_list) { | ||
| 842 | if (batchbuffer[0] == NULL && | ||
| 843 | bbaddr >= obj->gtt_offset && | ||
| 844 | bbaddr < obj->gtt_offset + obj->base.size) | ||
| 845 | batchbuffer[0] = obj; | ||
| 846 | |||
| 847 | if (batchbuffer[1] == NULL && | ||
| 848 | error->acthd >= obj->gtt_offset && | ||
| 849 | error->acthd < obj->gtt_offset + obj->base.size) | ||
| 850 | batchbuffer[1] = obj; | ||
| 851 | |||
| 852 | if (batchbuffer[0] && batchbuffer[1]) | ||
| 853 | break; | ||
| 854 | } | ||
| 855 | } | ||
| 856 | if (batchbuffer[0] == NULL || batchbuffer[1] == NULL) { | ||
| 857 | list_for_each_entry(obj, &dev_priv->mm.inactive_list, mm_list) { | ||
| 858 | if (batchbuffer[0] == NULL && | ||
| 859 | bbaddr >= obj->gtt_offset && | ||
| 860 | bbaddr < obj->gtt_offset + obj->base.size) | ||
| 861 | batchbuffer[0] = obj; | ||
| 862 | |||
| 863 | if (batchbuffer[1] == NULL && | ||
| 864 | error->acthd >= obj->gtt_offset && | ||
| 865 | error->acthd < obj->gtt_offset + obj->base.size) | ||
| 866 | batchbuffer[1] = obj; | ||
| 867 | |||
| 868 | if (batchbuffer[0] && batchbuffer[1]) | ||
| 869 | break; | ||
| 870 | } | ||
| 871 | } | ||
| 872 | |||
| 873 | /* We need to copy these to an anonymous buffer as the simplest | ||
| 874 | * method to avoid being overwritten by userspace. | ||
| 875 | */ | ||
| 876 | error->batchbuffer[0] = i915_error_object_create(dev, batchbuffer[0]); | ||
| 877 | if (batchbuffer[1] != batchbuffer[0]) | ||
| 878 | error->batchbuffer[1] = i915_error_object_create(dev, batchbuffer[1]); | ||
| 879 | else | ||
| 880 | error->batchbuffer[1] = NULL; | ||
| 881 | 817 | ||
| 882 | /* Record the ringbuffer */ | 818 | /* Record the ringbuffer */ |
| 883 | error->ringbuffer = i915_error_object_create(dev, | 819 | error->ringbuffer = i915_error_object_create(dev_priv, |
| 884 | dev_priv->ring[RCS].obj); | 820 | dev_priv->ring[RCS].obj); |
| 885 | 821 | ||
| 886 | /* Record buffers on the active and pinned lists. */ | 822 | /* Record buffers on the active and pinned lists. */ |
| 887 | error->active_bo = NULL; | 823 | error->active_bo = NULL; |
| 888 | error->pinned_bo = NULL; | 824 | error->pinned_bo = NULL; |
| 889 | 825 | ||
| 890 | error->active_bo_count = count; | 826 | i = 0; |
| 827 | list_for_each_entry(obj, &dev_priv->mm.active_list, mm_list) | ||
| 828 | i++; | ||
| 829 | error->active_bo_count = i; | ||
| 891 | list_for_each_entry(obj, &dev_priv->mm.pinned_list, mm_list) | 830 | list_for_each_entry(obj, &dev_priv->mm.pinned_list, mm_list) |
| 892 | count++; | 831 | i++; |
| 893 | error->pinned_bo_count = count - error->active_bo_count; | 832 | error->pinned_bo_count = i - error->active_bo_count; |
| 894 | 833 | ||
| 895 | if (count) { | 834 | if (i) { |
| 896 | error->active_bo = kmalloc(sizeof(*error->active_bo)*count, | 835 | error->active_bo = kmalloc(sizeof(*error->active_bo)*i, |
| 897 | GFP_ATOMIC); | 836 | GFP_ATOMIC); |
| 898 | if (error->active_bo) | 837 | if (error->active_bo) |
| 899 | error->pinned_bo = | 838 | error->pinned_bo = |
| @@ -1673,11 +1612,6 @@ static int ironlake_irq_postinstall(struct drm_device *dev) | |||
| 1673 | 1612 | ||
| 1674 | I915_WRITE(GTIIR, I915_READ(GTIIR)); | 1613 | I915_WRITE(GTIIR, I915_READ(GTIIR)); |
| 1675 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); | 1614 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); |
| 1676 | if (IS_GEN6(dev)) { | ||
| 1677 | I915_WRITE(GEN6_RENDER_IMR, ~GEN6_RENDER_USER_INTERRUPT); | ||
| 1678 | I915_WRITE(GEN6_BSD_IMR, ~GEN6_BSD_USER_INTERRUPT); | ||
| 1679 | I915_WRITE(GEN6_BLITTER_IMR, ~GEN6_BLITTER_USER_INTERRUPT); | ||
| 1680 | } | ||
| 1681 | 1615 | ||
| 1682 | if (IS_GEN6(dev)) | 1616 | if (IS_GEN6(dev)) |
| 1683 | render_irqs = | 1617 | render_irqs = |
| @@ -1698,6 +1632,9 @@ static int ironlake_irq_postinstall(struct drm_device *dev) | |||
| 1698 | } else { | 1632 | } else { |
| 1699 | hotplug_mask = SDE_CRT_HOTPLUG | SDE_PORTB_HOTPLUG | | 1633 | hotplug_mask = SDE_CRT_HOTPLUG | SDE_PORTB_HOTPLUG | |
| 1700 | SDE_PORTC_HOTPLUG | SDE_PORTD_HOTPLUG; | 1634 | SDE_PORTC_HOTPLUG | SDE_PORTD_HOTPLUG; |
| 1635 | hotplug_mask |= SDE_AUX_MASK | SDE_FDI_MASK | SDE_TRANS_MASK; | ||
| 1636 | I915_WRITE(FDI_RXA_IMR, 0); | ||
| 1637 | I915_WRITE(FDI_RXB_IMR, 0); | ||
| 1701 | } | 1638 | } |
| 1702 | 1639 | ||
| 1703 | dev_priv->pch_irq_mask = ~hotplug_mask; | 1640 | dev_priv->pch_irq_mask = ~hotplug_mask; |
diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h index 8f948a6fbc1c..40a407f41f61 100644 --- a/drivers/gpu/drm/i915/i915_reg.h +++ b/drivers/gpu/drm/i915/i915_reg.h | |||
| @@ -145,6 +145,8 @@ | |||
| 145 | #define MI_END_SCENE (1 << 4) /* flush binner and incr scene count */ | 145 | #define MI_END_SCENE (1 << 4) /* flush binner and incr scene count */ |
| 146 | #define MI_INVALIDATE_ISP (1 << 5) /* invalidate indirect state pointers */ | 146 | #define MI_INVALIDATE_ISP (1 << 5) /* invalidate indirect state pointers */ |
| 147 | #define MI_BATCH_BUFFER_END MI_INSTR(0x0a, 0) | 147 | #define MI_BATCH_BUFFER_END MI_INSTR(0x0a, 0) |
| 148 | #define MI_SUSPEND_FLUSH MI_INSTR(0x0b, 0) | ||
| 149 | #define MI_SUSPEND_FLUSH_EN (1<<0) | ||
| 148 | #define MI_REPORT_HEAD MI_INSTR(0x07, 0) | 150 | #define MI_REPORT_HEAD MI_INSTR(0x07, 0) |
| 149 | #define MI_OVERLAY_FLIP MI_INSTR(0x11,0) | 151 | #define MI_OVERLAY_FLIP MI_INSTR(0x11,0) |
| 150 | #define MI_OVERLAY_CONTINUE (0x0<<21) | 152 | #define MI_OVERLAY_CONTINUE (0x0<<21) |
| @@ -159,6 +161,7 @@ | |||
| 159 | #define MI_MM_SPACE_PHYSICAL (0<<8) | 161 | #define MI_MM_SPACE_PHYSICAL (0<<8) |
| 160 | #define MI_SAVE_EXT_STATE_EN (1<<3) | 162 | #define MI_SAVE_EXT_STATE_EN (1<<3) |
| 161 | #define MI_RESTORE_EXT_STATE_EN (1<<2) | 163 | #define MI_RESTORE_EXT_STATE_EN (1<<2) |
| 164 | #define MI_FORCE_RESTORE (1<<1) | ||
| 162 | #define MI_RESTORE_INHIBIT (1<<0) | 165 | #define MI_RESTORE_INHIBIT (1<<0) |
| 163 | #define MI_STORE_DWORD_IMM MI_INSTR(0x20, 1) | 166 | #define MI_STORE_DWORD_IMM MI_INSTR(0x20, 1) |
| 164 | #define MI_MEM_VIRTUAL (1 << 22) /* 965+ only */ | 167 | #define MI_MEM_VIRTUAL (1 << 22) /* 965+ only */ |
| @@ -288,6 +291,7 @@ | |||
| 288 | #define RING_HWS_PGA_GEN6(base) ((base)+0x2080) | 291 | #define RING_HWS_PGA_GEN6(base) ((base)+0x2080) |
| 289 | #define RING_ACTHD(base) ((base)+0x74) | 292 | #define RING_ACTHD(base) ((base)+0x74) |
| 290 | #define RING_NOPID(base) ((base)+0x94) | 293 | #define RING_NOPID(base) ((base)+0x94) |
| 294 | #define RING_IMR(base) ((base)+0xa8) | ||
| 291 | #define TAIL_ADDR 0x001FFFF8 | 295 | #define TAIL_ADDR 0x001FFFF8 |
| 292 | #define HEAD_WRAP_COUNT 0xFFE00000 | 296 | #define HEAD_WRAP_COUNT 0xFFE00000 |
| 293 | #define HEAD_WRAP_ONE 0x00200000 | 297 | #define HEAD_WRAP_ONE 0x00200000 |
| @@ -1130,9 +1134,50 @@ | |||
| 1130 | #define RCBMINAVG 0x111a0 | 1134 | #define RCBMINAVG 0x111a0 |
| 1131 | #define RCUPEI 0x111b0 | 1135 | #define RCUPEI 0x111b0 |
| 1132 | #define RCDNEI 0x111b4 | 1136 | #define RCDNEI 0x111b4 |
| 1133 | #define MCHBAR_RENDER_STANDBY 0x111b8 | 1137 | #define RSTDBYCTL 0x111b8 |
| 1134 | #define RCX_SW_EXIT (1<<23) | 1138 | #define RS1EN (1<<31) |
| 1135 | #define RSX_STATUS_MASK 0x00700000 | 1139 | #define RS2EN (1<<30) |
| 1140 | #define RS3EN (1<<29) | ||
| 1141 | #define D3RS3EN (1<<28) /* Display D3 imlies RS3 */ | ||
| 1142 | #define SWPROMORSX (1<<27) /* RSx promotion timers ignored */ | ||
| 1143 | #define RCWAKERW (1<<26) /* Resetwarn from PCH causes wakeup */ | ||
| 1144 | #define DPRSLPVREN (1<<25) /* Fast voltage ramp enable */ | ||
| 1145 | #define GFXTGHYST (1<<24) /* Hysteresis to allow trunk gating */ | ||
| 1146 | #define RCX_SW_EXIT (1<<23) /* Leave RSx and prevent re-entry */ | ||
| 1147 | #define RSX_STATUS_MASK (7<<20) | ||
| 1148 | #define RSX_STATUS_ON (0<<20) | ||
| 1149 | #define RSX_STATUS_RC1 (1<<20) | ||
| 1150 | #define RSX_STATUS_RC1E (2<<20) | ||
| 1151 | #define RSX_STATUS_RS1 (3<<20) | ||
| 1152 | #define RSX_STATUS_RS2 (4<<20) /* aka rc6 */ | ||
| 1153 | #define RSX_STATUS_RSVD (5<<20) /* deep rc6 unsupported on ilk */ | ||
| 1154 | #define RSX_STATUS_RS3 (6<<20) /* rs3 unsupported on ilk */ | ||
| 1155 | #define RSX_STATUS_RSVD2 (7<<20) | ||
| 1156 | #define UWRCRSXE (1<<19) /* wake counter limit prevents rsx */ | ||
| 1157 | #define RSCRP (1<<18) /* rs requests control on rs1/2 reqs */ | ||
| 1158 | #define JRSC (1<<17) /* rsx coupled to cpu c-state */ | ||
| 1159 | #define RS2INC0 (1<<16) /* allow rs2 in cpu c0 */ | ||
| 1160 | #define RS1CONTSAV_MASK (3<<14) | ||
| 1161 | #define RS1CONTSAV_NO_RS1 (0<<14) /* rs1 doesn't save/restore context */ | ||
| 1162 | #define RS1CONTSAV_RSVD (1<<14) | ||
| 1163 | #define RS1CONTSAV_SAVE_RS1 (2<<14) /* rs1 saves context */ | ||
| 1164 | #define RS1CONTSAV_FULL_RS1 (3<<14) /* rs1 saves and restores context */ | ||
| 1165 | #define NORMSLEXLAT_MASK (3<<12) | ||
| 1166 | #define SLOW_RS123 (0<<12) | ||
| 1167 | #define SLOW_RS23 (1<<12) | ||
| 1168 | #define SLOW_RS3 (2<<12) | ||
| 1169 | #define NORMAL_RS123 (3<<12) | ||
| 1170 | #define RCMODE_TIMEOUT (1<<11) /* 0 is eval interval method */ | ||
| 1171 | #define IMPROMOEN (1<<10) /* promo is immediate or delayed until next idle interval (only for timeout method above) */ | ||
| 1172 | #define RCENTSYNC (1<<9) /* rs coupled to cpu c-state (3/6/7) */ | ||
| 1173 | #define STATELOCK (1<<7) /* locked to rs_cstate if 0 */ | ||
| 1174 | #define RS_CSTATE_MASK (3<<4) | ||
| 1175 | #define RS_CSTATE_C367_RS1 (0<<4) | ||
| 1176 | #define RS_CSTATE_C36_RS1_C7_RS2 (1<<4) | ||
| 1177 | #define RS_CSTATE_RSVD (2<<4) | ||
| 1178 | #define RS_CSTATE_C367_RS2 (3<<4) | ||
| 1179 | #define REDSAVES (1<<3) /* no context save if was idle during rs0 */ | ||
| 1180 | #define REDRESTORES (1<<2) /* no restore if was idle during rs0 */ | ||
| 1136 | #define VIDCTL 0x111c0 | 1181 | #define VIDCTL 0x111c0 |
| 1137 | #define VIDSTS 0x111c8 | 1182 | #define VIDSTS 0x111c8 |
| 1138 | #define VIDSTART 0x111cc /* 8 bits */ | 1183 | #define VIDSTART 0x111cc /* 8 bits */ |
| @@ -2345,8 +2390,13 @@ | |||
| 2345 | 2390 | ||
| 2346 | /* Memory latency timer register */ | 2391 | /* Memory latency timer register */ |
| 2347 | #define MLTR_ILK 0x11222 | 2392 | #define MLTR_ILK 0x11222 |
| 2393 | #define MLTR_WM1_SHIFT 0 | ||
| 2394 | #define MLTR_WM2_SHIFT 8 | ||
| 2348 | /* the unit of memory self-refresh latency time is 0.5us */ | 2395 | /* the unit of memory self-refresh latency time is 0.5us */ |
| 2349 | #define ILK_SRLT_MASK 0x3f | 2396 | #define ILK_SRLT_MASK 0x3f |
| 2397 | #define ILK_LATENCY(shift) (I915_READ(MLTR_ILK) >> (shift) & ILK_SRLT_MASK) | ||
| 2398 | #define ILK_READ_WM1_LATENCY() ILK_LATENCY(MLTR_WM1_SHIFT) | ||
| 2399 | #define ILK_READ_WM2_LATENCY() ILK_LATENCY(MLTR_WM2_SHIFT) | ||
| 2350 | 2400 | ||
| 2351 | /* define the fifo size on Ironlake */ | 2401 | /* define the fifo size on Ironlake */ |
| 2352 | #define ILK_DISPLAY_FIFO 128 | 2402 | #define ILK_DISPLAY_FIFO 128 |
| @@ -2728,12 +2778,41 @@ | |||
| 2728 | /* PCH */ | 2778 | /* PCH */ |
| 2729 | 2779 | ||
| 2730 | /* south display engine interrupt */ | 2780 | /* south display engine interrupt */ |
| 2781 | #define SDE_AUDIO_POWER_D (1 << 27) | ||
| 2782 | #define SDE_AUDIO_POWER_C (1 << 26) | ||
| 2783 | #define SDE_AUDIO_POWER_B (1 << 25) | ||
| 2784 | #define SDE_AUDIO_POWER_SHIFT (25) | ||
| 2785 | #define SDE_AUDIO_POWER_MASK (7 << SDE_AUDIO_POWER_SHIFT) | ||
| 2786 | #define SDE_GMBUS (1 << 24) | ||
| 2787 | #define SDE_AUDIO_HDCP_TRANSB (1 << 23) | ||
| 2788 | #define SDE_AUDIO_HDCP_TRANSA (1 << 22) | ||
| 2789 | #define SDE_AUDIO_HDCP_MASK (3 << 22) | ||
| 2790 | #define SDE_AUDIO_TRANSB (1 << 21) | ||
| 2791 | #define SDE_AUDIO_TRANSA (1 << 20) | ||
| 2792 | #define SDE_AUDIO_TRANS_MASK (3 << 20) | ||
| 2793 | #define SDE_POISON (1 << 19) | ||
| 2794 | /* 18 reserved */ | ||
| 2795 | #define SDE_FDI_RXB (1 << 17) | ||
| 2796 | #define SDE_FDI_RXA (1 << 16) | ||
| 2797 | #define SDE_FDI_MASK (3 << 16) | ||
| 2798 | #define SDE_AUXD (1 << 15) | ||
| 2799 | #define SDE_AUXC (1 << 14) | ||
| 2800 | #define SDE_AUXB (1 << 13) | ||
| 2801 | #define SDE_AUX_MASK (7 << 13) | ||
| 2802 | /* 12 reserved */ | ||
| 2731 | #define SDE_CRT_HOTPLUG (1 << 11) | 2803 | #define SDE_CRT_HOTPLUG (1 << 11) |
| 2732 | #define SDE_PORTD_HOTPLUG (1 << 10) | 2804 | #define SDE_PORTD_HOTPLUG (1 << 10) |
| 2733 | #define SDE_PORTC_HOTPLUG (1 << 9) | 2805 | #define SDE_PORTC_HOTPLUG (1 << 9) |
| 2734 | #define SDE_PORTB_HOTPLUG (1 << 8) | 2806 | #define SDE_PORTB_HOTPLUG (1 << 8) |
| 2735 | #define SDE_SDVOB_HOTPLUG (1 << 6) | 2807 | #define SDE_SDVOB_HOTPLUG (1 << 6) |
| 2736 | #define SDE_HOTPLUG_MASK (0xf << 8) | 2808 | #define SDE_HOTPLUG_MASK (0xf << 8) |
| 2809 | #define SDE_TRANSB_CRC_DONE (1 << 5) | ||
| 2810 | #define SDE_TRANSB_CRC_ERR (1 << 4) | ||
| 2811 | #define SDE_TRANSB_FIFO_UNDER (1 << 3) | ||
| 2812 | #define SDE_TRANSA_CRC_DONE (1 << 2) | ||
| 2813 | #define SDE_TRANSA_CRC_ERR (1 << 1) | ||
| 2814 | #define SDE_TRANSA_FIFO_UNDER (1 << 0) | ||
| 2815 | #define SDE_TRANS_MASK (0x3f) | ||
| 2737 | /* CPT */ | 2816 | /* CPT */ |
| 2738 | #define SDE_CRT_HOTPLUG_CPT (1 << 19) | 2817 | #define SDE_CRT_HOTPLUG_CPT (1 << 19) |
| 2739 | #define SDE_PORTD_HOTPLUG_CPT (1 << 23) | 2818 | #define SDE_PORTD_HOTPLUG_CPT (1 << 23) |
| @@ -3174,10 +3253,11 @@ | |||
| 3174 | #define EDP_LINK_TRAIN_600MV_3_5DB_SNB_A (0x01<<22) | 3253 | #define EDP_LINK_TRAIN_600MV_3_5DB_SNB_A (0x01<<22) |
| 3175 | #define EDP_LINK_TRAIN_800MV_0DB_SNB_A (0x0<<22) | 3254 | #define EDP_LINK_TRAIN_800MV_0DB_SNB_A (0x0<<22) |
| 3176 | /* SNB B-stepping */ | 3255 | /* SNB B-stepping */ |
| 3177 | #define EDP_LINK_TRAIN_400MV_0DB_SNB_B (0x0<<22) | 3256 | #define EDP_LINK_TRAIN_400_600MV_0DB_SNB_B (0x0<<22) |
| 3178 | #define EDP_LINK_TRAIN_400MV_6DB_SNB_B (0x3a<<22) | 3257 | #define EDP_LINK_TRAIN_400MV_3_5DB_SNB_B (0x1<<22) |
| 3179 | #define EDP_LINK_TRAIN_600MV_3_5DB_SNB_B (0x39<<22) | 3258 | #define EDP_LINK_TRAIN_400_600MV_6DB_SNB_B (0x3a<<22) |
| 3180 | #define EDP_LINK_TRAIN_800MV_0DB_SNB_B (0x38<<22) | 3259 | #define EDP_LINK_TRAIN_600_800MV_3_5DB_SNB_B (0x39<<22) |
| 3260 | #define EDP_LINK_TRAIN_800_1200MV_0DB_SNB_B (0x38<<22) | ||
| 3181 | #define EDP_LINK_TRAIN_VOL_EMP_MASK_SNB (0x3f<<22) | 3261 | #define EDP_LINK_TRAIN_VOL_EMP_MASK_SNB (0x3f<<22) |
| 3182 | 3262 | ||
| 3183 | #define FORCEWAKE 0xA18C | 3263 | #define FORCEWAKE 0xA18C |
| @@ -3239,6 +3319,7 @@ | |||
| 3239 | 3319 | ||
| 3240 | #define GEN6_PCODE_MAILBOX 0x138124 | 3320 | #define GEN6_PCODE_MAILBOX 0x138124 |
| 3241 | #define GEN6_PCODE_READY (1<<31) | 3321 | #define GEN6_PCODE_READY (1<<31) |
| 3322 | #define GEN6_READ_OC_PARAMS 0xc | ||
| 3242 | #define GEN6_PCODE_WRITE_MIN_FREQ_TABLE 0x9 | 3323 | #define GEN6_PCODE_WRITE_MIN_FREQ_TABLE 0x9 |
| 3243 | #define GEN6_PCODE_DATA 0x138128 | 3324 | #define GEN6_PCODE_DATA 0x138128 |
| 3244 | 3325 | ||
diff --git a/drivers/gpu/drm/i915/i915_suspend.c b/drivers/gpu/drm/i915/i915_suspend.c index 410772466fa7..0521ecf26017 100644 --- a/drivers/gpu/drm/i915/i915_suspend.c +++ b/drivers/gpu/drm/i915/i915_suspend.c | |||
| @@ -740,7 +740,7 @@ void i915_restore_display(struct drm_device *dev) | |||
| 740 | I915_WRITE(PCH_PP_OFF_DELAYS, dev_priv->savePP_OFF_DELAYS); | 740 | I915_WRITE(PCH_PP_OFF_DELAYS, dev_priv->savePP_OFF_DELAYS); |
| 741 | I915_WRITE(PCH_PP_DIVISOR, dev_priv->savePP_DIVISOR); | 741 | I915_WRITE(PCH_PP_DIVISOR, dev_priv->savePP_DIVISOR); |
| 742 | I915_WRITE(PCH_PP_CONTROL, dev_priv->savePP_CONTROL); | 742 | I915_WRITE(PCH_PP_CONTROL, dev_priv->savePP_CONTROL); |
| 743 | I915_WRITE(MCHBAR_RENDER_STANDBY, | 743 | I915_WRITE(RSTDBYCTL, |
| 744 | dev_priv->saveMCHBAR_RENDER_STANDBY); | 744 | dev_priv->saveMCHBAR_RENDER_STANDBY); |
| 745 | } else { | 745 | } else { |
| 746 | I915_WRITE(PFIT_PGM_RATIOS, dev_priv->savePFIT_PGM_RATIOS); | 746 | I915_WRITE(PFIT_PGM_RATIOS, dev_priv->savePFIT_PGM_RATIOS); |
| @@ -811,7 +811,7 @@ int i915_save_state(struct drm_device *dev) | |||
| 811 | dev_priv->saveFDI_RXA_IMR = I915_READ(FDI_RXA_IMR); | 811 | dev_priv->saveFDI_RXA_IMR = I915_READ(FDI_RXA_IMR); |
| 812 | dev_priv->saveFDI_RXB_IMR = I915_READ(FDI_RXB_IMR); | 812 | dev_priv->saveFDI_RXB_IMR = I915_READ(FDI_RXB_IMR); |
| 813 | dev_priv->saveMCHBAR_RENDER_STANDBY = | 813 | dev_priv->saveMCHBAR_RENDER_STANDBY = |
| 814 | I915_READ(MCHBAR_RENDER_STANDBY); | 814 | I915_READ(RSTDBYCTL); |
| 815 | } else { | 815 | } else { |
| 816 | dev_priv->saveIER = I915_READ(IER); | 816 | dev_priv->saveIER = I915_READ(IER); |
| 817 | dev_priv->saveIMR = I915_READ(IMR); | 817 | dev_priv->saveIMR = I915_READ(IMR); |
| @@ -822,10 +822,6 @@ int i915_save_state(struct drm_device *dev) | |||
| 822 | if (IS_GEN6(dev)) | 822 | if (IS_GEN6(dev)) |
| 823 | gen6_disable_rps(dev); | 823 | gen6_disable_rps(dev); |
| 824 | 824 | ||
| 825 | /* XXX disabling the clock gating breaks suspend on gm45 | ||
| 826 | intel_disable_clock_gating(dev); | ||
| 827 | */ | ||
| 828 | |||
| 829 | /* Cache mode state */ | 825 | /* Cache mode state */ |
| 830 | dev_priv->saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); | 826 | dev_priv->saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); |
| 831 | 827 | ||
diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c index 8df574316063..17035b87ee46 100644 --- a/drivers/gpu/drm/i915/intel_crt.c +++ b/drivers/gpu/drm/i915/intel_crt.c | |||
| @@ -30,6 +30,7 @@ | |||
| 30 | #include "drm.h" | 30 | #include "drm.h" |
| 31 | #include "drm_crtc.h" | 31 | #include "drm_crtc.h" |
| 32 | #include "drm_crtc_helper.h" | 32 | #include "drm_crtc_helper.h" |
| 33 | #include "drm_edid.h" | ||
| 33 | #include "intel_drv.h" | 34 | #include "intel_drv.h" |
| 34 | #include "i915_drm.h" | 35 | #include "i915_drm.h" |
| 35 | #include "i915_drv.h" | 36 | #include "i915_drv.h" |
| @@ -287,8 +288,9 @@ static bool intel_crt_ddc_probe(struct drm_i915_private *dev_priv, int ddc_bus) | |||
| 287 | return i2c_transfer(&dev_priv->gmbus[ddc_bus].adapter, msgs, 1) == 1; | 288 | return i2c_transfer(&dev_priv->gmbus[ddc_bus].adapter, msgs, 1) == 1; |
| 288 | } | 289 | } |
| 289 | 290 | ||
| 290 | static bool intel_crt_detect_ddc(struct intel_crt *crt) | 291 | static bool intel_crt_detect_ddc(struct drm_connector *connector) |
| 291 | { | 292 | { |
| 293 | struct intel_crt *crt = intel_attached_crt(connector); | ||
| 292 | struct drm_i915_private *dev_priv = crt->base.base.dev->dev_private; | 294 | struct drm_i915_private *dev_priv = crt->base.base.dev->dev_private; |
| 293 | 295 | ||
| 294 | /* CRT should always be at 0, but check anyway */ | 296 | /* CRT should always be at 0, but check anyway */ |
| @@ -301,8 +303,26 @@ static bool intel_crt_detect_ddc(struct intel_crt *crt) | |||
| 301 | } | 303 | } |
| 302 | 304 | ||
| 303 | if (intel_ddc_probe(&crt->base, dev_priv->crt_ddc_pin)) { | 305 | if (intel_ddc_probe(&crt->base, dev_priv->crt_ddc_pin)) { |
| 304 | DRM_DEBUG_KMS("CRT detected via DDC:0x50 [EDID]\n"); | 306 | struct edid *edid; |
| 305 | return true; | 307 | bool is_digital = false; |
| 308 | |||
| 309 | edid = drm_get_edid(connector, | ||
| 310 | &dev_priv->gmbus[dev_priv->crt_ddc_pin].adapter); | ||
| 311 | /* | ||
| 312 | * This may be a DVI-I connector with a shared DDC | ||
| 313 | * link between analog and digital outputs, so we | ||
| 314 | * have to check the EDID input spec of the attached device. | ||
| 315 | */ | ||
| 316 | if (edid != NULL) { | ||
| 317 | is_digital = edid->input & DRM_EDID_INPUT_DIGITAL; | ||
| 318 | connector->display_info.raw_edid = NULL; | ||
| 319 | kfree(edid); | ||
| 320 | } | ||
| 321 | |||
| 322 | if (!is_digital) { | ||
| 323 | DRM_DEBUG_KMS("CRT detected via DDC:0x50 [EDID]\n"); | ||
| 324 | return true; | ||
| 325 | } | ||
| 306 | } | 326 | } |
| 307 | 327 | ||
| 308 | return false; | 328 | return false; |
| @@ -458,7 +478,7 @@ intel_crt_detect(struct drm_connector *connector, bool force) | |||
| 458 | } | 478 | } |
| 459 | } | 479 | } |
| 460 | 480 | ||
| 461 | if (intel_crt_detect_ddc(crt)) | 481 | if (intel_crt_detect_ddc(connector)) |
| 462 | return connector_status_connected; | 482 | return connector_status_connected; |
| 463 | 483 | ||
| 464 | if (!force) | 484 | if (!force) |
| @@ -472,7 +492,7 @@ intel_crt_detect(struct drm_connector *connector, bool force) | |||
| 472 | crtc = intel_get_load_detect_pipe(&crt->base, connector, | 492 | crtc = intel_get_load_detect_pipe(&crt->base, connector, |
| 473 | NULL, &dpms_mode); | 493 | NULL, &dpms_mode); |
| 474 | if (crtc) { | 494 | if (crtc) { |
| 475 | if (intel_crt_detect_ddc(crt)) | 495 | if (intel_crt_detect_ddc(connector)) |
| 476 | status = connector_status_connected; | 496 | status = connector_status_connected; |
| 477 | else | 497 | else |
| 478 | status = intel_crt_load_detect(crtc, crt); | 498 | status = intel_crt_load_detect(crtc, crt); |
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index 0abe79fb6385..25d96889d7d2 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
| @@ -3418,15 +3418,16 @@ static void i830_update_wm(struct drm_device *dev, int planea_clock, int unused, | |||
| 3418 | static bool ironlake_compute_wm0(struct drm_device *dev, | 3418 | static bool ironlake_compute_wm0(struct drm_device *dev, |
| 3419 | int pipe, | 3419 | int pipe, |
| 3420 | const struct intel_watermark_params *display, | 3420 | const struct intel_watermark_params *display, |
| 3421 | int display_latency, | 3421 | int display_latency_ns, |
| 3422 | const struct intel_watermark_params *cursor, | 3422 | const struct intel_watermark_params *cursor, |
| 3423 | int cursor_latency, | 3423 | int cursor_latency_ns, |
| 3424 | int *plane_wm, | 3424 | int *plane_wm, |
| 3425 | int *cursor_wm) | 3425 | int *cursor_wm) |
| 3426 | { | 3426 | { |
| 3427 | struct drm_crtc *crtc; | 3427 | struct drm_crtc *crtc; |
| 3428 | int htotal, hdisplay, clock, pixel_size = 0; | 3428 | int htotal, hdisplay, clock, pixel_size; |
| 3429 | int line_time_us, line_count, entries; | 3429 | int line_time_us, line_count; |
| 3430 | int entries, tlb_miss; | ||
| 3430 | 3431 | ||
| 3431 | crtc = intel_get_crtc_for_pipe(dev, pipe); | 3432 | crtc = intel_get_crtc_for_pipe(dev, pipe); |
| 3432 | if (crtc->fb == NULL || !crtc->enabled) | 3433 | if (crtc->fb == NULL || !crtc->enabled) |
| @@ -3438,7 +3439,10 @@ static bool ironlake_compute_wm0(struct drm_device *dev, | |||
| 3438 | pixel_size = crtc->fb->bits_per_pixel / 8; | 3439 | pixel_size = crtc->fb->bits_per_pixel / 8; |
| 3439 | 3440 | ||
| 3440 | /* Use the small buffer method to calculate plane watermark */ | 3441 | /* Use the small buffer method to calculate plane watermark */ |
| 3441 | entries = ((clock * pixel_size / 1000) * display_latency * 100) / 1000; | 3442 | entries = ((clock * pixel_size / 1000) * display_latency_ns) / 1000; |
| 3443 | tlb_miss = display->fifo_size*display->cacheline_size - hdisplay * 8; | ||
| 3444 | if (tlb_miss > 0) | ||
| 3445 | entries += tlb_miss; | ||
| 3442 | entries = DIV_ROUND_UP(entries, display->cacheline_size); | 3446 | entries = DIV_ROUND_UP(entries, display->cacheline_size); |
| 3443 | *plane_wm = entries + display->guard_size; | 3447 | *plane_wm = entries + display->guard_size; |
| 3444 | if (*plane_wm > (int)display->max_wm) | 3448 | if (*plane_wm > (int)display->max_wm) |
| @@ -3446,8 +3450,11 @@ static bool ironlake_compute_wm0(struct drm_device *dev, | |||
| 3446 | 3450 | ||
| 3447 | /* Use the large buffer method to calculate cursor watermark */ | 3451 | /* Use the large buffer method to calculate cursor watermark */ |
| 3448 | line_time_us = ((htotal * 1000) / clock); | 3452 | line_time_us = ((htotal * 1000) / clock); |
| 3449 | line_count = (cursor_latency * 100 / line_time_us + 1000) / 1000; | 3453 | line_count = (cursor_latency_ns / line_time_us + 1000) / 1000; |
| 3450 | entries = line_count * 64 * pixel_size; | 3454 | entries = line_count * 64 * pixel_size; |
| 3455 | tlb_miss = cursor->fifo_size*cursor->cacheline_size - hdisplay * 8; | ||
| 3456 | if (tlb_miss > 0) | ||
| 3457 | entries += tlb_miss; | ||
| 3451 | entries = DIV_ROUND_UP(entries, cursor->cacheline_size); | 3458 | entries = DIV_ROUND_UP(entries, cursor->cacheline_size); |
| 3452 | *cursor_wm = entries + cursor->guard_size; | 3459 | *cursor_wm = entries + cursor->guard_size; |
| 3453 | if (*cursor_wm > (int)cursor->max_wm) | 3460 | if (*cursor_wm > (int)cursor->max_wm) |
| @@ -3456,113 +3463,17 @@ static bool ironlake_compute_wm0(struct drm_device *dev, | |||
| 3456 | return true; | 3463 | return true; |
| 3457 | } | 3464 | } |
| 3458 | 3465 | ||
| 3459 | static void ironlake_update_wm(struct drm_device *dev, | ||
| 3460 | int planea_clock, int planeb_clock, | ||
| 3461 | int sr_hdisplay, int sr_htotal, | ||
| 3462 | int pixel_size) | ||
| 3463 | { | ||
| 3464 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 3465 | int plane_wm, cursor_wm, enabled; | ||
| 3466 | int tmp; | ||
| 3467 | |||
| 3468 | enabled = 0; | ||
| 3469 | if (ironlake_compute_wm0(dev, 0, | ||
| 3470 | &ironlake_display_wm_info, | ||
| 3471 | ILK_LP0_PLANE_LATENCY, | ||
| 3472 | &ironlake_cursor_wm_info, | ||
| 3473 | ILK_LP0_CURSOR_LATENCY, | ||
| 3474 | &plane_wm, &cursor_wm)) { | ||
| 3475 | I915_WRITE(WM0_PIPEA_ILK, | ||
| 3476 | (plane_wm << WM0_PIPE_PLANE_SHIFT) | cursor_wm); | ||
| 3477 | DRM_DEBUG_KMS("FIFO watermarks For pipe A -" | ||
| 3478 | " plane %d, " "cursor: %d\n", | ||
| 3479 | plane_wm, cursor_wm); | ||
| 3480 | enabled++; | ||
| 3481 | } | ||
| 3482 | |||
| 3483 | if (ironlake_compute_wm0(dev, 1, | ||
| 3484 | &ironlake_display_wm_info, | ||
| 3485 | ILK_LP0_PLANE_LATENCY, | ||
| 3486 | &ironlake_cursor_wm_info, | ||
| 3487 | ILK_LP0_CURSOR_LATENCY, | ||
| 3488 | &plane_wm, &cursor_wm)) { | ||
| 3489 | I915_WRITE(WM0_PIPEB_ILK, | ||
| 3490 | (plane_wm << WM0_PIPE_PLANE_SHIFT) | cursor_wm); | ||
| 3491 | DRM_DEBUG_KMS("FIFO watermarks For pipe B -" | ||
| 3492 | " plane %d, cursor: %d\n", | ||
| 3493 | plane_wm, cursor_wm); | ||
| 3494 | enabled++; | ||
| 3495 | } | ||
| 3496 | |||
| 3497 | /* | ||
| 3498 | * Calculate and update the self-refresh watermark only when one | ||
| 3499 | * display plane is used. | ||
| 3500 | */ | ||
| 3501 | tmp = 0; | ||
| 3502 | if (enabled == 1) { | ||
| 3503 | unsigned long line_time_us; | ||
| 3504 | int small, large, plane_fbc; | ||
| 3505 | int sr_clock, entries; | ||
| 3506 | int line_count, line_size; | ||
| 3507 | /* Read the self-refresh latency. The unit is 0.5us */ | ||
| 3508 | int ilk_sr_latency = I915_READ(MLTR_ILK) & ILK_SRLT_MASK; | ||
| 3509 | |||
| 3510 | sr_clock = planea_clock ? planea_clock : planeb_clock; | ||
| 3511 | line_time_us = (sr_htotal * 1000) / sr_clock; | ||
| 3512 | |||
| 3513 | /* Use ns/us then divide to preserve precision */ | ||
| 3514 | line_count = ((ilk_sr_latency * 500) / line_time_us + 1000) | ||
| 3515 | / 1000; | ||
| 3516 | line_size = sr_hdisplay * pixel_size; | ||
| 3517 | |||
| 3518 | /* Use the minimum of the small and large buffer method for primary */ | ||
| 3519 | small = ((sr_clock * pixel_size / 1000) * (ilk_sr_latency * 500)) / 1000; | ||
| 3520 | large = line_count * line_size; | ||
| 3521 | |||
| 3522 | entries = DIV_ROUND_UP(min(small, large), | ||
| 3523 | ironlake_display_srwm_info.cacheline_size); | ||
| 3524 | |||
| 3525 | plane_fbc = entries * 64; | ||
| 3526 | plane_fbc = DIV_ROUND_UP(plane_fbc, line_size); | ||
| 3527 | |||
| 3528 | plane_wm = entries + ironlake_display_srwm_info.guard_size; | ||
| 3529 | if (plane_wm > (int)ironlake_display_srwm_info.max_wm) | ||
| 3530 | plane_wm = ironlake_display_srwm_info.max_wm; | ||
| 3531 | |||
| 3532 | /* calculate the self-refresh watermark for display cursor */ | ||
| 3533 | entries = line_count * pixel_size * 64; | ||
| 3534 | entries = DIV_ROUND_UP(entries, | ||
| 3535 | ironlake_cursor_srwm_info.cacheline_size); | ||
| 3536 | |||
| 3537 | cursor_wm = entries + ironlake_cursor_srwm_info.guard_size; | ||
| 3538 | if (cursor_wm > (int)ironlake_cursor_srwm_info.max_wm) | ||
| 3539 | cursor_wm = ironlake_cursor_srwm_info.max_wm; | ||
| 3540 | |||
| 3541 | /* configure watermark and enable self-refresh */ | ||
| 3542 | tmp = (WM1_LP_SR_EN | | ||
| 3543 | (ilk_sr_latency << WM1_LP_LATENCY_SHIFT) | | ||
| 3544 | (plane_fbc << WM1_LP_FBC_SHIFT) | | ||
| 3545 | (plane_wm << WM1_LP_SR_SHIFT) | | ||
| 3546 | cursor_wm); | ||
| 3547 | DRM_DEBUG_KMS("self-refresh watermark: display plane %d, fbc lines %d," | ||
| 3548 | " cursor %d\n", plane_wm, plane_fbc, cursor_wm); | ||
| 3549 | } | ||
| 3550 | I915_WRITE(WM1_LP_ILK, tmp); | ||
| 3551 | /* XXX setup WM2 and WM3 */ | ||
| 3552 | } | ||
| 3553 | |||
| 3554 | /* | 3466 | /* |
| 3555 | * Check the wm result. | 3467 | * Check the wm result. |
| 3556 | * | 3468 | * |
| 3557 | * If any calculated watermark values is larger than the maximum value that | 3469 | * If any calculated watermark values is larger than the maximum value that |
| 3558 | * can be programmed into the associated watermark register, that watermark | 3470 | * can be programmed into the associated watermark register, that watermark |
| 3559 | * must be disabled. | 3471 | * must be disabled. |
| 3560 | * | ||
| 3561 | * Also return true if all of those watermark values is 0, which is set by | ||
| 3562 | * sandybridge_compute_srwm, to indicate the latency is ZERO. | ||
| 3563 | */ | 3472 | */ |
| 3564 | static bool sandybridge_check_srwm(struct drm_device *dev, int level, | 3473 | static bool ironlake_check_srwm(struct drm_device *dev, int level, |
| 3565 | int fbc_wm, int display_wm, int cursor_wm) | 3474 | int fbc_wm, int display_wm, int cursor_wm, |
| 3475 | const struct intel_watermark_params *display, | ||
| 3476 | const struct intel_watermark_params *cursor) | ||
| 3566 | { | 3477 | { |
| 3567 | struct drm_i915_private *dev_priv = dev->dev_private; | 3478 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 3568 | 3479 | ||
| @@ -3571,7 +3482,7 @@ static bool sandybridge_check_srwm(struct drm_device *dev, int level, | |||
| 3571 | 3482 | ||
| 3572 | if (fbc_wm > SNB_FBC_MAX_SRWM) { | 3483 | if (fbc_wm > SNB_FBC_MAX_SRWM) { |
| 3573 | DRM_DEBUG_KMS("fbc watermark(%d) is too large(%d), disabling wm%d+\n", | 3484 | DRM_DEBUG_KMS("fbc watermark(%d) is too large(%d), disabling wm%d+\n", |
| 3574 | fbc_wm, SNB_FBC_MAX_SRWM, level); | 3485 | fbc_wm, SNB_FBC_MAX_SRWM, level); |
| 3575 | 3486 | ||
| 3576 | /* fbc has it's own way to disable FBC WM */ | 3487 | /* fbc has it's own way to disable FBC WM */ |
| 3577 | I915_WRITE(DISP_ARB_CTL, | 3488 | I915_WRITE(DISP_ARB_CTL, |
| @@ -3579,15 +3490,15 @@ static bool sandybridge_check_srwm(struct drm_device *dev, int level, | |||
| 3579 | return false; | 3490 | return false; |
| 3580 | } | 3491 | } |
| 3581 | 3492 | ||
| 3582 | if (display_wm > SNB_DISPLAY_MAX_SRWM) { | 3493 | if (display_wm > display->max_wm) { |
| 3583 | DRM_DEBUG_KMS("display watermark(%d) is too large(%d), disabling wm%d+\n", | 3494 | DRM_DEBUG_KMS("display watermark(%d) is too large(%d), disabling wm%d+\n", |
| 3584 | display_wm, SNB_DISPLAY_MAX_SRWM, level); | 3495 | display_wm, SNB_DISPLAY_MAX_SRWM, level); |
| 3585 | return false; | 3496 | return false; |
| 3586 | } | 3497 | } |
| 3587 | 3498 | ||
| 3588 | if (cursor_wm > SNB_CURSOR_MAX_SRWM) { | 3499 | if (cursor_wm > cursor->max_wm) { |
| 3589 | DRM_DEBUG_KMS("cursor watermark(%d) is too large(%d), disabling wm%d+\n", | 3500 | DRM_DEBUG_KMS("cursor watermark(%d) is too large(%d), disabling wm%d+\n", |
| 3590 | cursor_wm, SNB_CURSOR_MAX_SRWM, level); | 3501 | cursor_wm, SNB_CURSOR_MAX_SRWM, level); |
| 3591 | return false; | 3502 | return false; |
| 3592 | } | 3503 | } |
| 3593 | 3504 | ||
| @@ -3602,16 +3513,18 @@ static bool sandybridge_check_srwm(struct drm_device *dev, int level, | |||
| 3602 | /* | 3513 | /* |
| 3603 | * Compute watermark values of WM[1-3], | 3514 | * Compute watermark values of WM[1-3], |
| 3604 | */ | 3515 | */ |
| 3605 | static bool sandybridge_compute_srwm(struct drm_device *dev, int level, | 3516 | static bool ironlake_compute_srwm(struct drm_device *dev, int level, |
| 3606 | int hdisplay, int htotal, int pixel_size, | 3517 | int hdisplay, int htotal, |
| 3607 | int clock, int latency_ns, int *fbc_wm, | 3518 | int pixel_size, int clock, int latency_ns, |
| 3608 | int *display_wm, int *cursor_wm) | 3519 | const struct intel_watermark_params *display, |
| 3520 | const struct intel_watermark_params *cursor, | ||
| 3521 | int *fbc_wm, int *display_wm, int *cursor_wm) | ||
| 3609 | { | 3522 | { |
| 3610 | 3523 | ||
| 3611 | unsigned long line_time_us; | 3524 | unsigned long line_time_us; |
| 3525 | int line_count, line_size; | ||
| 3612 | int small, large; | 3526 | int small, large; |
| 3613 | int entries; | 3527 | int entries; |
| 3614 | int line_count, line_size; | ||
| 3615 | 3528 | ||
| 3616 | if (!latency_ns) { | 3529 | if (!latency_ns) { |
| 3617 | *fbc_wm = *display_wm = *cursor_wm = 0; | 3530 | *fbc_wm = *display_wm = *cursor_wm = 0; |
| @@ -3626,24 +3539,110 @@ static bool sandybridge_compute_srwm(struct drm_device *dev, int level, | |||
| 3626 | small = ((clock * pixel_size / 1000) * latency_ns) / 1000; | 3539 | small = ((clock * pixel_size / 1000) * latency_ns) / 1000; |
| 3627 | large = line_count * line_size; | 3540 | large = line_count * line_size; |
| 3628 | 3541 | ||
| 3629 | entries = DIV_ROUND_UP(min(small, large), | 3542 | entries = DIV_ROUND_UP(min(small, large), display->cacheline_size); |
| 3630 | sandybridge_display_srwm_info.cacheline_size); | 3543 | *display_wm = entries + display->guard_size; |
| 3631 | *display_wm = entries + sandybridge_display_srwm_info.guard_size; | ||
| 3632 | 3544 | ||
| 3633 | /* | 3545 | /* |
| 3634 | * Spec said: | 3546 | * Spec says: |
| 3635 | * FBC WM = ((Final Primary WM * 64) / number of bytes per line) + 2 | 3547 | * FBC WM = ((Final Primary WM * 64) / number of bytes per line) + 2 |
| 3636 | */ | 3548 | */ |
| 3637 | *fbc_wm = DIV_ROUND_UP(*display_wm * 64, line_size) + 2; | 3549 | *fbc_wm = DIV_ROUND_UP(*display_wm * 64, line_size) + 2; |
| 3638 | 3550 | ||
| 3639 | /* calculate the self-refresh watermark for display cursor */ | 3551 | /* calculate the self-refresh watermark for display cursor */ |
| 3640 | entries = line_count * pixel_size * 64; | 3552 | entries = line_count * pixel_size * 64; |
| 3641 | entries = DIV_ROUND_UP(entries, | 3553 | entries = DIV_ROUND_UP(entries, cursor->cacheline_size); |
| 3642 | sandybridge_cursor_srwm_info.cacheline_size); | 3554 | *cursor_wm = entries + cursor->guard_size; |
| 3643 | *cursor_wm = entries + sandybridge_cursor_srwm_info.guard_size; | ||
| 3644 | 3555 | ||
| 3645 | return sandybridge_check_srwm(dev, level, | 3556 | return ironlake_check_srwm(dev, level, |
| 3646 | *fbc_wm, *display_wm, *cursor_wm); | 3557 | *fbc_wm, *display_wm, *cursor_wm, |
| 3558 | display, cursor); | ||
| 3559 | } | ||
| 3560 | |||
| 3561 | static void ironlake_update_wm(struct drm_device *dev, | ||
| 3562 | int planea_clock, int planeb_clock, | ||
| 3563 | int hdisplay, int htotal, | ||
| 3564 | int pixel_size) | ||
| 3565 | { | ||
| 3566 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 3567 | int fbc_wm, plane_wm, cursor_wm, enabled; | ||
| 3568 | int clock; | ||
| 3569 | |||
| 3570 | enabled = 0; | ||
| 3571 | if (ironlake_compute_wm0(dev, 0, | ||
| 3572 | &ironlake_display_wm_info, | ||
| 3573 | ILK_LP0_PLANE_LATENCY, | ||
| 3574 | &ironlake_cursor_wm_info, | ||
| 3575 | ILK_LP0_CURSOR_LATENCY, | ||
| 3576 | &plane_wm, &cursor_wm)) { | ||
| 3577 | I915_WRITE(WM0_PIPEA_ILK, | ||
| 3578 | (plane_wm << WM0_PIPE_PLANE_SHIFT) | cursor_wm); | ||
| 3579 | DRM_DEBUG_KMS("FIFO watermarks For pipe A -" | ||
| 3580 | " plane %d, " "cursor: %d\n", | ||
| 3581 | plane_wm, cursor_wm); | ||
| 3582 | enabled++; | ||
| 3583 | } | ||
| 3584 | |||
| 3585 | if (ironlake_compute_wm0(dev, 1, | ||
| 3586 | &ironlake_display_wm_info, | ||
| 3587 | ILK_LP0_PLANE_LATENCY, | ||
| 3588 | &ironlake_cursor_wm_info, | ||
| 3589 | ILK_LP0_CURSOR_LATENCY, | ||
| 3590 | &plane_wm, &cursor_wm)) { | ||
| 3591 | I915_WRITE(WM0_PIPEB_ILK, | ||
| 3592 | (plane_wm << WM0_PIPE_PLANE_SHIFT) | cursor_wm); | ||
| 3593 | DRM_DEBUG_KMS("FIFO watermarks For pipe B -" | ||
| 3594 | " plane %d, cursor: %d\n", | ||
| 3595 | plane_wm, cursor_wm); | ||
| 3596 | enabled++; | ||
| 3597 | } | ||
| 3598 | |||
| 3599 | /* | ||
| 3600 | * Calculate and update the self-refresh watermark only when one | ||
| 3601 | * display plane is used. | ||
| 3602 | */ | ||
| 3603 | I915_WRITE(WM3_LP_ILK, 0); | ||
| 3604 | I915_WRITE(WM2_LP_ILK, 0); | ||
| 3605 | I915_WRITE(WM1_LP_ILK, 0); | ||
| 3606 | |||
| 3607 | if (enabled != 1) | ||
| 3608 | return; | ||
| 3609 | |||
| 3610 | clock = planea_clock ? planea_clock : planeb_clock; | ||
| 3611 | |||
| 3612 | /* WM1 */ | ||
| 3613 | if (!ironlake_compute_srwm(dev, 1, hdisplay, htotal, pixel_size, | ||
| 3614 | clock, ILK_READ_WM1_LATENCY() * 500, | ||
| 3615 | &ironlake_display_srwm_info, | ||
| 3616 | &ironlake_cursor_srwm_info, | ||
| 3617 | &fbc_wm, &plane_wm, &cursor_wm)) | ||
| 3618 | return; | ||
| 3619 | |||
| 3620 | I915_WRITE(WM1_LP_ILK, | ||
| 3621 | WM1_LP_SR_EN | | ||
| 3622 | (ILK_READ_WM1_LATENCY() << WM1_LP_LATENCY_SHIFT) | | ||
| 3623 | (fbc_wm << WM1_LP_FBC_SHIFT) | | ||
| 3624 | (plane_wm << WM1_LP_SR_SHIFT) | | ||
| 3625 | cursor_wm); | ||
| 3626 | |||
| 3627 | /* WM2 */ | ||
| 3628 | if (!ironlake_compute_srwm(dev, 2, hdisplay, htotal, pixel_size, | ||
| 3629 | clock, ILK_READ_WM2_LATENCY() * 500, | ||
| 3630 | &ironlake_display_srwm_info, | ||
| 3631 | &ironlake_cursor_srwm_info, | ||
| 3632 | &fbc_wm, &plane_wm, &cursor_wm)) | ||
| 3633 | return; | ||
| 3634 | |||
| 3635 | I915_WRITE(WM2_LP_ILK, | ||
| 3636 | WM2_LP_EN | | ||
| 3637 | (ILK_READ_WM2_LATENCY() << WM1_LP_LATENCY_SHIFT) | | ||
| 3638 | (fbc_wm << WM1_LP_FBC_SHIFT) | | ||
| 3639 | (plane_wm << WM1_LP_SR_SHIFT) | | ||
| 3640 | cursor_wm); | ||
| 3641 | |||
| 3642 | /* | ||
| 3643 | * WM3 is unsupported on ILK, probably because we don't have latency | ||
| 3644 | * data for that power state | ||
| 3645 | */ | ||
| 3647 | } | 3646 | } |
| 3648 | 3647 | ||
| 3649 | static void sandybridge_update_wm(struct drm_device *dev, | 3648 | static void sandybridge_update_wm(struct drm_device *dev, |
| @@ -3652,7 +3651,7 @@ static void sandybridge_update_wm(struct drm_device *dev, | |||
| 3652 | int pixel_size) | 3651 | int pixel_size) |
| 3653 | { | 3652 | { |
| 3654 | struct drm_i915_private *dev_priv = dev->dev_private; | 3653 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 3655 | int latency = SNB_READ_WM0_LATENCY(); | 3654 | int latency = SNB_READ_WM0_LATENCY() * 100; /* In unit 0.1us */ |
| 3656 | int fbc_wm, plane_wm, cursor_wm, enabled; | 3655 | int fbc_wm, plane_wm, cursor_wm, enabled; |
| 3657 | int clock; | 3656 | int clock; |
| 3658 | 3657 | ||
| @@ -3701,9 +3700,11 @@ static void sandybridge_update_wm(struct drm_device *dev, | |||
| 3701 | clock = planea_clock ? planea_clock : planeb_clock; | 3700 | clock = planea_clock ? planea_clock : planeb_clock; |
| 3702 | 3701 | ||
| 3703 | /* WM1 */ | 3702 | /* WM1 */ |
| 3704 | if (!sandybridge_compute_srwm(dev, 1, hdisplay, htotal, pixel_size, | 3703 | if (!ironlake_compute_srwm(dev, 1, hdisplay, htotal, pixel_size, |
| 3705 | clock, SNB_READ_WM1_LATENCY() * 500, | 3704 | clock, SNB_READ_WM1_LATENCY() * 500, |
| 3706 | &fbc_wm, &plane_wm, &cursor_wm)) | 3705 | &sandybridge_display_srwm_info, |
| 3706 | &sandybridge_cursor_srwm_info, | ||
| 3707 | &fbc_wm, &plane_wm, &cursor_wm)) | ||
| 3707 | return; | 3708 | return; |
| 3708 | 3709 | ||
| 3709 | I915_WRITE(WM1_LP_ILK, | 3710 | I915_WRITE(WM1_LP_ILK, |
| @@ -3714,10 +3715,12 @@ static void sandybridge_update_wm(struct drm_device *dev, | |||
| 3714 | cursor_wm); | 3715 | cursor_wm); |
| 3715 | 3716 | ||
| 3716 | /* WM2 */ | 3717 | /* WM2 */ |
| 3717 | if (!sandybridge_compute_srwm(dev, 2, | 3718 | if (!ironlake_compute_srwm(dev, 2, |
| 3718 | hdisplay, htotal, pixel_size, | 3719 | hdisplay, htotal, pixel_size, |
| 3719 | clock, SNB_READ_WM2_LATENCY() * 500, | 3720 | clock, SNB_READ_WM2_LATENCY() * 500, |
| 3720 | &fbc_wm, &plane_wm, &cursor_wm)) | 3721 | &sandybridge_display_srwm_info, |
| 3722 | &sandybridge_cursor_srwm_info, | ||
| 3723 | &fbc_wm, &plane_wm, &cursor_wm)) | ||
| 3721 | return; | 3724 | return; |
| 3722 | 3725 | ||
| 3723 | I915_WRITE(WM2_LP_ILK, | 3726 | I915_WRITE(WM2_LP_ILK, |
| @@ -3728,10 +3731,12 @@ static void sandybridge_update_wm(struct drm_device *dev, | |||
| 3728 | cursor_wm); | 3731 | cursor_wm); |
| 3729 | 3732 | ||
| 3730 | /* WM3 */ | 3733 | /* WM3 */ |
| 3731 | if (!sandybridge_compute_srwm(dev, 3, | 3734 | if (!ironlake_compute_srwm(dev, 3, |
| 3732 | hdisplay, htotal, pixel_size, | 3735 | hdisplay, htotal, pixel_size, |
| 3733 | clock, SNB_READ_WM3_LATENCY() * 500, | 3736 | clock, SNB_READ_WM3_LATENCY() * 500, |
| 3734 | &fbc_wm, &plane_wm, &cursor_wm)) | 3737 | &sandybridge_display_srwm_info, |
| 3738 | &sandybridge_cursor_srwm_info, | ||
| 3739 | &fbc_wm, &plane_wm, &cursor_wm)) | ||
| 3735 | return; | 3740 | return; |
| 3736 | 3741 | ||
| 3737 | I915_WRITE(WM3_LP_ILK, | 3742 | I915_WRITE(WM3_LP_ILK, |
| @@ -3951,7 +3956,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3951 | int lane = 0, link_bw, bpp; | 3956 | int lane = 0, link_bw, bpp; |
| 3952 | /* CPU eDP doesn't require FDI link, so just set DP M/N | 3957 | /* CPU eDP doesn't require FDI link, so just set DP M/N |
| 3953 | according to current link config */ | 3958 | according to current link config */ |
| 3954 | if (has_edp_encoder && !intel_encoder_is_pch_edp(&encoder->base)) { | 3959 | if (has_edp_encoder && !intel_encoder_is_pch_edp(&has_edp_encoder->base)) { |
| 3955 | target_clock = mode->clock; | 3960 | target_clock = mode->clock; |
| 3956 | intel_edp_link_config(has_edp_encoder, | 3961 | intel_edp_link_config(has_edp_encoder, |
| 3957 | &lane, &link_bw); | 3962 | &lane, &link_bw); |
| @@ -5038,8 +5043,8 @@ static void intel_increase_pllclock(struct drm_crtc *crtc) | |||
| 5038 | drm_i915_private_t *dev_priv = dev->dev_private; | 5043 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 5039 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 5044 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
| 5040 | int pipe = intel_crtc->pipe; | 5045 | int pipe = intel_crtc->pipe; |
| 5041 | int dpll_reg = (pipe == 0) ? DPLL_A : DPLL_B; | 5046 | int dpll_reg = DPLL(pipe); |
| 5042 | int dpll = I915_READ(dpll_reg); | 5047 | int dpll; |
| 5043 | 5048 | ||
| 5044 | if (HAS_PCH_SPLIT(dev)) | 5049 | if (HAS_PCH_SPLIT(dev)) |
| 5045 | return; | 5050 | return; |
| @@ -5047,17 +5052,19 @@ static void intel_increase_pllclock(struct drm_crtc *crtc) | |||
| 5047 | if (!dev_priv->lvds_downclock_avail) | 5052 | if (!dev_priv->lvds_downclock_avail) |
| 5048 | return; | 5053 | return; |
| 5049 | 5054 | ||
| 5055 | dpll = I915_READ(dpll_reg); | ||
| 5050 | if (!HAS_PIPE_CXSR(dev) && (dpll & DISPLAY_RATE_SELECT_FPA1)) { | 5056 | if (!HAS_PIPE_CXSR(dev) && (dpll & DISPLAY_RATE_SELECT_FPA1)) { |
| 5051 | DRM_DEBUG_DRIVER("upclocking LVDS\n"); | 5057 | DRM_DEBUG_DRIVER("upclocking LVDS\n"); |
| 5052 | 5058 | ||
| 5053 | /* Unlock panel regs */ | 5059 | /* Unlock panel regs */ |
| 5054 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) | | 5060 | I915_WRITE(PP_CONTROL, |
| 5055 | PANEL_UNLOCK_REGS); | 5061 | I915_READ(PP_CONTROL) | PANEL_UNLOCK_REGS); |
| 5056 | 5062 | ||
| 5057 | dpll &= ~DISPLAY_RATE_SELECT_FPA1; | 5063 | dpll &= ~DISPLAY_RATE_SELECT_FPA1; |
| 5058 | I915_WRITE(dpll_reg, dpll); | 5064 | I915_WRITE(dpll_reg, dpll); |
| 5059 | dpll = I915_READ(dpll_reg); | 5065 | POSTING_READ(dpll_reg); |
| 5060 | intel_wait_for_vblank(dev, pipe); | 5066 | intel_wait_for_vblank(dev, pipe); |
| 5067 | |||
| 5061 | dpll = I915_READ(dpll_reg); | 5068 | dpll = I915_READ(dpll_reg); |
| 5062 | if (dpll & DISPLAY_RATE_SELECT_FPA1) | 5069 | if (dpll & DISPLAY_RATE_SELECT_FPA1) |
| 5063 | DRM_DEBUG_DRIVER("failed to upclock LVDS!\n"); | 5070 | DRM_DEBUG_DRIVER("failed to upclock LVDS!\n"); |
| @@ -5802,6 +5809,8 @@ static void intel_setup_outputs(struct drm_device *dev) | |||
| 5802 | encoder->base.possible_clones = | 5809 | encoder->base.possible_clones = |
| 5803 | intel_encoder_clones(dev, encoder->clone_mask); | 5810 | intel_encoder_clones(dev, encoder->clone_mask); |
| 5804 | } | 5811 | } |
| 5812 | |||
| 5813 | intel_panel_setup_backlight(dev); | ||
| 5805 | } | 5814 | } |
| 5806 | 5815 | ||
| 5807 | static void intel_user_framebuffer_destroy(struct drm_framebuffer *fb) | 5816 | static void intel_user_framebuffer_destroy(struct drm_framebuffer *fb) |
| @@ -6145,6 +6154,10 @@ void intel_init_emon(struct drm_device *dev) | |||
| 6145 | 6154 | ||
| 6146 | void gen6_enable_rps(struct drm_i915_private *dev_priv) | 6155 | void gen6_enable_rps(struct drm_i915_private *dev_priv) |
| 6147 | { | 6156 | { |
| 6157 | u32 rp_state_cap = I915_READ(GEN6_RP_STATE_CAP); | ||
| 6158 | u32 gt_perf_status = I915_READ(GEN6_GT_PERF_STATUS); | ||
| 6159 | u32 pcu_mbox; | ||
| 6160 | int cur_freq, min_freq, max_freq; | ||
| 6148 | int i; | 6161 | int i; |
| 6149 | 6162 | ||
| 6150 | /* Here begins a magic sequence of register writes to enable | 6163 | /* Here begins a magic sequence of register writes to enable |
| @@ -6216,6 +6229,29 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
| 6216 | 500)) | 6229 | 500)) |
| 6217 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); | 6230 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); |
| 6218 | 6231 | ||
| 6232 | min_freq = (rp_state_cap & 0xff0000) >> 16; | ||
| 6233 | max_freq = rp_state_cap & 0xff; | ||
| 6234 | cur_freq = (gt_perf_status & 0xff00) >> 8; | ||
| 6235 | |||
| 6236 | /* Check for overclock support */ | ||
| 6237 | if (wait_for((I915_READ(GEN6_PCODE_MAILBOX) & GEN6_PCODE_READY) == 0, | ||
| 6238 | 500)) | ||
| 6239 | DRM_ERROR("timeout waiting for pcode mailbox to become idle\n"); | ||
| 6240 | I915_WRITE(GEN6_PCODE_MAILBOX, GEN6_READ_OC_PARAMS); | ||
| 6241 | pcu_mbox = I915_READ(GEN6_PCODE_DATA); | ||
| 6242 | if (wait_for((I915_READ(GEN6_PCODE_MAILBOX) & GEN6_PCODE_READY) == 0, | ||
| 6243 | 500)) | ||
| 6244 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); | ||
| 6245 | if (pcu_mbox & (1<<31)) { /* OC supported */ | ||
| 6246 | max_freq = pcu_mbox & 0xff; | ||
| 6247 | DRM_DEBUG_DRIVER("overclocking supported, adjusting frequency max to %dMHz\n", pcu_mbox * 100); | ||
| 6248 | } | ||
| 6249 | |||
| 6250 | /* In units of 100MHz */ | ||
| 6251 | dev_priv->max_delay = max_freq; | ||
| 6252 | dev_priv->min_delay = min_freq; | ||
| 6253 | dev_priv->cur_delay = cur_freq; | ||
| 6254 | |||
| 6219 | /* requires MSI enabled */ | 6255 | /* requires MSI enabled */ |
| 6220 | I915_WRITE(GEN6_PMIER, | 6256 | I915_WRITE(GEN6_PMIER, |
| 6221 | GEN6_PM_MBOX_EVENT | | 6257 | GEN6_PM_MBOX_EVENT | |
| @@ -6386,42 +6422,6 @@ void intel_enable_clock_gating(struct drm_device *dev) | |||
| 6386 | } else if (IS_I830(dev)) { | 6422 | } else if (IS_I830(dev)) { |
| 6387 | I915_WRITE(DSPCLK_GATE_D, OVRUNIT_CLOCK_GATE_DISABLE); | 6423 | I915_WRITE(DSPCLK_GATE_D, OVRUNIT_CLOCK_GATE_DISABLE); |
| 6388 | } | 6424 | } |
| 6389 | |||
| 6390 | /* | ||
| 6391 | * GPU can automatically power down the render unit if given a page | ||
| 6392 | * to save state. | ||
| 6393 | */ | ||
| 6394 | if (IS_IRONLAKE_M(dev) && 0) { /* XXX causes a failure during suspend */ | ||
| 6395 | if (dev_priv->renderctx == NULL) | ||
| 6396 | dev_priv->renderctx = intel_alloc_context_page(dev); | ||
| 6397 | if (dev_priv->renderctx) { | ||
| 6398 | struct drm_i915_gem_object *obj = dev_priv->renderctx; | ||
| 6399 | if (BEGIN_LP_RING(4) == 0) { | ||
| 6400 | OUT_RING(MI_SET_CONTEXT); | ||
| 6401 | OUT_RING(obj->gtt_offset | | ||
| 6402 | MI_MM_SPACE_GTT | | ||
| 6403 | MI_SAVE_EXT_STATE_EN | | ||
| 6404 | MI_RESTORE_EXT_STATE_EN | | ||
| 6405 | MI_RESTORE_INHIBIT); | ||
| 6406 | OUT_RING(MI_NOOP); | ||
| 6407 | OUT_RING(MI_FLUSH); | ||
| 6408 | ADVANCE_LP_RING(); | ||
| 6409 | } | ||
| 6410 | } else | ||
| 6411 | DRM_DEBUG_KMS("Failed to allocate render context." | ||
| 6412 | "Disable RC6\n"); | ||
| 6413 | } | ||
| 6414 | |||
| 6415 | if (IS_GEN4(dev) && IS_MOBILE(dev)) { | ||
| 6416 | if (dev_priv->pwrctx == NULL) | ||
| 6417 | dev_priv->pwrctx = intel_alloc_context_page(dev); | ||
| 6418 | if (dev_priv->pwrctx) { | ||
| 6419 | struct drm_i915_gem_object *obj = dev_priv->pwrctx; | ||
| 6420 | I915_WRITE(PWRCTXA, obj->gtt_offset | PWRCTX_EN); | ||
| 6421 | I915_WRITE(MCHBAR_RENDER_STANDBY, | ||
| 6422 | I915_READ(MCHBAR_RENDER_STANDBY) & ~RCX_SW_EXIT); | ||
| 6423 | } | ||
| 6424 | } | ||
| 6425 | } | 6425 | } |
| 6426 | 6426 | ||
| 6427 | void intel_disable_clock_gating(struct drm_device *dev) | 6427 | void intel_disable_clock_gating(struct drm_device *dev) |
| @@ -6451,6 +6451,57 @@ void intel_disable_clock_gating(struct drm_device *dev) | |||
| 6451 | } | 6451 | } |
| 6452 | } | 6452 | } |
| 6453 | 6453 | ||
| 6454 | static void ironlake_disable_rc6(struct drm_device *dev) | ||
| 6455 | { | ||
| 6456 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 6457 | |||
| 6458 | /* Wake the GPU, prevent RC6, then restore RSTDBYCTL */ | ||
| 6459 | I915_WRITE(RSTDBYCTL, I915_READ(RSTDBYCTL) | RCX_SW_EXIT); | ||
| 6460 | wait_for(((I915_READ(RSTDBYCTL) & RSX_STATUS_MASK) == RSX_STATUS_ON), | ||
| 6461 | 10); | ||
| 6462 | POSTING_READ(CCID); | ||
| 6463 | I915_WRITE(PWRCTXA, 0); | ||
| 6464 | POSTING_READ(PWRCTXA); | ||
| 6465 | I915_WRITE(RSTDBYCTL, I915_READ(RSTDBYCTL) & ~RCX_SW_EXIT); | ||
| 6466 | POSTING_READ(RSTDBYCTL); | ||
| 6467 | i915_gem_object_unpin(dev_priv->renderctx); | ||
| 6468 | drm_gem_object_unreference(&dev_priv->renderctx->base); | ||
| 6469 | dev_priv->renderctx = NULL; | ||
| 6470 | i915_gem_object_unpin(dev_priv->pwrctx); | ||
| 6471 | drm_gem_object_unreference(&dev_priv->pwrctx->base); | ||
| 6472 | dev_priv->pwrctx = NULL; | ||
| 6473 | } | ||
| 6474 | |||
| 6475 | void ironlake_enable_rc6(struct drm_device *dev) | ||
| 6476 | { | ||
| 6477 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 6478 | int ret; | ||
| 6479 | |||
| 6480 | /* | ||
| 6481 | * GPU can automatically power down the render unit if given a page | ||
| 6482 | * to save state. | ||
| 6483 | */ | ||
| 6484 | ret = BEGIN_LP_RING(6); | ||
| 6485 | if (ret) { | ||
| 6486 | ironlake_disable_rc6(dev); | ||
| 6487 | return; | ||
| 6488 | } | ||
| 6489 | OUT_RING(MI_SUSPEND_FLUSH | MI_SUSPEND_FLUSH_EN); | ||
| 6490 | OUT_RING(MI_SET_CONTEXT); | ||
| 6491 | OUT_RING(dev_priv->renderctx->gtt_offset | | ||
| 6492 | MI_MM_SPACE_GTT | | ||
| 6493 | MI_SAVE_EXT_STATE_EN | | ||
| 6494 | MI_RESTORE_EXT_STATE_EN | | ||
| 6495 | MI_RESTORE_INHIBIT); | ||
| 6496 | OUT_RING(MI_SUSPEND_FLUSH); | ||
| 6497 | OUT_RING(MI_NOOP); | ||
| 6498 | OUT_RING(MI_FLUSH); | ||
| 6499 | ADVANCE_LP_RING(); | ||
| 6500 | |||
| 6501 | I915_WRITE(PWRCTXA, dev_priv->pwrctx->gtt_offset | PWRCTX_EN); | ||
| 6502 | I915_WRITE(RSTDBYCTL, I915_READ(RSTDBYCTL) & ~RCX_SW_EXIT); | ||
| 6503 | } | ||
| 6504 | |||
| 6454 | /* Set up chip specific display functions */ | 6505 | /* Set up chip specific display functions */ |
| 6455 | static void intel_init_display(struct drm_device *dev) | 6506 | static void intel_init_display(struct drm_device *dev) |
| 6456 | { | 6507 | { |
| @@ -6665,12 +6716,7 @@ void intel_modeset_init(struct drm_device *dev) | |||
| 6665 | dev->mode_config.max_width = 8192; | 6716 | dev->mode_config.max_width = 8192; |
| 6666 | dev->mode_config.max_height = 8192; | 6717 | dev->mode_config.max_height = 8192; |
| 6667 | } | 6718 | } |
| 6668 | 6719 | dev->mode_config.fb_base = dev->agp->base; | |
| 6669 | /* set memory base */ | ||
| 6670 | if (IS_GEN2(dev)) | ||
| 6671 | dev->mode_config.fb_base = pci_resource_start(dev->pdev, 0); | ||
| 6672 | else | ||
| 6673 | dev->mode_config.fb_base = pci_resource_start(dev->pdev, 2); | ||
| 6674 | 6720 | ||
| 6675 | if (IS_MOBILE(dev) || !IS_GEN2(dev)) | 6721 | if (IS_MOBILE(dev) || !IS_GEN2(dev)) |
| 6676 | dev_priv->num_pipe = 2; | 6722 | dev_priv->num_pipe = 2; |
| @@ -6698,6 +6744,21 @@ void intel_modeset_init(struct drm_device *dev) | |||
| 6698 | if (IS_GEN6(dev)) | 6744 | if (IS_GEN6(dev)) |
| 6699 | gen6_enable_rps(dev_priv); | 6745 | gen6_enable_rps(dev_priv); |
| 6700 | 6746 | ||
| 6747 | if (IS_IRONLAKE_M(dev)) { | ||
| 6748 | dev_priv->renderctx = intel_alloc_context_page(dev); | ||
| 6749 | if (!dev_priv->renderctx) | ||
| 6750 | goto skip_rc6; | ||
| 6751 | dev_priv->pwrctx = intel_alloc_context_page(dev); | ||
| 6752 | if (!dev_priv->pwrctx) { | ||
| 6753 | i915_gem_object_unpin(dev_priv->renderctx); | ||
| 6754 | drm_gem_object_unreference(&dev_priv->renderctx->base); | ||
| 6755 | dev_priv->renderctx = NULL; | ||
| 6756 | goto skip_rc6; | ||
| 6757 | } | ||
| 6758 | ironlake_enable_rc6(dev); | ||
| 6759 | } | ||
| 6760 | |||
| 6761 | skip_rc6: | ||
| 6701 | INIT_WORK(&dev_priv->idle_work, intel_idle_update); | 6762 | INIT_WORK(&dev_priv->idle_work, intel_idle_update); |
| 6702 | setup_timer(&dev_priv->idle_timer, intel_gpu_idle_timer, | 6763 | setup_timer(&dev_priv->idle_timer, intel_gpu_idle_timer, |
| 6703 | (unsigned long)dev); | 6764 | (unsigned long)dev); |
| @@ -6734,7 +6795,8 @@ void intel_modeset_cleanup(struct drm_device *dev) | |||
| 6734 | if (IS_GEN6(dev)) | 6795 | if (IS_GEN6(dev)) |
| 6735 | gen6_disable_rps(dev); | 6796 | gen6_disable_rps(dev); |
| 6736 | 6797 | ||
| 6737 | intel_disable_clock_gating(dev); | 6798 | if (IS_IRONLAKE_M(dev)) |
| 6799 | ironlake_disable_rc6(dev); | ||
| 6738 | 6800 | ||
| 6739 | mutex_unlock(&dev->struct_mutex); | 6801 | mutex_unlock(&dev->struct_mutex); |
| 6740 | 6802 | ||
diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c index 1dc60408d5b8..1f4242b682c8 100644 --- a/drivers/gpu/drm/i915/intel_dp.c +++ b/drivers/gpu/drm/i915/intel_dp.c | |||
| @@ -1153,18 +1153,27 @@ intel_dp_signal_levels(uint8_t train_set, int lane_count) | |||
| 1153 | static uint32_t | 1153 | static uint32_t |
| 1154 | intel_gen6_edp_signal_levels(uint8_t train_set) | 1154 | intel_gen6_edp_signal_levels(uint8_t train_set) |
| 1155 | { | 1155 | { |
| 1156 | switch (train_set & (DP_TRAIN_VOLTAGE_SWING_MASK|DP_TRAIN_PRE_EMPHASIS_MASK)) { | 1156 | int signal_levels = train_set & (DP_TRAIN_VOLTAGE_SWING_MASK | |
| 1157 | DP_TRAIN_PRE_EMPHASIS_MASK); | ||
| 1158 | switch (signal_levels) { | ||
| 1157 | case DP_TRAIN_VOLTAGE_SWING_400 | DP_TRAIN_PRE_EMPHASIS_0: | 1159 | case DP_TRAIN_VOLTAGE_SWING_400 | DP_TRAIN_PRE_EMPHASIS_0: |
| 1158 | return EDP_LINK_TRAIN_400MV_0DB_SNB_B; | 1160 | case DP_TRAIN_VOLTAGE_SWING_600 | DP_TRAIN_PRE_EMPHASIS_0: |
| 1161 | return EDP_LINK_TRAIN_400_600MV_0DB_SNB_B; | ||
| 1162 | case DP_TRAIN_VOLTAGE_SWING_400 | DP_TRAIN_PRE_EMPHASIS_3_5: | ||
| 1163 | return EDP_LINK_TRAIN_400MV_3_5DB_SNB_B; | ||
| 1159 | case DP_TRAIN_VOLTAGE_SWING_400 | DP_TRAIN_PRE_EMPHASIS_6: | 1164 | case DP_TRAIN_VOLTAGE_SWING_400 | DP_TRAIN_PRE_EMPHASIS_6: |
| 1160 | return EDP_LINK_TRAIN_400MV_6DB_SNB_B; | 1165 | case DP_TRAIN_VOLTAGE_SWING_600 | DP_TRAIN_PRE_EMPHASIS_6: |
| 1166 | return EDP_LINK_TRAIN_400_600MV_6DB_SNB_B; | ||
| 1161 | case DP_TRAIN_VOLTAGE_SWING_600 | DP_TRAIN_PRE_EMPHASIS_3_5: | 1167 | case DP_TRAIN_VOLTAGE_SWING_600 | DP_TRAIN_PRE_EMPHASIS_3_5: |
| 1162 | return EDP_LINK_TRAIN_600MV_3_5DB_SNB_B; | 1168 | case DP_TRAIN_VOLTAGE_SWING_800 | DP_TRAIN_PRE_EMPHASIS_3_5: |
| 1169 | return EDP_LINK_TRAIN_600_800MV_3_5DB_SNB_B; | ||
| 1163 | case DP_TRAIN_VOLTAGE_SWING_800 | DP_TRAIN_PRE_EMPHASIS_0: | 1170 | case DP_TRAIN_VOLTAGE_SWING_800 | DP_TRAIN_PRE_EMPHASIS_0: |
| 1164 | return EDP_LINK_TRAIN_800MV_0DB_SNB_B; | 1171 | case DP_TRAIN_VOLTAGE_SWING_1200 | DP_TRAIN_PRE_EMPHASIS_0: |
| 1172 | return EDP_LINK_TRAIN_800_1200MV_0DB_SNB_B; | ||
| 1165 | default: | 1173 | default: |
| 1166 | DRM_DEBUG_KMS("Unsupported voltage swing/pre-emphasis level\n"); | 1174 | DRM_DEBUG_KMS("Unsupported voltage swing/pre-emphasis level:" |
| 1167 | return EDP_LINK_TRAIN_400MV_0DB_SNB_B; | 1175 | "0x%x\n", signal_levels); |
| 1176 | return EDP_LINK_TRAIN_400_600MV_0DB_SNB_B; | ||
| 1168 | } | 1177 | } |
| 1169 | } | 1178 | } |
| 1170 | 1179 | ||
| @@ -1334,17 +1343,24 @@ intel_dp_complete_link_train(struct intel_dp *intel_dp) | |||
| 1334 | struct drm_device *dev = intel_dp->base.base.dev; | 1343 | struct drm_device *dev = intel_dp->base.base.dev; |
| 1335 | struct drm_i915_private *dev_priv = dev->dev_private; | 1344 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1336 | bool channel_eq = false; | 1345 | bool channel_eq = false; |
| 1337 | int tries; | 1346 | int tries, cr_tries; |
| 1338 | u32 reg; | 1347 | u32 reg; |
| 1339 | uint32_t DP = intel_dp->DP; | 1348 | uint32_t DP = intel_dp->DP; |
| 1340 | 1349 | ||
| 1341 | /* channel equalization */ | 1350 | /* channel equalization */ |
| 1342 | tries = 0; | 1351 | tries = 0; |
| 1352 | cr_tries = 0; | ||
| 1343 | channel_eq = false; | 1353 | channel_eq = false; |
| 1344 | for (;;) { | 1354 | for (;;) { |
| 1345 | /* Use intel_dp->train_set[0] to set the voltage and pre emphasis values */ | 1355 | /* Use intel_dp->train_set[0] to set the voltage and pre emphasis values */ |
| 1346 | uint32_t signal_levels; | 1356 | uint32_t signal_levels; |
| 1347 | 1357 | ||
| 1358 | if (cr_tries > 5) { | ||
| 1359 | DRM_ERROR("failed to train DP, aborting\n"); | ||
| 1360 | intel_dp_link_down(intel_dp); | ||
| 1361 | break; | ||
| 1362 | } | ||
| 1363 | |||
| 1348 | if (IS_GEN6(dev) && is_edp(intel_dp)) { | 1364 | if (IS_GEN6(dev) && is_edp(intel_dp)) { |
| 1349 | signal_levels = intel_gen6_edp_signal_levels(intel_dp->train_set[0]); | 1365 | signal_levels = intel_gen6_edp_signal_levels(intel_dp->train_set[0]); |
| 1350 | DP = (DP & ~EDP_LINK_TRAIN_VOL_EMP_MASK_SNB) | signal_levels; | 1366 | DP = (DP & ~EDP_LINK_TRAIN_VOL_EMP_MASK_SNB) | signal_levels; |
| @@ -1367,14 +1383,26 @@ intel_dp_complete_link_train(struct intel_dp *intel_dp) | |||
| 1367 | if (!intel_dp_get_link_status(intel_dp)) | 1383 | if (!intel_dp_get_link_status(intel_dp)) |
| 1368 | break; | 1384 | break; |
| 1369 | 1385 | ||
| 1386 | /* Make sure clock is still ok */ | ||
| 1387 | if (!intel_clock_recovery_ok(intel_dp->link_status, intel_dp->lane_count)) { | ||
| 1388 | intel_dp_start_link_train(intel_dp); | ||
| 1389 | cr_tries++; | ||
| 1390 | continue; | ||
| 1391 | } | ||
| 1392 | |||
| 1370 | if (intel_channel_eq_ok(intel_dp)) { | 1393 | if (intel_channel_eq_ok(intel_dp)) { |
| 1371 | channel_eq = true; | 1394 | channel_eq = true; |
| 1372 | break; | 1395 | break; |
| 1373 | } | 1396 | } |
| 1374 | 1397 | ||
| 1375 | /* Try 5 times */ | 1398 | /* Try 5 times, then try clock recovery if that fails */ |
| 1376 | if (tries > 5) | 1399 | if (tries > 5) { |
| 1377 | break; | 1400 | intel_dp_link_down(intel_dp); |
| 1401 | intel_dp_start_link_train(intel_dp); | ||
| 1402 | tries = 0; | ||
| 1403 | cr_tries++; | ||
| 1404 | continue; | ||
| 1405 | } | ||
| 1378 | 1406 | ||
| 1379 | /* Compute new intel_dp->train_set as requested by target */ | 1407 | /* Compute new intel_dp->train_set as requested by target */ |
| 1380 | intel_get_adjust_train(intel_dp); | 1408 | intel_get_adjust_train(intel_dp); |
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h index d782ad9fd6db..74db2557d644 100644 --- a/drivers/gpu/drm/i915/intel_drv.h +++ b/drivers/gpu/drm/i915/intel_drv.h | |||
| @@ -257,6 +257,9 @@ extern void intel_pch_panel_fitting(struct drm_device *dev, | |||
| 257 | extern u32 intel_panel_get_max_backlight(struct drm_device *dev); | 257 | extern u32 intel_panel_get_max_backlight(struct drm_device *dev); |
| 258 | extern u32 intel_panel_get_backlight(struct drm_device *dev); | 258 | extern u32 intel_panel_get_backlight(struct drm_device *dev); |
| 259 | extern void intel_panel_set_backlight(struct drm_device *dev, u32 level); | 259 | extern void intel_panel_set_backlight(struct drm_device *dev, u32 level); |
| 260 | extern void intel_panel_setup_backlight(struct drm_device *dev); | ||
| 261 | extern void intel_panel_enable_backlight(struct drm_device *dev); | ||
| 262 | extern void intel_panel_disable_backlight(struct drm_device *dev); | ||
| 260 | 263 | ||
| 261 | extern void intel_crtc_load_lut(struct drm_crtc *crtc); | 264 | extern void intel_crtc_load_lut(struct drm_crtc *crtc); |
| 262 | extern void intel_encoder_prepare (struct drm_encoder *encoder); | 265 | extern void intel_encoder_prepare (struct drm_encoder *encoder); |
diff --git a/drivers/gpu/drm/i915/intel_fb.c b/drivers/gpu/drm/i915/intel_fb.c index 701e830d0012..ee145a257287 100644 --- a/drivers/gpu/drm/i915/intel_fb.c +++ b/drivers/gpu/drm/i915/intel_fb.c | |||
| @@ -62,6 +62,7 @@ static int intelfb_create(struct intel_fbdev *ifbdev, | |||
| 62 | struct drm_fb_helper_surface_size *sizes) | 62 | struct drm_fb_helper_surface_size *sizes) |
| 63 | { | 63 | { |
| 64 | struct drm_device *dev = ifbdev->helper.dev; | 64 | struct drm_device *dev = ifbdev->helper.dev; |
| 65 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 65 | struct fb_info *info; | 66 | struct fb_info *info; |
| 66 | struct drm_framebuffer *fb; | 67 | struct drm_framebuffer *fb; |
| 67 | struct drm_mode_fb_cmd mode_cmd; | 68 | struct drm_mode_fb_cmd mode_cmd; |
| @@ -77,7 +78,7 @@ static int intelfb_create(struct intel_fbdev *ifbdev, | |||
| 77 | mode_cmd.height = sizes->surface_height; | 78 | mode_cmd.height = sizes->surface_height; |
| 78 | 79 | ||
| 79 | mode_cmd.bpp = sizes->surface_bpp; | 80 | mode_cmd.bpp = sizes->surface_bpp; |
| 80 | mode_cmd.pitch = ALIGN(mode_cmd.width * ((mode_cmd.bpp + 1) / 8), 64); | 81 | mode_cmd.pitch = ALIGN(mode_cmd.width * ((mode_cmd.bpp + 7) / 8), 64); |
| 81 | mode_cmd.depth = sizes->surface_depth; | 82 | mode_cmd.depth = sizes->surface_depth; |
| 82 | 83 | ||
| 83 | size = mode_cmd.pitch * mode_cmd.height; | 84 | size = mode_cmd.pitch * mode_cmd.height; |
| @@ -120,6 +121,11 @@ static int intelfb_create(struct intel_fbdev *ifbdev, | |||
| 120 | info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT; | 121 | info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT; |
| 121 | info->fbops = &intelfb_ops; | 122 | info->fbops = &intelfb_ops; |
| 122 | 123 | ||
| 124 | ret = fb_alloc_cmap(&info->cmap, 256, 0); | ||
| 125 | if (ret) { | ||
| 126 | ret = -ENOMEM; | ||
| 127 | goto out_unpin; | ||
| 128 | } | ||
| 123 | /* setup aperture base/size for vesafb takeover */ | 129 | /* setup aperture base/size for vesafb takeover */ |
| 124 | info->apertures = alloc_apertures(1); | 130 | info->apertures = alloc_apertures(1); |
| 125 | if (!info->apertures) { | 131 | if (!info->apertures) { |
| @@ -127,10 +133,8 @@ static int intelfb_create(struct intel_fbdev *ifbdev, | |||
| 127 | goto out_unpin; | 133 | goto out_unpin; |
| 128 | } | 134 | } |
| 129 | info->apertures->ranges[0].base = dev->mode_config.fb_base; | 135 | info->apertures->ranges[0].base = dev->mode_config.fb_base; |
| 130 | if (!IS_GEN2(dev)) | 136 | info->apertures->ranges[0].size = |
| 131 | info->apertures->ranges[0].size = pci_resource_len(dev->pdev, 2); | 137 | dev_priv->mm.gtt->gtt_mappable_entries << PAGE_SHIFT; |
| 132 | else | ||
| 133 | info->apertures->ranges[0].size = pci_resource_len(dev->pdev, 0); | ||
| 134 | 138 | ||
| 135 | info->fix.smem_start = dev->mode_config.fb_base + obj->gtt_offset; | 139 | info->fix.smem_start = dev->mode_config.fb_base + obj->gtt_offset; |
| 136 | info->fix.smem_len = size; | 140 | info->fix.smem_len = size; |
| @@ -140,12 +144,6 @@ static int intelfb_create(struct intel_fbdev *ifbdev, | |||
| 140 | ret = -ENOSPC; | 144 | ret = -ENOSPC; |
| 141 | goto out_unpin; | 145 | goto out_unpin; |
| 142 | } | 146 | } |
| 143 | |||
| 144 | ret = fb_alloc_cmap(&info->cmap, 256, 0); | ||
| 145 | if (ret) { | ||
| 146 | ret = -ENOMEM; | ||
| 147 | goto out_unpin; | ||
| 148 | } | ||
| 149 | info->screen_size = size; | 147 | info->screen_size = size; |
| 150 | 148 | ||
| 151 | // memset(info->screen_base, 0, size); | 149 | // memset(info->screen_base, 0, size); |
diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c index aa2307080be2..8f4f6bd33ee9 100644 --- a/drivers/gpu/drm/i915/intel_lvds.c +++ b/drivers/gpu/drm/i915/intel_lvds.c | |||
| @@ -106,7 +106,7 @@ static void intel_lvds_enable(struct intel_lvds *intel_lvds) | |||
| 106 | I915_WRITE(ctl_reg, I915_READ(ctl_reg) | POWER_TARGET_ON); | 106 | I915_WRITE(ctl_reg, I915_READ(ctl_reg) | POWER_TARGET_ON); |
| 107 | POSTING_READ(lvds_reg); | 107 | POSTING_READ(lvds_reg); |
| 108 | 108 | ||
| 109 | intel_panel_set_backlight(dev, dev_priv->backlight_level); | 109 | intel_panel_enable_backlight(dev); |
| 110 | } | 110 | } |
| 111 | 111 | ||
| 112 | static void intel_lvds_disable(struct intel_lvds *intel_lvds) | 112 | static void intel_lvds_disable(struct intel_lvds *intel_lvds) |
| @@ -123,8 +123,7 @@ static void intel_lvds_disable(struct intel_lvds *intel_lvds) | |||
| 123 | lvds_reg = LVDS; | 123 | lvds_reg = LVDS; |
| 124 | } | 124 | } |
| 125 | 125 | ||
| 126 | dev_priv->backlight_level = intel_panel_get_backlight(dev); | 126 | intel_panel_disable_backlight(dev); |
| 127 | intel_panel_set_backlight(dev, 0); | ||
| 128 | 127 | ||
| 129 | I915_WRITE(ctl_reg, I915_READ(ctl_reg) & ~POWER_TARGET_ON); | 128 | I915_WRITE(ctl_reg, I915_READ(ctl_reg) & ~POWER_TARGET_ON); |
| 130 | 129 | ||
| @@ -375,6 +374,10 @@ static bool intel_lvds_mode_fixup(struct drm_encoder *encoder, | |||
| 375 | } | 374 | } |
| 376 | 375 | ||
| 377 | out: | 376 | out: |
| 377 | if ((pfit_control & PFIT_ENABLE) == 0) { | ||
| 378 | pfit_control = 0; | ||
| 379 | pfit_pgm_ratios = 0; | ||
| 380 | } | ||
| 378 | if (pfit_control != intel_lvds->pfit_control || | 381 | if (pfit_control != intel_lvds->pfit_control || |
| 379 | pfit_pgm_ratios != intel_lvds->pfit_pgm_ratios) { | 382 | pfit_pgm_ratios != intel_lvds->pfit_pgm_ratios) { |
| 380 | intel_lvds->pfit_control = pfit_control; | 383 | intel_lvds->pfit_control = pfit_control; |
| @@ -398,8 +401,6 @@ static void intel_lvds_prepare(struct drm_encoder *encoder) | |||
| 398 | struct drm_i915_private *dev_priv = dev->dev_private; | 401 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 399 | struct intel_lvds *intel_lvds = to_intel_lvds(encoder); | 402 | struct intel_lvds *intel_lvds = to_intel_lvds(encoder); |
| 400 | 403 | ||
| 401 | dev_priv->backlight_level = intel_panel_get_backlight(dev); | ||
| 402 | |||
| 403 | /* We try to do the minimum that is necessary in order to unlock | 404 | /* We try to do the minimum that is necessary in order to unlock |
| 404 | * the registers for mode setting. | 405 | * the registers for mode setting. |
| 405 | * | 406 | * |
| @@ -430,9 +431,6 @@ static void intel_lvds_commit(struct drm_encoder *encoder) | |||
| 430 | struct drm_i915_private *dev_priv = dev->dev_private; | 431 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 431 | struct intel_lvds *intel_lvds = to_intel_lvds(encoder); | 432 | struct intel_lvds *intel_lvds = to_intel_lvds(encoder); |
| 432 | 433 | ||
| 433 | if (dev_priv->backlight_level == 0) | ||
| 434 | dev_priv->backlight_level = intel_panel_get_max_backlight(dev); | ||
| 435 | |||
| 436 | /* Undo any unlocking done in prepare to prevent accidental | 434 | /* Undo any unlocking done in prepare to prevent accidental |
| 437 | * adjustment of the registers. | 435 | * adjustment of the registers. |
| 438 | */ | 436 | */ |
diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c index 7350ec2515c6..e00d200df3db 100644 --- a/drivers/gpu/drm/i915/intel_panel.c +++ b/drivers/gpu/drm/i915/intel_panel.c | |||
| @@ -250,3 +250,34 @@ void intel_panel_set_backlight(struct drm_device *dev, u32 level) | |||
| 250 | tmp &= ~BACKLIGHT_DUTY_CYCLE_MASK; | 250 | tmp &= ~BACKLIGHT_DUTY_CYCLE_MASK; |
| 251 | I915_WRITE(BLC_PWM_CTL, tmp | level); | 251 | I915_WRITE(BLC_PWM_CTL, tmp | level); |
| 252 | } | 252 | } |
| 253 | |||
| 254 | void intel_panel_disable_backlight(struct drm_device *dev) | ||
| 255 | { | ||
| 256 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 257 | |||
| 258 | if (dev_priv->backlight_enabled) { | ||
| 259 | dev_priv->backlight_level = intel_panel_get_backlight(dev); | ||
| 260 | dev_priv->backlight_enabled = false; | ||
| 261 | } | ||
| 262 | |||
| 263 | intel_panel_set_backlight(dev, 0); | ||
| 264 | } | ||
| 265 | |||
| 266 | void intel_panel_enable_backlight(struct drm_device *dev) | ||
| 267 | { | ||
| 268 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 269 | |||
| 270 | if (dev_priv->backlight_level == 0) | ||
| 271 | dev_priv->backlight_level = intel_panel_get_max_backlight(dev); | ||
| 272 | |||
| 273 | intel_panel_set_backlight(dev, dev_priv->backlight_level); | ||
| 274 | dev_priv->backlight_enabled = true; | ||
| 275 | } | ||
| 276 | |||
| 277 | void intel_panel_setup_backlight(struct drm_device *dev) | ||
| 278 | { | ||
| 279 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 280 | |||
| 281 | dev_priv->backlight_level = intel_panel_get_max_backlight(dev); | ||
| 282 | dev_priv->backlight_enabled = dev_priv->backlight_level != 0; | ||
| 283 | } | ||
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c index 56bc95c056dd..03e337072517 100644 --- a/drivers/gpu/drm/i915/intel_ringbuffer.c +++ b/drivers/gpu/drm/i915/intel_ringbuffer.c | |||
| @@ -48,7 +48,7 @@ static u32 i915_gem_get_seqno(struct drm_device *dev) | |||
| 48 | return seqno; | 48 | return seqno; |
| 49 | } | 49 | } |
| 50 | 50 | ||
| 51 | static void | 51 | static int |
| 52 | render_ring_flush(struct intel_ring_buffer *ring, | 52 | render_ring_flush(struct intel_ring_buffer *ring, |
| 53 | u32 invalidate_domains, | 53 | u32 invalidate_domains, |
| 54 | u32 flush_domains) | 54 | u32 flush_domains) |
| @@ -56,6 +56,7 @@ render_ring_flush(struct intel_ring_buffer *ring, | |||
| 56 | struct drm_device *dev = ring->dev; | 56 | struct drm_device *dev = ring->dev; |
| 57 | drm_i915_private_t *dev_priv = dev->dev_private; | 57 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 58 | u32 cmd; | 58 | u32 cmd; |
| 59 | int ret; | ||
| 59 | 60 | ||
| 60 | #if WATCH_EXEC | 61 | #if WATCH_EXEC |
| 61 | DRM_INFO("%s: invalidate %08x flush %08x\n", __func__, | 62 | DRM_INFO("%s: invalidate %08x flush %08x\n", __func__, |
| @@ -116,12 +117,16 @@ render_ring_flush(struct intel_ring_buffer *ring, | |||
| 116 | #if WATCH_EXEC | 117 | #if WATCH_EXEC |
| 117 | DRM_INFO("%s: queue flush %08x to ring\n", __func__, cmd); | 118 | DRM_INFO("%s: queue flush %08x to ring\n", __func__, cmd); |
| 118 | #endif | 119 | #endif |
| 119 | if (intel_ring_begin(ring, 2) == 0) { | 120 | ret = intel_ring_begin(ring, 2); |
| 120 | intel_ring_emit(ring, cmd); | 121 | if (ret) |
| 121 | intel_ring_emit(ring, MI_NOOP); | 122 | return ret; |
| 122 | intel_ring_advance(ring); | 123 | |
| 123 | } | 124 | intel_ring_emit(ring, cmd); |
| 125 | intel_ring_emit(ring, MI_NOOP); | ||
| 126 | intel_ring_advance(ring); | ||
| 124 | } | 127 | } |
| 128 | |||
| 129 | return 0; | ||
| 125 | } | 130 | } |
| 126 | 131 | ||
| 127 | static void ring_write_tail(struct intel_ring_buffer *ring, | 132 | static void ring_write_tail(struct intel_ring_buffer *ring, |
| @@ -480,26 +485,56 @@ pc_render_get_seqno(struct intel_ring_buffer *ring) | |||
| 480 | return pc->cpu_page[0]; | 485 | return pc->cpu_page[0]; |
| 481 | } | 486 | } |
| 482 | 487 | ||
| 488 | static void | ||
| 489 | ironlake_enable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 490 | { | ||
| 491 | dev_priv->gt_irq_mask &= ~mask; | ||
| 492 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); | ||
| 493 | POSTING_READ(GTIMR); | ||
| 494 | } | ||
| 495 | |||
| 496 | static void | ||
| 497 | ironlake_disable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 498 | { | ||
| 499 | dev_priv->gt_irq_mask |= mask; | ||
| 500 | I915_WRITE(GTIMR, dev_priv->gt_irq_mask); | ||
| 501 | POSTING_READ(GTIMR); | ||
| 502 | } | ||
| 503 | |||
| 504 | static void | ||
| 505 | i915_enable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 506 | { | ||
| 507 | dev_priv->irq_mask &= ~mask; | ||
| 508 | I915_WRITE(IMR, dev_priv->irq_mask); | ||
| 509 | POSTING_READ(IMR); | ||
| 510 | } | ||
| 511 | |||
| 512 | static void | ||
| 513 | i915_disable_irq(drm_i915_private_t *dev_priv, u32 mask) | ||
| 514 | { | ||
| 515 | dev_priv->irq_mask |= mask; | ||
| 516 | I915_WRITE(IMR, dev_priv->irq_mask); | ||
| 517 | POSTING_READ(IMR); | ||
| 518 | } | ||
| 519 | |||
| 483 | static bool | 520 | static bool |
| 484 | render_ring_get_irq(struct intel_ring_buffer *ring) | 521 | render_ring_get_irq(struct intel_ring_buffer *ring) |
| 485 | { | 522 | { |
| 486 | struct drm_device *dev = ring->dev; | 523 | struct drm_device *dev = ring->dev; |
| 524 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 487 | 525 | ||
| 488 | if (!dev->irq_enabled) | 526 | if (!dev->irq_enabled) |
| 489 | return false; | 527 | return false; |
| 490 | 528 | ||
| 491 | if (atomic_inc_return(&ring->irq_refcount) == 1) { | 529 | spin_lock(&ring->irq_lock); |
| 492 | drm_i915_private_t *dev_priv = dev->dev_private; | 530 | if (ring->irq_refcount++ == 0) { |
| 493 | unsigned long irqflags; | ||
| 494 | |||
| 495 | spin_lock_irqsave(&dev_priv->irq_lock, irqflags); | ||
| 496 | if (HAS_PCH_SPLIT(dev)) | 531 | if (HAS_PCH_SPLIT(dev)) |
| 497 | ironlake_enable_graphics_irq(dev_priv, | 532 | ironlake_enable_irq(dev_priv, |
| 498 | GT_PIPE_NOTIFY | GT_USER_INTERRUPT); | 533 | GT_PIPE_NOTIFY | GT_USER_INTERRUPT); |
| 499 | else | 534 | else |
| 500 | i915_enable_irq(dev_priv, I915_USER_INTERRUPT); | 535 | i915_enable_irq(dev_priv, I915_USER_INTERRUPT); |
| 501 | spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); | ||
| 502 | } | 536 | } |
| 537 | spin_unlock(&ring->irq_lock); | ||
| 503 | 538 | ||
| 504 | return true; | 539 | return true; |
| 505 | } | 540 | } |
| @@ -508,20 +543,18 @@ static void | |||
| 508 | render_ring_put_irq(struct intel_ring_buffer *ring) | 543 | render_ring_put_irq(struct intel_ring_buffer *ring) |
| 509 | { | 544 | { |
| 510 | struct drm_device *dev = ring->dev; | 545 | struct drm_device *dev = ring->dev; |
| 546 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 511 | 547 | ||
| 512 | if (atomic_dec_and_test(&ring->irq_refcount)) { | 548 | spin_lock(&ring->irq_lock); |
| 513 | drm_i915_private_t *dev_priv = dev->dev_private; | 549 | if (--ring->irq_refcount == 0) { |
| 514 | unsigned long irqflags; | ||
| 515 | |||
| 516 | spin_lock_irqsave(&dev_priv->irq_lock, irqflags); | ||
| 517 | if (HAS_PCH_SPLIT(dev)) | 550 | if (HAS_PCH_SPLIT(dev)) |
| 518 | ironlake_disable_graphics_irq(dev_priv, | 551 | ironlake_disable_irq(dev_priv, |
| 519 | GT_USER_INTERRUPT | | 552 | GT_USER_INTERRUPT | |
| 520 | GT_PIPE_NOTIFY); | 553 | GT_PIPE_NOTIFY); |
| 521 | else | 554 | else |
| 522 | i915_disable_irq(dev_priv, I915_USER_INTERRUPT); | 555 | i915_disable_irq(dev_priv, I915_USER_INTERRUPT); |
| 523 | spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); | ||
| 524 | } | 556 | } |
| 557 | spin_unlock(&ring->irq_lock); | ||
| 525 | } | 558 | } |
| 526 | 559 | ||
| 527 | void intel_ring_setup_status_page(struct intel_ring_buffer *ring) | 560 | void intel_ring_setup_status_page(struct intel_ring_buffer *ring) |
| @@ -534,19 +567,24 @@ void intel_ring_setup_status_page(struct intel_ring_buffer *ring) | |||
| 534 | POSTING_READ(mmio); | 567 | POSTING_READ(mmio); |
| 535 | } | 568 | } |
| 536 | 569 | ||
| 537 | static void | 570 | static int |
| 538 | bsd_ring_flush(struct intel_ring_buffer *ring, | 571 | bsd_ring_flush(struct intel_ring_buffer *ring, |
| 539 | u32 invalidate_domains, | 572 | u32 invalidate_domains, |
| 540 | u32 flush_domains) | 573 | u32 flush_domains) |
| 541 | { | 574 | { |
| 575 | int ret; | ||
| 576 | |||
| 542 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) | 577 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) |
| 543 | return; | 578 | return 0; |
| 544 | 579 | ||
| 545 | if (intel_ring_begin(ring, 2) == 0) { | 580 | ret = intel_ring_begin(ring, 2); |
| 546 | intel_ring_emit(ring, MI_FLUSH); | 581 | if (ret) |
| 547 | intel_ring_emit(ring, MI_NOOP); | 582 | return ret; |
| 548 | intel_ring_advance(ring); | 583 | |
| 549 | } | 584 | intel_ring_emit(ring, MI_FLUSH); |
| 585 | intel_ring_emit(ring, MI_NOOP); | ||
| 586 | intel_ring_advance(ring); | ||
| 587 | return 0; | ||
| 550 | } | 588 | } |
| 551 | 589 | ||
| 552 | static int | 590 | static int |
| @@ -577,18 +615,15 @@ static bool | |||
| 577 | ring_get_irq(struct intel_ring_buffer *ring, u32 flag) | 615 | ring_get_irq(struct intel_ring_buffer *ring, u32 flag) |
| 578 | { | 616 | { |
| 579 | struct drm_device *dev = ring->dev; | 617 | struct drm_device *dev = ring->dev; |
| 618 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 580 | 619 | ||
| 581 | if (!dev->irq_enabled) | 620 | if (!dev->irq_enabled) |
| 582 | return false; | 621 | return false; |
| 583 | 622 | ||
| 584 | if (atomic_inc_return(&ring->irq_refcount) == 1) { | 623 | spin_lock(&ring->irq_lock); |
| 585 | drm_i915_private_t *dev_priv = dev->dev_private; | 624 | if (ring->irq_refcount++ == 0) |
| 586 | unsigned long irqflags; | 625 | ironlake_enable_irq(dev_priv, flag); |
| 587 | 626 | spin_unlock(&ring->irq_lock); | |
| 588 | spin_lock_irqsave(&dev_priv->irq_lock, irqflags); | ||
| 589 | ironlake_enable_graphics_irq(dev_priv, flag); | ||
| 590 | spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); | ||
| 591 | } | ||
| 592 | 627 | ||
| 593 | return true; | 628 | return true; |
| 594 | } | 629 | } |
| @@ -597,15 +632,47 @@ static void | |||
| 597 | ring_put_irq(struct intel_ring_buffer *ring, u32 flag) | 632 | ring_put_irq(struct intel_ring_buffer *ring, u32 flag) |
| 598 | { | 633 | { |
| 599 | struct drm_device *dev = ring->dev; | 634 | struct drm_device *dev = ring->dev; |
| 635 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 600 | 636 | ||
| 601 | if (atomic_dec_and_test(&ring->irq_refcount)) { | 637 | spin_lock(&ring->irq_lock); |
| 602 | drm_i915_private_t *dev_priv = dev->dev_private; | 638 | if (--ring->irq_refcount == 0) |
| 603 | unsigned long irqflags; | 639 | ironlake_disable_irq(dev_priv, flag); |
| 640 | spin_unlock(&ring->irq_lock); | ||
| 641 | } | ||
| 604 | 642 | ||
| 605 | spin_lock_irqsave(&dev_priv->irq_lock, irqflags); | 643 | static bool |
| 606 | ironlake_disable_graphics_irq(dev_priv, flag); | 644 | gen6_ring_get_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) |
| 607 | spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); | 645 | { |
| 646 | struct drm_device *dev = ring->dev; | ||
| 647 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 648 | |||
| 649 | if (!dev->irq_enabled) | ||
| 650 | return false; | ||
| 651 | |||
| 652 | spin_lock(&ring->irq_lock); | ||
| 653 | if (ring->irq_refcount++ == 0) { | ||
| 654 | ring->irq_mask &= ~rflag; | ||
| 655 | I915_WRITE_IMR(ring, ring->irq_mask); | ||
| 656 | ironlake_enable_irq(dev_priv, gflag); | ||
| 608 | } | 657 | } |
| 658 | spin_unlock(&ring->irq_lock); | ||
| 659 | |||
| 660 | return true; | ||
| 661 | } | ||
| 662 | |||
| 663 | static void | ||
| 664 | gen6_ring_put_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag) | ||
| 665 | { | ||
| 666 | struct drm_device *dev = ring->dev; | ||
| 667 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 668 | |||
| 669 | spin_lock(&ring->irq_lock); | ||
| 670 | if (--ring->irq_refcount == 0) { | ||
| 671 | ring->irq_mask |= rflag; | ||
| 672 | I915_WRITE_IMR(ring, ring->irq_mask); | ||
| 673 | ironlake_disable_irq(dev_priv, gflag); | ||
| 674 | } | ||
| 675 | spin_unlock(&ring->irq_lock); | ||
| 609 | } | 676 | } |
| 610 | 677 | ||
| 611 | static bool | 678 | static bool |
| @@ -748,6 +815,9 @@ int intel_init_ring_buffer(struct drm_device *dev, | |||
| 748 | INIT_LIST_HEAD(&ring->request_list); | 815 | INIT_LIST_HEAD(&ring->request_list); |
| 749 | INIT_LIST_HEAD(&ring->gpu_write_list); | 816 | INIT_LIST_HEAD(&ring->gpu_write_list); |
| 750 | 817 | ||
| 818 | spin_lock_init(&ring->irq_lock); | ||
| 819 | ring->irq_mask = ~0; | ||
| 820 | |||
| 751 | if (I915_NEED_GFX_HWS(dev)) { | 821 | if (I915_NEED_GFX_HWS(dev)) { |
| 752 | ret = init_status_page(ring); | 822 | ret = init_status_page(ring); |
| 753 | if (ret) | 823 | if (ret) |
| @@ -785,6 +855,14 @@ int intel_init_ring_buffer(struct drm_device *dev, | |||
| 785 | if (ret) | 855 | if (ret) |
| 786 | goto err_unmap; | 856 | goto err_unmap; |
| 787 | 857 | ||
| 858 | /* Workaround an erratum on the i830 which causes a hang if | ||
| 859 | * the TAIL pointer points to within the last 2 cachelines | ||
| 860 | * of the buffer. | ||
| 861 | */ | ||
| 862 | ring->effective_size = ring->size; | ||
| 863 | if (IS_I830(ring->dev)) | ||
| 864 | ring->effective_size -= 128; | ||
| 865 | |||
| 788 | return 0; | 866 | return 0; |
| 789 | 867 | ||
| 790 | err_unmap: | 868 | err_unmap: |
| @@ -827,8 +905,7 @@ void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) | |||
| 827 | static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring) | 905 | static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring) |
| 828 | { | 906 | { |
| 829 | unsigned int *virt; | 907 | unsigned int *virt; |
| 830 | int rem; | 908 | int rem = ring->size - ring->tail; |
| 831 | rem = ring->size - ring->tail; | ||
| 832 | 909 | ||
| 833 | if (ring->space < rem) { | 910 | if (ring->space < rem) { |
| 834 | int ret = intel_wait_ring_buffer(ring, rem); | 911 | int ret = intel_wait_ring_buffer(ring, rem); |
| @@ -895,7 +972,7 @@ int intel_ring_begin(struct intel_ring_buffer *ring, | |||
| 895 | int n = 4*num_dwords; | 972 | int n = 4*num_dwords; |
| 896 | int ret; | 973 | int ret; |
| 897 | 974 | ||
| 898 | if (unlikely(ring->tail + n > ring->size)) { | 975 | if (unlikely(ring->tail + n > ring->effective_size)) { |
| 899 | ret = intel_wrap_ring_buffer(ring); | 976 | ret = intel_wrap_ring_buffer(ring); |
| 900 | if (unlikely(ret)) | 977 | if (unlikely(ret)) |
| 901 | return ret; | 978 | return ret; |
| @@ -973,20 +1050,25 @@ static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, | |||
| 973 | GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_ENABLE); | 1050 | GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_ENABLE); |
| 974 | } | 1051 | } |
| 975 | 1052 | ||
| 976 | static void gen6_ring_flush(struct intel_ring_buffer *ring, | 1053 | static int gen6_ring_flush(struct intel_ring_buffer *ring, |
| 977 | u32 invalidate_domains, | 1054 | u32 invalidate_domains, |
| 978 | u32 flush_domains) | 1055 | u32 flush_domains) |
| 979 | { | 1056 | { |
| 1057 | int ret; | ||
| 1058 | |||
| 980 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) | 1059 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) |
| 981 | return; | 1060 | return 0; |
| 982 | 1061 | ||
| 983 | if (intel_ring_begin(ring, 4) == 0) { | 1062 | ret = intel_ring_begin(ring, 4); |
| 984 | intel_ring_emit(ring, MI_FLUSH_DW); | 1063 | if (ret) |
| 985 | intel_ring_emit(ring, 0); | 1064 | return ret; |
| 986 | intel_ring_emit(ring, 0); | 1065 | |
| 987 | intel_ring_emit(ring, 0); | 1066 | intel_ring_emit(ring, MI_FLUSH_DW); |
| 988 | intel_ring_advance(ring); | 1067 | intel_ring_emit(ring, 0); |
| 989 | } | 1068 | intel_ring_emit(ring, 0); |
| 1069 | intel_ring_emit(ring, 0); | ||
| 1070 | intel_ring_advance(ring); | ||
| 1071 | return 0; | ||
| 990 | } | 1072 | } |
| 991 | 1073 | ||
| 992 | static int | 1074 | static int |
| @@ -1008,15 +1090,35 @@ gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, | |||
| 1008 | } | 1090 | } |
| 1009 | 1091 | ||
| 1010 | static bool | 1092 | static bool |
| 1093 | gen6_render_ring_get_irq(struct intel_ring_buffer *ring) | ||
| 1094 | { | ||
| 1095 | return gen6_ring_get_irq(ring, | ||
| 1096 | GT_USER_INTERRUPT, | ||
| 1097 | GEN6_RENDER_USER_INTERRUPT); | ||
| 1098 | } | ||
| 1099 | |||
| 1100 | static void | ||
| 1101 | gen6_render_ring_put_irq(struct intel_ring_buffer *ring) | ||
| 1102 | { | ||
| 1103 | return gen6_ring_put_irq(ring, | ||
| 1104 | GT_USER_INTERRUPT, | ||
| 1105 | GEN6_RENDER_USER_INTERRUPT); | ||
| 1106 | } | ||
| 1107 | |||
| 1108 | static bool | ||
| 1011 | gen6_bsd_ring_get_irq(struct intel_ring_buffer *ring) | 1109 | gen6_bsd_ring_get_irq(struct intel_ring_buffer *ring) |
| 1012 | { | 1110 | { |
| 1013 | return ring_get_irq(ring, GT_GEN6_BSD_USER_INTERRUPT); | 1111 | return gen6_ring_get_irq(ring, |
| 1112 | GT_GEN6_BSD_USER_INTERRUPT, | ||
| 1113 | GEN6_BSD_USER_INTERRUPT); | ||
| 1014 | } | 1114 | } |
| 1015 | 1115 | ||
| 1016 | static void | 1116 | static void |
| 1017 | gen6_bsd_ring_put_irq(struct intel_ring_buffer *ring) | 1117 | gen6_bsd_ring_put_irq(struct intel_ring_buffer *ring) |
| 1018 | { | 1118 | { |
| 1019 | ring_put_irq(ring, GT_GEN6_BSD_USER_INTERRUPT); | 1119 | return gen6_ring_put_irq(ring, |
| 1120 | GT_GEN6_BSD_USER_INTERRUPT, | ||
| 1121 | GEN6_BSD_USER_INTERRUPT); | ||
| 1020 | } | 1122 | } |
| 1021 | 1123 | ||
| 1022 | /* ring buffer for Video Codec for Gen6+ */ | 1124 | /* ring buffer for Video Codec for Gen6+ */ |
| @@ -1040,13 +1142,17 @@ static const struct intel_ring_buffer gen6_bsd_ring = { | |||
| 1040 | static bool | 1142 | static bool |
| 1041 | blt_ring_get_irq(struct intel_ring_buffer *ring) | 1143 | blt_ring_get_irq(struct intel_ring_buffer *ring) |
| 1042 | { | 1144 | { |
| 1043 | return ring_get_irq(ring, GT_BLT_USER_INTERRUPT); | 1145 | return gen6_ring_get_irq(ring, |
| 1146 | GT_BLT_USER_INTERRUPT, | ||
| 1147 | GEN6_BLITTER_USER_INTERRUPT); | ||
| 1044 | } | 1148 | } |
| 1045 | 1149 | ||
| 1046 | static void | 1150 | static void |
| 1047 | blt_ring_put_irq(struct intel_ring_buffer *ring) | 1151 | blt_ring_put_irq(struct intel_ring_buffer *ring) |
| 1048 | { | 1152 | { |
| 1049 | ring_put_irq(ring, GT_BLT_USER_INTERRUPT); | 1153 | gen6_ring_put_irq(ring, |
| 1154 | GT_BLT_USER_INTERRUPT, | ||
| 1155 | GEN6_BLITTER_USER_INTERRUPT); | ||
| 1050 | } | 1156 | } |
| 1051 | 1157 | ||
| 1052 | 1158 | ||
| @@ -1115,20 +1221,25 @@ static int blt_ring_begin(struct intel_ring_buffer *ring, | |||
| 1115 | return intel_ring_begin(ring, 4); | 1221 | return intel_ring_begin(ring, 4); |
| 1116 | } | 1222 | } |
| 1117 | 1223 | ||
| 1118 | static void blt_ring_flush(struct intel_ring_buffer *ring, | 1224 | static int blt_ring_flush(struct intel_ring_buffer *ring, |
| 1119 | u32 invalidate_domains, | 1225 | u32 invalidate_domains, |
| 1120 | u32 flush_domains) | 1226 | u32 flush_domains) |
| 1121 | { | 1227 | { |
| 1228 | int ret; | ||
| 1229 | |||
| 1122 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) | 1230 | if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0) |
| 1123 | return; | 1231 | return 0; |
| 1124 | 1232 | ||
| 1125 | if (blt_ring_begin(ring, 4) == 0) { | 1233 | ret = blt_ring_begin(ring, 4); |
| 1126 | intel_ring_emit(ring, MI_FLUSH_DW); | 1234 | if (ret) |
| 1127 | intel_ring_emit(ring, 0); | 1235 | return ret; |
| 1128 | intel_ring_emit(ring, 0); | 1236 | |
| 1129 | intel_ring_emit(ring, 0); | 1237 | intel_ring_emit(ring, MI_FLUSH_DW); |
| 1130 | intel_ring_advance(ring); | 1238 | intel_ring_emit(ring, 0); |
| 1131 | } | 1239 | intel_ring_emit(ring, 0); |
| 1240 | intel_ring_emit(ring, 0); | ||
| 1241 | intel_ring_advance(ring); | ||
| 1242 | return 0; | ||
| 1132 | } | 1243 | } |
| 1133 | 1244 | ||
| 1134 | static void blt_ring_cleanup(struct intel_ring_buffer *ring) | 1245 | static void blt_ring_cleanup(struct intel_ring_buffer *ring) |
| @@ -1165,6 +1276,8 @@ int intel_init_render_ring_buffer(struct drm_device *dev) | |||
| 1165 | *ring = render_ring; | 1276 | *ring = render_ring; |
| 1166 | if (INTEL_INFO(dev)->gen >= 6) { | 1277 | if (INTEL_INFO(dev)->gen >= 6) { |
| 1167 | ring->add_request = gen6_add_request; | 1278 | ring->add_request = gen6_add_request; |
| 1279 | ring->irq_get = gen6_render_ring_get_irq; | ||
| 1280 | ring->irq_put = gen6_render_ring_put_irq; | ||
| 1168 | } else if (IS_GEN5(dev)) { | 1281 | } else if (IS_GEN5(dev)) { |
| 1169 | ring->add_request = pc_render_add_request; | 1282 | ring->add_request = pc_render_add_request; |
| 1170 | ring->get_seqno = pc_render_get_seqno; | 1283 | ring->get_seqno = pc_render_get_seqno; |
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h index 8e2e357ad6ee..be9087e4c9be 100644 --- a/drivers/gpu/drm/i915/intel_ringbuffer.h +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h | |||
| @@ -16,21 +16,24 @@ struct intel_hw_status_page { | |||
| 16 | 16 | ||
| 17 | #define I915_RING_READ(reg) i915_safe_read(dev_priv, reg) | 17 | #define I915_RING_READ(reg) i915_safe_read(dev_priv, reg) |
| 18 | 18 | ||
| 19 | #define I915_READ_TAIL(ring) I915_RING_READ(RING_TAIL(ring->mmio_base)) | 19 | #define I915_READ_TAIL(ring) I915_RING_READ(RING_TAIL((ring)->mmio_base)) |
| 20 | #define I915_WRITE_TAIL(ring, val) I915_WRITE(RING_TAIL(ring->mmio_base), val) | 20 | #define I915_WRITE_TAIL(ring, val) I915_WRITE(RING_TAIL((ring)->mmio_base), val) |
| 21 | 21 | ||
| 22 | #define I915_READ_START(ring) I915_RING_READ(RING_START(ring->mmio_base)) | 22 | #define I915_READ_START(ring) I915_RING_READ(RING_START((ring)->mmio_base)) |
| 23 | #define I915_WRITE_START(ring, val) I915_WRITE(RING_START(ring->mmio_base), val) | 23 | #define I915_WRITE_START(ring, val) I915_WRITE(RING_START((ring)->mmio_base), val) |
| 24 | 24 | ||
| 25 | #define I915_READ_HEAD(ring) I915_RING_READ(RING_HEAD(ring->mmio_base)) | 25 | #define I915_READ_HEAD(ring) I915_RING_READ(RING_HEAD((ring)->mmio_base)) |
| 26 | #define I915_WRITE_HEAD(ring, val) I915_WRITE(RING_HEAD(ring->mmio_base), val) | 26 | #define I915_WRITE_HEAD(ring, val) I915_WRITE(RING_HEAD((ring)->mmio_base), val) |
| 27 | 27 | ||
| 28 | #define I915_READ_CTL(ring) I915_RING_READ(RING_CTL(ring->mmio_base)) | 28 | #define I915_READ_CTL(ring) I915_RING_READ(RING_CTL((ring)->mmio_base)) |
| 29 | #define I915_WRITE_CTL(ring, val) I915_WRITE(RING_CTL(ring->mmio_base), val) | 29 | #define I915_WRITE_CTL(ring, val) I915_WRITE(RING_CTL((ring)->mmio_base), val) |
| 30 | 30 | ||
| 31 | #define I915_READ_NOPID(ring) I915_RING_READ(RING_NOPID(ring->mmio_base)) | 31 | #define I915_WRITE_IMR(ring, val) I915_WRITE(RING_IMR((ring)->mmio_base), val) |
| 32 | #define I915_READ_SYNC_0(ring) I915_RING_READ(RING_SYNC_0(ring->mmio_base)) | 32 | #define I915_READ_IMR(ring) I915_RING_READ(RING_IMR((ring)->mmio_base)) |
| 33 | #define I915_READ_SYNC_1(ring) I915_RING_READ(RING_SYNC_1(ring->mmio_base)) | 33 | |
| 34 | #define I915_READ_NOPID(ring) I915_RING_READ(RING_NOPID((ring)->mmio_base)) | ||
| 35 | #define I915_READ_SYNC_0(ring) I915_RING_READ(RING_SYNC_0((ring)->mmio_base)) | ||
| 36 | #define I915_READ_SYNC_1(ring) I915_RING_READ(RING_SYNC_1((ring)->mmio_base)) | ||
| 34 | 37 | ||
| 35 | struct intel_ring_buffer { | 38 | struct intel_ring_buffer { |
| 36 | const char *name; | 39 | const char *name; |
| @@ -49,12 +52,15 @@ struct intel_ring_buffer { | |||
| 49 | u32 tail; | 52 | u32 tail; |
| 50 | int space; | 53 | int space; |
| 51 | int size; | 54 | int size; |
| 55 | int effective_size; | ||
| 52 | struct intel_hw_status_page status_page; | 56 | struct intel_hw_status_page status_page; |
| 53 | 57 | ||
| 58 | spinlock_t irq_lock; | ||
| 59 | u32 irq_refcount; | ||
| 60 | u32 irq_mask; | ||
| 54 | u32 irq_seqno; /* last seq seem at irq time */ | 61 | u32 irq_seqno; /* last seq seem at irq time */ |
| 55 | u32 waiting_seqno; | 62 | u32 waiting_seqno; |
| 56 | u32 sync_seqno[I915_NUM_RINGS-1]; | 63 | u32 sync_seqno[I915_NUM_RINGS-1]; |
| 57 | atomic_t irq_refcount; | ||
| 58 | bool __must_check (*irq_get)(struct intel_ring_buffer *ring); | 64 | bool __must_check (*irq_get)(struct intel_ring_buffer *ring); |
| 59 | void (*irq_put)(struct intel_ring_buffer *ring); | 65 | void (*irq_put)(struct intel_ring_buffer *ring); |
| 60 | 66 | ||
| @@ -62,9 +68,9 @@ struct intel_ring_buffer { | |||
| 62 | 68 | ||
| 63 | void (*write_tail)(struct intel_ring_buffer *ring, | 69 | void (*write_tail)(struct intel_ring_buffer *ring, |
| 64 | u32 value); | 70 | u32 value); |
| 65 | void (*flush)(struct intel_ring_buffer *ring, | 71 | int __must_check (*flush)(struct intel_ring_buffer *ring, |
| 66 | u32 invalidate_domains, | 72 | u32 invalidate_domains, |
| 67 | u32 flush_domains); | 73 | u32 flush_domains); |
| 68 | int (*add_request)(struct intel_ring_buffer *ring, | 74 | int (*add_request)(struct intel_ring_buffer *ring, |
| 69 | u32 *seqno); | 75 | u32 *seqno); |
| 70 | u32 (*get_seqno)(struct intel_ring_buffer *ring); | 76 | u32 (*get_seqno)(struct intel_ring_buffer *ring); |
diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c index 9d0af36a13ec..45cd37652a37 100644 --- a/drivers/gpu/drm/i915/intel_sdvo.c +++ b/drivers/gpu/drm/i915/intel_sdvo.c | |||
| @@ -1024,9 +1024,13 @@ static void intel_sdvo_mode_set(struct drm_encoder *encoder, | |||
| 1024 | if (!intel_sdvo_set_target_input(intel_sdvo)) | 1024 | if (!intel_sdvo_set_target_input(intel_sdvo)) |
| 1025 | return; | 1025 | return; |
| 1026 | 1026 | ||
| 1027 | if (intel_sdvo->has_hdmi_monitor && | 1027 | if (intel_sdvo->has_hdmi_monitor) { |
| 1028 | !intel_sdvo_set_avi_infoframe(intel_sdvo)) | 1028 | intel_sdvo_set_encode(intel_sdvo, SDVO_ENCODE_HDMI); |
| 1029 | return; | 1029 | intel_sdvo_set_colorimetry(intel_sdvo, |
| 1030 | SDVO_COLORIMETRY_RGB256); | ||
| 1031 | intel_sdvo_set_avi_infoframe(intel_sdvo); | ||
| 1032 | } else | ||
| 1033 | intel_sdvo_set_encode(intel_sdvo, SDVO_ENCODE_DVI); | ||
| 1030 | 1034 | ||
| 1031 | if (intel_sdvo->is_tv && | 1035 | if (intel_sdvo->is_tv && |
| 1032 | !intel_sdvo_set_tv_format(intel_sdvo)) | 1036 | !intel_sdvo_set_tv_format(intel_sdvo)) |
| @@ -1398,6 +1402,9 @@ intel_sdvo_detect(struct drm_connector *connector, bool force) | |||
| 1398 | 1402 | ||
| 1399 | intel_sdvo->attached_output = response; | 1403 | intel_sdvo->attached_output = response; |
| 1400 | 1404 | ||
| 1405 | intel_sdvo->has_hdmi_monitor = false; | ||
| 1406 | intel_sdvo->has_hdmi_audio = false; | ||
| 1407 | |||
| 1401 | if ((intel_sdvo_connector->output_flag & response) == 0) | 1408 | if ((intel_sdvo_connector->output_flag & response) == 0) |
| 1402 | ret = connector_status_disconnected; | 1409 | ret = connector_status_disconnected; |
| 1403 | else if (response & SDVO_TMDS_MASK) | 1410 | else if (response & SDVO_TMDS_MASK) |
| @@ -1922,20 +1929,7 @@ intel_sdvo_select_i2c_bus(struct drm_i915_private *dev_priv, | |||
| 1922 | static bool | 1929 | static bool |
| 1923 | intel_sdvo_is_hdmi_connector(struct intel_sdvo *intel_sdvo, int device) | 1930 | intel_sdvo_is_hdmi_connector(struct intel_sdvo *intel_sdvo, int device) |
| 1924 | { | 1931 | { |
| 1925 | int is_hdmi; | 1932 | return intel_sdvo_check_supp_encode(intel_sdvo); |
| 1926 | |||
| 1927 | if (!intel_sdvo_check_supp_encode(intel_sdvo)) | ||
| 1928 | return false; | ||
| 1929 | |||
| 1930 | if (!intel_sdvo_set_target_output(intel_sdvo, | ||
| 1931 | device == 0 ? SDVO_OUTPUT_TMDS0 : SDVO_OUTPUT_TMDS1)) | ||
| 1932 | return false; | ||
| 1933 | |||
| 1934 | is_hdmi = 0; | ||
| 1935 | if (!intel_sdvo_get_value(intel_sdvo, SDVO_CMD_GET_ENCODE, &is_hdmi, 1)) | ||
| 1936 | return false; | ||
| 1937 | |||
| 1938 | return !!is_hdmi; | ||
| 1939 | } | 1933 | } |
| 1940 | 1934 | ||
| 1941 | static u8 | 1935 | static u8 |
| @@ -2037,12 +2031,7 @@ intel_sdvo_dvi_init(struct intel_sdvo *intel_sdvo, int device) | |||
| 2037 | connector->connector_type = DRM_MODE_CONNECTOR_DVID; | 2031 | connector->connector_type = DRM_MODE_CONNECTOR_DVID; |
| 2038 | 2032 | ||
| 2039 | if (intel_sdvo_is_hdmi_connector(intel_sdvo, device)) { | 2033 | if (intel_sdvo_is_hdmi_connector(intel_sdvo, device)) { |
| 2040 | /* enable hdmi encoding mode if supported */ | ||
| 2041 | intel_sdvo_set_encode(intel_sdvo, SDVO_ENCODE_HDMI); | ||
| 2042 | intel_sdvo_set_colorimetry(intel_sdvo, | ||
| 2043 | SDVO_COLORIMETRY_RGB256); | ||
| 2044 | connector->connector_type = DRM_MODE_CONNECTOR_HDMIA; | 2034 | connector->connector_type = DRM_MODE_CONNECTOR_HDMIA; |
| 2045 | |||
| 2046 | intel_sdvo->is_hdmi = true; | 2035 | intel_sdvo->is_hdmi = true; |
| 2047 | } | 2036 | } |
| 2048 | intel_sdvo->base.clone_mask = ((1 << INTEL_SDVO_NON_TV_CLONE_BIT) | | 2037 | intel_sdvo->base.clone_mask = ((1 << INTEL_SDVO_NON_TV_CLONE_BIT) | |
