diff options
Diffstat (limited to 'drivers/gpu/drm/i915')
-rw-r--r-- | drivers/gpu/drm/i915/i915_debugfs.c | 2 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 16 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_irq.c | 37 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_display.c | 56 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_dp.c | 24 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_i2c.c | 21 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_lvds.c | 8 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_pm.c | 64 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_sdvo.c | 12 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_sdvo_regs.h | 5 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/intel_tv.c | 53 |
11 files changed, 197 insertions, 101 deletions
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index eb2b3c25b9e1..5363e9c66c27 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c | |||
@@ -2032,6 +2032,8 @@ void i915_debugfs_cleanup(struct drm_minor *minor) | |||
2032 | 1, minor); | 2032 | 1, minor); |
2033 | drm_debugfs_remove_files((struct drm_info_list *) &i915_ring_stop_fops, | 2033 | drm_debugfs_remove_files((struct drm_info_list *) &i915_ring_stop_fops, |
2034 | 1, minor); | 2034 | 1, minor); |
2035 | drm_debugfs_remove_files((struct drm_info_list *) &i915_error_state_fops, | ||
2036 | 1, minor); | ||
2035 | } | 2037 | } |
2036 | 2038 | ||
2037 | #endif /* CONFIG_DEBUG_FS */ | 2039 | #endif /* CONFIG_DEBUG_FS */ |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index c1e5c66553df..288d7b8f49ae 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
@@ -2063,10 +2063,8 @@ i915_gem_object_unbind(struct drm_i915_gem_object *obj) | |||
2063 | if (obj->gtt_space == NULL) | 2063 | if (obj->gtt_space == NULL) |
2064 | return 0; | 2064 | return 0; |
2065 | 2065 | ||
2066 | if (obj->pin_count != 0) { | 2066 | if (obj->pin_count) |
2067 | DRM_ERROR("Attempting to unbind pinned buffer\n"); | 2067 | return -EBUSY; |
2068 | return -EINVAL; | ||
2069 | } | ||
2070 | 2068 | ||
2071 | ret = i915_gem_object_finish_gpu(obj); | 2069 | ret = i915_gem_object_finish_gpu(obj); |
2072 | if (ret) | 2070 | if (ret) |
@@ -3293,6 +3291,7 @@ struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, | |||
3293 | struct drm_i915_private *dev_priv = dev->dev_private; | 3291 | struct drm_i915_private *dev_priv = dev->dev_private; |
3294 | struct drm_i915_gem_object *obj; | 3292 | struct drm_i915_gem_object *obj; |
3295 | struct address_space *mapping; | 3293 | struct address_space *mapping; |
3294 | u32 mask; | ||
3296 | 3295 | ||
3297 | obj = kzalloc(sizeof(*obj), GFP_KERNEL); | 3296 | obj = kzalloc(sizeof(*obj), GFP_KERNEL); |
3298 | if (obj == NULL) | 3297 | if (obj == NULL) |
@@ -3303,8 +3302,15 @@ struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, | |||
3303 | return NULL; | 3302 | return NULL; |
3304 | } | 3303 | } |
3305 | 3304 | ||
3305 | mask = GFP_HIGHUSER | __GFP_RECLAIMABLE; | ||
3306 | if (IS_CRESTLINE(dev) || IS_BROADWATER(dev)) { | ||
3307 | /* 965gm cannot relocate objects above 4GiB. */ | ||
3308 | mask &= ~__GFP_HIGHMEM; | ||
3309 | mask |= __GFP_DMA32; | ||
3310 | } | ||
3311 | |||
3306 | mapping = obj->base.filp->f_path.dentry->d_inode->i_mapping; | 3312 | mapping = obj->base.filp->f_path.dentry->d_inode->i_mapping; |
3307 | mapping_set_gfp_mask(mapping, GFP_HIGHUSER | __GFP_RECLAIMABLE); | 3313 | mapping_set_gfp_mask(mapping, mask); |
3308 | 3314 | ||
3309 | i915_gem_info_add_obj(dev_priv, size); | 3315 | i915_gem_info_add_obj(dev_priv, size); |
3310 | 3316 | ||
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index cc4a63307611..1417660a93ec 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
@@ -350,8 +350,8 @@ static void gen6_pm_rps_work(struct work_struct *work) | |||
350 | { | 350 | { |
351 | drm_i915_private_t *dev_priv = container_of(work, drm_i915_private_t, | 351 | drm_i915_private_t *dev_priv = container_of(work, drm_i915_private_t, |
352 | rps_work); | 352 | rps_work); |
353 | u8 new_delay = dev_priv->cur_delay; | ||
354 | u32 pm_iir, pm_imr; | 353 | u32 pm_iir, pm_imr; |
354 | u8 new_delay; | ||
355 | 355 | ||
356 | spin_lock_irq(&dev_priv->rps_lock); | 356 | spin_lock_irq(&dev_priv->rps_lock); |
357 | pm_iir = dev_priv->pm_iir; | 357 | pm_iir = dev_priv->pm_iir; |
@@ -360,41 +360,18 @@ static void gen6_pm_rps_work(struct work_struct *work) | |||
360 | I915_WRITE(GEN6_PMIMR, 0); | 360 | I915_WRITE(GEN6_PMIMR, 0); |
361 | spin_unlock_irq(&dev_priv->rps_lock); | 361 | spin_unlock_irq(&dev_priv->rps_lock); |
362 | 362 | ||
363 | if (!pm_iir) | 363 | if ((pm_iir & GEN6_PM_DEFERRED_EVENTS) == 0) |
364 | return; | 364 | return; |
365 | 365 | ||
366 | mutex_lock(&dev_priv->dev->struct_mutex); | 366 | mutex_lock(&dev_priv->dev->struct_mutex); |
367 | if (pm_iir & GEN6_PM_RP_UP_THRESHOLD) { | 367 | |
368 | if (dev_priv->cur_delay != dev_priv->max_delay) | 368 | if (pm_iir & GEN6_PM_RP_UP_THRESHOLD) |
369 | new_delay = dev_priv->cur_delay + 1; | 369 | new_delay = dev_priv->cur_delay + 1; |
370 | if (new_delay > dev_priv->max_delay) | 370 | else |
371 | new_delay = dev_priv->max_delay; | 371 | new_delay = dev_priv->cur_delay - 1; |
372 | } else if (pm_iir & (GEN6_PM_RP_DOWN_THRESHOLD | GEN6_PM_RP_DOWN_TIMEOUT)) { | ||
373 | gen6_gt_force_wake_get(dev_priv); | ||
374 | if (dev_priv->cur_delay != dev_priv->min_delay) | ||
375 | new_delay = dev_priv->cur_delay - 1; | ||
376 | if (new_delay < dev_priv->min_delay) { | ||
377 | new_delay = dev_priv->min_delay; | ||
378 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, | ||
379 | I915_READ(GEN6_RP_INTERRUPT_LIMITS) | | ||
380 | ((new_delay << 16) & 0x3f0000)); | ||
381 | } else { | ||
382 | /* Make sure we continue to get down interrupts | ||
383 | * until we hit the minimum frequency */ | ||
384 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, | ||
385 | I915_READ(GEN6_RP_INTERRUPT_LIMITS) & ~0x3f0000); | ||
386 | } | ||
387 | gen6_gt_force_wake_put(dev_priv); | ||
388 | } | ||
389 | 372 | ||
390 | gen6_set_rps(dev_priv->dev, new_delay); | 373 | gen6_set_rps(dev_priv->dev, new_delay); |
391 | dev_priv->cur_delay = new_delay; | ||
392 | 374 | ||
393 | /* | ||
394 | * rps_lock not held here because clearing is non-destructive. There is | ||
395 | * an *extremely* unlikely race with gen6_rps_enable() that is prevented | ||
396 | * by holding struct_mutex for the duration of the write. | ||
397 | */ | ||
398 | mutex_unlock(&dev_priv->dev->struct_mutex); | 375 | mutex_unlock(&dev_priv->dev->struct_mutex); |
399 | } | 376 | } |
400 | 377 | ||
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index ee61ad1e642b..914789420906 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
@@ -910,9 +910,10 @@ static void assert_pll(struct drm_i915_private *dev_priv, | |||
910 | 910 | ||
911 | /* For ILK+ */ | 911 | /* For ILK+ */ |
912 | static void assert_pch_pll(struct drm_i915_private *dev_priv, | 912 | static void assert_pch_pll(struct drm_i915_private *dev_priv, |
913 | struct intel_crtc *intel_crtc, bool state) | 913 | struct intel_pch_pll *pll, |
914 | struct intel_crtc *crtc, | ||
915 | bool state) | ||
914 | { | 916 | { |
915 | int reg; | ||
916 | u32 val; | 917 | u32 val; |
917 | bool cur_state; | 918 | bool cur_state; |
918 | 919 | ||
@@ -921,30 +922,37 @@ static void assert_pch_pll(struct drm_i915_private *dev_priv, | |||
921 | return; | 922 | return; |
922 | } | 923 | } |
923 | 924 | ||
924 | if (!intel_crtc->pch_pll) { | 925 | if (WARN (!pll, |
925 | WARN(1, "asserting PCH PLL enabled with no PLL\n"); | 926 | "asserting PCH PLL %s with no PLL\n", state_string(state))) |
926 | return; | 927 | return; |
927 | } | ||
928 | 928 | ||
929 | if (HAS_PCH_CPT(dev_priv->dev)) { | 929 | val = I915_READ(pll->pll_reg); |
930 | cur_state = !!(val & DPLL_VCO_ENABLE); | ||
931 | WARN(cur_state != state, | ||
932 | "PCH PLL state for reg %x assertion failure (expected %s, current %s), val=%08x\n", | ||
933 | pll->pll_reg, state_string(state), state_string(cur_state), val); | ||
934 | |||
935 | /* Make sure the selected PLL is correctly attached to the transcoder */ | ||
936 | if (crtc && HAS_PCH_CPT(dev_priv->dev)) { | ||
930 | u32 pch_dpll; | 937 | u32 pch_dpll; |
931 | 938 | ||
932 | pch_dpll = I915_READ(PCH_DPLL_SEL); | 939 | pch_dpll = I915_READ(PCH_DPLL_SEL); |
933 | 940 | cur_state = pll->pll_reg == _PCH_DPLL_B; | |
934 | /* Make sure the selected PLL is enabled to the transcoder */ | 941 | if (!WARN(((pch_dpll >> (4 * crtc->pipe)) & 1) != cur_state, |
935 | WARN(!((pch_dpll >> (4 * intel_crtc->pipe)) & 8), | 942 | "PLL[%d] not attached to this transcoder %d: %08x\n", |
936 | "transcoder %d PLL not enabled\n", intel_crtc->pipe); | 943 | cur_state, crtc->pipe, pch_dpll)) { |
944 | cur_state = !!(val >> (4*crtc->pipe + 3)); | ||
945 | WARN(cur_state != state, | ||
946 | "PLL[%d] not %s on this transcoder %d: %08x\n", | ||
947 | pll->pll_reg == _PCH_DPLL_B, | ||
948 | state_string(state), | ||
949 | crtc->pipe, | ||
950 | val); | ||
951 | } | ||
937 | } | 952 | } |
938 | |||
939 | reg = intel_crtc->pch_pll->pll_reg; | ||
940 | val = I915_READ(reg); | ||
941 | cur_state = !!(val & DPLL_VCO_ENABLE); | ||
942 | WARN(cur_state != state, | ||
943 | "PCH PLL state assertion failure (expected %s, current %s)\n", | ||
944 | state_string(state), state_string(cur_state)); | ||
945 | } | 953 | } |
946 | #define assert_pch_pll_enabled(d, p) assert_pch_pll(d, p, true) | 954 | #define assert_pch_pll_enabled(d, p, c) assert_pch_pll(d, p, c, true) |
947 | #define assert_pch_pll_disabled(d, p) assert_pch_pll(d, p, false) | 955 | #define assert_pch_pll_disabled(d, p, c) assert_pch_pll(d, p, c, false) |
948 | 956 | ||
949 | static void assert_fdi_tx(struct drm_i915_private *dev_priv, | 957 | static void assert_fdi_tx(struct drm_i915_private *dev_priv, |
950 | enum pipe pipe, bool state) | 958 | enum pipe pipe, bool state) |
@@ -1424,7 +1432,7 @@ static void intel_enable_pch_pll(struct intel_crtc *intel_crtc) | |||
1424 | assert_pch_refclk_enabled(dev_priv); | 1432 | assert_pch_refclk_enabled(dev_priv); |
1425 | 1433 | ||
1426 | if (pll->active++ && pll->on) { | 1434 | if (pll->active++ && pll->on) { |
1427 | assert_pch_pll_enabled(dev_priv, intel_crtc); | 1435 | assert_pch_pll_enabled(dev_priv, pll, NULL); |
1428 | return; | 1436 | return; |
1429 | } | 1437 | } |
1430 | 1438 | ||
@@ -1460,12 +1468,12 @@ static void intel_disable_pch_pll(struct intel_crtc *intel_crtc) | |||
1460 | intel_crtc->base.base.id); | 1468 | intel_crtc->base.base.id); |
1461 | 1469 | ||
1462 | if (WARN_ON(pll->active == 0)) { | 1470 | if (WARN_ON(pll->active == 0)) { |
1463 | assert_pch_pll_disabled(dev_priv, intel_crtc); | 1471 | assert_pch_pll_disabled(dev_priv, pll, NULL); |
1464 | return; | 1472 | return; |
1465 | } | 1473 | } |
1466 | 1474 | ||
1467 | if (--pll->active) { | 1475 | if (--pll->active) { |
1468 | assert_pch_pll_enabled(dev_priv, intel_crtc); | 1476 | assert_pch_pll_enabled(dev_priv, pll, NULL); |
1469 | return; | 1477 | return; |
1470 | } | 1478 | } |
1471 | 1479 | ||
@@ -1495,7 +1503,9 @@ static void intel_enable_transcoder(struct drm_i915_private *dev_priv, | |||
1495 | BUG_ON(dev_priv->info->gen < 5); | 1503 | BUG_ON(dev_priv->info->gen < 5); |
1496 | 1504 | ||
1497 | /* Make sure PCH DPLL is enabled */ | 1505 | /* Make sure PCH DPLL is enabled */ |
1498 | assert_pch_pll_enabled(dev_priv, to_intel_crtc(crtc)); | 1506 | assert_pch_pll_enabled(dev_priv, |
1507 | to_intel_crtc(crtc)->pch_pll, | ||
1508 | to_intel_crtc(crtc)); | ||
1499 | 1509 | ||
1500 | /* FDI must be feeding us bits for PCH ports */ | 1510 | /* FDI must be feeding us bits for PCH ports */ |
1501 | assert_fdi_tx_enabled(dev_priv, pipe); | 1511 | assert_fdi_tx_enabled(dev_priv, pipe); |
diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c index 71c7096e3869..296cfc201a81 100644 --- a/drivers/gpu/drm/i915/intel_dp.c +++ b/drivers/gpu/drm/i915/intel_dp.c | |||
@@ -266,6 +266,9 @@ intel_dp_mode_valid(struct drm_connector *connector, | |||
266 | if (mode->clock < 10000) | 266 | if (mode->clock < 10000) |
267 | return MODE_CLOCK_LOW; | 267 | return MODE_CLOCK_LOW; |
268 | 268 | ||
269 | if (mode->flags & DRM_MODE_FLAG_DBLCLK) | ||
270 | return MODE_H_ILLEGAL; | ||
271 | |||
269 | return MODE_OK; | 272 | return MODE_OK; |
270 | } | 273 | } |
271 | 274 | ||
@@ -702,6 +705,9 @@ intel_dp_mode_fixup(struct drm_encoder *encoder, struct drm_display_mode *mode, | |||
702 | mode->clock = intel_dp->panel_fixed_mode->clock; | 705 | mode->clock = intel_dp->panel_fixed_mode->clock; |
703 | } | 706 | } |
704 | 707 | ||
708 | if (mode->flags & DRM_MODE_FLAG_DBLCLK) | ||
709 | return false; | ||
710 | |||
705 | DRM_DEBUG_KMS("DP link computation with max lane count %i " | 711 | DRM_DEBUG_KMS("DP link computation with max lane count %i " |
706 | "max bw %02x pixel clock %iKHz\n", | 712 | "max bw %02x pixel clock %iKHz\n", |
707 | max_lane_count, bws[max_clock], mode->clock); | 713 | max_lane_count, bws[max_clock], mode->clock); |
@@ -1154,11 +1160,10 @@ static void ironlake_edp_panel_off(struct intel_dp *intel_dp) | |||
1154 | 1160 | ||
1155 | DRM_DEBUG_KMS("Turn eDP power off\n"); | 1161 | DRM_DEBUG_KMS("Turn eDP power off\n"); |
1156 | 1162 | ||
1157 | WARN(intel_dp->want_panel_vdd, "Cannot turn power off while VDD is on\n"); | 1163 | WARN(!intel_dp->want_panel_vdd, "Need VDD to turn off panel\n"); |
1158 | ironlake_panel_vdd_off_sync(intel_dp); /* finish any pending work */ | ||
1159 | 1164 | ||
1160 | pp = ironlake_get_pp_control(dev_priv); | 1165 | pp = ironlake_get_pp_control(dev_priv); |
1161 | pp &= ~(POWER_TARGET_ON | EDP_FORCE_VDD | PANEL_POWER_RESET | EDP_BLC_ENABLE); | 1166 | pp &= ~(POWER_TARGET_ON | PANEL_POWER_RESET | EDP_BLC_ENABLE); |
1162 | I915_WRITE(PCH_PP_CONTROL, pp); | 1167 | I915_WRITE(PCH_PP_CONTROL, pp); |
1163 | POSTING_READ(PCH_PP_CONTROL); | 1168 | POSTING_READ(PCH_PP_CONTROL); |
1164 | 1169 | ||
@@ -1266,18 +1271,16 @@ static void intel_dp_prepare(struct drm_encoder *encoder) | |||
1266 | { | 1271 | { |
1267 | struct intel_dp *intel_dp = enc_to_intel_dp(encoder); | 1272 | struct intel_dp *intel_dp = enc_to_intel_dp(encoder); |
1268 | 1273 | ||
1274 | |||
1275 | /* Make sure the panel is off before trying to change the mode. But also | ||
1276 | * ensure that we have vdd while we switch off the panel. */ | ||
1277 | ironlake_edp_panel_vdd_on(intel_dp); | ||
1269 | ironlake_edp_backlight_off(intel_dp); | 1278 | ironlake_edp_backlight_off(intel_dp); |
1270 | ironlake_edp_panel_off(intel_dp); | 1279 | ironlake_edp_panel_off(intel_dp); |
1271 | 1280 | ||
1272 | /* Wake up the sink first */ | ||
1273 | ironlake_edp_panel_vdd_on(intel_dp); | ||
1274 | intel_dp_sink_dpms(intel_dp, DRM_MODE_DPMS_ON); | 1281 | intel_dp_sink_dpms(intel_dp, DRM_MODE_DPMS_ON); |
1275 | intel_dp_link_down(intel_dp); | 1282 | intel_dp_link_down(intel_dp); |
1276 | ironlake_edp_panel_vdd_off(intel_dp, false); | 1283 | ironlake_edp_panel_vdd_off(intel_dp, false); |
1277 | |||
1278 | /* Make sure the panel is off before trying to | ||
1279 | * change the mode | ||
1280 | */ | ||
1281 | } | 1284 | } |
1282 | 1285 | ||
1283 | static void intel_dp_commit(struct drm_encoder *encoder) | 1286 | static void intel_dp_commit(struct drm_encoder *encoder) |
@@ -1309,10 +1312,11 @@ intel_dp_dpms(struct drm_encoder *encoder, int mode) | |||
1309 | uint32_t dp_reg = I915_READ(intel_dp->output_reg); | 1312 | uint32_t dp_reg = I915_READ(intel_dp->output_reg); |
1310 | 1313 | ||
1311 | if (mode != DRM_MODE_DPMS_ON) { | 1314 | if (mode != DRM_MODE_DPMS_ON) { |
1315 | /* Switching the panel off requires vdd. */ | ||
1316 | ironlake_edp_panel_vdd_on(intel_dp); | ||
1312 | ironlake_edp_backlight_off(intel_dp); | 1317 | ironlake_edp_backlight_off(intel_dp); |
1313 | ironlake_edp_panel_off(intel_dp); | 1318 | ironlake_edp_panel_off(intel_dp); |
1314 | 1319 | ||
1315 | ironlake_edp_panel_vdd_on(intel_dp); | ||
1316 | intel_dp_sink_dpms(intel_dp, mode); | 1320 | intel_dp_sink_dpms(intel_dp, mode); |
1317 | intel_dp_link_down(intel_dp); | 1321 | intel_dp_link_down(intel_dp); |
1318 | ironlake_edp_panel_vdd_off(intel_dp, false); | 1322 | ironlake_edp_panel_vdd_off(intel_dp, false); |
diff --git a/drivers/gpu/drm/i915/intel_i2c.c b/drivers/gpu/drm/i915/intel_i2c.c index 4a9707dd0f9c..1991a4408cf9 100644 --- a/drivers/gpu/drm/i915/intel_i2c.c +++ b/drivers/gpu/drm/i915/intel_i2c.c | |||
@@ -396,11 +396,22 @@ clear_err: | |||
396 | * Wait for bus to IDLE before clearing NAK. | 396 | * Wait for bus to IDLE before clearing NAK. |
397 | * If we clear the NAK while bus is still active, then it will stay | 397 | * If we clear the NAK while bus is still active, then it will stay |
398 | * active and the next transaction may fail. | 398 | * active and the next transaction may fail. |
399 | * | ||
400 | * If no ACK is received during the address phase of a transaction, the | ||
401 | * adapter must report -ENXIO. It is not clear what to return if no ACK | ||
402 | * is received at other times. But we have to be careful to not return | ||
403 | * spurious -ENXIO because that will prevent i2c and drm edid functions | ||
404 | * from retrying. So return -ENXIO only when gmbus properly quiescents - | ||
405 | * timing out seems to happen when there _is_ a ddc chip present, but | ||
406 | * it's slow responding and only answers on the 2nd retry. | ||
399 | */ | 407 | */ |
408 | ret = -ENXIO; | ||
400 | if (wait_for((I915_READ(GMBUS2 + reg_offset) & GMBUS_ACTIVE) == 0, | 409 | if (wait_for((I915_READ(GMBUS2 + reg_offset) & GMBUS_ACTIVE) == 0, |
401 | 10)) | 410 | 10)) { |
402 | DRM_DEBUG_KMS("GMBUS [%s] timed out after NAK\n", | 411 | DRM_DEBUG_KMS("GMBUS [%s] timed out after NAK\n", |
403 | adapter->name); | 412 | adapter->name); |
413 | ret = -ETIMEDOUT; | ||
414 | } | ||
404 | 415 | ||
405 | /* Toggle the Software Clear Interrupt bit. This has the effect | 416 | /* Toggle the Software Clear Interrupt bit. This has the effect |
406 | * of resetting the GMBUS controller and so clearing the | 417 | * of resetting the GMBUS controller and so clearing the |
@@ -414,14 +425,6 @@ clear_err: | |||
414 | adapter->name, msgs[i].addr, | 425 | adapter->name, msgs[i].addr, |
415 | (msgs[i].flags & I2C_M_RD) ? 'r' : 'w', msgs[i].len); | 426 | (msgs[i].flags & I2C_M_RD) ? 'r' : 'w', msgs[i].len); |
416 | 427 | ||
417 | /* | ||
418 | * If no ACK is received during the address phase of a transaction, | ||
419 | * the adapter must report -ENXIO. | ||
420 | * It is not clear what to return if no ACK is received at other times. | ||
421 | * So, we always return -ENXIO in all NAK cases, to ensure we send | ||
422 | * it at least during the one case that is specified. | ||
423 | */ | ||
424 | ret = -ENXIO; | ||
425 | goto out; | 428 | goto out; |
426 | 429 | ||
427 | timeout: | 430 | timeout: |
diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c index 9dee82350def..08eb04c787e8 100644 --- a/drivers/gpu/drm/i915/intel_lvds.c +++ b/drivers/gpu/drm/i915/intel_lvds.c | |||
@@ -747,6 +747,14 @@ static const struct dmi_system_id intel_no_lvds[] = { | |||
747 | }, | 747 | }, |
748 | { | 748 | { |
749 | .callback = intel_no_lvds_dmi_callback, | 749 | .callback = intel_no_lvds_dmi_callback, |
750 | .ident = "Hewlett-Packard HP t5740e Thin Client", | ||
751 | .matches = { | ||
752 | DMI_MATCH(DMI_BOARD_VENDOR, "Hewlett-Packard"), | ||
753 | DMI_MATCH(DMI_PRODUCT_NAME, "HP t5740e Thin Client"), | ||
754 | }, | ||
755 | }, | ||
756 | { | ||
757 | .callback = intel_no_lvds_dmi_callback, | ||
750 | .ident = "Hewlett-Packard t5745", | 758 | .ident = "Hewlett-Packard t5745", |
751 | .matches = { | 759 | .matches = { |
752 | DMI_MATCH(DMI_BOARD_VENDOR, "Hewlett-Packard"), | 760 | DMI_MATCH(DMI_BOARD_VENDOR, "Hewlett-Packard"), |
diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c index 8e79ff67ec98..d0ce2a5b1d3f 100644 --- a/drivers/gpu/drm/i915/intel_pm.c +++ b/drivers/gpu/drm/i915/intel_pm.c | |||
@@ -2270,10 +2270,33 @@ void ironlake_disable_drps(struct drm_device *dev) | |||
2270 | void gen6_set_rps(struct drm_device *dev, u8 val) | 2270 | void gen6_set_rps(struct drm_device *dev, u8 val) |
2271 | { | 2271 | { |
2272 | struct drm_i915_private *dev_priv = dev->dev_private; | 2272 | struct drm_i915_private *dev_priv = dev->dev_private; |
2273 | u32 swreq; | 2273 | u32 limits; |
2274 | 2274 | ||
2275 | swreq = (val & 0x3ff) << 25; | 2275 | limits = 0; |
2276 | I915_WRITE(GEN6_RPNSWREQ, swreq); | 2276 | if (val >= dev_priv->max_delay) |
2277 | val = dev_priv->max_delay; | ||
2278 | else | ||
2279 | limits |= dev_priv->max_delay << 24; | ||
2280 | |||
2281 | if (val <= dev_priv->min_delay) | ||
2282 | val = dev_priv->min_delay; | ||
2283 | else | ||
2284 | limits |= dev_priv->min_delay << 16; | ||
2285 | |||
2286 | if (val == dev_priv->cur_delay) | ||
2287 | return; | ||
2288 | |||
2289 | I915_WRITE(GEN6_RPNSWREQ, | ||
2290 | GEN6_FREQUENCY(val) | | ||
2291 | GEN6_OFFSET(0) | | ||
2292 | GEN6_AGGRESSIVE_TURBO); | ||
2293 | |||
2294 | /* Make sure we continue to get interrupts | ||
2295 | * until we hit the minimum or maximum frequencies. | ||
2296 | */ | ||
2297 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, limits); | ||
2298 | |||
2299 | dev_priv->cur_delay = val; | ||
2277 | } | 2300 | } |
2278 | 2301 | ||
2279 | void gen6_disable_rps(struct drm_device *dev) | 2302 | void gen6_disable_rps(struct drm_device *dev) |
@@ -2327,11 +2350,10 @@ int intel_enable_rc6(const struct drm_device *dev) | |||
2327 | void gen6_enable_rps(struct drm_i915_private *dev_priv) | 2350 | void gen6_enable_rps(struct drm_i915_private *dev_priv) |
2328 | { | 2351 | { |
2329 | struct intel_ring_buffer *ring; | 2352 | struct intel_ring_buffer *ring; |
2330 | u32 rp_state_cap = I915_READ(GEN6_RP_STATE_CAP); | 2353 | u32 rp_state_cap; |
2331 | u32 gt_perf_status = I915_READ(GEN6_GT_PERF_STATUS); | 2354 | u32 gt_perf_status; |
2332 | u32 pcu_mbox, rc6_mask = 0; | 2355 | u32 pcu_mbox, rc6_mask = 0; |
2333 | u32 gtfifodbg; | 2356 | u32 gtfifodbg; |
2334 | int cur_freq, min_freq, max_freq; | ||
2335 | int rc6_mode; | 2357 | int rc6_mode; |
2336 | int i; | 2358 | int i; |
2337 | 2359 | ||
@@ -2352,6 +2374,14 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
2352 | 2374 | ||
2353 | gen6_gt_force_wake_get(dev_priv); | 2375 | gen6_gt_force_wake_get(dev_priv); |
2354 | 2376 | ||
2377 | rp_state_cap = I915_READ(GEN6_RP_STATE_CAP); | ||
2378 | gt_perf_status = I915_READ(GEN6_GT_PERF_STATUS); | ||
2379 | |||
2380 | /* In units of 100MHz */ | ||
2381 | dev_priv->max_delay = rp_state_cap & 0xff; | ||
2382 | dev_priv->min_delay = (rp_state_cap & 0xff0000) >> 16; | ||
2383 | dev_priv->cur_delay = 0; | ||
2384 | |||
2355 | /* disable the counters and set deterministic thresholds */ | 2385 | /* disable the counters and set deterministic thresholds */ |
2356 | I915_WRITE(GEN6_RC_CONTROL, 0); | 2386 | I915_WRITE(GEN6_RC_CONTROL, 0); |
2357 | 2387 | ||
@@ -2399,8 +2429,8 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
2399 | 2429 | ||
2400 | I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000); | 2430 | I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000); |
2401 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, | 2431 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, |
2402 | 18 << 24 | | 2432 | dev_priv->max_delay << 24 | |
2403 | 6 << 16); | 2433 | dev_priv->min_delay << 16); |
2404 | I915_WRITE(GEN6_RP_UP_THRESHOLD, 10000); | 2434 | I915_WRITE(GEN6_RP_UP_THRESHOLD, 10000); |
2405 | I915_WRITE(GEN6_RP_DOWN_THRESHOLD, 1000000); | 2435 | I915_WRITE(GEN6_RP_DOWN_THRESHOLD, 1000000); |
2406 | I915_WRITE(GEN6_RP_UP_EI, 100000); | 2436 | I915_WRITE(GEN6_RP_UP_EI, 100000); |
@@ -2408,7 +2438,7 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
2408 | I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 10); | 2438 | I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 10); |
2409 | I915_WRITE(GEN6_RP_CONTROL, | 2439 | I915_WRITE(GEN6_RP_CONTROL, |
2410 | GEN6_RP_MEDIA_TURBO | | 2440 | GEN6_RP_MEDIA_TURBO | |
2411 | GEN6_RP_MEDIA_HW_MODE | | 2441 | GEN6_RP_MEDIA_HW_NORMAL_MODE | |
2412 | GEN6_RP_MEDIA_IS_GFX | | 2442 | GEN6_RP_MEDIA_IS_GFX | |
2413 | GEN6_RP_ENABLE | | 2443 | GEN6_RP_ENABLE | |
2414 | GEN6_RP_UP_BUSY_AVG | | 2444 | GEN6_RP_UP_BUSY_AVG | |
@@ -2426,10 +2456,6 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
2426 | 500)) | 2456 | 500)) |
2427 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); | 2457 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); |
2428 | 2458 | ||
2429 | min_freq = (rp_state_cap & 0xff0000) >> 16; | ||
2430 | max_freq = rp_state_cap & 0xff; | ||
2431 | cur_freq = (gt_perf_status & 0xff00) >> 8; | ||
2432 | |||
2433 | /* Check for overclock support */ | 2459 | /* Check for overclock support */ |
2434 | if (wait_for((I915_READ(GEN6_PCODE_MAILBOX) & GEN6_PCODE_READY) == 0, | 2460 | if (wait_for((I915_READ(GEN6_PCODE_MAILBOX) & GEN6_PCODE_READY) == 0, |
2435 | 500)) | 2461 | 500)) |
@@ -2440,14 +2466,11 @@ void gen6_enable_rps(struct drm_i915_private *dev_priv) | |||
2440 | 500)) | 2466 | 500)) |
2441 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); | 2467 | DRM_ERROR("timeout waiting for pcode mailbox to finish\n"); |
2442 | if (pcu_mbox & (1<<31)) { /* OC supported */ | 2468 | if (pcu_mbox & (1<<31)) { /* OC supported */ |
2443 | max_freq = pcu_mbox & 0xff; | 2469 | dev_priv->max_delay = pcu_mbox & 0xff; |
2444 | DRM_DEBUG_DRIVER("overclocking supported, adjusting frequency max to %dMHz\n", pcu_mbox * 50); | 2470 | DRM_DEBUG_DRIVER("overclocking supported, adjusting frequency max to %dMHz\n", pcu_mbox * 50); |
2445 | } | 2471 | } |
2446 | 2472 | ||
2447 | /* In units of 100MHz */ | 2473 | gen6_set_rps(dev_priv->dev, (gt_perf_status & 0xff00) >> 8); |
2448 | dev_priv->max_delay = max_freq; | ||
2449 | dev_priv->min_delay = min_freq; | ||
2450 | dev_priv->cur_delay = cur_freq; | ||
2451 | 2474 | ||
2452 | /* requires MSI enabled */ | 2475 | /* requires MSI enabled */ |
2453 | I915_WRITE(GEN6_PMIER, | 2476 | I915_WRITE(GEN6_PMIER, |
@@ -3580,8 +3603,9 @@ static void gen6_sanitize_pm(struct drm_device *dev) | |||
3580 | limits |= (dev_priv->min_delay & 0x3f) << 16; | 3603 | limits |= (dev_priv->min_delay & 0x3f) << 16; |
3581 | 3604 | ||
3582 | if (old != limits) { | 3605 | if (old != limits) { |
3583 | DRM_ERROR("Power management discrepancy: GEN6_RP_INTERRUPT_LIMITS expected %08x, was %08x\n", | 3606 | /* Note that the known failure case is to read back 0. */ |
3584 | limits, old); | 3607 | DRM_DEBUG_DRIVER("Power management discrepancy: GEN6_RP_INTERRUPT_LIMITS " |
3608 | "expected %08x, was %08x\n", limits, old); | ||
3585 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, limits); | 3609 | I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, limits); |
3586 | } | 3610 | } |
3587 | 3611 | ||
diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c index a949b73880c8..b6a9d45fc3c6 100644 --- a/drivers/gpu/drm/i915/intel_sdvo.c +++ b/drivers/gpu/drm/i915/intel_sdvo.c | |||
@@ -783,10 +783,12 @@ static void intel_sdvo_get_dtd_from_mode(struct intel_sdvo_dtd *dtd, | |||
783 | ((v_sync_len & 0x30) >> 4); | 783 | ((v_sync_len & 0x30) >> 4); |
784 | 784 | ||
785 | dtd->part2.dtd_flags = 0x18; | 785 | dtd->part2.dtd_flags = 0x18; |
786 | if (mode->flags & DRM_MODE_FLAG_INTERLACE) | ||
787 | dtd->part2.dtd_flags |= DTD_FLAG_INTERLACE; | ||
786 | if (mode->flags & DRM_MODE_FLAG_PHSYNC) | 788 | if (mode->flags & DRM_MODE_FLAG_PHSYNC) |
787 | dtd->part2.dtd_flags |= 0x2; | 789 | dtd->part2.dtd_flags |= DTD_FLAG_HSYNC_POSITIVE; |
788 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) | 790 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) |
789 | dtd->part2.dtd_flags |= 0x4; | 791 | dtd->part2.dtd_flags |= DTD_FLAG_VSYNC_POSITIVE; |
790 | 792 | ||
791 | dtd->part2.sdvo_flags = 0; | 793 | dtd->part2.sdvo_flags = 0; |
792 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; | 794 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; |
@@ -820,9 +822,11 @@ static void intel_sdvo_get_mode_from_dtd(struct drm_display_mode * mode, | |||
820 | mode->clock = dtd->part1.clock * 10; | 822 | mode->clock = dtd->part1.clock * 10; |
821 | 823 | ||
822 | mode->flags &= ~(DRM_MODE_FLAG_PHSYNC | DRM_MODE_FLAG_PVSYNC); | 824 | mode->flags &= ~(DRM_MODE_FLAG_PHSYNC | DRM_MODE_FLAG_PVSYNC); |
823 | if (dtd->part2.dtd_flags & 0x2) | 825 | if (dtd->part2.dtd_flags & DTD_FLAG_INTERLACE) |
826 | mode->flags |= DRM_MODE_FLAG_INTERLACE; | ||
827 | if (dtd->part2.dtd_flags & DTD_FLAG_HSYNC_POSITIVE) | ||
824 | mode->flags |= DRM_MODE_FLAG_PHSYNC; | 828 | mode->flags |= DRM_MODE_FLAG_PHSYNC; |
825 | if (dtd->part2.dtd_flags & 0x4) | 829 | if (dtd->part2.dtd_flags & DTD_FLAG_VSYNC_POSITIVE) |
826 | mode->flags |= DRM_MODE_FLAG_PVSYNC; | 830 | mode->flags |= DRM_MODE_FLAG_PVSYNC; |
827 | } | 831 | } |
828 | 832 | ||
diff --git a/drivers/gpu/drm/i915/intel_sdvo_regs.h b/drivers/gpu/drm/i915/intel_sdvo_regs.h index 6b7b22f4d63e..9d030142ee43 100644 --- a/drivers/gpu/drm/i915/intel_sdvo_regs.h +++ b/drivers/gpu/drm/i915/intel_sdvo_regs.h | |||
@@ -61,6 +61,11 @@ struct intel_sdvo_caps { | |||
61 | u16 output_flags; | 61 | u16 output_flags; |
62 | } __attribute__((packed)); | 62 | } __attribute__((packed)); |
63 | 63 | ||
64 | /* Note: SDVO detailed timing flags match EDID misc flags. */ | ||
65 | #define DTD_FLAG_HSYNC_POSITIVE (1 << 1) | ||
66 | #define DTD_FLAG_VSYNC_POSITIVE (1 << 2) | ||
67 | #define DTD_FLAG_INTERLACE (1 << 7) | ||
68 | |||
64 | /** This matches the EDID DTD structure, more or less */ | 69 | /** This matches the EDID DTD structure, more or less */ |
65 | struct intel_sdvo_dtd { | 70 | struct intel_sdvo_dtd { |
66 | struct { | 71 | struct { |
diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c index 3346612d2953..a233a51fd7e6 100644 --- a/drivers/gpu/drm/i915/intel_tv.c +++ b/drivers/gpu/drm/i915/intel_tv.c | |||
@@ -674,6 +674,54 @@ static const struct tv_mode tv_modes[] = { | |||
674 | .filter_table = filter_table, | 674 | .filter_table = filter_table, |
675 | }, | 675 | }, |
676 | { | 676 | { |
677 | .name = "480p", | ||
678 | .clock = 107520, | ||
679 | .refresh = 59940, | ||
680 | .oversample = TV_OVERSAMPLE_4X, | ||
681 | .component_only = 1, | ||
682 | |||
683 | .hsync_end = 64, .hblank_end = 122, | ||
684 | .hblank_start = 842, .htotal = 857, | ||
685 | |||
686 | .progressive = true, .trilevel_sync = false, | ||
687 | |||
688 | .vsync_start_f1 = 12, .vsync_start_f2 = 12, | ||
689 | .vsync_len = 12, | ||
690 | |||
691 | .veq_ena = false, | ||
692 | |||
693 | .vi_end_f1 = 44, .vi_end_f2 = 44, | ||
694 | .nbr_end = 479, | ||
695 | |||
696 | .burst_ena = false, | ||
697 | |||
698 | .filter_table = filter_table, | ||
699 | }, | ||
700 | { | ||
701 | .name = "576p", | ||
702 | .clock = 107520, | ||
703 | .refresh = 50000, | ||
704 | .oversample = TV_OVERSAMPLE_4X, | ||
705 | .component_only = 1, | ||
706 | |||
707 | .hsync_end = 64, .hblank_end = 139, | ||
708 | .hblank_start = 859, .htotal = 863, | ||
709 | |||
710 | .progressive = true, .trilevel_sync = false, | ||
711 | |||
712 | .vsync_start_f1 = 10, .vsync_start_f2 = 10, | ||
713 | .vsync_len = 10, | ||
714 | |||
715 | .veq_ena = false, | ||
716 | |||
717 | .vi_end_f1 = 48, .vi_end_f2 = 48, | ||
718 | .nbr_end = 575, | ||
719 | |||
720 | .burst_ena = false, | ||
721 | |||
722 | .filter_table = filter_table, | ||
723 | }, | ||
724 | { | ||
677 | .name = "720p@60Hz", | 725 | .name = "720p@60Hz", |
678 | .clock = 148800, | 726 | .clock = 148800, |
679 | .refresh = 60000, | 727 | .refresh = 60000, |
@@ -1194,6 +1242,11 @@ intel_tv_detect_type(struct intel_tv *intel_tv, | |||
1194 | 1242 | ||
1195 | I915_WRITE(TV_DAC, save_tv_dac & ~TVDAC_STATE_CHG_EN); | 1243 | I915_WRITE(TV_DAC, save_tv_dac & ~TVDAC_STATE_CHG_EN); |
1196 | I915_WRITE(TV_CTL, save_tv_ctl); | 1244 | I915_WRITE(TV_CTL, save_tv_ctl); |
1245 | POSTING_READ(TV_CTL); | ||
1246 | |||
1247 | /* For unknown reasons the hw barfs if we don't do this vblank wait. */ | ||
1248 | intel_wait_for_vblank(intel_tv->base.base.dev, | ||
1249 | to_intel_crtc(intel_tv->base.base.crtc)->pipe); | ||
1197 | 1250 | ||
1198 | /* Restore interrupt config */ | 1251 | /* Restore interrupt config */ |
1199 | if (connector->polled & DRM_CONNECTOR_POLL_HPD) { | 1252 | if (connector->polled & DRM_CONNECTOR_POLL_HPD) { |