diff options
Diffstat (limited to 'drivers/gpu')
96 files changed, 1434 insertions, 607 deletions
diff --git a/drivers/gpu/drm/ast/ast_drv.h b/drivers/gpu/drm/ast/ast_drv.h index 796dbb212a41..8492b68e873c 100644 --- a/drivers/gpu/drm/ast/ast_drv.h +++ b/drivers/gpu/drm/ast/ast_drv.h | |||
@@ -177,7 +177,7 @@ uint8_t ast_get_index_reg_mask(struct ast_private *ast, | |||
177 | 177 | ||
178 | static inline void ast_open_key(struct ast_private *ast) | 178 | static inline void ast_open_key(struct ast_private *ast) |
179 | { | 179 | { |
180 | ast_set_index_reg_mask(ast, AST_IO_CRTC_PORT, 0xA1, 0xFF, 0x04); | 180 | ast_set_index_reg(ast, AST_IO_CRTC_PORT, 0x80, 0xA8); |
181 | } | 181 | } |
182 | 182 | ||
183 | #define AST_VIDMEM_SIZE_8M 0x00800000 | 183 | #define AST_VIDMEM_SIZE_8M 0x00800000 |
diff --git a/drivers/gpu/drm/drm_context.c b/drivers/gpu/drm/drm_context.c index b4fb86d89850..224ff965bcf7 100644 --- a/drivers/gpu/drm/drm_context.c +++ b/drivers/gpu/drm/drm_context.c | |||
@@ -42,6 +42,10 @@ | |||
42 | 42 | ||
43 | #include <drm/drmP.h> | 43 | #include <drm/drmP.h> |
44 | 44 | ||
45 | /******************************************************************/ | ||
46 | /** \name Context bitmap support */ | ||
47 | /*@{*/ | ||
48 | |||
45 | /** | 49 | /** |
46 | * Free a handle from the context bitmap. | 50 | * Free a handle from the context bitmap. |
47 | * | 51 | * |
@@ -52,48 +56,13 @@ | |||
52 | * in drm_device::ctx_idr, while holding the drm_device::struct_mutex | 56 | * in drm_device::ctx_idr, while holding the drm_device::struct_mutex |
53 | * lock. | 57 | * lock. |
54 | */ | 58 | */ |
55 | static void drm_ctxbitmap_free(struct drm_device * dev, int ctx_handle) | 59 | void drm_ctxbitmap_free(struct drm_device * dev, int ctx_handle) |
56 | { | 60 | { |
57 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
58 | return; | ||
59 | |||
60 | mutex_lock(&dev->struct_mutex); | 61 | mutex_lock(&dev->struct_mutex); |
61 | idr_remove(&dev->ctx_idr, ctx_handle); | 62 | idr_remove(&dev->ctx_idr, ctx_handle); |
62 | mutex_unlock(&dev->struct_mutex); | 63 | mutex_unlock(&dev->struct_mutex); |
63 | } | 64 | } |
64 | 65 | ||
65 | /******************************************************************/ | ||
66 | /** \name Context bitmap support */ | ||
67 | /*@{*/ | ||
68 | |||
69 | void drm_legacy_ctxbitmap_release(struct drm_device *dev, | ||
70 | struct drm_file *file_priv) | ||
71 | { | ||
72 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
73 | return; | ||
74 | |||
75 | mutex_lock(&dev->ctxlist_mutex); | ||
76 | if (!list_empty(&dev->ctxlist)) { | ||
77 | struct drm_ctx_list *pos, *n; | ||
78 | |||
79 | list_for_each_entry_safe(pos, n, &dev->ctxlist, head) { | ||
80 | if (pos->tag == file_priv && | ||
81 | pos->handle != DRM_KERNEL_CONTEXT) { | ||
82 | if (dev->driver->context_dtor) | ||
83 | dev->driver->context_dtor(dev, | ||
84 | pos->handle); | ||
85 | |||
86 | drm_ctxbitmap_free(dev, pos->handle); | ||
87 | |||
88 | list_del(&pos->head); | ||
89 | kfree(pos); | ||
90 | --dev->ctx_count; | ||
91 | } | ||
92 | } | ||
93 | } | ||
94 | mutex_unlock(&dev->ctxlist_mutex); | ||
95 | } | ||
96 | |||
97 | /** | 66 | /** |
98 | * Context bitmap allocation. | 67 | * Context bitmap allocation. |
99 | * | 68 | * |
@@ -121,12 +90,10 @@ static int drm_ctxbitmap_next(struct drm_device * dev) | |||
121 | * | 90 | * |
122 | * Initialise the drm_device::ctx_idr | 91 | * Initialise the drm_device::ctx_idr |
123 | */ | 92 | */ |
124 | void drm_legacy_ctxbitmap_init(struct drm_device * dev) | 93 | int drm_ctxbitmap_init(struct drm_device * dev) |
125 | { | 94 | { |
126 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
127 | return; | ||
128 | |||
129 | idr_init(&dev->ctx_idr); | 95 | idr_init(&dev->ctx_idr); |
96 | return 0; | ||
130 | } | 97 | } |
131 | 98 | ||
132 | /** | 99 | /** |
@@ -137,7 +104,7 @@ void drm_legacy_ctxbitmap_init(struct drm_device * dev) | |||
137 | * Free all idr members using drm_ctx_sarea_free helper function | 104 | * Free all idr members using drm_ctx_sarea_free helper function |
138 | * while holding the drm_device::struct_mutex lock. | 105 | * while holding the drm_device::struct_mutex lock. |
139 | */ | 106 | */ |
140 | void drm_legacy_ctxbitmap_cleanup(struct drm_device * dev) | 107 | void drm_ctxbitmap_cleanup(struct drm_device * dev) |
141 | { | 108 | { |
142 | mutex_lock(&dev->struct_mutex); | 109 | mutex_lock(&dev->struct_mutex); |
143 | idr_destroy(&dev->ctx_idr); | 110 | idr_destroy(&dev->ctx_idr); |
@@ -169,9 +136,6 @@ int drm_getsareactx(struct drm_device *dev, void *data, | |||
169 | struct drm_local_map *map; | 136 | struct drm_local_map *map; |
170 | struct drm_map_list *_entry; | 137 | struct drm_map_list *_entry; |
171 | 138 | ||
172 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
173 | return -EINVAL; | ||
174 | |||
175 | mutex_lock(&dev->struct_mutex); | 139 | mutex_lock(&dev->struct_mutex); |
176 | 140 | ||
177 | map = idr_find(&dev->ctx_idr, request->ctx_id); | 141 | map = idr_find(&dev->ctx_idr, request->ctx_id); |
@@ -216,9 +180,6 @@ int drm_setsareactx(struct drm_device *dev, void *data, | |||
216 | struct drm_local_map *map = NULL; | 180 | struct drm_local_map *map = NULL; |
217 | struct drm_map_list *r_list = NULL; | 181 | struct drm_map_list *r_list = NULL; |
218 | 182 | ||
219 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
220 | return -EINVAL; | ||
221 | |||
222 | mutex_lock(&dev->struct_mutex); | 183 | mutex_lock(&dev->struct_mutex); |
223 | list_for_each_entry(r_list, &dev->maplist, head) { | 184 | list_for_each_entry(r_list, &dev->maplist, head) { |
224 | if (r_list->map | 185 | if (r_list->map |
@@ -319,9 +280,6 @@ int drm_resctx(struct drm_device *dev, void *data, | |||
319 | struct drm_ctx ctx; | 280 | struct drm_ctx ctx; |
320 | int i; | 281 | int i; |
321 | 282 | ||
322 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
323 | return -EINVAL; | ||
324 | |||
325 | if (res->count >= DRM_RESERVED_CONTEXTS) { | 283 | if (res->count >= DRM_RESERVED_CONTEXTS) { |
326 | memset(&ctx, 0, sizeof(ctx)); | 284 | memset(&ctx, 0, sizeof(ctx)); |
327 | for (i = 0; i < DRM_RESERVED_CONTEXTS; i++) { | 285 | for (i = 0; i < DRM_RESERVED_CONTEXTS; i++) { |
@@ -352,9 +310,6 @@ int drm_addctx(struct drm_device *dev, void *data, | |||
352 | struct drm_ctx_list *ctx_entry; | 310 | struct drm_ctx_list *ctx_entry; |
353 | struct drm_ctx *ctx = data; | 311 | struct drm_ctx *ctx = data; |
354 | 312 | ||
355 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
356 | return -EINVAL; | ||
357 | |||
358 | ctx->handle = drm_ctxbitmap_next(dev); | 313 | ctx->handle = drm_ctxbitmap_next(dev); |
359 | if (ctx->handle == DRM_KERNEL_CONTEXT) { | 314 | if (ctx->handle == DRM_KERNEL_CONTEXT) { |
360 | /* Skip kernel's context and get a new one. */ | 315 | /* Skip kernel's context and get a new one. */ |
@@ -398,9 +353,6 @@ int drm_getctx(struct drm_device *dev, void *data, struct drm_file *file_priv) | |||
398 | { | 353 | { |
399 | struct drm_ctx *ctx = data; | 354 | struct drm_ctx *ctx = data; |
400 | 355 | ||
401 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
402 | return -EINVAL; | ||
403 | |||
404 | /* This is 0, because we don't handle any context flags */ | 356 | /* This is 0, because we don't handle any context flags */ |
405 | ctx->flags = 0; | 357 | ctx->flags = 0; |
406 | 358 | ||
@@ -423,9 +375,6 @@ int drm_switchctx(struct drm_device *dev, void *data, | |||
423 | { | 375 | { |
424 | struct drm_ctx *ctx = data; | 376 | struct drm_ctx *ctx = data; |
425 | 377 | ||
426 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
427 | return -EINVAL; | ||
428 | |||
429 | DRM_DEBUG("%d\n", ctx->handle); | 378 | DRM_DEBUG("%d\n", ctx->handle); |
430 | return drm_context_switch(dev, dev->last_context, ctx->handle); | 379 | return drm_context_switch(dev, dev->last_context, ctx->handle); |
431 | } | 380 | } |
@@ -446,9 +395,6 @@ int drm_newctx(struct drm_device *dev, void *data, | |||
446 | { | 395 | { |
447 | struct drm_ctx *ctx = data; | 396 | struct drm_ctx *ctx = data; |
448 | 397 | ||
449 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
450 | return -EINVAL; | ||
451 | |||
452 | DRM_DEBUG("%d\n", ctx->handle); | 398 | DRM_DEBUG("%d\n", ctx->handle); |
453 | drm_context_switch_complete(dev, file_priv, ctx->handle); | 399 | drm_context_switch_complete(dev, file_priv, ctx->handle); |
454 | 400 | ||
@@ -471,9 +417,6 @@ int drm_rmctx(struct drm_device *dev, void *data, | |||
471 | { | 417 | { |
472 | struct drm_ctx *ctx = data; | 418 | struct drm_ctx *ctx = data; |
473 | 419 | ||
474 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
475 | return -EINVAL; | ||
476 | |||
477 | DRM_DEBUG("%d\n", ctx->handle); | 420 | DRM_DEBUG("%d\n", ctx->handle); |
478 | if (ctx->handle != DRM_KERNEL_CONTEXT) { | 421 | if (ctx->handle != DRM_KERNEL_CONTEXT) { |
479 | if (dev->driver->context_dtor) | 422 | if (dev->driver->context_dtor) |
diff --git a/drivers/gpu/drm/drm_edid.c b/drivers/gpu/drm/drm_edid.c index 1688ff500513..830f7501cb4d 100644 --- a/drivers/gpu/drm/drm_edid.c +++ b/drivers/gpu/drm/drm_edid.c | |||
@@ -2925,6 +2925,8 @@ int drm_edid_to_speaker_allocation(struct edid *edid, u8 **sadb) | |||
2925 | /* Speaker Allocation Data Block */ | 2925 | /* Speaker Allocation Data Block */ |
2926 | if (dbl == 3) { | 2926 | if (dbl == 3) { |
2927 | *sadb = kmalloc(dbl, GFP_KERNEL); | 2927 | *sadb = kmalloc(dbl, GFP_KERNEL); |
2928 | if (!*sadb) | ||
2929 | return -ENOMEM; | ||
2928 | memcpy(*sadb, &db[1], dbl); | 2930 | memcpy(*sadb, &db[1], dbl); |
2929 | count = dbl; | 2931 | count = dbl; |
2930 | break; | 2932 | break; |
diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c index 4be8e09a32ef..3f84277d7036 100644 --- a/drivers/gpu/drm/drm_fops.c +++ b/drivers/gpu/drm/drm_fops.c | |||
@@ -439,7 +439,26 @@ int drm_release(struct inode *inode, struct file *filp) | |||
439 | if (dev->driver->driver_features & DRIVER_GEM) | 439 | if (dev->driver->driver_features & DRIVER_GEM) |
440 | drm_gem_release(dev, file_priv); | 440 | drm_gem_release(dev, file_priv); |
441 | 441 | ||
442 | drm_legacy_ctxbitmap_release(dev, file_priv); | 442 | mutex_lock(&dev->ctxlist_mutex); |
443 | if (!list_empty(&dev->ctxlist)) { | ||
444 | struct drm_ctx_list *pos, *n; | ||
445 | |||
446 | list_for_each_entry_safe(pos, n, &dev->ctxlist, head) { | ||
447 | if (pos->tag == file_priv && | ||
448 | pos->handle != DRM_KERNEL_CONTEXT) { | ||
449 | if (dev->driver->context_dtor) | ||
450 | dev->driver->context_dtor(dev, | ||
451 | pos->handle); | ||
452 | |||
453 | drm_ctxbitmap_free(dev, pos->handle); | ||
454 | |||
455 | list_del(&pos->head); | ||
456 | kfree(pos); | ||
457 | --dev->ctx_count; | ||
458 | } | ||
459 | } | ||
460 | } | ||
461 | mutex_unlock(&dev->ctxlist_mutex); | ||
443 | 462 | ||
444 | mutex_lock(&dev->struct_mutex); | 463 | mutex_lock(&dev->struct_mutex); |
445 | 464 | ||
diff --git a/drivers/gpu/drm/drm_stub.c b/drivers/gpu/drm/drm_stub.c index e7eb0276f7f1..39d864576be4 100644 --- a/drivers/gpu/drm/drm_stub.c +++ b/drivers/gpu/drm/drm_stub.c | |||
@@ -292,7 +292,13 @@ int drm_fill_in_dev(struct drm_device *dev, | |||
292 | goto error_out_unreg; | 292 | goto error_out_unreg; |
293 | } | 293 | } |
294 | 294 | ||
295 | drm_legacy_ctxbitmap_init(dev); | 295 | |
296 | |||
297 | retcode = drm_ctxbitmap_init(dev); | ||
298 | if (retcode) { | ||
299 | DRM_ERROR("Cannot allocate memory for context bitmap.\n"); | ||
300 | goto error_out_unreg; | ||
301 | } | ||
296 | 302 | ||
297 | if (driver->driver_features & DRIVER_GEM) { | 303 | if (driver->driver_features & DRIVER_GEM) { |
298 | retcode = drm_gem_init(dev); | 304 | retcode = drm_gem_init(dev); |
@@ -446,7 +452,7 @@ void drm_put_dev(struct drm_device *dev) | |||
446 | drm_rmmap(dev, r_list->map); | 452 | drm_rmmap(dev, r_list->map); |
447 | drm_ht_remove(&dev->map_hash); | 453 | drm_ht_remove(&dev->map_hash); |
448 | 454 | ||
449 | drm_legacy_ctxbitmap_cleanup(dev); | 455 | drm_ctxbitmap_cleanup(dev); |
450 | 456 | ||
451 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | 457 | if (drm_core_check_feature(dev, DRIVER_MODESET)) |
452 | drm_put_minor(&dev->control); | 458 | drm_put_minor(&dev->control); |
diff --git a/drivers/gpu/drm/exynos/Kconfig b/drivers/gpu/drm/exynos/Kconfig index 4752f223e5b2..45b6ef595965 100644 --- a/drivers/gpu/drm/exynos/Kconfig +++ b/drivers/gpu/drm/exynos/Kconfig | |||
@@ -56,7 +56,7 @@ config DRM_EXYNOS_IPP | |||
56 | 56 | ||
57 | config DRM_EXYNOS_FIMC | 57 | config DRM_EXYNOS_FIMC |
58 | bool "Exynos DRM FIMC" | 58 | bool "Exynos DRM FIMC" |
59 | depends on DRM_EXYNOS_IPP && MFD_SYSCON && OF | 59 | depends on DRM_EXYNOS_IPP && MFD_SYSCON |
60 | help | 60 | help |
61 | Choose this option if you want to use Exynos FIMC for DRM. | 61 | Choose this option if you want to use Exynos FIMC for DRM. |
62 | 62 | ||
diff --git a/drivers/gpu/drm/exynos/exynos_drm_buf.c b/drivers/gpu/drm/exynos/exynos_drm_buf.c index 3445a0f3a6b2..9c8088462c26 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_buf.c +++ b/drivers/gpu/drm/exynos/exynos_drm_buf.c | |||
@@ -63,7 +63,8 @@ static int lowlevel_buffer_allocate(struct drm_device *dev, | |||
63 | return -ENOMEM; | 63 | return -ENOMEM; |
64 | } | 64 | } |
65 | 65 | ||
66 | buf->kvaddr = dma_alloc_attrs(dev->dev, buf->size, | 66 | buf->kvaddr = (void __iomem *)dma_alloc_attrs(dev->dev, |
67 | buf->size, | ||
67 | &buf->dma_addr, GFP_KERNEL, | 68 | &buf->dma_addr, GFP_KERNEL, |
68 | &buf->dma_attrs); | 69 | &buf->dma_attrs); |
69 | if (!buf->kvaddr) { | 70 | if (!buf->kvaddr) { |
@@ -90,9 +91,9 @@ static int lowlevel_buffer_allocate(struct drm_device *dev, | |||
90 | } | 91 | } |
91 | 92 | ||
92 | buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages); | 93 | buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages); |
93 | if (!buf->sgt) { | 94 | if (IS_ERR(buf->sgt)) { |
94 | DRM_ERROR("failed to get sg table.\n"); | 95 | DRM_ERROR("failed to get sg table.\n"); |
95 | ret = -ENOMEM; | 96 | ret = PTR_ERR(buf->sgt); |
96 | goto err_free_attrs; | 97 | goto err_free_attrs; |
97 | } | 98 | } |
98 | 99 | ||
diff --git a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c index 78e868bcf1ec..e7c2f2d07f19 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c +++ b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c | |||
@@ -99,12 +99,13 @@ static int exynos_drm_fbdev_update(struct drm_fb_helper *helper, | |||
99 | if (is_drm_iommu_supported(dev)) { | 99 | if (is_drm_iommu_supported(dev)) { |
100 | unsigned int nr_pages = buffer->size >> PAGE_SHIFT; | 100 | unsigned int nr_pages = buffer->size >> PAGE_SHIFT; |
101 | 101 | ||
102 | buffer->kvaddr = vmap(buffer->pages, nr_pages, VM_MAP, | 102 | buffer->kvaddr = (void __iomem *) vmap(buffer->pages, |
103 | nr_pages, VM_MAP, | ||
103 | pgprot_writecombine(PAGE_KERNEL)); | 104 | pgprot_writecombine(PAGE_KERNEL)); |
104 | } else { | 105 | } else { |
105 | phys_addr_t dma_addr = buffer->dma_addr; | 106 | phys_addr_t dma_addr = buffer->dma_addr; |
106 | if (dma_addr) | 107 | if (dma_addr) |
107 | buffer->kvaddr = phys_to_virt(dma_addr); | 108 | buffer->kvaddr = (void __iomem *)phys_to_virt(dma_addr); |
108 | else | 109 | else |
109 | buffer->kvaddr = (void __iomem *)NULL; | 110 | buffer->kvaddr = (void __iomem *)NULL; |
110 | } | 111 | } |
diff --git a/drivers/gpu/drm/gma500/gtt.c b/drivers/gpu/drm/gma500/gtt.c index 92babac362ec..2db731f00930 100644 --- a/drivers/gpu/drm/gma500/gtt.c +++ b/drivers/gpu/drm/gma500/gtt.c | |||
@@ -204,6 +204,7 @@ static int psb_gtt_attach_pages(struct gtt_range *gt) | |||
204 | if (IS_ERR(pages)) | 204 | if (IS_ERR(pages)) |
205 | return PTR_ERR(pages); | 205 | return PTR_ERR(pages); |
206 | 206 | ||
207 | gt->npage = gt->gem.size / PAGE_SIZE; | ||
207 | gt->pages = pages; | 208 | gt->pages = pages; |
208 | 209 | ||
209 | return 0; | 210 | return 0; |
diff --git a/drivers/gpu/drm/i2c/tda998x_drv.c b/drivers/gpu/drm/i2c/tda998x_drv.c index b1f8fc69023f..60e84043aa34 100644 --- a/drivers/gpu/drm/i2c/tda998x_drv.c +++ b/drivers/gpu/drm/i2c/tda998x_drv.c | |||
@@ -707,8 +707,7 @@ tda998x_encoder_dpms(struct drm_encoder *encoder, int mode) | |||
707 | reg_write(encoder, REG_VIP_CNTRL_2, priv->vip_cntrl_2); | 707 | reg_write(encoder, REG_VIP_CNTRL_2, priv->vip_cntrl_2); |
708 | break; | 708 | break; |
709 | case DRM_MODE_DPMS_OFF: | 709 | case DRM_MODE_DPMS_OFF: |
710 | /* disable audio and video ports */ | 710 | /* disable video ports */ |
711 | reg_write(encoder, REG_ENA_AP, 0x00); | ||
712 | reg_write(encoder, REG_ENA_VP_0, 0x00); | 711 | reg_write(encoder, REG_ENA_VP_0, 0x00); |
713 | reg_write(encoder, REG_ENA_VP_1, 0x00); | 712 | reg_write(encoder, REG_ENA_VP_1, 0x00); |
714 | reg_write(encoder, REG_ENA_VP_2, 0x00); | 713 | reg_write(encoder, REG_ENA_VP_2, 0x00); |
diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c index c27a21034a5e..d5c784d48671 100644 --- a/drivers/gpu/drm/i915/i915_dma.c +++ b/drivers/gpu/drm/i915/i915_dma.c | |||
@@ -1290,12 +1290,9 @@ static int i915_load_modeset_init(struct drm_device *dev) | |||
1290 | * then we do not take part in VGA arbitration and the | 1290 | * then we do not take part in VGA arbitration and the |
1291 | * vga_client_register() fails with -ENODEV. | 1291 | * vga_client_register() fails with -ENODEV. |
1292 | */ | 1292 | */ |
1293 | if (!HAS_PCH_SPLIT(dev)) { | 1293 | ret = vga_client_register(dev->pdev, dev, NULL, i915_vga_set_decode); |
1294 | ret = vga_client_register(dev->pdev, dev, NULL, | 1294 | if (ret && ret != -ENODEV) |
1295 | i915_vga_set_decode); | 1295 | goto out; |
1296 | if (ret && ret != -ENODEV) | ||
1297 | goto out; | ||
1298 | } | ||
1299 | 1296 | ||
1300 | intel_register_dsm_handler(); | 1297 | intel_register_dsm_handler(); |
1301 | 1298 | ||
@@ -1351,12 +1348,6 @@ static int i915_load_modeset_init(struct drm_device *dev) | |||
1351 | */ | 1348 | */ |
1352 | intel_fbdev_initial_config(dev); | 1349 | intel_fbdev_initial_config(dev); |
1353 | 1350 | ||
1354 | /* | ||
1355 | * Must do this after fbcon init so that | ||
1356 | * vgacon_save_screen() works during the handover. | ||
1357 | */ | ||
1358 | i915_disable_vga_mem(dev); | ||
1359 | |||
1360 | /* Only enable hotplug handling once the fbdev is fully set up. */ | 1351 | /* Only enable hotplug handling once the fbdev is fully set up. */ |
1361 | dev_priv->enable_hotplug_processing = true; | 1352 | dev_priv->enable_hotplug_processing = true; |
1362 | 1353 | ||
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 8507c6d1e642..cdfb9da0e4ce 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
@@ -1392,14 +1392,11 @@ out: | |||
1392 | if (i915_terminally_wedged(&dev_priv->gpu_error)) | 1392 | if (i915_terminally_wedged(&dev_priv->gpu_error)) |
1393 | return VM_FAULT_SIGBUS; | 1393 | return VM_FAULT_SIGBUS; |
1394 | case -EAGAIN: | 1394 | case -EAGAIN: |
1395 | /* Give the error handler a chance to run and move the | 1395 | /* |
1396 | * objects off the GPU active list. Next time we service the | 1396 | * EAGAIN means the gpu is hung and we'll wait for the error |
1397 | * fault, we should be able to transition the page into the | 1397 | * handler to reset everything when re-faulting in |
1398 | * GTT without touching the GPU (and so avoid further | 1398 | * i915_mutex_lock_interruptible. |
1399 | * EIO/EGAIN). If the GPU is wedged, then there is no issue | ||
1400 | * with coherency, just lost writes. | ||
1401 | */ | 1399 | */ |
1402 | set_need_resched(); | ||
1403 | case 0: | 1400 | case 0: |
1404 | case -ERESTARTSYS: | 1401 | case -ERESTARTSYS: |
1405 | case -EINTR: | 1402 | case -EINTR: |
@@ -4803,10 +4800,10 @@ i915_gem_inactive_count(struct shrinker *shrinker, struct shrink_control *sc) | |||
4803 | 4800 | ||
4804 | if (!mutex_trylock(&dev->struct_mutex)) { | 4801 | if (!mutex_trylock(&dev->struct_mutex)) { |
4805 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) | 4802 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) |
4806 | return SHRINK_STOP; | 4803 | return 0; |
4807 | 4804 | ||
4808 | if (dev_priv->mm.shrinker_no_lock_stealing) | 4805 | if (dev_priv->mm.shrinker_no_lock_stealing) |
4809 | return SHRINK_STOP; | 4806 | return 0; |
4810 | 4807 | ||
4811 | unlock = false; | 4808 | unlock = false; |
4812 | } | 4809 | } |
@@ -4904,10 +4901,10 @@ i915_gem_inactive_scan(struct shrinker *shrinker, struct shrink_control *sc) | |||
4904 | 4901 | ||
4905 | if (!mutex_trylock(&dev->struct_mutex)) { | 4902 | if (!mutex_trylock(&dev->struct_mutex)) { |
4906 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) | 4903 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) |
4907 | return 0; | 4904 | return SHRINK_STOP; |
4908 | 4905 | ||
4909 | if (dev_priv->mm.shrinker_no_lock_stealing) | 4906 | if (dev_priv->mm.shrinker_no_lock_stealing) |
4910 | return 0; | 4907 | return SHRINK_STOP; |
4911 | 4908 | ||
4912 | unlock = false; | 4909 | unlock = false; |
4913 | } | 4910 | } |
diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c index aba9d7498996..dae364f0028c 100644 --- a/drivers/gpu/drm/i915/i915_gpu_error.c +++ b/drivers/gpu/drm/i915/i915_gpu_error.c | |||
@@ -143,8 +143,10 @@ static void i915_error_vprintf(struct drm_i915_error_state_buf *e, | |||
143 | 143 | ||
144 | /* Seek the first printf which is hits start position */ | 144 | /* Seek the first printf which is hits start position */ |
145 | if (e->pos < e->start) { | 145 | if (e->pos < e->start) { |
146 | len = vsnprintf(NULL, 0, f, args); | 146 | va_list tmp; |
147 | if (!__i915_error_seek(e, len)) | 147 | |
148 | va_copy(tmp, args); | ||
149 | if (!__i915_error_seek(e, vsnprintf(NULL, 0, f, tmp))) | ||
148 | return; | 150 | return; |
149 | } | 151 | } |
150 | 152 | ||
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index 83cce0cdb769..4b91228fd9bd 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
@@ -1469,6 +1469,34 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg) | |||
1469 | return ret; | 1469 | return ret; |
1470 | } | 1470 | } |
1471 | 1471 | ||
1472 | static void i915_error_wake_up(struct drm_i915_private *dev_priv, | ||
1473 | bool reset_completed) | ||
1474 | { | ||
1475 | struct intel_ring_buffer *ring; | ||
1476 | int i; | ||
1477 | |||
1478 | /* | ||
1479 | * Notify all waiters for GPU completion events that reset state has | ||
1480 | * been changed, and that they need to restart their wait after | ||
1481 | * checking for potential errors (and bail out to drop locks if there is | ||
1482 | * a gpu reset pending so that i915_error_work_func can acquire them). | ||
1483 | */ | ||
1484 | |||
1485 | /* Wake up __wait_seqno, potentially holding dev->struct_mutex. */ | ||
1486 | for_each_ring(ring, dev_priv, i) | ||
1487 | wake_up_all(&ring->irq_queue); | ||
1488 | |||
1489 | /* Wake up intel_crtc_wait_for_pending_flips, holding crtc->mutex. */ | ||
1490 | wake_up_all(&dev_priv->pending_flip_queue); | ||
1491 | |||
1492 | /* | ||
1493 | * Signal tasks blocked in i915_gem_wait_for_error that the pending | ||
1494 | * reset state is cleared. | ||
1495 | */ | ||
1496 | if (reset_completed) | ||
1497 | wake_up_all(&dev_priv->gpu_error.reset_queue); | ||
1498 | } | ||
1499 | |||
1472 | /** | 1500 | /** |
1473 | * i915_error_work_func - do process context error handling work | 1501 | * i915_error_work_func - do process context error handling work |
1474 | * @work: work struct | 1502 | * @work: work struct |
@@ -1483,11 +1511,10 @@ static void i915_error_work_func(struct work_struct *work) | |||
1483 | drm_i915_private_t *dev_priv = container_of(error, drm_i915_private_t, | 1511 | drm_i915_private_t *dev_priv = container_of(error, drm_i915_private_t, |
1484 | gpu_error); | 1512 | gpu_error); |
1485 | struct drm_device *dev = dev_priv->dev; | 1513 | struct drm_device *dev = dev_priv->dev; |
1486 | struct intel_ring_buffer *ring; | ||
1487 | char *error_event[] = { I915_ERROR_UEVENT "=1", NULL }; | 1514 | char *error_event[] = { I915_ERROR_UEVENT "=1", NULL }; |
1488 | char *reset_event[] = { I915_RESET_UEVENT "=1", NULL }; | 1515 | char *reset_event[] = { I915_RESET_UEVENT "=1", NULL }; |
1489 | char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL }; | 1516 | char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL }; |
1490 | int i, ret; | 1517 | int ret; |
1491 | 1518 | ||
1492 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); | 1519 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); |
1493 | 1520 | ||
@@ -1506,8 +1533,16 @@ static void i915_error_work_func(struct work_struct *work) | |||
1506 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, | 1533 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, |
1507 | reset_event); | 1534 | reset_event); |
1508 | 1535 | ||
1536 | /* | ||
1537 | * All state reset _must_ be completed before we update the | ||
1538 | * reset counter, for otherwise waiters might miss the reset | ||
1539 | * pending state and not properly drop locks, resulting in | ||
1540 | * deadlocks with the reset work. | ||
1541 | */ | ||
1509 | ret = i915_reset(dev); | 1542 | ret = i915_reset(dev); |
1510 | 1543 | ||
1544 | intel_display_handle_reset(dev); | ||
1545 | |||
1511 | if (ret == 0) { | 1546 | if (ret == 0) { |
1512 | /* | 1547 | /* |
1513 | * After all the gem state is reset, increment the reset | 1548 | * After all the gem state is reset, increment the reset |
@@ -1528,12 +1563,11 @@ static void i915_error_work_func(struct work_struct *work) | |||
1528 | atomic_set(&error->reset_counter, I915_WEDGED); | 1563 | atomic_set(&error->reset_counter, I915_WEDGED); |
1529 | } | 1564 | } |
1530 | 1565 | ||
1531 | for_each_ring(ring, dev_priv, i) | 1566 | /* |
1532 | wake_up_all(&ring->irq_queue); | 1567 | * Note: The wake_up also serves as a memory barrier so that |
1533 | 1568 | * waiters see the update value of the reset counter atomic_t. | |
1534 | intel_display_handle_reset(dev); | 1569 | */ |
1535 | 1570 | i915_error_wake_up(dev_priv, true); | |
1536 | wake_up_all(&dev_priv->gpu_error.reset_queue); | ||
1537 | } | 1571 | } |
1538 | } | 1572 | } |
1539 | 1573 | ||
@@ -1642,8 +1676,6 @@ static void i915_report_and_clear_eir(struct drm_device *dev) | |||
1642 | void i915_handle_error(struct drm_device *dev, bool wedged) | 1676 | void i915_handle_error(struct drm_device *dev, bool wedged) |
1643 | { | 1677 | { |
1644 | struct drm_i915_private *dev_priv = dev->dev_private; | 1678 | struct drm_i915_private *dev_priv = dev->dev_private; |
1645 | struct intel_ring_buffer *ring; | ||
1646 | int i; | ||
1647 | 1679 | ||
1648 | i915_capture_error_state(dev); | 1680 | i915_capture_error_state(dev); |
1649 | i915_report_and_clear_eir(dev); | 1681 | i915_report_and_clear_eir(dev); |
@@ -1653,11 +1685,19 @@ void i915_handle_error(struct drm_device *dev, bool wedged) | |||
1653 | &dev_priv->gpu_error.reset_counter); | 1685 | &dev_priv->gpu_error.reset_counter); |
1654 | 1686 | ||
1655 | /* | 1687 | /* |
1656 | * Wakeup waiting processes so that the reset work item | 1688 | * Wakeup waiting processes so that the reset work function |
1657 | * doesn't deadlock trying to grab various locks. | 1689 | * i915_error_work_func doesn't deadlock trying to grab various |
1690 | * locks. By bumping the reset counter first, the woken | ||
1691 | * processes will see a reset in progress and back off, | ||
1692 | * releasing their locks and then wait for the reset completion. | ||
1693 | * We must do this for _all_ gpu waiters that might hold locks | ||
1694 | * that the reset work needs to acquire. | ||
1695 | * | ||
1696 | * Note: The wake_up serves as the required memory barrier to | ||
1697 | * ensure that the waiters see the updated value of the reset | ||
1698 | * counter atomic_t. | ||
1658 | */ | 1699 | */ |
1659 | for_each_ring(ring, dev_priv, i) | 1700 | i915_error_wake_up(dev_priv, false); |
1660 | wake_up_all(&ring->irq_queue); | ||
1661 | } | 1701 | } |
1662 | 1702 | ||
1663 | /* | 1703 | /* |
diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h index c159e1a6810f..38f96f65d87a 100644 --- a/drivers/gpu/drm/i915/i915_reg.h +++ b/drivers/gpu/drm/i915/i915_reg.h | |||
@@ -3881,6 +3881,9 @@ | |||
3881 | #define GEN7_SQ_CHICKEN_MBCUNIT_CONFIG 0x9030 | 3881 | #define GEN7_SQ_CHICKEN_MBCUNIT_CONFIG 0x9030 |
3882 | #define GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB (1<<11) | 3882 | #define GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB (1<<11) |
3883 | 3883 | ||
3884 | #define HSW_SCRATCH1 0xb038 | ||
3885 | #define HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE (1<<27) | ||
3886 | |||
3884 | #define HSW_FUSE_STRAP 0x42014 | 3887 | #define HSW_FUSE_STRAP 0x42014 |
3885 | #define HSW_CDCLK_LIMIT (1 << 24) | 3888 | #define HSW_CDCLK_LIMIT (1 << 24) |
3886 | 3889 | ||
@@ -4728,6 +4731,9 @@ | |||
4728 | #define GEN7_ROW_CHICKEN2_GT2 0xf4f4 | 4731 | #define GEN7_ROW_CHICKEN2_GT2 0xf4f4 |
4729 | #define DOP_CLOCK_GATING_DISABLE (1<<0) | 4732 | #define DOP_CLOCK_GATING_DISABLE (1<<0) |
4730 | 4733 | ||
4734 | #define HSW_ROW_CHICKEN3 0xe49c | ||
4735 | #define HSW_ROW_CHICKEN3_L3_GLOBAL_ATOMICS_DISABLE (1 << 6) | ||
4736 | |||
4731 | #define G4X_AUD_VID_DID (dev_priv->info->display_mmio_offset + 0x62020) | 4737 | #define G4X_AUD_VID_DID (dev_priv->info->display_mmio_offset + 0x62020) |
4732 | #define INTEL_AUDIO_DEVCL 0x808629FB | 4738 | #define INTEL_AUDIO_DEVCL 0x808629FB |
4733 | #define INTEL_AUDIO_DEVBLC 0x80862801 | 4739 | #define INTEL_AUDIO_DEVBLC 0x80862801 |
diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c index 63aca49d11a8..63de2701b974 100644 --- a/drivers/gpu/drm/i915/intel_ddi.c +++ b/drivers/gpu/drm/i915/intel_ddi.c | |||
@@ -778,7 +778,7 @@ void intel_ddi_enable_transcoder_func(struct drm_crtc *crtc) | |||
778 | /* Can only use the always-on power well for eDP when | 778 | /* Can only use the always-on power well for eDP when |
779 | * not using the panel fitter, and when not using motion | 779 | * not using the panel fitter, and when not using motion |
780 | * blur mitigation (which we don't support). */ | 780 | * blur mitigation (which we don't support). */ |
781 | if (intel_crtc->config.pch_pfit.size) | 781 | if (intel_crtc->config.pch_pfit.enabled) |
782 | temp |= TRANS_DDI_EDP_INPUT_A_ONOFF; | 782 | temp |= TRANS_DDI_EDP_INPUT_A_ONOFF; |
783 | else | 783 | else |
784 | temp |= TRANS_DDI_EDP_INPUT_A_ON; | 784 | temp |= TRANS_DDI_EDP_INPUT_A_ON; |
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index 2489d0b4c7d2..581fb4b2f766 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
@@ -2249,7 +2249,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
2249 | I915_WRITE(PIPESRC(intel_crtc->pipe), | 2249 | I915_WRITE(PIPESRC(intel_crtc->pipe), |
2250 | ((crtc->mode.hdisplay - 1) << 16) | | 2250 | ((crtc->mode.hdisplay - 1) << 16) | |
2251 | (crtc->mode.vdisplay - 1)); | 2251 | (crtc->mode.vdisplay - 1)); |
2252 | if (!intel_crtc->config.pch_pfit.size && | 2252 | if (!intel_crtc->config.pch_pfit.enabled && |
2253 | (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS) || | 2253 | (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS) || |
2254 | intel_pipe_has_type(crtc, INTEL_OUTPUT_EDP))) { | 2254 | intel_pipe_has_type(crtc, INTEL_OUTPUT_EDP))) { |
2255 | I915_WRITE(PF_CTL(intel_crtc->pipe), 0); | 2255 | I915_WRITE(PF_CTL(intel_crtc->pipe), 0); |
@@ -3203,7 +3203,7 @@ static void ironlake_pfit_enable(struct intel_crtc *crtc) | |||
3203 | struct drm_i915_private *dev_priv = dev->dev_private; | 3203 | struct drm_i915_private *dev_priv = dev->dev_private; |
3204 | int pipe = crtc->pipe; | 3204 | int pipe = crtc->pipe; |
3205 | 3205 | ||
3206 | if (crtc->config.pch_pfit.size) { | 3206 | if (crtc->config.pch_pfit.enabled) { |
3207 | /* Force use of hard-coded filter coefficients | 3207 | /* Force use of hard-coded filter coefficients |
3208 | * as some pre-programmed values are broken, | 3208 | * as some pre-programmed values are broken, |
3209 | * e.g. x201. | 3209 | * e.g. x201. |
@@ -3428,7 +3428,7 @@ static void ironlake_pfit_disable(struct intel_crtc *crtc) | |||
3428 | 3428 | ||
3429 | /* To avoid upsetting the power well on haswell only disable the pfit if | 3429 | /* To avoid upsetting the power well on haswell only disable the pfit if |
3430 | * it's in use. The hw state code will make sure we get this right. */ | 3430 | * it's in use. The hw state code will make sure we get this right. */ |
3431 | if (crtc->config.pch_pfit.size) { | 3431 | if (crtc->config.pch_pfit.enabled) { |
3432 | I915_WRITE(PF_CTL(pipe), 0); | 3432 | I915_WRITE(PF_CTL(pipe), 0); |
3433 | I915_WRITE(PF_WIN_POS(pipe), 0); | 3433 | I915_WRITE(PF_WIN_POS(pipe), 0); |
3434 | I915_WRITE(PF_WIN_SZ(pipe), 0); | 3434 | I915_WRITE(PF_WIN_SZ(pipe), 0); |
@@ -3941,8 +3941,6 @@ static void intel_connector_check_state(struct intel_connector *connector) | |||
3941 | * consider. */ | 3941 | * consider. */ |
3942 | void intel_connector_dpms(struct drm_connector *connector, int mode) | 3942 | void intel_connector_dpms(struct drm_connector *connector, int mode) |
3943 | { | 3943 | { |
3944 | struct intel_encoder *encoder = intel_attached_encoder(connector); | ||
3945 | |||
3946 | /* All the simple cases only support two dpms states. */ | 3944 | /* All the simple cases only support two dpms states. */ |
3947 | if (mode != DRM_MODE_DPMS_ON) | 3945 | if (mode != DRM_MODE_DPMS_ON) |
3948 | mode = DRM_MODE_DPMS_OFF; | 3946 | mode = DRM_MODE_DPMS_OFF; |
@@ -3953,10 +3951,8 @@ void intel_connector_dpms(struct drm_connector *connector, int mode) | |||
3953 | connector->dpms = mode; | 3951 | connector->dpms = mode; |
3954 | 3952 | ||
3955 | /* Only need to change hw state when actually enabled */ | 3953 | /* Only need to change hw state when actually enabled */ |
3956 | if (encoder->base.crtc) | 3954 | if (connector->encoder) |
3957 | intel_encoder_dpms(encoder, mode); | 3955 | intel_encoder_dpms(to_intel_encoder(connector->encoder), mode); |
3958 | else | ||
3959 | WARN_ON(encoder->connectors_active != false); | ||
3960 | 3956 | ||
3961 | intel_modeset_check_state(connector->dev); | 3957 | intel_modeset_check_state(connector->dev); |
3962 | } | 3958 | } |
@@ -4775,6 +4771,10 @@ static void i9xx_set_pipeconf(struct intel_crtc *intel_crtc) | |||
4775 | 4771 | ||
4776 | pipeconf = 0; | 4772 | pipeconf = 0; |
4777 | 4773 | ||
4774 | if (dev_priv->quirks & QUIRK_PIPEA_FORCE && | ||
4775 | I915_READ(PIPECONF(intel_crtc->pipe)) & PIPECONF_ENABLE) | ||
4776 | pipeconf |= PIPECONF_ENABLE; | ||
4777 | |||
4778 | if (intel_crtc->pipe == 0 && INTEL_INFO(dev)->gen < 4) { | 4778 | if (intel_crtc->pipe == 0 && INTEL_INFO(dev)->gen < 4) { |
4779 | /* Enable pixel doubling when the dot clock is > 90% of the (display) | 4779 | /* Enable pixel doubling when the dot clock is > 90% of the (display) |
4780 | * core speed. | 4780 | * core speed. |
@@ -4877,9 +4877,6 @@ static int i9xx_crtc_mode_set(struct drm_crtc *crtc, | |||
4877 | return -EINVAL; | 4877 | return -EINVAL; |
4878 | } | 4878 | } |
4879 | 4879 | ||
4880 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
4881 | intel_crtc_update_cursor(crtc, true); | ||
4882 | |||
4883 | if (is_lvds && dev_priv->lvds_downclock_avail) { | 4880 | if (is_lvds && dev_priv->lvds_downclock_avail) { |
4884 | /* | 4881 | /* |
4885 | * Ensure we match the reduced clock's P to the target clock. | 4882 | * Ensure we match the reduced clock's P to the target clock. |
@@ -5768,9 +5765,6 @@ static int ironlake_crtc_mode_set(struct drm_crtc *crtc, | |||
5768 | intel_crtc->config.dpll.p2 = clock.p2; | 5765 | intel_crtc->config.dpll.p2 = clock.p2; |
5769 | } | 5766 | } |
5770 | 5767 | ||
5771 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
5772 | intel_crtc_update_cursor(crtc, true); | ||
5773 | |||
5774 | /* CPU eDP is the only output that doesn't need a PCH PLL of its own. */ | 5768 | /* CPU eDP is the only output that doesn't need a PCH PLL of its own. */ |
5775 | if (intel_crtc->config.has_pch_encoder) { | 5769 | if (intel_crtc->config.has_pch_encoder) { |
5776 | fp = i9xx_dpll_compute_fp(&intel_crtc->config.dpll); | 5770 | fp = i9xx_dpll_compute_fp(&intel_crtc->config.dpll); |
@@ -5859,6 +5853,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc, | |||
5859 | tmp = I915_READ(PF_CTL(crtc->pipe)); | 5853 | tmp = I915_READ(PF_CTL(crtc->pipe)); |
5860 | 5854 | ||
5861 | if (tmp & PF_ENABLE) { | 5855 | if (tmp & PF_ENABLE) { |
5856 | pipe_config->pch_pfit.enabled = true; | ||
5862 | pipe_config->pch_pfit.pos = I915_READ(PF_WIN_POS(crtc->pipe)); | 5857 | pipe_config->pch_pfit.pos = I915_READ(PF_WIN_POS(crtc->pipe)); |
5863 | pipe_config->pch_pfit.size = I915_READ(PF_WIN_SZ(crtc->pipe)); | 5858 | pipe_config->pch_pfit.size = I915_READ(PF_WIN_SZ(crtc->pipe)); |
5864 | 5859 | ||
@@ -6236,7 +6231,7 @@ static void haswell_modeset_global_resources(struct drm_device *dev) | |||
6236 | if (!crtc->base.enabled) | 6231 | if (!crtc->base.enabled) |
6237 | continue; | 6232 | continue; |
6238 | 6233 | ||
6239 | if (crtc->pipe != PIPE_A || crtc->config.pch_pfit.size || | 6234 | if (crtc->pipe != PIPE_A || crtc->config.pch_pfit.enabled || |
6240 | crtc->config.cpu_transcoder != TRANSCODER_EDP) | 6235 | crtc->config.cpu_transcoder != TRANSCODER_EDP) |
6241 | enable = true; | 6236 | enable = true; |
6242 | } | 6237 | } |
@@ -6259,9 +6254,6 @@ static int haswell_crtc_mode_set(struct drm_crtc *crtc, | |||
6259 | if (!intel_ddi_pll_mode_set(crtc)) | 6254 | if (!intel_ddi_pll_mode_set(crtc)) |
6260 | return -EINVAL; | 6255 | return -EINVAL; |
6261 | 6256 | ||
6262 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
6263 | intel_crtc_update_cursor(crtc, true); | ||
6264 | |||
6265 | if (intel_crtc->config.has_dp_encoder) | 6257 | if (intel_crtc->config.has_dp_encoder) |
6266 | intel_dp_set_m_n(intel_crtc); | 6258 | intel_dp_set_m_n(intel_crtc); |
6267 | 6259 | ||
@@ -6494,15 +6486,15 @@ static void haswell_write_eld(struct drm_connector *connector, | |||
6494 | 6486 | ||
6495 | /* Set ELD valid state */ | 6487 | /* Set ELD valid state */ |
6496 | tmp = I915_READ(aud_cntrl_st2); | 6488 | tmp = I915_READ(aud_cntrl_st2); |
6497 | DRM_DEBUG_DRIVER("HDMI audio: pin eld vld status=0x%8x\n", tmp); | 6489 | DRM_DEBUG_DRIVER("HDMI audio: pin eld vld status=0x%08x\n", tmp); |
6498 | tmp |= (AUDIO_ELD_VALID_A << (pipe * 4)); | 6490 | tmp |= (AUDIO_ELD_VALID_A << (pipe * 4)); |
6499 | I915_WRITE(aud_cntrl_st2, tmp); | 6491 | I915_WRITE(aud_cntrl_st2, tmp); |
6500 | tmp = I915_READ(aud_cntrl_st2); | 6492 | tmp = I915_READ(aud_cntrl_st2); |
6501 | DRM_DEBUG_DRIVER("HDMI audio: eld vld status=0x%8x\n", tmp); | 6493 | DRM_DEBUG_DRIVER("HDMI audio: eld vld status=0x%08x\n", tmp); |
6502 | 6494 | ||
6503 | /* Enable HDMI mode */ | 6495 | /* Enable HDMI mode */ |
6504 | tmp = I915_READ(aud_config); | 6496 | tmp = I915_READ(aud_config); |
6505 | DRM_DEBUG_DRIVER("HDMI audio: audio conf: 0x%8x\n", tmp); | 6497 | DRM_DEBUG_DRIVER("HDMI audio: audio conf: 0x%08x\n", tmp); |
6506 | /* clear N_programing_enable and N_value_index */ | 6498 | /* clear N_programing_enable and N_value_index */ |
6507 | tmp &= ~(AUD_CONFIG_N_VALUE_INDEX | AUD_CONFIG_N_PROG_ENABLE); | 6499 | tmp &= ~(AUD_CONFIG_N_VALUE_INDEX | AUD_CONFIG_N_PROG_ENABLE); |
6508 | I915_WRITE(aud_config, tmp); | 6500 | I915_WRITE(aud_config, tmp); |
@@ -6937,7 +6929,8 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc, | |||
6937 | intel_crtc->cursor_width = width; | 6929 | intel_crtc->cursor_width = width; |
6938 | intel_crtc->cursor_height = height; | 6930 | intel_crtc->cursor_height = height; |
6939 | 6931 | ||
6940 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | 6932 | if (intel_crtc->active) |
6933 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | ||
6941 | 6934 | ||
6942 | return 0; | 6935 | return 0; |
6943 | fail_unpin: | 6936 | fail_unpin: |
@@ -6956,7 +6949,8 @@ static int intel_crtc_cursor_move(struct drm_crtc *crtc, int x, int y) | |||
6956 | intel_crtc->cursor_x = x; | 6949 | intel_crtc->cursor_x = x; |
6957 | intel_crtc->cursor_y = y; | 6950 | intel_crtc->cursor_y = y; |
6958 | 6951 | ||
6959 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | 6952 | if (intel_crtc->active) |
6953 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | ||
6960 | 6954 | ||
6961 | return 0; | 6955 | return 0; |
6962 | } | 6956 | } |
@@ -8205,9 +8199,10 @@ static void intel_dump_pipe_config(struct intel_crtc *crtc, | |||
8205 | pipe_config->gmch_pfit.control, | 8199 | pipe_config->gmch_pfit.control, |
8206 | pipe_config->gmch_pfit.pgm_ratios, | 8200 | pipe_config->gmch_pfit.pgm_ratios, |
8207 | pipe_config->gmch_pfit.lvds_border_bits); | 8201 | pipe_config->gmch_pfit.lvds_border_bits); |
8208 | DRM_DEBUG_KMS("pch pfit: pos: 0x%08x, size: 0x%08x\n", | 8202 | DRM_DEBUG_KMS("pch pfit: pos: 0x%08x, size: 0x%08x, %s\n", |
8209 | pipe_config->pch_pfit.pos, | 8203 | pipe_config->pch_pfit.pos, |
8210 | pipe_config->pch_pfit.size); | 8204 | pipe_config->pch_pfit.size, |
8205 | pipe_config->pch_pfit.enabled ? "enabled" : "disabled"); | ||
8211 | DRM_DEBUG_KMS("ips: %i\n", pipe_config->ips_enabled); | 8206 | DRM_DEBUG_KMS("ips: %i\n", pipe_config->ips_enabled); |
8212 | } | 8207 | } |
8213 | 8208 | ||
@@ -8603,8 +8598,11 @@ intel_pipe_config_compare(struct drm_device *dev, | |||
8603 | if (INTEL_INFO(dev)->gen < 4) | 8598 | if (INTEL_INFO(dev)->gen < 4) |
8604 | PIPE_CONF_CHECK_I(gmch_pfit.pgm_ratios); | 8599 | PIPE_CONF_CHECK_I(gmch_pfit.pgm_ratios); |
8605 | PIPE_CONF_CHECK_I(gmch_pfit.lvds_border_bits); | 8600 | PIPE_CONF_CHECK_I(gmch_pfit.lvds_border_bits); |
8606 | PIPE_CONF_CHECK_I(pch_pfit.pos); | 8601 | PIPE_CONF_CHECK_I(pch_pfit.enabled); |
8607 | PIPE_CONF_CHECK_I(pch_pfit.size); | 8602 | if (current_config->pch_pfit.enabled) { |
8603 | PIPE_CONF_CHECK_I(pch_pfit.pos); | ||
8604 | PIPE_CONF_CHECK_I(pch_pfit.size); | ||
8605 | } | ||
8608 | 8606 | ||
8609 | PIPE_CONF_CHECK_I(ips_enabled); | 8607 | PIPE_CONF_CHECK_I(ips_enabled); |
8610 | 8608 | ||
@@ -10047,33 +10045,6 @@ static void i915_disable_vga(struct drm_device *dev) | |||
10047 | POSTING_READ(vga_reg); | 10045 | POSTING_READ(vga_reg); |
10048 | } | 10046 | } |
10049 | 10047 | ||
10050 | static void i915_enable_vga_mem(struct drm_device *dev) | ||
10051 | { | ||
10052 | /* Enable VGA memory on Intel HD */ | ||
10053 | if (HAS_PCH_SPLIT(dev)) { | ||
10054 | vga_get_uninterruptible(dev->pdev, VGA_RSRC_LEGACY_IO); | ||
10055 | outb(inb(VGA_MSR_READ) | VGA_MSR_MEM_EN, VGA_MSR_WRITE); | ||
10056 | vga_set_legacy_decoding(dev->pdev, VGA_RSRC_LEGACY_IO | | ||
10057 | VGA_RSRC_LEGACY_MEM | | ||
10058 | VGA_RSRC_NORMAL_IO | | ||
10059 | VGA_RSRC_NORMAL_MEM); | ||
10060 | vga_put(dev->pdev, VGA_RSRC_LEGACY_IO); | ||
10061 | } | ||
10062 | } | ||
10063 | |||
10064 | void i915_disable_vga_mem(struct drm_device *dev) | ||
10065 | { | ||
10066 | /* Disable VGA memory on Intel HD */ | ||
10067 | if (HAS_PCH_SPLIT(dev)) { | ||
10068 | vga_get_uninterruptible(dev->pdev, VGA_RSRC_LEGACY_IO); | ||
10069 | outb(inb(VGA_MSR_READ) & ~VGA_MSR_MEM_EN, VGA_MSR_WRITE); | ||
10070 | vga_set_legacy_decoding(dev->pdev, VGA_RSRC_LEGACY_IO | | ||
10071 | VGA_RSRC_NORMAL_IO | | ||
10072 | VGA_RSRC_NORMAL_MEM); | ||
10073 | vga_put(dev->pdev, VGA_RSRC_LEGACY_IO); | ||
10074 | } | ||
10075 | } | ||
10076 | |||
10077 | void intel_modeset_init_hw(struct drm_device *dev) | 10048 | void intel_modeset_init_hw(struct drm_device *dev) |
10078 | { | 10049 | { |
10079 | intel_init_power_well(dev); | 10050 | intel_init_power_well(dev); |
@@ -10352,7 +10323,6 @@ void i915_redisable_vga(struct drm_device *dev) | |||
10352 | if (I915_READ(vga_reg) != VGA_DISP_DISABLE) { | 10323 | if (I915_READ(vga_reg) != VGA_DISP_DISABLE) { |
10353 | DRM_DEBUG_KMS("Something enabled VGA plane, disabling it\n"); | 10324 | DRM_DEBUG_KMS("Something enabled VGA plane, disabling it\n"); |
10354 | i915_disable_vga(dev); | 10325 | i915_disable_vga(dev); |
10355 | i915_disable_vga_mem(dev); | ||
10356 | } | 10326 | } |
10357 | } | 10327 | } |
10358 | 10328 | ||
@@ -10566,8 +10536,6 @@ void intel_modeset_cleanup(struct drm_device *dev) | |||
10566 | 10536 | ||
10567 | intel_disable_fbc(dev); | 10537 | intel_disable_fbc(dev); |
10568 | 10538 | ||
10569 | i915_enable_vga_mem(dev); | ||
10570 | |||
10571 | intel_disable_gt_powersave(dev); | 10539 | intel_disable_gt_powersave(dev); |
10572 | 10540 | ||
10573 | ironlake_teardown_rc6(dev); | 10541 | ironlake_teardown_rc6(dev); |
diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c index 2151d13772b8..2c555f91bfae 100644 --- a/drivers/gpu/drm/i915/intel_dp.c +++ b/drivers/gpu/drm/i915/intel_dp.c | |||
@@ -588,7 +588,18 @@ intel_dp_i2c_aux_ch(struct i2c_adapter *adapter, int mode, | |||
588 | DRM_DEBUG_KMS("aux_ch native nack\n"); | 588 | DRM_DEBUG_KMS("aux_ch native nack\n"); |
589 | return -EREMOTEIO; | 589 | return -EREMOTEIO; |
590 | case AUX_NATIVE_REPLY_DEFER: | 590 | case AUX_NATIVE_REPLY_DEFER: |
591 | udelay(100); | 591 | /* |
592 | * For now, just give more slack to branch devices. We | ||
593 | * could check the DPCD for I2C bit rate capabilities, | ||
594 | * and if available, adjust the interval. We could also | ||
595 | * be more careful with DP-to-Legacy adapters where a | ||
596 | * long legacy cable may force very low I2C bit rates. | ||
597 | */ | ||
598 | if (intel_dp->dpcd[DP_DOWNSTREAMPORT_PRESENT] & | ||
599 | DP_DWN_STRM_PORT_PRESENT) | ||
600 | usleep_range(500, 600); | ||
601 | else | ||
602 | usleep_range(300, 400); | ||
592 | continue; | 603 | continue; |
593 | default: | 604 | default: |
594 | DRM_ERROR("aux_ch invalid native reply 0x%02x\n", | 605 | DRM_ERROR("aux_ch invalid native reply 0x%02x\n", |
@@ -1456,7 +1467,7 @@ static void intel_edp_psr_setup(struct intel_dp *intel_dp) | |||
1456 | 1467 | ||
1457 | /* Avoid continuous PSR exit by masking memup and hpd */ | 1468 | /* Avoid continuous PSR exit by masking memup and hpd */ |
1458 | I915_WRITE(EDP_PSR_DEBUG_CTL, EDP_PSR_DEBUG_MASK_MEMUP | | 1469 | I915_WRITE(EDP_PSR_DEBUG_CTL, EDP_PSR_DEBUG_MASK_MEMUP | |
1459 | EDP_PSR_DEBUG_MASK_HPD); | 1470 | EDP_PSR_DEBUG_MASK_HPD | EDP_PSR_DEBUG_MASK_LPSP); |
1460 | 1471 | ||
1461 | intel_dp->psr_setup_done = true; | 1472 | intel_dp->psr_setup_done = true; |
1462 | } | 1473 | } |
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h index a47799e832c6..9b7b68fd5d47 100644 --- a/drivers/gpu/drm/i915/intel_drv.h +++ b/drivers/gpu/drm/i915/intel_drv.h | |||
@@ -280,6 +280,7 @@ struct intel_crtc_config { | |||
280 | struct { | 280 | struct { |
281 | u32 pos; | 281 | u32 pos; |
282 | u32 size; | 282 | u32 size; |
283 | bool enabled; | ||
283 | } pch_pfit; | 284 | } pch_pfit; |
284 | 285 | ||
285 | /* FDI configuration, only valid if has_pch_encoder is set. */ | 286 | /* FDI configuration, only valid if has_pch_encoder is set. */ |
@@ -792,6 +793,5 @@ extern void hsw_pc8_disable_interrupts(struct drm_device *dev); | |||
792 | extern void hsw_pc8_restore_interrupts(struct drm_device *dev); | 793 | extern void hsw_pc8_restore_interrupts(struct drm_device *dev); |
793 | extern void intel_aux_display_runtime_get(struct drm_i915_private *dev_priv); | 794 | extern void intel_aux_display_runtime_get(struct drm_i915_private *dev_priv); |
794 | extern void intel_aux_display_runtime_put(struct drm_i915_private *dev_priv); | 795 | extern void intel_aux_display_runtime_put(struct drm_i915_private *dev_priv); |
795 | extern void i915_disable_vga_mem(struct drm_device *dev); | ||
796 | 796 | ||
797 | #endif /* __INTEL_DRV_H__ */ | 797 | #endif /* __INTEL_DRV_H__ */ |
diff --git a/drivers/gpu/drm/i915/intel_dvo.c b/drivers/gpu/drm/i915/intel_dvo.c index 406303b509c1..7fa7df546c1e 100644 --- a/drivers/gpu/drm/i915/intel_dvo.c +++ b/drivers/gpu/drm/i915/intel_dvo.c | |||
@@ -263,6 +263,8 @@ static bool intel_dvo_compute_config(struct intel_encoder *encoder, | |||
263 | C(vtotal); | 263 | C(vtotal); |
264 | C(clock); | 264 | C(clock); |
265 | #undef C | 265 | #undef C |
266 | |||
267 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
266 | } | 268 | } |
267 | 269 | ||
268 | if (intel_dvo->dev.dev_ops->mode_fixup) | 270 | if (intel_dvo->dev.dev_ops->mode_fixup) |
diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c index 42114ecbae0e..293564a2896a 100644 --- a/drivers/gpu/drm/i915/intel_panel.c +++ b/drivers/gpu/drm/i915/intel_panel.c | |||
@@ -112,6 +112,7 @@ intel_pch_panel_fitting(struct intel_crtc *intel_crtc, | |||
112 | done: | 112 | done: |
113 | pipe_config->pch_pfit.pos = (x << 16) | y; | 113 | pipe_config->pch_pfit.pos = (x << 16) | y; |
114 | pipe_config->pch_pfit.size = (width << 16) | height; | 114 | pipe_config->pch_pfit.size = (width << 16) | height; |
115 | pipe_config->pch_pfit.enabled = pipe_config->pch_pfit.size != 0; | ||
115 | } | 116 | } |
116 | 117 | ||
117 | static void | 118 | static void |
diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c index 0c115cc4899f..f4c5e95b2d6f 100644 --- a/drivers/gpu/drm/i915/intel_pm.c +++ b/drivers/gpu/drm/i915/intel_pm.c | |||
@@ -2096,16 +2096,16 @@ static uint32_t ilk_pipe_pixel_rate(struct drm_device *dev, | |||
2096 | struct drm_crtc *crtc) | 2096 | struct drm_crtc *crtc) |
2097 | { | 2097 | { |
2098 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 2098 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
2099 | uint32_t pixel_rate, pfit_size; | 2099 | uint32_t pixel_rate; |
2100 | 2100 | ||
2101 | pixel_rate = intel_crtc->config.adjusted_mode.clock; | 2101 | pixel_rate = intel_crtc->config.adjusted_mode.clock; |
2102 | 2102 | ||
2103 | /* We only use IF-ID interlacing. If we ever use PF-ID we'll need to | 2103 | /* We only use IF-ID interlacing. If we ever use PF-ID we'll need to |
2104 | * adjust the pixel_rate here. */ | 2104 | * adjust the pixel_rate here. */ |
2105 | 2105 | ||
2106 | pfit_size = intel_crtc->config.pch_pfit.size; | 2106 | if (intel_crtc->config.pch_pfit.enabled) { |
2107 | if (pfit_size) { | ||
2108 | uint64_t pipe_w, pipe_h, pfit_w, pfit_h; | 2107 | uint64_t pipe_w, pipe_h, pfit_w, pfit_h; |
2108 | uint32_t pfit_size = intel_crtc->config.pch_pfit.size; | ||
2109 | 2109 | ||
2110 | pipe_w = intel_crtc->config.requested_mode.hdisplay; | 2110 | pipe_w = intel_crtc->config.requested_mode.hdisplay; |
2111 | pipe_h = intel_crtc->config.requested_mode.vdisplay; | 2111 | pipe_h = intel_crtc->config.requested_mode.vdisplay; |
@@ -3864,8 +3864,6 @@ static void valleyview_enable_rps(struct drm_device *dev) | |||
3864 | dev_priv->rps.rpe_delay), | 3864 | dev_priv->rps.rpe_delay), |
3865 | dev_priv->rps.rpe_delay); | 3865 | dev_priv->rps.rpe_delay); |
3866 | 3866 | ||
3867 | INIT_DELAYED_WORK(&dev_priv->rps.vlv_work, vlv_rps_timer_work); | ||
3868 | |||
3869 | valleyview_set_rps(dev_priv->dev, dev_priv->rps.rpe_delay); | 3867 | valleyview_set_rps(dev_priv->dev, dev_priv->rps.rpe_delay); |
3870 | 3868 | ||
3871 | gen6_enable_rps_interrupts(dev); | 3869 | gen6_enable_rps_interrupts(dev); |
@@ -4955,6 +4953,11 @@ static void haswell_init_clock_gating(struct drm_device *dev) | |||
4955 | I915_WRITE(GEN7_L3_CHICKEN_MODE_REGISTER, | 4953 | I915_WRITE(GEN7_L3_CHICKEN_MODE_REGISTER, |
4956 | GEN7_WA_L3_CHICKEN_MODE); | 4954 | GEN7_WA_L3_CHICKEN_MODE); |
4957 | 4955 | ||
4956 | /* L3 caching of data atomics doesn't work -- disable it. */ | ||
4957 | I915_WRITE(HSW_SCRATCH1, HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE); | ||
4958 | I915_WRITE(HSW_ROW_CHICKEN3, | ||
4959 | _MASKED_BIT_ENABLE(HSW_ROW_CHICKEN3_L3_GLOBAL_ATOMICS_DISABLE)); | ||
4960 | |||
4958 | /* This is required by WaCatErrorRejectionIssue:hsw */ | 4961 | /* This is required by WaCatErrorRejectionIssue:hsw */ |
4959 | I915_WRITE(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG, | 4962 | I915_WRITE(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG, |
4960 | I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) | | 4963 | I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) | |
@@ -5681,5 +5684,7 @@ void intel_pm_init(struct drm_device *dev) | |||
5681 | 5684 | ||
5682 | INIT_DELAYED_WORK(&dev_priv->rps.delayed_resume_work, | 5685 | INIT_DELAYED_WORK(&dev_priv->rps.delayed_resume_work, |
5683 | intel_gen6_powersave_work); | 5686 | intel_gen6_powersave_work); |
5687 | |||
5688 | INIT_DELAYED_WORK(&dev_priv->rps.vlv_work, vlv_rps_timer_work); | ||
5684 | } | 5689 | } |
5685 | 5690 | ||
diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c index 85037b9d4934..49482fd5b76c 100644 --- a/drivers/gpu/drm/i915/intel_sdvo.c +++ b/drivers/gpu/drm/i915/intel_sdvo.c | |||
@@ -788,6 +788,8 @@ static void intel_sdvo_get_dtd_from_mode(struct intel_sdvo_dtd *dtd, | |||
788 | uint16_t h_sync_offset, v_sync_offset; | 788 | uint16_t h_sync_offset, v_sync_offset; |
789 | int mode_clock; | 789 | int mode_clock; |
790 | 790 | ||
791 | memset(dtd, 0, sizeof(*dtd)); | ||
792 | |||
791 | width = mode->hdisplay; | 793 | width = mode->hdisplay; |
792 | height = mode->vdisplay; | 794 | height = mode->vdisplay; |
793 | 795 | ||
@@ -830,44 +832,51 @@ static void intel_sdvo_get_dtd_from_mode(struct intel_sdvo_dtd *dtd, | |||
830 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) | 832 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) |
831 | dtd->part2.dtd_flags |= DTD_FLAG_VSYNC_POSITIVE; | 833 | dtd->part2.dtd_flags |= DTD_FLAG_VSYNC_POSITIVE; |
832 | 834 | ||
833 | dtd->part2.sdvo_flags = 0; | ||
834 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; | 835 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; |
835 | dtd->part2.reserved = 0; | ||
836 | } | 836 | } |
837 | 837 | ||
838 | static void intel_sdvo_get_mode_from_dtd(struct drm_display_mode * mode, | 838 | static void intel_sdvo_get_mode_from_dtd(struct drm_display_mode *pmode, |
839 | const struct intel_sdvo_dtd *dtd) | 839 | const struct intel_sdvo_dtd *dtd) |
840 | { | 840 | { |
841 | mode->hdisplay = dtd->part1.h_active; | 841 | struct drm_display_mode mode = {}; |
842 | mode->hdisplay += ((dtd->part1.h_high >> 4) & 0x0f) << 8; | 842 | |
843 | mode->hsync_start = mode->hdisplay + dtd->part2.h_sync_off; | 843 | mode.hdisplay = dtd->part1.h_active; |
844 | mode->hsync_start += (dtd->part2.sync_off_width_high & 0xc0) << 2; | 844 | mode.hdisplay += ((dtd->part1.h_high >> 4) & 0x0f) << 8; |
845 | mode->hsync_end = mode->hsync_start + dtd->part2.h_sync_width; | 845 | mode.hsync_start = mode.hdisplay + dtd->part2.h_sync_off; |
846 | mode->hsync_end += (dtd->part2.sync_off_width_high & 0x30) << 4; | 846 | mode.hsync_start += (dtd->part2.sync_off_width_high & 0xc0) << 2; |
847 | mode->htotal = mode->hdisplay + dtd->part1.h_blank; | 847 | mode.hsync_end = mode.hsync_start + dtd->part2.h_sync_width; |
848 | mode->htotal += (dtd->part1.h_high & 0xf) << 8; | 848 | mode.hsync_end += (dtd->part2.sync_off_width_high & 0x30) << 4; |
849 | 849 | mode.htotal = mode.hdisplay + dtd->part1.h_blank; | |
850 | mode->vdisplay = dtd->part1.v_active; | 850 | mode.htotal += (dtd->part1.h_high & 0xf) << 8; |
851 | mode->vdisplay += ((dtd->part1.v_high >> 4) & 0x0f) << 8; | 851 | |
852 | mode->vsync_start = mode->vdisplay; | 852 | mode.vdisplay = dtd->part1.v_active; |
853 | mode->vsync_start += (dtd->part2.v_sync_off_width >> 4) & 0xf; | 853 | mode.vdisplay += ((dtd->part1.v_high >> 4) & 0x0f) << 8; |
854 | mode->vsync_start += (dtd->part2.sync_off_width_high & 0x0c) << 2; | 854 | mode.vsync_start = mode.vdisplay; |
855 | mode->vsync_start += dtd->part2.v_sync_off_high & 0xc0; | 855 | mode.vsync_start += (dtd->part2.v_sync_off_width >> 4) & 0xf; |
856 | mode->vsync_end = mode->vsync_start + | 856 | mode.vsync_start += (dtd->part2.sync_off_width_high & 0x0c) << 2; |
857 | mode.vsync_start += dtd->part2.v_sync_off_high & 0xc0; | ||
858 | mode.vsync_end = mode.vsync_start + | ||
857 | (dtd->part2.v_sync_off_width & 0xf); | 859 | (dtd->part2.v_sync_off_width & 0xf); |
858 | mode->vsync_end += (dtd->part2.sync_off_width_high & 0x3) << 4; | 860 | mode.vsync_end += (dtd->part2.sync_off_width_high & 0x3) << 4; |
859 | mode->vtotal = mode->vdisplay + dtd->part1.v_blank; | 861 | mode.vtotal = mode.vdisplay + dtd->part1.v_blank; |
860 | mode->vtotal += (dtd->part1.v_high & 0xf) << 8; | 862 | mode.vtotal += (dtd->part1.v_high & 0xf) << 8; |
861 | 863 | ||
862 | mode->clock = dtd->part1.clock * 10; | 864 | mode.clock = dtd->part1.clock * 10; |
863 | 865 | ||
864 | mode->flags &= ~(DRM_MODE_FLAG_PHSYNC | DRM_MODE_FLAG_PVSYNC); | ||
865 | if (dtd->part2.dtd_flags & DTD_FLAG_INTERLACE) | 866 | if (dtd->part2.dtd_flags & DTD_FLAG_INTERLACE) |
866 | mode->flags |= DRM_MODE_FLAG_INTERLACE; | 867 | mode.flags |= DRM_MODE_FLAG_INTERLACE; |
867 | if (dtd->part2.dtd_flags & DTD_FLAG_HSYNC_POSITIVE) | 868 | if (dtd->part2.dtd_flags & DTD_FLAG_HSYNC_POSITIVE) |
868 | mode->flags |= DRM_MODE_FLAG_PHSYNC; | 869 | mode.flags |= DRM_MODE_FLAG_PHSYNC; |
870 | else | ||
871 | mode.flags |= DRM_MODE_FLAG_NHSYNC; | ||
869 | if (dtd->part2.dtd_flags & DTD_FLAG_VSYNC_POSITIVE) | 872 | if (dtd->part2.dtd_flags & DTD_FLAG_VSYNC_POSITIVE) |
870 | mode->flags |= DRM_MODE_FLAG_PVSYNC; | 873 | mode.flags |= DRM_MODE_FLAG_PVSYNC; |
874 | else | ||
875 | mode.flags |= DRM_MODE_FLAG_NVSYNC; | ||
876 | |||
877 | drm_mode_set_crtcinfo(&mode, 0); | ||
878 | |||
879 | drm_mode_copy(pmode, &mode); | ||
871 | } | 880 | } |
872 | 881 | ||
873 | static bool intel_sdvo_check_supp_encode(struct intel_sdvo *intel_sdvo) | 882 | static bool intel_sdvo_check_supp_encode(struct intel_sdvo *intel_sdvo) |
diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c index f2c6d7909ae2..dd6f84bf6c22 100644 --- a/drivers/gpu/drm/i915/intel_tv.c +++ b/drivers/gpu/drm/i915/intel_tv.c | |||
@@ -916,6 +916,14 @@ intel_tv_compute_config(struct intel_encoder *encoder, | |||
916 | DRM_DEBUG_KMS("forcing bpc to 8 for TV\n"); | 916 | DRM_DEBUG_KMS("forcing bpc to 8 for TV\n"); |
917 | pipe_config->pipe_bpp = 8*3; | 917 | pipe_config->pipe_bpp = 8*3; |
918 | 918 | ||
919 | /* TV has it's own notion of sync and other mode flags, so clear them. */ | ||
920 | pipe_config->adjusted_mode.flags = 0; | ||
921 | |||
922 | /* | ||
923 | * FIXME: We don't check whether the input mode is actually what we want | ||
924 | * or whether userspace is doing something stupid. | ||
925 | */ | ||
926 | |||
919 | return true; | 927 | return true; |
920 | } | 928 | } |
921 | 929 | ||
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c index a60584763b61..a0b9d8a95b16 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c | |||
@@ -124,6 +124,8 @@ void adreno_recover(struct msm_gpu *gpu) | |||
124 | 124 | ||
125 | /* reset completed fence seqno, just discard anything pending: */ | 125 | /* reset completed fence seqno, just discard anything pending: */ |
126 | adreno_gpu->memptrs->fence = gpu->submitted_fence; | 126 | adreno_gpu->memptrs->fence = gpu->submitted_fence; |
127 | adreno_gpu->memptrs->rptr = 0; | ||
128 | adreno_gpu->memptrs->wptr = 0; | ||
127 | 129 | ||
128 | gpu->funcs->pm_resume(gpu); | 130 | gpu->funcs->pm_resume(gpu); |
129 | ret = gpu->funcs->hw_init(gpu); | 131 | ret = gpu->funcs->hw_init(gpu); |
@@ -229,7 +231,7 @@ void adreno_idle(struct msm_gpu *gpu) | |||
229 | return; | 231 | return; |
230 | } while(time_before(jiffies, t)); | 232 | } while(time_before(jiffies, t)); |
231 | 233 | ||
232 | DRM_ERROR("timeout waiting for %s to drain ringbuffer!\n", gpu->name); | 234 | DRM_ERROR("%s: timeout waiting to drain ringbuffer!\n", gpu->name); |
233 | 235 | ||
234 | /* TODO maybe we need to reset GPU here to recover from hang? */ | 236 | /* TODO maybe we need to reset GPU here to recover from hang? */ |
235 | } | 237 | } |
@@ -256,11 +258,17 @@ void adreno_wait_ring(struct msm_gpu *gpu, uint32_t ndwords) | |||
256 | { | 258 | { |
257 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | 259 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); |
258 | uint32_t freedwords; | 260 | uint32_t freedwords; |
261 | unsigned long t = jiffies + ADRENO_IDLE_TIMEOUT; | ||
259 | do { | 262 | do { |
260 | uint32_t size = gpu->rb->size / 4; | 263 | uint32_t size = gpu->rb->size / 4; |
261 | uint32_t wptr = get_wptr(gpu->rb); | 264 | uint32_t wptr = get_wptr(gpu->rb); |
262 | uint32_t rptr = adreno_gpu->memptrs->rptr; | 265 | uint32_t rptr = adreno_gpu->memptrs->rptr; |
263 | freedwords = (rptr + (size - 1) - wptr) % size; | 266 | freedwords = (rptr + (size - 1) - wptr) % size; |
267 | |||
268 | if (time_after(jiffies, t)) { | ||
269 | DRM_ERROR("%s: timeout waiting for ringbuffer space\n", gpu->name); | ||
270 | break; | ||
271 | } | ||
264 | } while(freedwords < ndwords); | 272 | } while(freedwords < ndwords); |
265 | } | 273 | } |
266 | 274 | ||
diff --git a/drivers/gpu/drm/msm/mdp4/mdp4_kms.c b/drivers/gpu/drm/msm/mdp4/mdp4_kms.c index 5db5bbaedae2..bc7fd11ad8be 100644 --- a/drivers/gpu/drm/msm/mdp4/mdp4_kms.c +++ b/drivers/gpu/drm/msm/mdp4/mdp4_kms.c | |||
@@ -19,8 +19,6 @@ | |||
19 | #include "msm_drv.h" | 19 | #include "msm_drv.h" |
20 | #include "mdp4_kms.h" | 20 | #include "mdp4_kms.h" |
21 | 21 | ||
22 | #include <mach/iommu.h> | ||
23 | |||
24 | static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev); | 22 | static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev); |
25 | 23 | ||
26 | static int mdp4_hw_init(struct msm_kms *kms) | 24 | static int mdp4_hw_init(struct msm_kms *kms) |
diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c index 864c9773636b..b3a2f1629041 100644 --- a/drivers/gpu/drm/msm/msm_drv.c +++ b/drivers/gpu/drm/msm/msm_drv.c | |||
@@ -18,8 +18,6 @@ | |||
18 | #include "msm_drv.h" | 18 | #include "msm_drv.h" |
19 | #include "msm_gpu.h" | 19 | #include "msm_gpu.h" |
20 | 20 | ||
21 | #include <mach/iommu.h> | ||
22 | |||
23 | static void msm_fb_output_poll_changed(struct drm_device *dev) | 21 | static void msm_fb_output_poll_changed(struct drm_device *dev) |
24 | { | 22 | { |
25 | struct msm_drm_private *priv = dev->dev_private; | 23 | struct msm_drm_private *priv = dev->dev_private; |
@@ -62,6 +60,8 @@ int msm_iommu_attach(struct drm_device *dev, struct iommu_domain *iommu, | |||
62 | int i, ret; | 60 | int i, ret; |
63 | 61 | ||
64 | for (i = 0; i < cnt; i++) { | 62 | for (i = 0; i < cnt; i++) { |
63 | /* TODO maybe some day msm iommu won't require this hack: */ | ||
64 | struct device *msm_iommu_get_ctx(const char *ctx_name); | ||
65 | struct device *ctx = msm_iommu_get_ctx(names[i]); | 65 | struct device *ctx = msm_iommu_get_ctx(names[i]); |
66 | if (!ctx) | 66 | if (!ctx) |
67 | continue; | 67 | continue; |
@@ -199,7 +199,7 @@ static int msm_load(struct drm_device *dev, unsigned long flags) | |||
199 | * imx drm driver on iMX5 | 199 | * imx drm driver on iMX5 |
200 | */ | 200 | */ |
201 | dev_err(dev->dev, "failed to load kms\n"); | 201 | dev_err(dev->dev, "failed to load kms\n"); |
202 | ret = PTR_ERR(priv->kms); | 202 | ret = PTR_ERR(kms); |
203 | goto fail; | 203 | goto fail; |
204 | } | 204 | } |
205 | 205 | ||
@@ -499,25 +499,41 @@ int msm_wait_fence_interruptable(struct drm_device *dev, uint32_t fence, | |||
499 | struct timespec *timeout) | 499 | struct timespec *timeout) |
500 | { | 500 | { |
501 | struct msm_drm_private *priv = dev->dev_private; | 501 | struct msm_drm_private *priv = dev->dev_private; |
502 | unsigned long timeout_jiffies = timespec_to_jiffies(timeout); | ||
503 | unsigned long start_jiffies = jiffies; | ||
504 | unsigned long remaining_jiffies; | ||
505 | int ret; | 502 | int ret; |
506 | 503 | ||
507 | if (time_after(start_jiffies, timeout_jiffies)) | 504 | if (!priv->gpu) |
508 | remaining_jiffies = 0; | 505 | return 0; |
509 | else | 506 | |
510 | remaining_jiffies = timeout_jiffies - start_jiffies; | 507 | if (fence > priv->gpu->submitted_fence) { |
511 | 508 | DRM_ERROR("waiting on invalid fence: %u (of %u)\n", | |
512 | ret = wait_event_interruptible_timeout(priv->fence_event, | 509 | fence, priv->gpu->submitted_fence); |
513 | priv->completed_fence >= fence, | 510 | return -EINVAL; |
514 | remaining_jiffies); | 511 | } |
515 | if (ret == 0) { | 512 | |
516 | DBG("timeout waiting for fence: %u (completed: %u)", | 513 | if (!timeout) { |
517 | fence, priv->completed_fence); | 514 | /* no-wait: */ |
518 | ret = -ETIMEDOUT; | 515 | ret = fence_completed(dev, fence) ? 0 : -EBUSY; |
519 | } else if (ret != -ERESTARTSYS) { | 516 | } else { |
520 | ret = 0; | 517 | unsigned long timeout_jiffies = timespec_to_jiffies(timeout); |
518 | unsigned long start_jiffies = jiffies; | ||
519 | unsigned long remaining_jiffies; | ||
520 | |||
521 | if (time_after(start_jiffies, timeout_jiffies)) | ||
522 | remaining_jiffies = 0; | ||
523 | else | ||
524 | remaining_jiffies = timeout_jiffies - start_jiffies; | ||
525 | |||
526 | ret = wait_event_interruptible_timeout(priv->fence_event, | ||
527 | fence_completed(dev, fence), | ||
528 | remaining_jiffies); | ||
529 | |||
530 | if (ret == 0) { | ||
531 | DBG("timeout waiting for fence: %u (completed: %u)", | ||
532 | fence, priv->completed_fence); | ||
533 | ret = -ETIMEDOUT; | ||
534 | } else if (ret != -ERESTARTSYS) { | ||
535 | ret = 0; | ||
536 | } | ||
521 | } | 537 | } |
522 | 538 | ||
523 | return ret; | 539 | return ret; |
@@ -681,7 +697,7 @@ static struct drm_driver msm_driver = { | |||
681 | .gem_vm_ops = &vm_ops, | 697 | .gem_vm_ops = &vm_ops, |
682 | .dumb_create = msm_gem_dumb_create, | 698 | .dumb_create = msm_gem_dumb_create, |
683 | .dumb_map_offset = msm_gem_dumb_map_offset, | 699 | .dumb_map_offset = msm_gem_dumb_map_offset, |
684 | .dumb_destroy = msm_gem_dumb_destroy, | 700 | .dumb_destroy = drm_gem_dumb_destroy, |
685 | #ifdef CONFIG_DEBUG_FS | 701 | #ifdef CONFIG_DEBUG_FS |
686 | .debugfs_init = msm_debugfs_init, | 702 | .debugfs_init = msm_debugfs_init, |
687 | .debugfs_cleanup = msm_debugfs_cleanup, | 703 | .debugfs_cleanup = msm_debugfs_cleanup, |
diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h index 80d75094bf0a..df8f1d084bc1 100644 --- a/drivers/gpu/drm/msm/msm_drv.h +++ b/drivers/gpu/drm/msm/msm_drv.h | |||
@@ -153,7 +153,7 @@ void *msm_gem_vaddr(struct drm_gem_object *obj); | |||
153 | int msm_gem_queue_inactive_work(struct drm_gem_object *obj, | 153 | int msm_gem_queue_inactive_work(struct drm_gem_object *obj, |
154 | struct work_struct *work); | 154 | struct work_struct *work); |
155 | void msm_gem_move_to_active(struct drm_gem_object *obj, | 155 | void msm_gem_move_to_active(struct drm_gem_object *obj, |
156 | struct msm_gpu *gpu, uint32_t fence); | 156 | struct msm_gpu *gpu, bool write, uint32_t fence); |
157 | void msm_gem_move_to_inactive(struct drm_gem_object *obj); | 157 | void msm_gem_move_to_inactive(struct drm_gem_object *obj); |
158 | int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, | 158 | int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, |
159 | struct timespec *timeout); | 159 | struct timespec *timeout); |
@@ -191,6 +191,12 @@ u32 msm_readl(const void __iomem *addr); | |||
191 | #define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) | 191 | #define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) |
192 | #define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) | 192 | #define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) |
193 | 193 | ||
194 | static inline bool fence_completed(struct drm_device *dev, uint32_t fence) | ||
195 | { | ||
196 | struct msm_drm_private *priv = dev->dev_private; | ||
197 | return priv->completed_fence >= fence; | ||
198 | } | ||
199 | |||
194 | static inline int align_pitch(int width, int bpp) | 200 | static inline int align_pitch(int width, int bpp) |
195 | { | 201 | { |
196 | int bytespp = (bpp + 7) / 8; | 202 | int bytespp = (bpp + 7) / 8; |
diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c index 6b5a6c8c7658..2bae46c66a30 100644 --- a/drivers/gpu/drm/msm/msm_gem.c +++ b/drivers/gpu/drm/msm/msm_gem.c | |||
@@ -40,9 +40,9 @@ static struct page **get_pages(struct drm_gem_object *obj) | |||
40 | } | 40 | } |
41 | 41 | ||
42 | msm_obj->sgt = drm_prime_pages_to_sg(p, npages); | 42 | msm_obj->sgt = drm_prime_pages_to_sg(p, npages); |
43 | if (!msm_obj->sgt) { | 43 | if (IS_ERR(msm_obj->sgt)) { |
44 | dev_err(dev->dev, "failed to allocate sgt\n"); | 44 | dev_err(dev->dev, "failed to allocate sgt\n"); |
45 | return ERR_PTR(-ENOMEM); | 45 | return ERR_CAST(msm_obj->sgt); |
46 | } | 46 | } |
47 | 47 | ||
48 | msm_obj->pages = p; | 48 | msm_obj->pages = p; |
@@ -159,7 +159,6 @@ out_unlock: | |||
159 | out: | 159 | out: |
160 | switch (ret) { | 160 | switch (ret) { |
161 | case -EAGAIN: | 161 | case -EAGAIN: |
162 | set_need_resched(); | ||
163 | case 0: | 162 | case 0: |
164 | case -ERESTARTSYS: | 163 | case -ERESTARTSYS: |
165 | case -EINTR: | 164 | case -EINTR: |
@@ -320,13 +319,6 @@ int msm_gem_dumb_create(struct drm_file *file, struct drm_device *dev, | |||
320 | MSM_BO_SCANOUT | MSM_BO_WC, &args->handle); | 319 | MSM_BO_SCANOUT | MSM_BO_WC, &args->handle); |
321 | } | 320 | } |
322 | 321 | ||
323 | int msm_gem_dumb_destroy(struct drm_file *file, struct drm_device *dev, | ||
324 | uint32_t handle) | ||
325 | { | ||
326 | /* No special work needed, drop the reference and see what falls out */ | ||
327 | return drm_gem_handle_delete(file, handle); | ||
328 | } | ||
329 | |||
330 | int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, | 322 | int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, |
331 | uint32_t handle, uint64_t *offset) | 323 | uint32_t handle, uint64_t *offset) |
332 | { | 324 | { |
@@ -393,11 +385,14 @@ int msm_gem_queue_inactive_work(struct drm_gem_object *obj, | |||
393 | } | 385 | } |
394 | 386 | ||
395 | void msm_gem_move_to_active(struct drm_gem_object *obj, | 387 | void msm_gem_move_to_active(struct drm_gem_object *obj, |
396 | struct msm_gpu *gpu, uint32_t fence) | 388 | struct msm_gpu *gpu, bool write, uint32_t fence) |
397 | { | 389 | { |
398 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 390 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
399 | msm_obj->gpu = gpu; | 391 | msm_obj->gpu = gpu; |
400 | msm_obj->fence = fence; | 392 | if (write) |
393 | msm_obj->write_fence = fence; | ||
394 | else | ||
395 | msm_obj->read_fence = fence; | ||
401 | list_del_init(&msm_obj->mm_list); | 396 | list_del_init(&msm_obj->mm_list); |
402 | list_add_tail(&msm_obj->mm_list, &gpu->active_list); | 397 | list_add_tail(&msm_obj->mm_list, &gpu->active_list); |
403 | } | 398 | } |
@@ -411,7 +406,8 @@ void msm_gem_move_to_inactive(struct drm_gem_object *obj) | |||
411 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); | 406 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); |
412 | 407 | ||
413 | msm_obj->gpu = NULL; | 408 | msm_obj->gpu = NULL; |
414 | msm_obj->fence = 0; | 409 | msm_obj->read_fence = 0; |
410 | msm_obj->write_fence = 0; | ||
415 | list_del_init(&msm_obj->mm_list); | 411 | list_del_init(&msm_obj->mm_list); |
416 | list_add_tail(&msm_obj->mm_list, &priv->inactive_list); | 412 | list_add_tail(&msm_obj->mm_list, &priv->inactive_list); |
417 | 413 | ||
@@ -433,8 +429,18 @@ int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, | |||
433 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 429 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
434 | int ret = 0; | 430 | int ret = 0; |
435 | 431 | ||
436 | if (is_active(msm_obj) && !(op & MSM_PREP_NOSYNC)) | 432 | if (is_active(msm_obj)) { |
437 | ret = msm_wait_fence_interruptable(dev, msm_obj->fence, timeout); | 433 | uint32_t fence = 0; |
434 | |||
435 | if (op & MSM_PREP_READ) | ||
436 | fence = msm_obj->write_fence; | ||
437 | if (op & MSM_PREP_WRITE) | ||
438 | fence = max(fence, msm_obj->read_fence); | ||
439 | if (op & MSM_PREP_NOSYNC) | ||
440 | timeout = NULL; | ||
441 | |||
442 | ret = msm_wait_fence_interruptable(dev, fence, timeout); | ||
443 | } | ||
438 | 444 | ||
439 | /* TODO cache maintenance */ | 445 | /* TODO cache maintenance */ |
440 | 446 | ||
@@ -455,9 +461,10 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m) | |||
455 | uint64_t off = drm_vma_node_start(&obj->vma_node); | 461 | uint64_t off = drm_vma_node_start(&obj->vma_node); |
456 | 462 | ||
457 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); | 463 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); |
458 | seq_printf(m, "%08x: %c(%d) %2d (%2d) %08llx %p %d\n", | 464 | seq_printf(m, "%08x: %c(r=%u,w=%u) %2d (%2d) %08llx %p %d\n", |
459 | msm_obj->flags, is_active(msm_obj) ? 'A' : 'I', | 465 | msm_obj->flags, is_active(msm_obj) ? 'A' : 'I', |
460 | msm_obj->fence, obj->name, obj->refcount.refcount.counter, | 466 | msm_obj->read_fence, msm_obj->write_fence, |
467 | obj->name, obj->refcount.refcount.counter, | ||
461 | off, msm_obj->vaddr, obj->size); | 468 | off, msm_obj->vaddr, obj->size); |
462 | } | 469 | } |
463 | 470 | ||
diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h index d746f13d283c..0676f32e2c6a 100644 --- a/drivers/gpu/drm/msm/msm_gem.h +++ b/drivers/gpu/drm/msm/msm_gem.h | |||
@@ -36,7 +36,7 @@ struct msm_gem_object { | |||
36 | */ | 36 | */ |
37 | struct list_head mm_list; | 37 | struct list_head mm_list; |
38 | struct msm_gpu *gpu; /* non-null if active */ | 38 | struct msm_gpu *gpu; /* non-null if active */ |
39 | uint32_t fence; | 39 | uint32_t read_fence, write_fence; |
40 | 40 | ||
41 | /* Transiently in the process of submit ioctl, objects associated | 41 | /* Transiently in the process of submit ioctl, objects associated |
42 | * with the submit are on submit->bo_list.. this only lasts for | 42 | * with the submit are on submit->bo_list.. this only lasts for |
diff --git a/drivers/gpu/drm/msm/msm_gem_submit.c b/drivers/gpu/drm/msm/msm_gem_submit.c index 3e1ef3a00f60..5281d4bc37f7 100644 --- a/drivers/gpu/drm/msm/msm_gem_submit.c +++ b/drivers/gpu/drm/msm/msm_gem_submit.c | |||
@@ -78,7 +78,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
78 | } | 78 | } |
79 | 79 | ||
80 | if (submit_bo.flags & BO_INVALID_FLAGS) { | 80 | if (submit_bo.flags & BO_INVALID_FLAGS) { |
81 | DBG("invalid flags: %x", submit_bo.flags); | 81 | DRM_ERROR("invalid flags: %x\n", submit_bo.flags); |
82 | ret = -EINVAL; | 82 | ret = -EINVAL; |
83 | goto out_unlock; | 83 | goto out_unlock; |
84 | } | 84 | } |
@@ -92,7 +92,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
92 | */ | 92 | */ |
93 | obj = idr_find(&file->object_idr, submit_bo.handle); | 93 | obj = idr_find(&file->object_idr, submit_bo.handle); |
94 | if (!obj) { | 94 | if (!obj) { |
95 | DBG("invalid handle %u at index %u", submit_bo.handle, i); | 95 | DRM_ERROR("invalid handle %u at index %u\n", submit_bo.handle, i); |
96 | ret = -EINVAL; | 96 | ret = -EINVAL; |
97 | goto out_unlock; | 97 | goto out_unlock; |
98 | } | 98 | } |
@@ -100,7 +100,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
100 | msm_obj = to_msm_bo(obj); | 100 | msm_obj = to_msm_bo(obj); |
101 | 101 | ||
102 | if (!list_empty(&msm_obj->submit_entry)) { | 102 | if (!list_empty(&msm_obj->submit_entry)) { |
103 | DBG("handle %u at index %u already on submit list", | 103 | DRM_ERROR("handle %u at index %u already on submit list\n", |
104 | submit_bo.handle, i); | 104 | submit_bo.handle, i); |
105 | ret = -EINVAL; | 105 | ret = -EINVAL; |
106 | goto out_unlock; | 106 | goto out_unlock; |
@@ -216,8 +216,9 @@ static int submit_bo(struct msm_gem_submit *submit, uint32_t idx, | |||
216 | struct msm_gem_object **obj, uint32_t *iova, bool *valid) | 216 | struct msm_gem_object **obj, uint32_t *iova, bool *valid) |
217 | { | 217 | { |
218 | if (idx >= submit->nr_bos) { | 218 | if (idx >= submit->nr_bos) { |
219 | DBG("invalid buffer index: %u (out of %u)", idx, submit->nr_bos); | 219 | DRM_ERROR("invalid buffer index: %u (out of %u)\n", |
220 | return EINVAL; | 220 | idx, submit->nr_bos); |
221 | return -EINVAL; | ||
221 | } | 222 | } |
222 | 223 | ||
223 | if (obj) | 224 | if (obj) |
@@ -239,7 +240,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
239 | int ret; | 240 | int ret; |
240 | 241 | ||
241 | if (offset % 4) { | 242 | if (offset % 4) { |
242 | DBG("non-aligned cmdstream buffer: %u", offset); | 243 | DRM_ERROR("non-aligned cmdstream buffer: %u\n", offset); |
243 | return -EINVAL; | 244 | return -EINVAL; |
244 | } | 245 | } |
245 | 246 | ||
@@ -266,7 +267,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
266 | return -EFAULT; | 267 | return -EFAULT; |
267 | 268 | ||
268 | if (submit_reloc.submit_offset % 4) { | 269 | if (submit_reloc.submit_offset % 4) { |
269 | DBG("non-aligned reloc offset: %u", | 270 | DRM_ERROR("non-aligned reloc offset: %u\n", |
270 | submit_reloc.submit_offset); | 271 | submit_reloc.submit_offset); |
271 | return -EINVAL; | 272 | return -EINVAL; |
272 | } | 273 | } |
@@ -276,7 +277,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
276 | 277 | ||
277 | if ((off >= (obj->base.size / 4)) || | 278 | if ((off >= (obj->base.size / 4)) || |
278 | (off < last_offset)) { | 279 | (off < last_offset)) { |
279 | DBG("invalid offset %u at reloc %u", off, i); | 280 | DRM_ERROR("invalid offset %u at reloc %u\n", off, i); |
280 | return -EINVAL; | 281 | return -EINVAL; |
281 | } | 282 | } |
282 | 283 | ||
@@ -374,14 +375,15 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data, | |||
374 | goto out; | 375 | goto out; |
375 | 376 | ||
376 | if (submit_cmd.size % 4) { | 377 | if (submit_cmd.size % 4) { |
377 | DBG("non-aligned cmdstream buffer size: %u", | 378 | DRM_ERROR("non-aligned cmdstream buffer size: %u\n", |
378 | submit_cmd.size); | 379 | submit_cmd.size); |
379 | ret = -EINVAL; | 380 | ret = -EINVAL; |
380 | goto out; | 381 | goto out; |
381 | } | 382 | } |
382 | 383 | ||
383 | if (submit_cmd.size >= msm_obj->base.size) { | 384 | if ((submit_cmd.size + submit_cmd.submit_offset) >= |
384 | DBG("invalid cmdstream size: %u", submit_cmd.size); | 385 | msm_obj->base.size) { |
386 | DRM_ERROR("invalid cmdstream size: %u\n", submit_cmd.size); | ||
385 | ret = -EINVAL; | 387 | ret = -EINVAL; |
386 | goto out; | 388 | goto out; |
387 | } | 389 | } |
diff --git a/drivers/gpu/drm/msm/msm_gpu.c b/drivers/gpu/drm/msm/msm_gpu.c index e1e1ec9321ff..3bab937965d1 100644 --- a/drivers/gpu/drm/msm/msm_gpu.c +++ b/drivers/gpu/drm/msm/msm_gpu.c | |||
@@ -29,13 +29,14 @@ | |||
29 | static void bs_init(struct msm_gpu *gpu, struct platform_device *pdev) | 29 | static void bs_init(struct msm_gpu *gpu, struct platform_device *pdev) |
30 | { | 30 | { |
31 | struct drm_device *dev = gpu->dev; | 31 | struct drm_device *dev = gpu->dev; |
32 | struct kgsl_device_platform_data *pdata = pdev->dev.platform_data; | 32 | struct kgsl_device_platform_data *pdata; |
33 | 33 | ||
34 | if (!pdev) { | 34 | if (!pdev) { |
35 | dev_err(dev->dev, "could not find dtv pdata\n"); | 35 | dev_err(dev->dev, "could not find dtv pdata\n"); |
36 | return; | 36 | return; |
37 | } | 37 | } |
38 | 38 | ||
39 | pdata = pdev->dev.platform_data; | ||
39 | if (pdata->bus_scale_table) { | 40 | if (pdata->bus_scale_table) { |
40 | gpu->bsc = msm_bus_scale_register_client(pdata->bus_scale_table); | 41 | gpu->bsc = msm_bus_scale_register_client(pdata->bus_scale_table); |
41 | DBG("bus scale client: %08x", gpu->bsc); | 42 | DBG("bus scale client: %08x", gpu->bsc); |
@@ -230,6 +231,8 @@ static void hangcheck_timer_reset(struct msm_gpu *gpu) | |||
230 | static void hangcheck_handler(unsigned long data) | 231 | static void hangcheck_handler(unsigned long data) |
231 | { | 232 | { |
232 | struct msm_gpu *gpu = (struct msm_gpu *)data; | 233 | struct msm_gpu *gpu = (struct msm_gpu *)data; |
234 | struct drm_device *dev = gpu->dev; | ||
235 | struct msm_drm_private *priv = dev->dev_private; | ||
233 | uint32_t fence = gpu->funcs->last_fence(gpu); | 236 | uint32_t fence = gpu->funcs->last_fence(gpu); |
234 | 237 | ||
235 | if (fence != gpu->hangcheck_fence) { | 238 | if (fence != gpu->hangcheck_fence) { |
@@ -237,14 +240,22 @@ static void hangcheck_handler(unsigned long data) | |||
237 | gpu->hangcheck_fence = fence; | 240 | gpu->hangcheck_fence = fence; |
238 | } else if (fence < gpu->submitted_fence) { | 241 | } else if (fence < gpu->submitted_fence) { |
239 | /* no progress and not done.. hung! */ | 242 | /* no progress and not done.. hung! */ |
240 | struct msm_drm_private *priv = gpu->dev->dev_private; | ||
241 | gpu->hangcheck_fence = fence; | 243 | gpu->hangcheck_fence = fence; |
244 | dev_err(dev->dev, "%s: hangcheck detected gpu lockup!\n", | ||
245 | gpu->name); | ||
246 | dev_err(dev->dev, "%s: completed fence: %u\n", | ||
247 | gpu->name, fence); | ||
248 | dev_err(dev->dev, "%s: submitted fence: %u\n", | ||
249 | gpu->name, gpu->submitted_fence); | ||
242 | queue_work(priv->wq, &gpu->recover_work); | 250 | queue_work(priv->wq, &gpu->recover_work); |
243 | } | 251 | } |
244 | 252 | ||
245 | /* if still more pending work, reset the hangcheck timer: */ | 253 | /* if still more pending work, reset the hangcheck timer: */ |
246 | if (gpu->submitted_fence > gpu->hangcheck_fence) | 254 | if (gpu->submitted_fence > gpu->hangcheck_fence) |
247 | hangcheck_timer_reset(gpu); | 255 | hangcheck_timer_reset(gpu); |
256 | |||
257 | /* workaround for missing irq: */ | ||
258 | queue_work(priv->wq, &gpu->retire_work); | ||
248 | } | 259 | } |
249 | 260 | ||
250 | /* | 261 | /* |
@@ -265,7 +276,8 @@ static void retire_worker(struct work_struct *work) | |||
265 | obj = list_first_entry(&gpu->active_list, | 276 | obj = list_first_entry(&gpu->active_list, |
266 | struct msm_gem_object, mm_list); | 277 | struct msm_gem_object, mm_list); |
267 | 278 | ||
268 | if (obj->fence <= fence) { | 279 | if ((obj->read_fence <= fence) && |
280 | (obj->write_fence <= fence)) { | ||
269 | /* move to inactive: */ | 281 | /* move to inactive: */ |
270 | msm_gem_move_to_inactive(&obj->base); | 282 | msm_gem_move_to_inactive(&obj->base); |
271 | msm_gem_put_iova(&obj->base, gpu->id); | 283 | msm_gem_put_iova(&obj->base, gpu->id); |
@@ -321,7 +333,11 @@ int msm_gpu_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, | |||
321 | submit->gpu->id, &iova); | 333 | submit->gpu->id, &iova); |
322 | } | 334 | } |
323 | 335 | ||
324 | msm_gem_move_to_active(&msm_obj->base, gpu, submit->fence); | 336 | if (submit->bos[i].flags & MSM_SUBMIT_BO_READ) |
337 | msm_gem_move_to_active(&msm_obj->base, gpu, false, submit->fence); | ||
338 | |||
339 | if (submit->bos[i].flags & MSM_SUBMIT_BO_WRITE) | ||
340 | msm_gem_move_to_active(&msm_obj->base, gpu, true, submit->fence); | ||
325 | } | 341 | } |
326 | hangcheck_timer_reset(gpu); | 342 | hangcheck_timer_reset(gpu); |
327 | mutex_unlock(&dev->struct_mutex); | 343 | mutex_unlock(&dev->struct_mutex); |
diff --git a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c index 2e11ea02cf87..57cda2a1437b 100644 --- a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c +++ b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c | |||
@@ -579,8 +579,22 @@ static void | |||
579 | init_reserved(struct nvbios_init *init) | 579 | init_reserved(struct nvbios_init *init) |
580 | { | 580 | { |
581 | u8 opcode = nv_ro08(init->bios, init->offset); | 581 | u8 opcode = nv_ro08(init->bios, init->offset); |
582 | trace("RESERVED\t0x%02x\n", opcode); | 582 | u8 length, i; |
583 | init->offset += 1; | 583 | |
584 | switch (opcode) { | ||
585 | case 0xaa: | ||
586 | length = 4; | ||
587 | break; | ||
588 | default: | ||
589 | length = 1; | ||
590 | break; | ||
591 | } | ||
592 | |||
593 | trace("RESERVED 0x%02x\t", opcode); | ||
594 | for (i = 1; i < length; i++) | ||
595 | cont(" 0x%02x", nv_ro08(init->bios, init->offset + i)); | ||
596 | cont("\n"); | ||
597 | init->offset += length; | ||
584 | } | 598 | } |
585 | 599 | ||
586 | /** | 600 | /** |
@@ -1437,7 +1451,7 @@ init_configure_mem(struct nvbios_init *init) | |||
1437 | data = init_rdvgai(init, 0x03c4, 0x01); | 1451 | data = init_rdvgai(init, 0x03c4, 0x01); |
1438 | init_wrvgai(init, 0x03c4, 0x01, data | 0x20); | 1452 | init_wrvgai(init, 0x03c4, 0x01, data | 0x20); |
1439 | 1453 | ||
1440 | while ((addr = nv_ro32(bios, sdata)) != 0xffffffff) { | 1454 | for (; (addr = nv_ro32(bios, sdata)) != 0xffffffff; sdata += 4) { |
1441 | switch (addr) { | 1455 | switch (addr) { |
1442 | case 0x10021c: /* CKE_NORMAL */ | 1456 | case 0x10021c: /* CKE_NORMAL */ |
1443 | case 0x1002d0: /* CMD_REFRESH */ | 1457 | case 0x1002d0: /* CMD_REFRESH */ |
@@ -2135,6 +2149,7 @@ static struct nvbios_init_opcode { | |||
2135 | [0x99] = { init_zm_auxch }, | 2149 | [0x99] = { init_zm_auxch }, |
2136 | [0x9a] = { init_i2c_long_if }, | 2150 | [0x9a] = { init_i2c_long_if }, |
2137 | [0xa9] = { init_gpio_ne }, | 2151 | [0xa9] = { init_gpio_ne }, |
2152 | [0xaa] = { init_reserved }, | ||
2138 | }; | 2153 | }; |
2139 | 2154 | ||
2140 | #define init_opcode_nr (sizeof(init_opcode) / sizeof(init_opcode[0])) | 2155 | #define init_opcode_nr (sizeof(init_opcode) / sizeof(init_opcode[0])) |
diff --git a/drivers/gpu/drm/nouveau/core/subdev/mc/base.c b/drivers/gpu/drm/nouveau/core/subdev/mc/base.c index 37712a6df923..e290cfa4acee 100644 --- a/drivers/gpu/drm/nouveau/core/subdev/mc/base.c +++ b/drivers/gpu/drm/nouveau/core/subdev/mc/base.c | |||
@@ -113,7 +113,7 @@ nouveau_mc_create_(struct nouveau_object *parent, struct nouveau_object *engine, | |||
113 | pmc->use_msi = false; | 113 | pmc->use_msi = false; |
114 | break; | 114 | break; |
115 | default: | 115 | default: |
116 | pmc->use_msi = nouveau_boolopt(device->cfgopt, "NvMSI", true); | 116 | pmc->use_msi = nouveau_boolopt(device->cfgopt, "NvMSI", false); |
117 | if (pmc->use_msi) { | 117 | if (pmc->use_msi) { |
118 | pmc->use_msi = pci_enable_msi(device->pdev) == 0; | 118 | pmc->use_msi = pci_enable_msi(device->pdev) == 0; |
119 | if (pmc->use_msi) { | 119 | if (pmc->use_msi) { |
diff --git a/drivers/gpu/drm/nouveau/nouveau_display.c b/drivers/gpu/drm/nouveau/nouveau_display.c index d2712e6e5d31..7848590f5568 100644 --- a/drivers/gpu/drm/nouveau/nouveau_display.c +++ b/drivers/gpu/drm/nouveau/nouveau_display.c | |||
@@ -278,7 +278,6 @@ nouveau_display_create(struct drm_device *dev) | |||
278 | { | 278 | { |
279 | struct nouveau_drm *drm = nouveau_drm(dev); | 279 | struct nouveau_drm *drm = nouveau_drm(dev); |
280 | struct nouveau_display *disp; | 280 | struct nouveau_display *disp; |
281 | u32 pclass = dev->pdev->class >> 8; | ||
282 | int ret, gen; | 281 | int ret, gen; |
283 | 282 | ||
284 | disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL); | 283 | disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL); |
@@ -340,29 +339,25 @@ nouveau_display_create(struct drm_device *dev) | |||
340 | drm_kms_helper_poll_init(dev); | 339 | drm_kms_helper_poll_init(dev); |
341 | drm_kms_helper_poll_disable(dev); | 340 | drm_kms_helper_poll_disable(dev); |
342 | 341 | ||
343 | if (nouveau_modeset == 1 || | 342 | if (drm->vbios.dcb.entries) { |
344 | (nouveau_modeset < 0 && pclass == PCI_CLASS_DISPLAY_VGA)) { | 343 | if (nv_device(drm->device)->card_type < NV_50) |
345 | if (drm->vbios.dcb.entries) { | 344 | ret = nv04_display_create(dev); |
346 | if (nv_device(drm->device)->card_type < NV_50) | 345 | else |
347 | ret = nv04_display_create(dev); | 346 | ret = nv50_display_create(dev); |
348 | else | 347 | } else { |
349 | ret = nv50_display_create(dev); | 348 | ret = 0; |
350 | } else { | 349 | } |
351 | ret = 0; | ||
352 | } | ||
353 | |||
354 | if (ret) | ||
355 | goto disp_create_err; | ||
356 | 350 | ||
357 | if (dev->mode_config.num_crtc) { | 351 | if (ret) |
358 | ret = drm_vblank_init(dev, dev->mode_config.num_crtc); | 352 | goto disp_create_err; |
359 | if (ret) | ||
360 | goto vblank_err; | ||
361 | } | ||
362 | 353 | ||
363 | nouveau_backlight_init(dev); | 354 | if (dev->mode_config.num_crtc) { |
355 | ret = drm_vblank_init(dev, dev->mode_config.num_crtc); | ||
356 | if (ret) | ||
357 | goto vblank_err; | ||
364 | } | 358 | } |
365 | 359 | ||
360 | nouveau_backlight_init(dev); | ||
366 | return 0; | 361 | return 0; |
367 | 362 | ||
368 | vblank_err: | 363 | vblank_err: |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fbcon.c b/drivers/gpu/drm/nouveau/nouveau_fbcon.c index 8f6d63d7edd3..a86ecf65c164 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fbcon.c +++ b/drivers/gpu/drm/nouveau/nouveau_fbcon.c | |||
@@ -454,7 +454,8 @@ nouveau_fbcon_init(struct drm_device *dev) | |||
454 | int preferred_bpp; | 454 | int preferred_bpp; |
455 | int ret; | 455 | int ret; |
456 | 456 | ||
457 | if (!dev->mode_config.num_crtc) | 457 | if (!dev->mode_config.num_crtc || |
458 | (dev->pdev->class >> 8) != PCI_CLASS_DISPLAY_VGA) | ||
458 | return 0; | 459 | return 0; |
459 | 460 | ||
460 | fbcon = kzalloc(sizeof(struct nouveau_fbdev), GFP_KERNEL); | 461 | fbcon = kzalloc(sizeof(struct nouveau_fbdev), GFP_KERNEL); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/drivers/gpu/drm/nouveau/nouveau_sgdma.c index ca5492ac2da5..0843ebc910d4 100644 --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c | |||
@@ -104,9 +104,7 @@ nouveau_sgdma_create_ttm(struct ttm_bo_device *bdev, | |||
104 | else | 104 | else |
105 | nvbe->ttm.ttm.func = &nv50_sgdma_backend; | 105 | nvbe->ttm.ttm.func = &nv50_sgdma_backend; |
106 | 106 | ||
107 | if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) { | 107 | if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) |
108 | kfree(nvbe); | ||
109 | return NULL; | 108 | return NULL; |
110 | } | ||
111 | return &nvbe->ttm.ttm; | 109 | return &nvbe->ttm.ttm; |
112 | } | 110 | } |
diff --git a/drivers/gpu/drm/radeon/atombios_encoders.c b/drivers/gpu/drm/radeon/atombios_encoders.c index dfac7965ea28..32923d2f6002 100644 --- a/drivers/gpu/drm/radeon/atombios_encoders.c +++ b/drivers/gpu/drm/radeon/atombios_encoders.c | |||
@@ -707,8 +707,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
707 | switch (connector->connector_type) { | 707 | switch (connector->connector_type) { |
708 | case DRM_MODE_CONNECTOR_DVII: | 708 | case DRM_MODE_CONNECTOR_DVII: |
709 | case DRM_MODE_CONNECTOR_HDMIB: /* HDMI-B is basically DL-DVI; analog works fine */ | 709 | case DRM_MODE_CONNECTOR_HDMIB: /* HDMI-B is basically DL-DVI; analog works fine */ |
710 | if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 710 | if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
711 | radeon_audio) | 711 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
712 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
712 | return ATOM_ENCODER_MODE_HDMI; | 713 | return ATOM_ENCODER_MODE_HDMI; |
713 | else if (radeon_connector->use_digital) | 714 | else if (radeon_connector->use_digital) |
714 | return ATOM_ENCODER_MODE_DVI; | 715 | return ATOM_ENCODER_MODE_DVI; |
@@ -718,8 +719,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
718 | case DRM_MODE_CONNECTOR_DVID: | 719 | case DRM_MODE_CONNECTOR_DVID: |
719 | case DRM_MODE_CONNECTOR_HDMIA: | 720 | case DRM_MODE_CONNECTOR_HDMIA: |
720 | default: | 721 | default: |
721 | if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 722 | if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
722 | radeon_audio) | 723 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
724 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
723 | return ATOM_ENCODER_MODE_HDMI; | 725 | return ATOM_ENCODER_MODE_HDMI; |
724 | else | 726 | else |
725 | return ATOM_ENCODER_MODE_DVI; | 727 | return ATOM_ENCODER_MODE_DVI; |
@@ -732,8 +734,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
732 | if ((dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) || | 734 | if ((dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) || |
733 | (dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_eDP)) | 735 | (dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_eDP)) |
734 | return ATOM_ENCODER_MODE_DP; | 736 | return ATOM_ENCODER_MODE_DP; |
735 | else if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 737 | else if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
736 | radeon_audio) | 738 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
739 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
737 | return ATOM_ENCODER_MODE_HDMI; | 740 | return ATOM_ENCODER_MODE_HDMI; |
738 | else | 741 | else |
739 | return ATOM_ENCODER_MODE_DVI; | 742 | return ATOM_ENCODER_MODE_DVI; |
@@ -1647,8 +1650,12 @@ radeon_atom_encoder_dpms_dig(struct drm_encoder *encoder, int mode) | |||
1647 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE, 0); | 1650 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE, 0); |
1648 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP, 0, 0); | 1651 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP, 0, 0); |
1649 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE, 0, 0); | 1652 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE, 0, 0); |
1650 | /* some early dce3.2 boards have a bug in their transmitter control table */ | 1653 | /* some dce3.x boards have a bug in their transmitter control table. |
1651 | if ((rdev->family != CHIP_RV710) && (rdev->family != CHIP_RV730)) | 1654 | * ACTION_ENABLE_OUTPUT can probably be dropped since ACTION_ENABLE |
1655 | * does the same thing and more. | ||
1656 | */ | ||
1657 | if ((rdev->family != CHIP_RV710) && (rdev->family != CHIP_RV730) && | ||
1658 | (rdev->family != CHIP_RS880)) | ||
1652 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE_OUTPUT, 0, 0); | 1659 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE_OUTPUT, 0, 0); |
1653 | } | 1660 | } |
1654 | if (ENCODER_MODE_IS_DP(atombios_get_encoder_mode(encoder)) && connector) { | 1661 | if (ENCODER_MODE_IS_DP(atombios_get_encoder_mode(encoder)) && connector) { |
diff --git a/drivers/gpu/drm/radeon/btc_dpm.c b/drivers/gpu/drm/radeon/btc_dpm.c index 084e69414fd1..9b6950d9b3c0 100644 --- a/drivers/gpu/drm/radeon/btc_dpm.c +++ b/drivers/gpu/drm/radeon/btc_dpm.c | |||
@@ -1168,6 +1168,23 @@ static const struct radeon_blacklist_clocks btc_blacklist_clocks[] = | |||
1168 | { 25000, 30000, RADEON_SCLK_UP } | 1168 | { 25000, 30000, RADEON_SCLK_UP } |
1169 | }; | 1169 | }; |
1170 | 1170 | ||
1171 | void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
1172 | u32 *max_clock) | ||
1173 | { | ||
1174 | u32 i, clock = 0; | ||
1175 | |||
1176 | if ((table == NULL) || (table->count == 0)) { | ||
1177 | *max_clock = clock; | ||
1178 | return; | ||
1179 | } | ||
1180 | |||
1181 | for (i = 0; i < table->count; i++) { | ||
1182 | if (clock < table->entries[i].clk) | ||
1183 | clock = table->entries[i].clk; | ||
1184 | } | ||
1185 | *max_clock = clock; | ||
1186 | } | ||
1187 | |||
1171 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, | 1188 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, |
1172 | u32 clock, u16 max_voltage, u16 *voltage) | 1189 | u32 clock, u16 max_voltage, u16 *voltage) |
1173 | { | 1190 | { |
@@ -1913,7 +1930,7 @@ static int btc_set_mc_special_registers(struct radeon_device *rdev, | |||
1913 | } | 1930 | } |
1914 | j++; | 1931 | j++; |
1915 | 1932 | ||
1916 | if (j > SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) | 1933 | if (j >= SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) |
1917 | return -EINVAL; | 1934 | return -EINVAL; |
1918 | 1935 | ||
1919 | tmp = RREG32(MC_PMG_CMD_MRS); | 1936 | tmp = RREG32(MC_PMG_CMD_MRS); |
@@ -1928,7 +1945,7 @@ static int btc_set_mc_special_registers(struct radeon_device *rdev, | |||
1928 | } | 1945 | } |
1929 | j++; | 1946 | j++; |
1930 | 1947 | ||
1931 | if (j > SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) | 1948 | if (j >= SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) |
1932 | return -EINVAL; | 1949 | return -EINVAL; |
1933 | break; | 1950 | break; |
1934 | case MC_SEQ_RESERVE_M >> 2: | 1951 | case MC_SEQ_RESERVE_M >> 2: |
@@ -1942,7 +1959,7 @@ static int btc_set_mc_special_registers(struct radeon_device *rdev, | |||
1942 | } | 1959 | } |
1943 | j++; | 1960 | j++; |
1944 | 1961 | ||
1945 | if (j > SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) | 1962 | if (j >= SMC_EVERGREEN_MC_REGISTER_ARRAY_SIZE) |
1946 | return -EINVAL; | 1963 | return -EINVAL; |
1947 | break; | 1964 | break; |
1948 | default: | 1965 | default: |
@@ -2080,6 +2097,7 @@ static void btc_apply_state_adjust_rules(struct radeon_device *rdev, | |||
2080 | bool disable_mclk_switching; | 2097 | bool disable_mclk_switching; |
2081 | u32 mclk, sclk; | 2098 | u32 mclk, sclk; |
2082 | u16 vddc, vddci; | 2099 | u16 vddc, vddci; |
2100 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
2083 | 2101 | ||
2084 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 2102 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
2085 | btc_dpm_vblank_too_short(rdev)) | 2103 | btc_dpm_vblank_too_short(rdev)) |
@@ -2121,6 +2139,39 @@ static void btc_apply_state_adjust_rules(struct radeon_device *rdev, | |||
2121 | ps->low.vddci = max_limits->vddci; | 2139 | ps->low.vddci = max_limits->vddci; |
2122 | } | 2140 | } |
2123 | 2141 | ||
2142 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
2143 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
2144 | &max_sclk_vddc); | ||
2145 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
2146 | &max_mclk_vddci); | ||
2147 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
2148 | &max_mclk_vddc); | ||
2149 | |||
2150 | if (max_sclk_vddc) { | ||
2151 | if (ps->low.sclk > max_sclk_vddc) | ||
2152 | ps->low.sclk = max_sclk_vddc; | ||
2153 | if (ps->medium.sclk > max_sclk_vddc) | ||
2154 | ps->medium.sclk = max_sclk_vddc; | ||
2155 | if (ps->high.sclk > max_sclk_vddc) | ||
2156 | ps->high.sclk = max_sclk_vddc; | ||
2157 | } | ||
2158 | if (max_mclk_vddci) { | ||
2159 | if (ps->low.mclk > max_mclk_vddci) | ||
2160 | ps->low.mclk = max_mclk_vddci; | ||
2161 | if (ps->medium.mclk > max_mclk_vddci) | ||
2162 | ps->medium.mclk = max_mclk_vddci; | ||
2163 | if (ps->high.mclk > max_mclk_vddci) | ||
2164 | ps->high.mclk = max_mclk_vddci; | ||
2165 | } | ||
2166 | if (max_mclk_vddc) { | ||
2167 | if (ps->low.mclk > max_mclk_vddc) | ||
2168 | ps->low.mclk = max_mclk_vddc; | ||
2169 | if (ps->medium.mclk > max_mclk_vddc) | ||
2170 | ps->medium.mclk = max_mclk_vddc; | ||
2171 | if (ps->high.mclk > max_mclk_vddc) | ||
2172 | ps->high.mclk = max_mclk_vddc; | ||
2173 | } | ||
2174 | |||
2124 | /* XXX validate the min clocks required for display */ | 2175 | /* XXX validate the min clocks required for display */ |
2125 | 2176 | ||
2126 | if (disable_mclk_switching) { | 2177 | if (disable_mclk_switching) { |
@@ -2340,12 +2391,6 @@ int btc_dpm_set_power_state(struct radeon_device *rdev) | |||
2340 | return ret; | 2391 | return ret; |
2341 | } | 2392 | } |
2342 | 2393 | ||
2343 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
2344 | if (ret) { | ||
2345 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
2346 | return ret; | ||
2347 | } | ||
2348 | |||
2349 | return 0; | 2394 | return 0; |
2350 | } | 2395 | } |
2351 | 2396 | ||
diff --git a/drivers/gpu/drm/radeon/btc_dpm.h b/drivers/gpu/drm/radeon/btc_dpm.h index 1a15e0e41950..3b6f12b7760b 100644 --- a/drivers/gpu/drm/radeon/btc_dpm.h +++ b/drivers/gpu/drm/radeon/btc_dpm.h | |||
@@ -46,6 +46,8 @@ void btc_adjust_clock_combinations(struct radeon_device *rdev, | |||
46 | struct rv7xx_pl *pl); | 46 | struct rv7xx_pl *pl); |
47 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, | 47 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, |
48 | u32 clock, u16 max_voltage, u16 *voltage); | 48 | u32 clock, u16 max_voltage, u16 *voltage); |
49 | void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
50 | u32 *max_clock); | ||
49 | void btc_apply_voltage_delta_rules(struct radeon_device *rdev, | 51 | void btc_apply_voltage_delta_rules(struct radeon_device *rdev, |
50 | u16 max_vddc, u16 max_vddci, | 52 | u16 max_vddc, u16 max_vddci, |
51 | u16 *vddc, u16 *vddci); | 53 | u16 *vddc, u16 *vddci); |
diff --git a/drivers/gpu/drm/radeon/ci_dpm.c b/drivers/gpu/drm/radeon/ci_dpm.c index 3cce533397c6..51e947a97edf 100644 --- a/drivers/gpu/drm/radeon/ci_dpm.c +++ b/drivers/gpu/drm/radeon/ci_dpm.c | |||
@@ -146,6 +146,8 @@ static const struct ci_pt_config_reg didt_config_ci[] = | |||
146 | }; | 146 | }; |
147 | 147 | ||
148 | extern u8 rv770_get_memory_module_index(struct radeon_device *rdev); | 148 | extern u8 rv770_get_memory_module_index(struct radeon_device *rdev); |
149 | extern void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
150 | u32 *max_clock); | ||
149 | extern int ni_copy_and_switch_arb_sets(struct radeon_device *rdev, | 151 | extern int ni_copy_and_switch_arb_sets(struct radeon_device *rdev, |
150 | u32 arb_freq_src, u32 arb_freq_dest); | 152 | u32 arb_freq_src, u32 arb_freq_dest); |
151 | extern u8 si_get_ddr3_mclk_frequency_ratio(u32 memory_clock); | 153 | extern u8 si_get_ddr3_mclk_frequency_ratio(u32 memory_clock); |
@@ -712,6 +714,7 @@ static void ci_apply_state_adjust_rules(struct radeon_device *rdev, | |||
712 | struct radeon_clock_and_voltage_limits *max_limits; | 714 | struct radeon_clock_and_voltage_limits *max_limits; |
713 | bool disable_mclk_switching; | 715 | bool disable_mclk_switching; |
714 | u32 sclk, mclk; | 716 | u32 sclk, mclk; |
717 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
715 | int i; | 718 | int i; |
716 | 719 | ||
717 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 720 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
@@ -739,6 +742,29 @@ static void ci_apply_state_adjust_rules(struct radeon_device *rdev, | |||
739 | } | 742 | } |
740 | } | 743 | } |
741 | 744 | ||
745 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
746 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
747 | &max_sclk_vddc); | ||
748 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
749 | &max_mclk_vddci); | ||
750 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
751 | &max_mclk_vddc); | ||
752 | |||
753 | for (i = 0; i < ps->performance_level_count; i++) { | ||
754 | if (max_sclk_vddc) { | ||
755 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
756 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
757 | } | ||
758 | if (max_mclk_vddci) { | ||
759 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
760 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
761 | } | ||
762 | if (max_mclk_vddc) { | ||
763 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
764 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
765 | } | ||
766 | } | ||
767 | |||
742 | /* XXX validate the min clocks required for display */ | 768 | /* XXX validate the min clocks required for display */ |
743 | 769 | ||
744 | if (disable_mclk_switching) { | 770 | if (disable_mclk_switching) { |
@@ -4748,12 +4774,6 @@ int ci_dpm_set_power_state(struct radeon_device *rdev) | |||
4748 | if (pi->pcie_performance_request) | 4774 | if (pi->pcie_performance_request) |
4749 | ci_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); | 4775 | ci_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); |
4750 | 4776 | ||
4751 | ret = ci_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
4752 | if (ret) { | ||
4753 | DRM_ERROR("ci_dpm_force_performance_level failed\n"); | ||
4754 | return ret; | ||
4755 | } | ||
4756 | |||
4757 | cik_update_cg(rdev, (RADEON_CG_BLOCK_GFX | | 4777 | cik_update_cg(rdev, (RADEON_CG_BLOCK_GFX | |
4758 | RADEON_CG_BLOCK_MC | | 4778 | RADEON_CG_BLOCK_MC | |
4759 | RADEON_CG_BLOCK_SDMA | | 4779 | RADEON_CG_BLOCK_SDMA | |
diff --git a/drivers/gpu/drm/radeon/ci_smc.c b/drivers/gpu/drm/radeon/ci_smc.c index 53b43dd3cf1e..252e10a41cf5 100644 --- a/drivers/gpu/drm/radeon/ci_smc.c +++ b/drivers/gpu/drm/radeon/ci_smc.c | |||
@@ -47,10 +47,11 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
47 | u32 smc_start_address, | 47 | u32 smc_start_address, |
48 | const u8 *src, u32 byte_count, u32 limit) | 48 | const u8 *src, u32 byte_count, u32 limit) |
49 | { | 49 | { |
50 | unsigned long flags; | ||
50 | u32 data, original_data; | 51 | u32 data, original_data; |
51 | u32 addr; | 52 | u32 addr; |
52 | u32 extra_shift; | 53 | u32 extra_shift; |
53 | int ret; | 54 | int ret = 0; |
54 | 55 | ||
55 | if (smc_start_address & 3) | 56 | if (smc_start_address & 3) |
56 | return -EINVAL; | 57 | return -EINVAL; |
@@ -59,13 +60,14 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
59 | 60 | ||
60 | addr = smc_start_address; | 61 | addr = smc_start_address; |
61 | 62 | ||
63 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
62 | while (byte_count >= 4) { | 64 | while (byte_count >= 4) { |
63 | /* SMC address space is BE */ | 65 | /* SMC address space is BE */ |
64 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 66 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
65 | 67 | ||
66 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 68 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
67 | if (ret) | 69 | if (ret) |
68 | return ret; | 70 | goto done; |
69 | 71 | ||
70 | WREG32(SMC_IND_DATA_0, data); | 72 | WREG32(SMC_IND_DATA_0, data); |
71 | 73 | ||
@@ -80,7 +82,7 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
80 | 82 | ||
81 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 83 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
82 | if (ret) | 84 | if (ret) |
83 | return ret; | 85 | goto done; |
84 | 86 | ||
85 | original_data = RREG32(SMC_IND_DATA_0); | 87 | original_data = RREG32(SMC_IND_DATA_0); |
86 | 88 | ||
@@ -97,11 +99,15 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
97 | 99 | ||
98 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 100 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
99 | if (ret) | 101 | if (ret) |
100 | return ret; | 102 | goto done; |
101 | 103 | ||
102 | WREG32(SMC_IND_DATA_0, data); | 104 | WREG32(SMC_IND_DATA_0, data); |
103 | } | 105 | } |
104 | return 0; | 106 | |
107 | done: | ||
108 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
109 | |||
110 | return ret; | ||
105 | } | 111 | } |
106 | 112 | ||
107 | void ci_start_smc(struct radeon_device *rdev) | 113 | void ci_start_smc(struct radeon_device *rdev) |
@@ -197,6 +203,7 @@ PPSMC_Result ci_wait_for_smc_inactive(struct radeon_device *rdev) | |||
197 | 203 | ||
198 | int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | 204 | int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) |
199 | { | 205 | { |
206 | unsigned long flags; | ||
200 | u32 ucode_start_address; | 207 | u32 ucode_start_address; |
201 | u32 ucode_size; | 208 | u32 ucode_size; |
202 | const u8 *src; | 209 | const u8 *src; |
@@ -219,6 +226,7 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
219 | return -EINVAL; | 226 | return -EINVAL; |
220 | 227 | ||
221 | src = (const u8 *)rdev->smc_fw->data; | 228 | src = (const u8 *)rdev->smc_fw->data; |
229 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
222 | WREG32(SMC_IND_INDEX_0, ucode_start_address); | 230 | WREG32(SMC_IND_INDEX_0, ucode_start_address); |
223 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); | 231 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); |
224 | while (ucode_size >= 4) { | 232 | while (ucode_size >= 4) { |
@@ -231,6 +239,7 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
231 | ucode_size -= 4; | 239 | ucode_size -= 4; |
232 | } | 240 | } |
233 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); | 241 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); |
242 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
234 | 243 | ||
235 | return 0; | 244 | return 0; |
236 | } | 245 | } |
@@ -238,25 +247,29 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
238 | int ci_read_smc_sram_dword(struct radeon_device *rdev, | 247 | int ci_read_smc_sram_dword(struct radeon_device *rdev, |
239 | u32 smc_address, u32 *value, u32 limit) | 248 | u32 smc_address, u32 *value, u32 limit) |
240 | { | 249 | { |
250 | unsigned long flags; | ||
241 | int ret; | 251 | int ret; |
242 | 252 | ||
253 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
243 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); | 254 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); |
244 | if (ret) | 255 | if (ret == 0) |
245 | return ret; | 256 | *value = RREG32(SMC_IND_DATA_0); |
257 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
246 | 258 | ||
247 | *value = RREG32(SMC_IND_DATA_0); | 259 | return ret; |
248 | return 0; | ||
249 | } | 260 | } |
250 | 261 | ||
251 | int ci_write_smc_sram_dword(struct radeon_device *rdev, | 262 | int ci_write_smc_sram_dword(struct radeon_device *rdev, |
252 | u32 smc_address, u32 value, u32 limit) | 263 | u32 smc_address, u32 value, u32 limit) |
253 | { | 264 | { |
265 | unsigned long flags; | ||
254 | int ret; | 266 | int ret; |
255 | 267 | ||
268 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
256 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); | 269 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); |
257 | if (ret) | 270 | if (ret == 0) |
258 | return ret; | 271 | WREG32(SMC_IND_DATA_0, value); |
272 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
259 | 273 | ||
260 | WREG32(SMC_IND_DATA_0, value); | 274 | return ret; |
261 | return 0; | ||
262 | } | 275 | } |
diff --git a/drivers/gpu/drm/radeon/cik.c b/drivers/gpu/drm/radeon/cik.c index a3bba0587276..b874ccdf52f7 100644 --- a/drivers/gpu/drm/radeon/cik.c +++ b/drivers/gpu/drm/radeon/cik.c | |||
@@ -77,6 +77,10 @@ static void cik_pcie_gen3_enable(struct radeon_device *rdev); | |||
77 | static void cik_program_aspm(struct radeon_device *rdev); | 77 | static void cik_program_aspm(struct radeon_device *rdev); |
78 | static void cik_init_pg(struct radeon_device *rdev); | 78 | static void cik_init_pg(struct radeon_device *rdev); |
79 | static void cik_init_cg(struct radeon_device *rdev); | 79 | static void cik_init_cg(struct radeon_device *rdev); |
80 | static void cik_fini_pg(struct radeon_device *rdev); | ||
81 | static void cik_fini_cg(struct radeon_device *rdev); | ||
82 | static void cik_enable_gui_idle_interrupt(struct radeon_device *rdev, | ||
83 | bool enable); | ||
80 | 84 | ||
81 | /* get temperature in millidegrees */ | 85 | /* get temperature in millidegrees */ |
82 | int ci_get_temp(struct radeon_device *rdev) | 86 | int ci_get_temp(struct radeon_device *rdev) |
@@ -120,20 +124,27 @@ int kv_get_temp(struct radeon_device *rdev) | |||
120 | */ | 124 | */ |
121 | u32 cik_pciep_rreg(struct radeon_device *rdev, u32 reg) | 125 | u32 cik_pciep_rreg(struct radeon_device *rdev, u32 reg) |
122 | { | 126 | { |
127 | unsigned long flags; | ||
123 | u32 r; | 128 | u32 r; |
124 | 129 | ||
130 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
125 | WREG32(PCIE_INDEX, reg); | 131 | WREG32(PCIE_INDEX, reg); |
126 | (void)RREG32(PCIE_INDEX); | 132 | (void)RREG32(PCIE_INDEX); |
127 | r = RREG32(PCIE_DATA); | 133 | r = RREG32(PCIE_DATA); |
134 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
128 | return r; | 135 | return r; |
129 | } | 136 | } |
130 | 137 | ||
131 | void cik_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 138 | void cik_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
132 | { | 139 | { |
140 | unsigned long flags; | ||
141 | |||
142 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
133 | WREG32(PCIE_INDEX, reg); | 143 | WREG32(PCIE_INDEX, reg); |
134 | (void)RREG32(PCIE_INDEX); | 144 | (void)RREG32(PCIE_INDEX); |
135 | WREG32(PCIE_DATA, v); | 145 | WREG32(PCIE_DATA, v); |
136 | (void)RREG32(PCIE_DATA); | 146 | (void)RREG32(PCIE_DATA); |
147 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
137 | } | 148 | } |
138 | 149 | ||
139 | static const u32 spectre_rlc_save_restore_register_list[] = | 150 | static const u32 spectre_rlc_save_restore_register_list[] = |
@@ -2722,7 +2733,8 @@ static void cik_gpu_init(struct radeon_device *rdev) | |||
2722 | } else if ((rdev->pdev->device == 0x1309) || | 2733 | } else if ((rdev->pdev->device == 0x1309) || |
2723 | (rdev->pdev->device == 0x130A) || | 2734 | (rdev->pdev->device == 0x130A) || |
2724 | (rdev->pdev->device == 0x130D) || | 2735 | (rdev->pdev->device == 0x130D) || |
2725 | (rdev->pdev->device == 0x1313)) { | 2736 | (rdev->pdev->device == 0x1313) || |
2737 | (rdev->pdev->device == 0x131D)) { | ||
2726 | rdev->config.cik.max_cu_per_sh = 6; | 2738 | rdev->config.cik.max_cu_per_sh = 6; |
2727 | rdev->config.cik.max_backends_per_se = 2; | 2739 | rdev->config.cik.max_backends_per_se = 2; |
2728 | } else if ((rdev->pdev->device == 0x1306) || | 2740 | } else if ((rdev->pdev->device == 0x1306) || |
@@ -2835,10 +2847,8 @@ static void cik_gpu_init(struct radeon_device *rdev) | |||
2835 | rdev->config.cik.tile_config |= (3 << 0); | 2847 | rdev->config.cik.tile_config |= (3 << 0); |
2836 | break; | 2848 | break; |
2837 | } | 2849 | } |
2838 | if ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) | 2850 | rdev->config.cik.tile_config |= |
2839 | rdev->config.cik.tile_config |= 1 << 4; | 2851 | ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) << 4; |
2840 | else | ||
2841 | rdev->config.cik.tile_config |= 0 << 4; | ||
2842 | rdev->config.cik.tile_config |= | 2852 | rdev->config.cik.tile_config |= |
2843 | ((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8; | 2853 | ((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8; |
2844 | rdev->config.cik.tile_config |= | 2854 | rdev->config.cik.tile_config |= |
@@ -4013,6 +4023,8 @@ static int cik_cp_resume(struct radeon_device *rdev) | |||
4013 | { | 4023 | { |
4014 | int r; | 4024 | int r; |
4015 | 4025 | ||
4026 | cik_enable_gui_idle_interrupt(rdev, false); | ||
4027 | |||
4016 | r = cik_cp_load_microcode(rdev); | 4028 | r = cik_cp_load_microcode(rdev); |
4017 | if (r) | 4029 | if (r) |
4018 | return r; | 4030 | return r; |
@@ -4024,6 +4036,8 @@ static int cik_cp_resume(struct radeon_device *rdev) | |||
4024 | if (r) | 4036 | if (r) |
4025 | return r; | 4037 | return r; |
4026 | 4038 | ||
4039 | cik_enable_gui_idle_interrupt(rdev, true); | ||
4040 | |||
4027 | return 0; | 4041 | return 0; |
4028 | } | 4042 | } |
4029 | 4043 | ||
@@ -4173,6 +4187,10 @@ static void cik_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask) | |||
4173 | dev_info(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n", | 4187 | dev_info(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n", |
4174 | RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS)); | 4188 | RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS)); |
4175 | 4189 | ||
4190 | /* disable CG/PG */ | ||
4191 | cik_fini_pg(rdev); | ||
4192 | cik_fini_cg(rdev); | ||
4193 | |||
4176 | /* stop the rlc */ | 4194 | /* stop the rlc */ |
4177 | cik_rlc_stop(rdev); | 4195 | cik_rlc_stop(rdev); |
4178 | 4196 | ||
@@ -4442,8 +4460,8 @@ static int cik_mc_init(struct radeon_device *rdev) | |||
4442 | rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0); | 4460 | rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0); |
4443 | rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0); | 4461 | rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0); |
4444 | /* size in MB on si */ | 4462 | /* size in MB on si */ |
4445 | rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024 * 1024; | 4463 | rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL; |
4446 | rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024 * 1024; | 4464 | rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL; |
4447 | rdev->mc.visible_vram_size = rdev->mc.aper_size; | 4465 | rdev->mc.visible_vram_size = rdev->mc.aper_size; |
4448 | si_vram_gtt_location(rdev, &rdev->mc); | 4466 | si_vram_gtt_location(rdev, &rdev->mc); |
4449 | radeon_update_bandwidth_info(rdev); | 4467 | radeon_update_bandwidth_info(rdev); |
@@ -4721,12 +4739,13 @@ static void cik_vm_decode_fault(struct radeon_device *rdev, | |||
4721 | u32 mc_id = (status & MEMORY_CLIENT_ID_MASK) >> MEMORY_CLIENT_ID_SHIFT; | 4739 | u32 mc_id = (status & MEMORY_CLIENT_ID_MASK) >> MEMORY_CLIENT_ID_SHIFT; |
4722 | u32 vmid = (status & FAULT_VMID_MASK) >> FAULT_VMID_SHIFT; | 4740 | u32 vmid = (status & FAULT_VMID_MASK) >> FAULT_VMID_SHIFT; |
4723 | u32 protections = (status & PROTECTIONS_MASK) >> PROTECTIONS_SHIFT; | 4741 | u32 protections = (status & PROTECTIONS_MASK) >> PROTECTIONS_SHIFT; |
4724 | char *block = (char *)&mc_client; | 4742 | char block[5] = { mc_client >> 24, (mc_client >> 16) & 0xff, |
4743 | (mc_client >> 8) & 0xff, mc_client & 0xff, 0 }; | ||
4725 | 4744 | ||
4726 | printk("VM fault (0x%02x, vmid %d) at page %u, %s from %s (%d)\n", | 4745 | printk("VM fault (0x%02x, vmid %d) at page %u, %s from '%s' (0x%08x) (%d)\n", |
4727 | protections, vmid, addr, | 4746 | protections, vmid, addr, |
4728 | (status & MEMORY_CLIENT_RW_MASK) ? "write" : "read", | 4747 | (status & MEMORY_CLIENT_RW_MASK) ? "write" : "read", |
4729 | block, mc_id); | 4748 | block, mc_client, mc_id); |
4730 | } | 4749 | } |
4731 | 4750 | ||
4732 | /** | 4751 | /** |
@@ -5376,7 +5395,9 @@ static void cik_enable_hdp_ls(struct radeon_device *rdev, | |||
5376 | void cik_update_cg(struct radeon_device *rdev, | 5395 | void cik_update_cg(struct radeon_device *rdev, |
5377 | u32 block, bool enable) | 5396 | u32 block, bool enable) |
5378 | { | 5397 | { |
5398 | |||
5379 | if (block & RADEON_CG_BLOCK_GFX) { | 5399 | if (block & RADEON_CG_BLOCK_GFX) { |
5400 | cik_enable_gui_idle_interrupt(rdev, false); | ||
5380 | /* order matters! */ | 5401 | /* order matters! */ |
5381 | if (enable) { | 5402 | if (enable) { |
5382 | cik_enable_mgcg(rdev, true); | 5403 | cik_enable_mgcg(rdev, true); |
@@ -5385,6 +5406,7 @@ void cik_update_cg(struct radeon_device *rdev, | |||
5385 | cik_enable_cgcg(rdev, false); | 5406 | cik_enable_cgcg(rdev, false); |
5386 | cik_enable_mgcg(rdev, false); | 5407 | cik_enable_mgcg(rdev, false); |
5387 | } | 5408 | } |
5409 | cik_enable_gui_idle_interrupt(rdev, true); | ||
5388 | } | 5410 | } |
5389 | 5411 | ||
5390 | if (block & RADEON_CG_BLOCK_MC) { | 5412 | if (block & RADEON_CG_BLOCK_MC) { |
@@ -5541,7 +5563,7 @@ static void cik_enable_gfx_cgpg(struct radeon_device *rdev, | |||
5541 | { | 5563 | { |
5542 | u32 data, orig; | 5564 | u32 data, orig; |
5543 | 5565 | ||
5544 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG)) { | 5566 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG)) { |
5545 | orig = data = RREG32(RLC_PG_CNTL); | 5567 | orig = data = RREG32(RLC_PG_CNTL); |
5546 | data |= GFX_PG_ENABLE; | 5568 | data |= GFX_PG_ENABLE; |
5547 | if (orig != data) | 5569 | if (orig != data) |
@@ -5805,7 +5827,7 @@ static void cik_init_pg(struct radeon_device *rdev) | |||
5805 | if (rdev->pg_flags) { | 5827 | if (rdev->pg_flags) { |
5806 | cik_enable_sck_slowdown_on_pu(rdev, true); | 5828 | cik_enable_sck_slowdown_on_pu(rdev, true); |
5807 | cik_enable_sck_slowdown_on_pd(rdev, true); | 5829 | cik_enable_sck_slowdown_on_pd(rdev, true); |
5808 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5830 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
5809 | cik_init_gfx_cgpg(rdev); | 5831 | cik_init_gfx_cgpg(rdev); |
5810 | cik_enable_cp_pg(rdev, true); | 5832 | cik_enable_cp_pg(rdev, true); |
5811 | cik_enable_gds_pg(rdev, true); | 5833 | cik_enable_gds_pg(rdev, true); |
@@ -5819,7 +5841,7 @@ static void cik_fini_pg(struct radeon_device *rdev) | |||
5819 | { | 5841 | { |
5820 | if (rdev->pg_flags) { | 5842 | if (rdev->pg_flags) { |
5821 | cik_update_gfx_pg(rdev, false); | 5843 | cik_update_gfx_pg(rdev, false); |
5822 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5844 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
5823 | cik_enable_cp_pg(rdev, false); | 5845 | cik_enable_cp_pg(rdev, false); |
5824 | cik_enable_gds_pg(rdev, false); | 5846 | cik_enable_gds_pg(rdev, false); |
5825 | } | 5847 | } |
@@ -5895,7 +5917,9 @@ static void cik_disable_interrupt_state(struct radeon_device *rdev) | |||
5895 | u32 tmp; | 5917 | u32 tmp; |
5896 | 5918 | ||
5897 | /* gfx ring */ | 5919 | /* gfx ring */ |
5898 | WREG32(CP_INT_CNTL_RING0, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | 5920 | tmp = RREG32(CP_INT_CNTL_RING0) & |
5921 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
5922 | WREG32(CP_INT_CNTL_RING0, tmp); | ||
5899 | /* sdma */ | 5923 | /* sdma */ |
5900 | tmp = RREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; | 5924 | tmp = RREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; |
5901 | WREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET, tmp); | 5925 | WREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET, tmp); |
@@ -6036,8 +6060,7 @@ static int cik_irq_init(struct radeon_device *rdev) | |||
6036 | */ | 6060 | */ |
6037 | int cik_irq_set(struct radeon_device *rdev) | 6061 | int cik_irq_set(struct radeon_device *rdev) |
6038 | { | 6062 | { |
6039 | u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE | | 6063 | u32 cp_int_cntl; |
6040 | PRIV_INSTR_INT_ENABLE | PRIV_REG_INT_ENABLE; | ||
6041 | u32 cp_m1p0, cp_m1p1, cp_m1p2, cp_m1p3; | 6064 | u32 cp_m1p0, cp_m1p1, cp_m1p2, cp_m1p3; |
6042 | u32 cp_m2p0, cp_m2p1, cp_m2p2, cp_m2p3; | 6065 | u32 cp_m2p0, cp_m2p1, cp_m2p2, cp_m2p3; |
6043 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; | 6066 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; |
@@ -6058,6 +6081,10 @@ int cik_irq_set(struct radeon_device *rdev) | |||
6058 | return 0; | 6081 | return 0; |
6059 | } | 6082 | } |
6060 | 6083 | ||
6084 | cp_int_cntl = RREG32(CP_INT_CNTL_RING0) & | ||
6085 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
6086 | cp_int_cntl |= PRIV_INSTR_INT_ENABLE | PRIV_REG_INT_ENABLE; | ||
6087 | |||
6061 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6088 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; |
6062 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6089 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; |
6063 | hpd3 = RREG32(DC_HPD3_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6090 | hpd3 = RREG32(DC_HPD3_INT_CONTROL) & ~DC_HPDx_INT_EN; |
diff --git a/drivers/gpu/drm/radeon/cypress_dpm.c b/drivers/gpu/drm/radeon/cypress_dpm.c index 95a66db08d9b..91bb470de0a3 100644 --- a/drivers/gpu/drm/radeon/cypress_dpm.c +++ b/drivers/gpu/drm/radeon/cypress_dpm.c | |||
@@ -2014,12 +2014,6 @@ int cypress_dpm_set_power_state(struct radeon_device *rdev) | |||
2014 | if (eg_pi->pcie_performance_request) | 2014 | if (eg_pi->pcie_performance_request) |
2015 | cypress_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); | 2015 | cypress_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); |
2016 | 2016 | ||
2017 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
2018 | if (ret) { | ||
2019 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
2020 | return ret; | ||
2021 | } | ||
2022 | |||
2023 | return 0; | 2017 | return 0; |
2024 | } | 2018 | } |
2025 | 2019 | ||
diff --git a/drivers/gpu/drm/radeon/dce6_afmt.c b/drivers/gpu/drm/radeon/dce6_afmt.c index 8953255e894b..85a69d2ea3d2 100644 --- a/drivers/gpu/drm/radeon/dce6_afmt.c +++ b/drivers/gpu/drm/radeon/dce6_afmt.c | |||
@@ -28,22 +28,30 @@ | |||
28 | static u32 dce6_endpoint_rreg(struct radeon_device *rdev, | 28 | static u32 dce6_endpoint_rreg(struct radeon_device *rdev, |
29 | u32 block_offset, u32 reg) | 29 | u32 block_offset, u32 reg) |
30 | { | 30 | { |
31 | unsigned long flags; | ||
31 | u32 r; | 32 | u32 r; |
32 | 33 | ||
34 | spin_lock_irqsave(&rdev->end_idx_lock, flags); | ||
33 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); | 35 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); |
34 | r = RREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset); | 36 | r = RREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset); |
37 | spin_unlock_irqrestore(&rdev->end_idx_lock, flags); | ||
38 | |||
35 | return r; | 39 | return r; |
36 | } | 40 | } |
37 | 41 | ||
38 | static void dce6_endpoint_wreg(struct radeon_device *rdev, | 42 | static void dce6_endpoint_wreg(struct radeon_device *rdev, |
39 | u32 block_offset, u32 reg, u32 v) | 43 | u32 block_offset, u32 reg, u32 v) |
40 | { | 44 | { |
45 | unsigned long flags; | ||
46 | |||
47 | spin_lock_irqsave(&rdev->end_idx_lock, flags); | ||
41 | if (ASIC_IS_DCE8(rdev)) | 48 | if (ASIC_IS_DCE8(rdev)) |
42 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); | 49 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); |
43 | else | 50 | else |
44 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, | 51 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, |
45 | AZ_ENDPOINT_REG_WRITE_EN | AZ_ENDPOINT_REG_INDEX(reg)); | 52 | AZ_ENDPOINT_REG_WRITE_EN | AZ_ENDPOINT_REG_INDEX(reg)); |
46 | WREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset, v); | 53 | WREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset, v); |
54 | spin_unlock_irqrestore(&rdev->end_idx_lock, flags); | ||
47 | } | 55 | } |
48 | 56 | ||
49 | #define RREG32_ENDPOINT(block, reg) dce6_endpoint_rreg(rdev, (block), (reg)) | 57 | #define RREG32_ENDPOINT(block, reg) dce6_endpoint_rreg(rdev, (block), (reg)) |
@@ -86,12 +94,12 @@ void dce6_afmt_select_pin(struct drm_encoder *encoder) | |||
86 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | 94 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); |
87 | struct radeon_encoder_atom_dig *dig = radeon_encoder->enc_priv; | 95 | struct radeon_encoder_atom_dig *dig = radeon_encoder->enc_priv; |
88 | u32 offset = dig->afmt->offset; | 96 | u32 offset = dig->afmt->offset; |
89 | u32 id = dig->afmt->pin->id; | ||
90 | 97 | ||
91 | if (!dig->afmt->pin) | 98 | if (!dig->afmt->pin) |
92 | return; | 99 | return; |
93 | 100 | ||
94 | WREG32(AFMT_AUDIO_SRC_CONTROL + offset, AFMT_AUDIO_SRC_SELECT(id)); | 101 | WREG32(AFMT_AUDIO_SRC_CONTROL + offset, |
102 | AFMT_AUDIO_SRC_SELECT(dig->afmt->pin->id)); | ||
95 | } | 103 | } |
96 | 104 | ||
97 | void dce6_afmt_write_speaker_allocation(struct drm_encoder *encoder) | 105 | void dce6_afmt_write_speaker_allocation(struct drm_encoder *encoder) |
diff --git a/drivers/gpu/drm/radeon/evergreen.c b/drivers/gpu/drm/radeon/evergreen.c index 555164e270a7..b5c67a99dda9 100644 --- a/drivers/gpu/drm/radeon/evergreen.c +++ b/drivers/gpu/drm/radeon/evergreen.c | |||
@@ -3131,7 +3131,7 @@ static void evergreen_gpu_init(struct radeon_device *rdev) | |||
3131 | rdev->config.evergreen.sx_max_export_size = 256; | 3131 | rdev->config.evergreen.sx_max_export_size = 256; |
3132 | rdev->config.evergreen.sx_max_export_pos_size = 64; | 3132 | rdev->config.evergreen.sx_max_export_pos_size = 64; |
3133 | rdev->config.evergreen.sx_max_export_smx_size = 192; | 3133 | rdev->config.evergreen.sx_max_export_smx_size = 192; |
3134 | rdev->config.evergreen.max_hw_contexts = 8; | 3134 | rdev->config.evergreen.max_hw_contexts = 4; |
3135 | rdev->config.evergreen.sq_num_cf_insts = 2; | 3135 | rdev->config.evergreen.sq_num_cf_insts = 2; |
3136 | 3136 | ||
3137 | rdev->config.evergreen.sc_prim_fifo_size = 0x40; | 3137 | rdev->config.evergreen.sc_prim_fifo_size = 0x40; |
diff --git a/drivers/gpu/drm/radeon/evergreen_hdmi.c b/drivers/gpu/drm/radeon/evergreen_hdmi.c index f71ce390aebe..f815c20640bd 100644 --- a/drivers/gpu/drm/radeon/evergreen_hdmi.c +++ b/drivers/gpu/drm/radeon/evergreen_hdmi.c | |||
@@ -288,8 +288,7 @@ void evergreen_hdmi_setmode(struct drm_encoder *encoder, struct drm_display_mode | |||
288 | /* fglrx clears sth in AFMT_AUDIO_PACKET_CONTROL2 here */ | 288 | /* fglrx clears sth in AFMT_AUDIO_PACKET_CONTROL2 here */ |
289 | 289 | ||
290 | WREG32(HDMI_ACR_PACKET_CONTROL + offset, | 290 | WREG32(HDMI_ACR_PACKET_CONTROL + offset, |
291 | HDMI_ACR_AUTO_SEND | /* allow hw to sent ACR packets when required */ | 291 | HDMI_ACR_AUTO_SEND); /* allow hw to sent ACR packets when required */ |
292 | HDMI_ACR_SOURCE); /* select SW CTS value */ | ||
293 | 292 | ||
294 | evergreen_hdmi_update_ACR(encoder, mode->clock); | 293 | evergreen_hdmi_update_ACR(encoder, mode->clock); |
295 | 294 | ||
diff --git a/drivers/gpu/drm/radeon/evergreend.h b/drivers/gpu/drm/radeon/evergreend.h index 8768fd6a1e27..4f6d2962767d 100644 --- a/drivers/gpu/drm/radeon/evergreend.h +++ b/drivers/gpu/drm/radeon/evergreend.h | |||
@@ -1501,7 +1501,7 @@ | |||
1501 | * 6. COMMAND [29:22] | BYTE_COUNT [20:0] | 1501 | * 6. COMMAND [29:22] | BYTE_COUNT [20:0] |
1502 | */ | 1502 | */ |
1503 | # define PACKET3_CP_DMA_DST_SEL(x) ((x) << 20) | 1503 | # define PACKET3_CP_DMA_DST_SEL(x) ((x) << 20) |
1504 | /* 0 - SRC_ADDR | 1504 | /* 0 - DST_ADDR |
1505 | * 1 - GDS | 1505 | * 1 - GDS |
1506 | */ | 1506 | */ |
1507 | # define PACKET3_CP_DMA_ENGINE(x) ((x) << 27) | 1507 | # define PACKET3_CP_DMA_ENGINE(x) ((x) << 27) |
@@ -1516,7 +1516,7 @@ | |||
1516 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) | 1516 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) |
1517 | /* COMMAND */ | 1517 | /* COMMAND */ |
1518 | # define PACKET3_CP_DMA_DIS_WC (1 << 21) | 1518 | # define PACKET3_CP_DMA_DIS_WC (1 << 21) |
1519 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 23) | 1519 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 22) |
1520 | /* 0 - none | 1520 | /* 0 - none |
1521 | * 1 - 8 in 16 | 1521 | * 1 - 8 in 16 |
1522 | * 2 - 8 in 32 | 1522 | * 2 - 8 in 32 |
diff --git a/drivers/gpu/drm/radeon/kv_dpm.c b/drivers/gpu/drm/radeon/kv_dpm.c index ecd60809db4e..71399065db04 100644 --- a/drivers/gpu/drm/radeon/kv_dpm.c +++ b/drivers/gpu/drm/radeon/kv_dpm.c | |||
@@ -40,6 +40,7 @@ static int kv_calculate_dpm_settings(struct radeon_device *rdev); | |||
40 | static void kv_enable_new_levels(struct radeon_device *rdev); | 40 | static void kv_enable_new_levels(struct radeon_device *rdev); |
41 | static void kv_program_nbps_index_settings(struct radeon_device *rdev, | 41 | static void kv_program_nbps_index_settings(struct radeon_device *rdev, |
42 | struct radeon_ps *new_rps); | 42 | struct radeon_ps *new_rps); |
43 | static int kv_set_enabled_level(struct radeon_device *rdev, u32 level); | ||
43 | static int kv_set_enabled_levels(struct radeon_device *rdev); | 44 | static int kv_set_enabled_levels(struct radeon_device *rdev); |
44 | static int kv_force_dpm_highest(struct radeon_device *rdev); | 45 | static int kv_force_dpm_highest(struct radeon_device *rdev); |
45 | static int kv_force_dpm_lowest(struct radeon_device *rdev); | 46 | static int kv_force_dpm_lowest(struct radeon_device *rdev); |
@@ -519,7 +520,7 @@ static int kv_set_dpm_boot_state(struct radeon_device *rdev) | |||
519 | 520 | ||
520 | static void kv_program_vc(struct radeon_device *rdev) | 521 | static void kv_program_vc(struct radeon_device *rdev) |
521 | { | 522 | { |
522 | WREG32_SMC(CG_FTV_0, 0x3FFFC000); | 523 | WREG32_SMC(CG_FTV_0, 0x3FFFC100); |
523 | } | 524 | } |
524 | 525 | ||
525 | static void kv_clear_vc(struct radeon_device *rdev) | 526 | static void kv_clear_vc(struct radeon_device *rdev) |
@@ -638,7 +639,10 @@ static int kv_force_lowest_valid(struct radeon_device *rdev) | |||
638 | 639 | ||
639 | static int kv_unforce_levels(struct radeon_device *rdev) | 640 | static int kv_unforce_levels(struct radeon_device *rdev) |
640 | { | 641 | { |
641 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_NoForcedLevel); | 642 | if (rdev->family == CHIP_KABINI) |
643 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_NoForcedLevel); | ||
644 | else | ||
645 | return kv_set_enabled_levels(rdev); | ||
642 | } | 646 | } |
643 | 647 | ||
644 | static int kv_update_sclk_t(struct radeon_device *rdev) | 648 | static int kv_update_sclk_t(struct radeon_device *rdev) |
@@ -667,9 +671,8 @@ static int kv_program_bootup_state(struct radeon_device *rdev) | |||
667 | &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk; | 671 | &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk; |
668 | 672 | ||
669 | if (table && table->count) { | 673 | if (table && table->count) { |
670 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 674 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
671 | if ((table->entries[i].clk == pi->boot_pl.sclk) || | 675 | if (table->entries[i].clk == pi->boot_pl.sclk) |
672 | (i == 0)) | ||
673 | break; | 676 | break; |
674 | } | 677 | } |
675 | 678 | ||
@@ -682,9 +685,8 @@ static int kv_program_bootup_state(struct radeon_device *rdev) | |||
682 | if (table->num_max_dpm_entries == 0) | 685 | if (table->num_max_dpm_entries == 0) |
683 | return -EINVAL; | 686 | return -EINVAL; |
684 | 687 | ||
685 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 688 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
686 | if ((table->entries[i].sclk_frequency == pi->boot_pl.sclk) || | 689 | if (table->entries[i].sclk_frequency == pi->boot_pl.sclk) |
687 | (i == 0)) | ||
688 | break; | 690 | break; |
689 | } | 691 | } |
690 | 692 | ||
@@ -1078,6 +1080,13 @@ static int kv_enable_ulv(struct radeon_device *rdev, bool enable) | |||
1078 | PPSMC_MSG_EnableULV : PPSMC_MSG_DisableULV); | 1080 | PPSMC_MSG_EnableULV : PPSMC_MSG_DisableULV); |
1079 | } | 1081 | } |
1080 | 1082 | ||
1083 | static void kv_reset_acp_boot_level(struct radeon_device *rdev) | ||
1084 | { | ||
1085 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
1086 | |||
1087 | pi->acp_boot_level = 0xff; | ||
1088 | } | ||
1089 | |||
1081 | static void kv_update_current_ps(struct radeon_device *rdev, | 1090 | static void kv_update_current_ps(struct radeon_device *rdev, |
1082 | struct radeon_ps *rps) | 1091 | struct radeon_ps *rps) |
1083 | { | 1092 | { |
@@ -1100,6 +1109,18 @@ static void kv_update_requested_ps(struct radeon_device *rdev, | |||
1100 | pi->requested_rps.ps_priv = &pi->requested_ps; | 1109 | pi->requested_rps.ps_priv = &pi->requested_ps; |
1101 | } | 1110 | } |
1102 | 1111 | ||
1112 | void kv_dpm_enable_bapm(struct radeon_device *rdev, bool enable) | ||
1113 | { | ||
1114 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
1115 | int ret; | ||
1116 | |||
1117 | if (pi->bapm_enable) { | ||
1118 | ret = kv_smc_bapm_enable(rdev, enable); | ||
1119 | if (ret) | ||
1120 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
1121 | } | ||
1122 | } | ||
1123 | |||
1103 | int kv_dpm_enable(struct radeon_device *rdev) | 1124 | int kv_dpm_enable(struct radeon_device *rdev) |
1104 | { | 1125 | { |
1105 | struct kv_power_info *pi = kv_get_pi(rdev); | 1126 | struct kv_power_info *pi = kv_get_pi(rdev); |
@@ -1192,6 +1213,8 @@ int kv_dpm_enable(struct radeon_device *rdev) | |||
1192 | return ret; | 1213 | return ret; |
1193 | } | 1214 | } |
1194 | 1215 | ||
1216 | kv_reset_acp_boot_level(rdev); | ||
1217 | |||
1195 | if (rdev->irq.installed && | 1218 | if (rdev->irq.installed && |
1196 | r600_is_internal_thermal_sensor(rdev->pm.int_thermal_type)) { | 1219 | r600_is_internal_thermal_sensor(rdev->pm.int_thermal_type)) { |
1197 | ret = kv_set_thermal_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX); | 1220 | ret = kv_set_thermal_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX); |
@@ -1203,6 +1226,12 @@ int kv_dpm_enable(struct radeon_device *rdev) | |||
1203 | radeon_irq_set(rdev); | 1226 | radeon_irq_set(rdev); |
1204 | } | 1227 | } |
1205 | 1228 | ||
1229 | ret = kv_smc_bapm_enable(rdev, false); | ||
1230 | if (ret) { | ||
1231 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
1232 | return ret; | ||
1233 | } | ||
1234 | |||
1206 | /* powerdown unused blocks for now */ | 1235 | /* powerdown unused blocks for now */ |
1207 | kv_dpm_powergate_acp(rdev, true); | 1236 | kv_dpm_powergate_acp(rdev, true); |
1208 | kv_dpm_powergate_samu(rdev, true); | 1237 | kv_dpm_powergate_samu(rdev, true); |
@@ -1226,6 +1255,8 @@ void kv_dpm_disable(struct radeon_device *rdev) | |||
1226 | RADEON_CG_BLOCK_BIF | | 1255 | RADEON_CG_BLOCK_BIF | |
1227 | RADEON_CG_BLOCK_HDP), false); | 1256 | RADEON_CG_BLOCK_HDP), false); |
1228 | 1257 | ||
1258 | kv_smc_bapm_enable(rdev, false); | ||
1259 | |||
1229 | /* powerup blocks */ | 1260 | /* powerup blocks */ |
1230 | kv_dpm_powergate_acp(rdev, false); | 1261 | kv_dpm_powergate_acp(rdev, false); |
1231 | kv_dpm_powergate_samu(rdev, false); | 1262 | kv_dpm_powergate_samu(rdev, false); |
@@ -1450,6 +1481,39 @@ static int kv_update_samu_dpm(struct radeon_device *rdev, bool gate) | |||
1450 | return kv_enable_samu_dpm(rdev, !gate); | 1481 | return kv_enable_samu_dpm(rdev, !gate); |
1451 | } | 1482 | } |
1452 | 1483 | ||
1484 | static u8 kv_get_acp_boot_level(struct radeon_device *rdev) | ||
1485 | { | ||
1486 | u8 i; | ||
1487 | struct radeon_clock_voltage_dependency_table *table = | ||
1488 | &rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table; | ||
1489 | |||
1490 | for (i = 0; i < table->count; i++) { | ||
1491 | if (table->entries[i].clk >= 0) /* XXX */ | ||
1492 | break; | ||
1493 | } | ||
1494 | |||
1495 | if (i >= table->count) | ||
1496 | i = table->count - 1; | ||
1497 | |||
1498 | return i; | ||
1499 | } | ||
1500 | |||
1501 | static void kv_update_acp_boot_level(struct radeon_device *rdev) | ||
1502 | { | ||
1503 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
1504 | u8 acp_boot_level; | ||
1505 | |||
1506 | if (!pi->caps_stable_p_state) { | ||
1507 | acp_boot_level = kv_get_acp_boot_level(rdev); | ||
1508 | if (acp_boot_level != pi->acp_boot_level) { | ||
1509 | pi->acp_boot_level = acp_boot_level; | ||
1510 | kv_send_msg_to_smc_with_parameter(rdev, | ||
1511 | PPSMC_MSG_ACPDPM_SetEnabledMask, | ||
1512 | (1 << pi->acp_boot_level)); | ||
1513 | } | ||
1514 | } | ||
1515 | } | ||
1516 | |||
1453 | static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) | 1517 | static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) |
1454 | { | 1518 | { |
1455 | struct kv_power_info *pi = kv_get_pi(rdev); | 1519 | struct kv_power_info *pi = kv_get_pi(rdev); |
@@ -1461,7 +1525,7 @@ static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) | |||
1461 | if (pi->caps_stable_p_state) | 1525 | if (pi->caps_stable_p_state) |
1462 | pi->acp_boot_level = table->count - 1; | 1526 | pi->acp_boot_level = table->count - 1; |
1463 | else | 1527 | else |
1464 | pi->acp_boot_level = 0; | 1528 | pi->acp_boot_level = kv_get_acp_boot_level(rdev); |
1465 | 1529 | ||
1466 | ret = kv_copy_bytes_to_smc(rdev, | 1530 | ret = kv_copy_bytes_to_smc(rdev, |
1467 | pi->dpm_table_start + | 1531 | pi->dpm_table_start + |
@@ -1588,13 +1652,11 @@ static void kv_set_valid_clock_range(struct radeon_device *rdev, | |||
1588 | } | 1652 | } |
1589 | } | 1653 | } |
1590 | 1654 | ||
1591 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 1655 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
1592 | if ((table->entries[i].clk <= new_ps->levels[new_ps->num_levels -1].sclk) || | 1656 | if (table->entries[i].clk <= new_ps->levels[new_ps->num_levels - 1].sclk) |
1593 | (i == 0)) { | ||
1594 | pi->highest_valid = i; | ||
1595 | break; | 1657 | break; |
1596 | } | ||
1597 | } | 1658 | } |
1659 | pi->highest_valid = i; | ||
1598 | 1660 | ||
1599 | if (pi->lowest_valid > pi->highest_valid) { | 1661 | if (pi->lowest_valid > pi->highest_valid) { |
1600 | if ((new_ps->levels[0].sclk - table->entries[pi->highest_valid].clk) > | 1662 | if ((new_ps->levels[0].sclk - table->entries[pi->highest_valid].clk) > |
@@ -1615,14 +1677,12 @@ static void kv_set_valid_clock_range(struct radeon_device *rdev, | |||
1615 | } | 1677 | } |
1616 | } | 1678 | } |
1617 | 1679 | ||
1618 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 1680 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
1619 | if (table->entries[i].sclk_frequency <= | 1681 | if (table->entries[i].sclk_frequency <= |
1620 | new_ps->levels[new_ps->num_levels - 1].sclk || | 1682 | new_ps->levels[new_ps->num_levels - 1].sclk) |
1621 | i == 0) { | ||
1622 | pi->highest_valid = i; | ||
1623 | break; | 1683 | break; |
1624 | } | ||
1625 | } | 1684 | } |
1685 | pi->highest_valid = i; | ||
1626 | 1686 | ||
1627 | if (pi->lowest_valid > pi->highest_valid) { | 1687 | if (pi->lowest_valid > pi->highest_valid) { |
1628 | if ((new_ps->levels[0].sclk - | 1688 | if ((new_ps->levels[0].sclk - |
@@ -1724,6 +1784,14 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
1724 | RADEON_CG_BLOCK_BIF | | 1784 | RADEON_CG_BLOCK_BIF | |
1725 | RADEON_CG_BLOCK_HDP), false); | 1785 | RADEON_CG_BLOCK_HDP), false); |
1726 | 1786 | ||
1787 | if (pi->bapm_enable) { | ||
1788 | ret = kv_smc_bapm_enable(rdev, rdev->pm.dpm.ac_power); | ||
1789 | if (ret) { | ||
1790 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
1791 | return ret; | ||
1792 | } | ||
1793 | } | ||
1794 | |||
1727 | if (rdev->family == CHIP_KABINI) { | 1795 | if (rdev->family == CHIP_KABINI) { |
1728 | if (pi->enable_dpm) { | 1796 | if (pi->enable_dpm) { |
1729 | kv_set_valid_clock_range(rdev, new_ps); | 1797 | kv_set_valid_clock_range(rdev, new_ps); |
@@ -1775,6 +1843,7 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
1775 | return ret; | 1843 | return ret; |
1776 | } | 1844 | } |
1777 | #endif | 1845 | #endif |
1846 | kv_update_acp_boot_level(rdev); | ||
1778 | kv_update_sclk_t(rdev); | 1847 | kv_update_sclk_t(rdev); |
1779 | kv_enable_nb_dpm(rdev); | 1848 | kv_enable_nb_dpm(rdev); |
1780 | } | 1849 | } |
@@ -1785,7 +1854,6 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
1785 | RADEON_CG_BLOCK_BIF | | 1854 | RADEON_CG_BLOCK_BIF | |
1786 | RADEON_CG_BLOCK_HDP), true); | 1855 | RADEON_CG_BLOCK_HDP), true); |
1787 | 1856 | ||
1788 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1789 | return 0; | 1857 | return 0; |
1790 | } | 1858 | } |
1791 | 1859 | ||
@@ -1806,12 +1874,23 @@ void kv_dpm_setup_asic(struct radeon_device *rdev) | |||
1806 | 1874 | ||
1807 | void kv_dpm_reset_asic(struct radeon_device *rdev) | 1875 | void kv_dpm_reset_asic(struct radeon_device *rdev) |
1808 | { | 1876 | { |
1809 | kv_force_lowest_valid(rdev); | 1877 | struct kv_power_info *pi = kv_get_pi(rdev); |
1810 | kv_init_graphics_levels(rdev); | 1878 | |
1811 | kv_program_bootup_state(rdev); | 1879 | if (rdev->family == CHIP_KABINI) { |
1812 | kv_upload_dpm_settings(rdev); | 1880 | kv_force_lowest_valid(rdev); |
1813 | kv_force_lowest_valid(rdev); | 1881 | kv_init_graphics_levels(rdev); |
1814 | kv_unforce_levels(rdev); | 1882 | kv_program_bootup_state(rdev); |
1883 | kv_upload_dpm_settings(rdev); | ||
1884 | kv_force_lowest_valid(rdev); | ||
1885 | kv_unforce_levels(rdev); | ||
1886 | } else { | ||
1887 | kv_init_graphics_levels(rdev); | ||
1888 | kv_program_bootup_state(rdev); | ||
1889 | kv_freeze_sclk_dpm(rdev, true); | ||
1890 | kv_upload_dpm_settings(rdev); | ||
1891 | kv_freeze_sclk_dpm(rdev, false); | ||
1892 | kv_set_enabled_level(rdev, pi->graphics_boot_level); | ||
1893 | } | ||
1815 | } | 1894 | } |
1816 | 1895 | ||
1817 | //XXX use sumo_dpm_display_configuration_changed | 1896 | //XXX use sumo_dpm_display_configuration_changed |
@@ -1871,12 +1950,15 @@ static int kv_force_dpm_highest(struct radeon_device *rdev) | |||
1871 | if (ret) | 1950 | if (ret) |
1872 | return ret; | 1951 | return ret; |
1873 | 1952 | ||
1874 | for (i = SMU7_MAX_LEVELS_GRAPHICS - 1; i >= 0; i--) { | 1953 | for (i = SMU7_MAX_LEVELS_GRAPHICS - 1; i > 0; i--) { |
1875 | if (enable_mask & (1 << i)) | 1954 | if (enable_mask & (1 << i)) |
1876 | break; | 1955 | break; |
1877 | } | 1956 | } |
1878 | 1957 | ||
1879 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | 1958 | if (rdev->family == CHIP_KABINI) |
1959 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | ||
1960 | else | ||
1961 | return kv_set_enabled_level(rdev, i); | ||
1880 | } | 1962 | } |
1881 | 1963 | ||
1882 | static int kv_force_dpm_lowest(struct radeon_device *rdev) | 1964 | static int kv_force_dpm_lowest(struct radeon_device *rdev) |
@@ -1893,7 +1975,10 @@ static int kv_force_dpm_lowest(struct radeon_device *rdev) | |||
1893 | break; | 1975 | break; |
1894 | } | 1976 | } |
1895 | 1977 | ||
1896 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | 1978 | if (rdev->family == CHIP_KABINI) |
1979 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | ||
1980 | else | ||
1981 | return kv_set_enabled_level(rdev, i); | ||
1897 | } | 1982 | } |
1898 | 1983 | ||
1899 | static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, | 1984 | static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, |
@@ -1911,9 +1996,9 @@ static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, | |||
1911 | if (!pi->caps_sclk_ds) | 1996 | if (!pi->caps_sclk_ds) |
1912 | return 0; | 1997 | return 0; |
1913 | 1998 | ||
1914 | for (i = KV_MAX_DEEPSLEEP_DIVIDER_ID; i <= 0; i--) { | 1999 | for (i = KV_MAX_DEEPSLEEP_DIVIDER_ID; i > 0; i--) { |
1915 | temp = sclk / sumo_get_sleep_divider_from_id(i); | 2000 | temp = sclk / sumo_get_sleep_divider_from_id(i); |
1916 | if ((temp >= min) || (i == 0)) | 2001 | if (temp >= min) |
1917 | break; | 2002 | break; |
1918 | } | 2003 | } |
1919 | 2004 | ||
@@ -2039,12 +2124,12 @@ static void kv_apply_state_adjust_rules(struct radeon_device *rdev, | |||
2039 | ps->dpmx_nb_ps_lo = 0x1; | 2124 | ps->dpmx_nb_ps_lo = 0x1; |
2040 | ps->dpmx_nb_ps_hi = 0x0; | 2125 | ps->dpmx_nb_ps_hi = 0x0; |
2041 | } else { | 2126 | } else { |
2042 | ps->dpm0_pg_nb_ps_lo = 0x1; | 2127 | ps->dpm0_pg_nb_ps_lo = 0x3; |
2043 | ps->dpm0_pg_nb_ps_hi = 0x0; | 2128 | ps->dpm0_pg_nb_ps_hi = 0x0; |
2044 | ps->dpmx_nb_ps_lo = 0x2; | 2129 | ps->dpmx_nb_ps_lo = 0x3; |
2045 | ps->dpmx_nb_ps_hi = 0x1; | 2130 | ps->dpmx_nb_ps_hi = 0x0; |
2046 | 2131 | ||
2047 | if (pi->sys_info.nb_dpm_enable && pi->battery_state) { | 2132 | if (pi->sys_info.nb_dpm_enable) { |
2048 | force_high = (mclk >= pi->sys_info.nbp_memory_clock[3]) || | 2133 | force_high = (mclk >= pi->sys_info.nbp_memory_clock[3]) || |
2049 | pi->video_start || (rdev->pm.dpm.new_active_crtc_count >= 3) || | 2134 | pi->video_start || (rdev->pm.dpm.new_active_crtc_count >= 3) || |
2050 | pi->disable_nb_ps3_in_battery; | 2135 | pi->disable_nb_ps3_in_battery; |
@@ -2210,6 +2295,15 @@ static void kv_enable_new_levels(struct radeon_device *rdev) | |||
2210 | } | 2295 | } |
2211 | } | 2296 | } |
2212 | 2297 | ||
2298 | static int kv_set_enabled_level(struct radeon_device *rdev, u32 level) | ||
2299 | { | ||
2300 | u32 new_mask = (1 << level); | ||
2301 | |||
2302 | return kv_send_msg_to_smc_with_parameter(rdev, | ||
2303 | PPSMC_MSG_SCLKDPM_SetEnabledMask, | ||
2304 | new_mask); | ||
2305 | } | ||
2306 | |||
2213 | static int kv_set_enabled_levels(struct radeon_device *rdev) | 2307 | static int kv_set_enabled_levels(struct radeon_device *rdev) |
2214 | { | 2308 | { |
2215 | struct kv_power_info *pi = kv_get_pi(rdev); | 2309 | struct kv_power_info *pi = kv_get_pi(rdev); |
diff --git a/drivers/gpu/drm/radeon/kv_dpm.h b/drivers/gpu/drm/radeon/kv_dpm.h index 32bb079572d7..8cef7525d7a8 100644 --- a/drivers/gpu/drm/radeon/kv_dpm.h +++ b/drivers/gpu/drm/radeon/kv_dpm.h | |||
@@ -192,6 +192,7 @@ int kv_send_msg_to_smc_with_parameter(struct radeon_device *rdev, | |||
192 | int kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 192 | int kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
193 | u32 *value, u32 limit); | 193 | u32 *value, u32 limit); |
194 | int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable); | 194 | int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable); |
195 | int kv_smc_bapm_enable(struct radeon_device *rdev, bool enable); | ||
195 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, | 196 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, |
196 | u32 smc_start_address, | 197 | u32 smc_start_address, |
197 | const u8 *src, u32 byte_count, u32 limit); | 198 | const u8 *src, u32 byte_count, u32 limit); |
diff --git a/drivers/gpu/drm/radeon/kv_smc.c b/drivers/gpu/drm/radeon/kv_smc.c index 34a226d7e34a..0000b59a6d05 100644 --- a/drivers/gpu/drm/radeon/kv_smc.c +++ b/drivers/gpu/drm/radeon/kv_smc.c | |||
@@ -107,6 +107,14 @@ int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable) | |||
107 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DPM_Disable); | 107 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DPM_Disable); |
108 | } | 108 | } |
109 | 109 | ||
110 | int kv_smc_bapm_enable(struct radeon_device *rdev, bool enable) | ||
111 | { | ||
112 | if (enable) | ||
113 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_EnableBAPM); | ||
114 | else | ||
115 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DisableBAPM); | ||
116 | } | ||
117 | |||
110 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, | 118 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, |
111 | u32 smc_start_address, | 119 | u32 smc_start_address, |
112 | const u8 *src, u32 byte_count, u32 limit) | 120 | const u8 *src, u32 byte_count, u32 limit) |
diff --git a/drivers/gpu/drm/radeon/ni_dpm.c b/drivers/gpu/drm/radeon/ni_dpm.c index f7b625c9e0e9..f26339028154 100644 --- a/drivers/gpu/drm/radeon/ni_dpm.c +++ b/drivers/gpu/drm/radeon/ni_dpm.c | |||
@@ -787,6 +787,7 @@ static void ni_apply_state_adjust_rules(struct radeon_device *rdev, | |||
787 | bool disable_mclk_switching; | 787 | bool disable_mclk_switching; |
788 | u32 mclk, sclk; | 788 | u32 mclk, sclk; |
789 | u16 vddc, vddci; | 789 | u16 vddc, vddci; |
790 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
790 | int i; | 791 | int i; |
791 | 792 | ||
792 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 793 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
@@ -813,6 +814,29 @@ static void ni_apply_state_adjust_rules(struct radeon_device *rdev, | |||
813 | } | 814 | } |
814 | } | 815 | } |
815 | 816 | ||
817 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
818 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
819 | &max_sclk_vddc); | ||
820 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
821 | &max_mclk_vddci); | ||
822 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
823 | &max_mclk_vddc); | ||
824 | |||
825 | for (i = 0; i < ps->performance_level_count; i++) { | ||
826 | if (max_sclk_vddc) { | ||
827 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
828 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
829 | } | ||
830 | if (max_mclk_vddci) { | ||
831 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
832 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
833 | } | ||
834 | if (max_mclk_vddc) { | ||
835 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
836 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
837 | } | ||
838 | } | ||
839 | |||
816 | /* XXX validate the min clocks required for display */ | 840 | /* XXX validate the min clocks required for display */ |
817 | 841 | ||
818 | if (disable_mclk_switching) { | 842 | if (disable_mclk_switching) { |
@@ -3865,12 +3889,6 @@ int ni_dpm_set_power_state(struct radeon_device *rdev) | |||
3865 | return ret; | 3889 | return ret; |
3866 | } | 3890 | } |
3867 | 3891 | ||
3868 | ret = ni_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
3869 | if (ret) { | ||
3870 | DRM_ERROR("ni_dpm_force_performance_level failed\n"); | ||
3871 | return ret; | ||
3872 | } | ||
3873 | |||
3874 | return 0; | 3892 | return 0; |
3875 | } | 3893 | } |
3876 | 3894 | ||
diff --git a/drivers/gpu/drm/radeon/ppsmc.h b/drivers/gpu/drm/radeon/ppsmc.h index 682842804bce..5670b8291285 100644 --- a/drivers/gpu/drm/radeon/ppsmc.h +++ b/drivers/gpu/drm/radeon/ppsmc.h | |||
@@ -163,6 +163,8 @@ typedef uint8_t PPSMC_Result; | |||
163 | #define PPSMC_MSG_VCEPowerON ((uint32_t) 0x10f) | 163 | #define PPSMC_MSG_VCEPowerON ((uint32_t) 0x10f) |
164 | #define PPSMC_MSG_DCE_RemoveVoltageAdjustment ((uint32_t) 0x11d) | 164 | #define PPSMC_MSG_DCE_RemoveVoltageAdjustment ((uint32_t) 0x11d) |
165 | #define PPSMC_MSG_DCE_AllowVoltageAdjustment ((uint32_t) 0x11e) | 165 | #define PPSMC_MSG_DCE_AllowVoltageAdjustment ((uint32_t) 0x11e) |
166 | #define PPSMC_MSG_EnableBAPM ((uint32_t) 0x120) | ||
167 | #define PPSMC_MSG_DisableBAPM ((uint32_t) 0x121) | ||
166 | #define PPSMC_MSG_UVD_DPM_Config ((uint32_t) 0x124) | 168 | #define PPSMC_MSG_UVD_DPM_Config ((uint32_t) 0x124) |
167 | 169 | ||
168 | 170 | ||
diff --git a/drivers/gpu/drm/radeon/r100.c b/drivers/gpu/drm/radeon/r100.c index 9fc61dd68bc0..d71333033b2b 100644 --- a/drivers/gpu/drm/radeon/r100.c +++ b/drivers/gpu/drm/radeon/r100.c | |||
@@ -2853,21 +2853,28 @@ static void r100_pll_errata_after_data(struct radeon_device *rdev) | |||
2853 | 2853 | ||
2854 | uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg) | 2854 | uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg) |
2855 | { | 2855 | { |
2856 | unsigned long flags; | ||
2856 | uint32_t data; | 2857 | uint32_t data; |
2857 | 2858 | ||
2859 | spin_lock_irqsave(&rdev->pll_idx_lock, flags); | ||
2858 | WREG8(RADEON_CLOCK_CNTL_INDEX, reg & 0x3f); | 2860 | WREG8(RADEON_CLOCK_CNTL_INDEX, reg & 0x3f); |
2859 | r100_pll_errata_after_index(rdev); | 2861 | r100_pll_errata_after_index(rdev); |
2860 | data = RREG32(RADEON_CLOCK_CNTL_DATA); | 2862 | data = RREG32(RADEON_CLOCK_CNTL_DATA); |
2861 | r100_pll_errata_after_data(rdev); | 2863 | r100_pll_errata_after_data(rdev); |
2864 | spin_unlock_irqrestore(&rdev->pll_idx_lock, flags); | ||
2862 | return data; | 2865 | return data; |
2863 | } | 2866 | } |
2864 | 2867 | ||
2865 | void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 2868 | void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
2866 | { | 2869 | { |
2870 | unsigned long flags; | ||
2871 | |||
2872 | spin_lock_irqsave(&rdev->pll_idx_lock, flags); | ||
2867 | WREG8(RADEON_CLOCK_CNTL_INDEX, ((reg & 0x3f) | RADEON_PLL_WR_EN)); | 2873 | WREG8(RADEON_CLOCK_CNTL_INDEX, ((reg & 0x3f) | RADEON_PLL_WR_EN)); |
2868 | r100_pll_errata_after_index(rdev); | 2874 | r100_pll_errata_after_index(rdev); |
2869 | WREG32(RADEON_CLOCK_CNTL_DATA, v); | 2875 | WREG32(RADEON_CLOCK_CNTL_DATA, v); |
2870 | r100_pll_errata_after_data(rdev); | 2876 | r100_pll_errata_after_data(rdev); |
2877 | spin_unlock_irqrestore(&rdev->pll_idx_lock, flags); | ||
2871 | } | 2878 | } |
2872 | 2879 | ||
2873 | static void r100_set_safe_registers(struct radeon_device *rdev) | 2880 | static void r100_set_safe_registers(struct radeon_device *rdev) |
@@ -2926,9 +2933,11 @@ static int r100_debugfs_cp_ring_info(struct seq_file *m, void *data) | |||
2926 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", rdp); | 2933 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", rdp); |
2927 | seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw); | 2934 | seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw); |
2928 | seq_printf(m, "%u dwords in ring\n", count); | 2935 | seq_printf(m, "%u dwords in ring\n", count); |
2929 | for (j = 0; j <= count; j++) { | 2936 | if (ring->ready) { |
2930 | i = (rdp + j) & ring->ptr_mask; | 2937 | for (j = 0; j <= count; j++) { |
2931 | seq_printf(m, "r[%04d]=0x%08x\n", i, ring->ring[i]); | 2938 | i = (rdp + j) & ring->ptr_mask; |
2939 | seq_printf(m, "r[%04d]=0x%08x\n", i, ring->ring[i]); | ||
2940 | } | ||
2932 | } | 2941 | } |
2933 | return 0; | 2942 | return 0; |
2934 | } | 2943 | } |
diff --git a/drivers/gpu/drm/radeon/r420.c b/drivers/gpu/drm/radeon/r420.c index 4e796ecf9ea4..6edf2b3a52b4 100644 --- a/drivers/gpu/drm/radeon/r420.c +++ b/drivers/gpu/drm/radeon/r420.c | |||
@@ -160,18 +160,25 @@ void r420_pipes_init(struct radeon_device *rdev) | |||
160 | 160 | ||
161 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) | 161 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) |
162 | { | 162 | { |
163 | unsigned long flags; | ||
163 | u32 r; | 164 | u32 r; |
164 | 165 | ||
166 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
165 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); | 167 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); |
166 | r = RREG32(R_0001FC_MC_IND_DATA); | 168 | r = RREG32(R_0001FC_MC_IND_DATA); |
169 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
167 | return r; | 170 | return r; |
168 | } | 171 | } |
169 | 172 | ||
170 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 173 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
171 | { | 174 | { |
175 | unsigned long flags; | ||
176 | |||
177 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
172 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | | 178 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | |
173 | S_0001F8_MC_IND_WR_EN(1)); | 179 | S_0001F8_MC_IND_WR_EN(1)); |
174 | WREG32(R_0001FC_MC_IND_DATA, v); | 180 | WREG32(R_0001FC_MC_IND_DATA, v); |
181 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
175 | } | 182 | } |
176 | 183 | ||
177 | static void r420_debugfs(struct radeon_device *rdev) | 184 | static void r420_debugfs(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index ea4d3734e6d9..2a1b1876b431 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c | |||
@@ -119,6 +119,11 @@ u32 r600_get_xclk(struct radeon_device *rdev) | |||
119 | return rdev->clock.spll.reference_freq; | 119 | return rdev->clock.spll.reference_freq; |
120 | } | 120 | } |
121 | 121 | ||
122 | int r600_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk) | ||
123 | { | ||
124 | return 0; | ||
125 | } | ||
126 | |||
122 | /* get temperature in millidegrees */ | 127 | /* get temperature in millidegrees */ |
123 | int rv6xx_get_temp(struct radeon_device *rdev) | 128 | int rv6xx_get_temp(struct radeon_device *rdev) |
124 | { | 129 | { |
@@ -1045,20 +1050,27 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev) | |||
1045 | 1050 | ||
1046 | uint32_t rs780_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 1051 | uint32_t rs780_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
1047 | { | 1052 | { |
1053 | unsigned long flags; | ||
1048 | uint32_t r; | 1054 | uint32_t r; |
1049 | 1055 | ||
1056 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
1050 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg)); | 1057 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg)); |
1051 | r = RREG32(R_0028FC_MC_DATA); | 1058 | r = RREG32(R_0028FC_MC_DATA); |
1052 | WREG32(R_0028F8_MC_INDEX, ~C_0028F8_MC_IND_ADDR); | 1059 | WREG32(R_0028F8_MC_INDEX, ~C_0028F8_MC_IND_ADDR); |
1060 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
1053 | return r; | 1061 | return r; |
1054 | } | 1062 | } |
1055 | 1063 | ||
1056 | void rs780_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 1064 | void rs780_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
1057 | { | 1065 | { |
1066 | unsigned long flags; | ||
1067 | |||
1068 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
1058 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg) | | 1069 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg) | |
1059 | S_0028F8_MC_IND_WR_EN(1)); | 1070 | S_0028F8_MC_IND_WR_EN(1)); |
1060 | WREG32(R_0028FC_MC_DATA, v); | 1071 | WREG32(R_0028FC_MC_DATA, v); |
1061 | WREG32(R_0028F8_MC_INDEX, 0x7F); | 1072 | WREG32(R_0028F8_MC_INDEX, 0x7F); |
1073 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
1062 | } | 1074 | } |
1063 | 1075 | ||
1064 | static void r600_mc_program(struct radeon_device *rdev) | 1076 | static void r600_mc_program(struct radeon_device *rdev) |
@@ -2092,20 +2104,27 @@ static void r600_gpu_init(struct radeon_device *rdev) | |||
2092 | */ | 2104 | */ |
2093 | u32 r600_pciep_rreg(struct radeon_device *rdev, u32 reg) | 2105 | u32 r600_pciep_rreg(struct radeon_device *rdev, u32 reg) |
2094 | { | 2106 | { |
2107 | unsigned long flags; | ||
2095 | u32 r; | 2108 | u32 r; |
2096 | 2109 | ||
2110 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
2097 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); | 2111 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); |
2098 | (void)RREG32(PCIE_PORT_INDEX); | 2112 | (void)RREG32(PCIE_PORT_INDEX); |
2099 | r = RREG32(PCIE_PORT_DATA); | 2113 | r = RREG32(PCIE_PORT_DATA); |
2114 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
2100 | return r; | 2115 | return r; |
2101 | } | 2116 | } |
2102 | 2117 | ||
2103 | void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2118 | void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2104 | { | 2119 | { |
2120 | unsigned long flags; | ||
2121 | |||
2122 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
2105 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); | 2123 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); |
2106 | (void)RREG32(PCIE_PORT_INDEX); | 2124 | (void)RREG32(PCIE_PORT_INDEX); |
2107 | WREG32(PCIE_PORT_DATA, (v)); | 2125 | WREG32(PCIE_PORT_DATA, (v)); |
2108 | (void)RREG32(PCIE_PORT_DATA); | 2126 | (void)RREG32(PCIE_PORT_DATA); |
2127 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
2109 | } | 2128 | } |
2110 | 2129 | ||
2111 | /* | 2130 | /* |
diff --git a/drivers/gpu/drm/radeon/r600_dpm.c b/drivers/gpu/drm/radeon/r600_dpm.c index fa0de46fcc0d..5513d8f06252 100644 --- a/drivers/gpu/drm/radeon/r600_dpm.c +++ b/drivers/gpu/drm/radeon/r600_dpm.c | |||
@@ -1084,7 +1084,7 @@ int r600_parse_extended_power_table(struct radeon_device *rdev) | |||
1084 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].dclk = | 1084 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].dclk = |
1085 | le16_to_cpu(uvd_clk->usDClkLow) | (uvd_clk->ucDClkHigh << 16); | 1085 | le16_to_cpu(uvd_clk->usDClkLow) | (uvd_clk->ucDClkHigh << 16); |
1086 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].v = | 1086 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].v = |
1087 | le16_to_cpu(limits->entries[i].usVoltage); | 1087 | le16_to_cpu(entry->usVoltage); |
1088 | entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) | 1088 | entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) |
1089 | ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); | 1089 | ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); |
1090 | } | 1090 | } |
@@ -1219,30 +1219,20 @@ int r600_parse_extended_power_table(struct radeon_device *rdev) | |||
1219 | 1219 | ||
1220 | void r600_free_extended_power_table(struct radeon_device *rdev) | 1220 | void r600_free_extended_power_table(struct radeon_device *rdev) |
1221 | { | 1221 | { |
1222 | if (rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries) | 1222 | struct radeon_dpm_dynamic_state *dyn_state = &rdev->pm.dpm.dyn_state; |
1223 | kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); | 1223 | |
1224 | if (rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries) | 1224 | kfree(dyn_state->vddc_dependency_on_sclk.entries); |
1225 | kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); | 1225 | kfree(dyn_state->vddci_dependency_on_mclk.entries); |
1226 | if (rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries) | 1226 | kfree(dyn_state->vddc_dependency_on_mclk.entries); |
1227 | kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries); | 1227 | kfree(dyn_state->mvdd_dependency_on_mclk.entries); |
1228 | if (rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries) | 1228 | kfree(dyn_state->cac_leakage_table.entries); |
1229 | kfree(rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries); | 1229 | kfree(dyn_state->phase_shedding_limits_table.entries); |
1230 | if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries) | 1230 | kfree(dyn_state->ppm_table); |
1231 | kfree(rdev->pm.dpm.dyn_state.cac_leakage_table.entries); | 1231 | kfree(dyn_state->cac_tdp_table); |
1232 | if (rdev->pm.dpm.dyn_state.phase_shedding_limits_table.entries) | 1232 | kfree(dyn_state->vce_clock_voltage_dependency_table.entries); |
1233 | kfree(rdev->pm.dpm.dyn_state.phase_shedding_limits_table.entries); | 1233 | kfree(dyn_state->uvd_clock_voltage_dependency_table.entries); |
1234 | if (rdev->pm.dpm.dyn_state.ppm_table) | 1234 | kfree(dyn_state->samu_clock_voltage_dependency_table.entries); |
1235 | kfree(rdev->pm.dpm.dyn_state.ppm_table); | 1235 | kfree(dyn_state->acp_clock_voltage_dependency_table.entries); |
1236 | if (rdev->pm.dpm.dyn_state.cac_tdp_table) | ||
1237 | kfree(rdev->pm.dpm.dyn_state.cac_tdp_table); | ||
1238 | if (rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries) | ||
1239 | kfree(rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries); | ||
1240 | if (rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries) | ||
1241 | kfree(rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries); | ||
1242 | if (rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries) | ||
1243 | kfree(rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries); | ||
1244 | if (rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries) | ||
1245 | kfree(rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries); | ||
1246 | } | 1236 | } |
1247 | 1237 | ||
1248 | enum radeon_pcie_gen r600_get_pcie_gen_support(struct radeon_device *rdev, | 1238 | enum radeon_pcie_gen r600_get_pcie_gen_support(struct radeon_device *rdev, |
diff --git a/drivers/gpu/drm/radeon/r600_hdmi.c b/drivers/gpu/drm/radeon/r600_hdmi.c index f443010ce90b..5b729319f27b 100644 --- a/drivers/gpu/drm/radeon/r600_hdmi.c +++ b/drivers/gpu/drm/radeon/r600_hdmi.c | |||
@@ -57,15 +57,15 @@ enum r600_hdmi_iec_status_bits { | |||
57 | static const struct radeon_hdmi_acr r600_hdmi_predefined_acr[] = { | 57 | static const struct radeon_hdmi_acr r600_hdmi_predefined_acr[] = { |
58 | /* 32kHz 44.1kHz 48kHz */ | 58 | /* 32kHz 44.1kHz 48kHz */ |
59 | /* Clock N CTS N CTS N CTS */ | 59 | /* Clock N CTS N CTS N CTS */ |
60 | { 25174, 4576, 28125, 7007, 31250, 6864, 28125 }, /* 25,20/1.001 MHz */ | 60 | { 25175, 4576, 28125, 7007, 31250, 6864, 28125 }, /* 25,20/1.001 MHz */ |
61 | { 25200, 4096, 25200, 6272, 28000, 6144, 25200 }, /* 25.20 MHz */ | 61 | { 25200, 4096, 25200, 6272, 28000, 6144, 25200 }, /* 25.20 MHz */ |
62 | { 27000, 4096, 27000, 6272, 30000, 6144, 27000 }, /* 27.00 MHz */ | 62 | { 27000, 4096, 27000, 6272, 30000, 6144, 27000 }, /* 27.00 MHz */ |
63 | { 27027, 4096, 27027, 6272, 30030, 6144, 27027 }, /* 27.00*1.001 MHz */ | 63 | { 27027, 4096, 27027, 6272, 30030, 6144, 27027 }, /* 27.00*1.001 MHz */ |
64 | { 54000, 4096, 54000, 6272, 60000, 6144, 54000 }, /* 54.00 MHz */ | 64 | { 54000, 4096, 54000, 6272, 60000, 6144, 54000 }, /* 54.00 MHz */ |
65 | { 54054, 4096, 54054, 6272, 60060, 6144, 54054 }, /* 54.00*1.001 MHz */ | 65 | { 54054, 4096, 54054, 6272, 60060, 6144, 54054 }, /* 54.00*1.001 MHz */ |
66 | { 74175, 11648, 210937, 17836, 234375, 11648, 140625 }, /* 74.25/1.001 MHz */ | 66 | { 74176, 11648, 210937, 17836, 234375, 11648, 140625 }, /* 74.25/1.001 MHz */ |
67 | { 74250, 4096, 74250, 6272, 82500, 6144, 74250 }, /* 74.25 MHz */ | 67 | { 74250, 4096, 74250, 6272, 82500, 6144, 74250 }, /* 74.25 MHz */ |
68 | { 148351, 11648, 421875, 8918, 234375, 5824, 140625 }, /* 148.50/1.001 MHz */ | 68 | { 148352, 11648, 421875, 8918, 234375, 5824, 140625 }, /* 148.50/1.001 MHz */ |
69 | { 148500, 4096, 148500, 6272, 165000, 6144, 148500 }, /* 148.50 MHz */ | 69 | { 148500, 4096, 148500, 6272, 165000, 6144, 148500 }, /* 148.50 MHz */ |
70 | { 0, 4096, 0, 6272, 0, 6144, 0 } /* Other */ | 70 | { 0, 4096, 0, 6272, 0, 6144, 0 } /* Other */ |
71 | }; | 71 | }; |
@@ -75,8 +75,15 @@ static const struct radeon_hdmi_acr r600_hdmi_predefined_acr[] = { | |||
75 | */ | 75 | */ |
76 | static void r600_hdmi_calc_cts(uint32_t clock, int *CTS, int N, int freq) | 76 | static void r600_hdmi_calc_cts(uint32_t clock, int *CTS, int N, int freq) |
77 | { | 77 | { |
78 | if (*CTS == 0) | 78 | u64 n; |
79 | *CTS = clock * N / (128 * freq) * 1000; | 79 | u32 d; |
80 | |||
81 | if (*CTS == 0) { | ||
82 | n = (u64)clock * (u64)N * 1000ULL; | ||
83 | d = 128 * freq; | ||
84 | do_div(n, d); | ||
85 | *CTS = n; | ||
86 | } | ||
80 | DRM_DEBUG("Using ACR timing N=%d CTS=%d for frequency %d\n", | 87 | DRM_DEBUG("Using ACR timing N=%d CTS=%d for frequency %d\n", |
81 | N, *CTS, freq); | 88 | N, *CTS, freq); |
82 | } | 89 | } |
@@ -257,10 +264,7 @@ void r600_audio_set_dto(struct drm_encoder *encoder, u32 clock) | |||
257 | * number (coefficient of two integer numbers. DCCG_AUDIO_DTOx_PHASE | 264 | * number (coefficient of two integer numbers. DCCG_AUDIO_DTOx_PHASE |
258 | * is the numerator, DCCG_AUDIO_DTOx_MODULE is the denominator | 265 | * is the numerator, DCCG_AUDIO_DTOx_MODULE is the denominator |
259 | */ | 266 | */ |
260 | if (ASIC_IS_DCE3(rdev)) { | 267 | if (ASIC_IS_DCE32(rdev)) { |
261 | /* according to the reg specs, this should DCE3.2 only, but in | ||
262 | * practice it seems to cover DCE3.0 as well. | ||
263 | */ | ||
264 | if (dig->dig_encoder == 0) { | 268 | if (dig->dig_encoder == 0) { |
265 | dto_cntl = RREG32(DCCG_AUDIO_DTO0_CNTL) & ~DCCG_AUDIO_DTO_WALLCLOCK_RATIO_MASK; | 269 | dto_cntl = RREG32(DCCG_AUDIO_DTO0_CNTL) & ~DCCG_AUDIO_DTO_WALLCLOCK_RATIO_MASK; |
266 | dto_cntl |= DCCG_AUDIO_DTO_WALLCLOCK_RATIO(wallclock_ratio); | 270 | dto_cntl |= DCCG_AUDIO_DTO_WALLCLOCK_RATIO(wallclock_ratio); |
@@ -276,8 +280,21 @@ void r600_audio_set_dto(struct drm_encoder *encoder, u32 clock) | |||
276 | WREG32(DCCG_AUDIO_DTO1_MODULE, dto_modulo); | 280 | WREG32(DCCG_AUDIO_DTO1_MODULE, dto_modulo); |
277 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ | 281 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ |
278 | } | 282 | } |
283 | } else if (ASIC_IS_DCE3(rdev)) { | ||
284 | /* according to the reg specs, this should DCE3.2 only, but in | ||
285 | * practice it seems to cover DCE3.0/3.1 as well. | ||
286 | */ | ||
287 | if (dig->dig_encoder == 0) { | ||
288 | WREG32(DCCG_AUDIO_DTO0_PHASE, base_rate * 100); | ||
289 | WREG32(DCCG_AUDIO_DTO0_MODULE, clock * 100); | ||
290 | WREG32(DCCG_AUDIO_DTO_SELECT, 0); /* select DTO0 */ | ||
291 | } else { | ||
292 | WREG32(DCCG_AUDIO_DTO1_PHASE, base_rate * 100); | ||
293 | WREG32(DCCG_AUDIO_DTO1_MODULE, clock * 100); | ||
294 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ | ||
295 | } | ||
279 | } else { | 296 | } else { |
280 | /* according to the reg specs, this should be DCE2.0 and DCE3.0 */ | 297 | /* according to the reg specs, this should be DCE2.0 and DCE3.0/3.1 */ |
281 | WREG32(AUDIO_DTO, AUDIO_DTO_PHASE(base_rate / 10) | | 298 | WREG32(AUDIO_DTO, AUDIO_DTO_PHASE(base_rate / 10) | |
282 | AUDIO_DTO_MODULE(clock / 10)); | 299 | AUDIO_DTO_MODULE(clock / 10)); |
283 | } | 300 | } |
@@ -434,8 +451,8 @@ void r600_hdmi_setmode(struct drm_encoder *encoder, struct drm_display_mode *mod | |||
434 | } | 451 | } |
435 | 452 | ||
436 | WREG32(HDMI0_ACR_PACKET_CONTROL + offset, | 453 | WREG32(HDMI0_ACR_PACKET_CONTROL + offset, |
437 | HDMI0_ACR_AUTO_SEND | /* allow hw to sent ACR packets when required */ | 454 | HDMI0_ACR_SOURCE | /* select SW CTS value - XXX verify that hw CTS works on all families */ |
438 | HDMI0_ACR_SOURCE); /* select SW CTS value */ | 455 | HDMI0_ACR_AUTO_SEND); /* allow hw to sent ACR packets when required */ |
439 | 456 | ||
440 | WREG32(HDMI0_VBI_PACKET_CONTROL + offset, | 457 | WREG32(HDMI0_VBI_PACKET_CONTROL + offset, |
441 | HDMI0_NULL_SEND | /* send null packets when required */ | 458 | HDMI0_NULL_SEND | /* send null packets when required */ |
diff --git a/drivers/gpu/drm/radeon/r600d.h b/drivers/gpu/drm/radeon/r600d.h index 454f90a849e4..7b3c7b5932c5 100644 --- a/drivers/gpu/drm/radeon/r600d.h +++ b/drivers/gpu/drm/radeon/r600d.h | |||
@@ -1040,7 +1040,7 @@ | |||
1040 | # define HDMI0_AVI_INFO_CONT (1 << 1) | 1040 | # define HDMI0_AVI_INFO_CONT (1 << 1) |
1041 | # define HDMI0_AUDIO_INFO_SEND (1 << 4) | 1041 | # define HDMI0_AUDIO_INFO_SEND (1 << 4) |
1042 | # define HDMI0_AUDIO_INFO_CONT (1 << 5) | 1042 | # define HDMI0_AUDIO_INFO_CONT (1 << 5) |
1043 | # define HDMI0_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hmdi regs */ | 1043 | # define HDMI0_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hdmi regs */ |
1044 | # define HDMI0_AUDIO_INFO_UPDATE (1 << 7) | 1044 | # define HDMI0_AUDIO_INFO_UPDATE (1 << 7) |
1045 | # define HDMI0_MPEG_INFO_SEND (1 << 8) | 1045 | # define HDMI0_MPEG_INFO_SEND (1 << 8) |
1046 | # define HDMI0_MPEG_INFO_CONT (1 << 9) | 1046 | # define HDMI0_MPEG_INFO_CONT (1 << 9) |
@@ -1523,7 +1523,7 @@ | |||
1523 | */ | 1523 | */ |
1524 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) | 1524 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) |
1525 | /* COMMAND */ | 1525 | /* COMMAND */ |
1526 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 23) | 1526 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 22) |
1527 | /* 0 - none | 1527 | /* 0 - none |
1528 | * 1 - 8 in 16 | 1528 | * 1 - 8 in 16 |
1529 | * 2 - 8 in 32 | 1529 | * 2 - 8 in 32 |
diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h index ff8b564ce2b2..a400ac1c4147 100644 --- a/drivers/gpu/drm/radeon/radeon.h +++ b/drivers/gpu/drm/radeon/radeon.h | |||
@@ -181,7 +181,7 @@ extern int radeon_aspm; | |||
181 | #define RADEON_CG_SUPPORT_HDP_MGCG (1 << 16) | 181 | #define RADEON_CG_SUPPORT_HDP_MGCG (1 << 16) |
182 | 182 | ||
183 | /* PG flags */ | 183 | /* PG flags */ |
184 | #define RADEON_PG_SUPPORT_GFX_CG (1 << 0) | 184 | #define RADEON_PG_SUPPORT_GFX_PG (1 << 0) |
185 | #define RADEON_PG_SUPPORT_GFX_SMG (1 << 1) | 185 | #define RADEON_PG_SUPPORT_GFX_SMG (1 << 1) |
186 | #define RADEON_PG_SUPPORT_GFX_DMG (1 << 2) | 186 | #define RADEON_PG_SUPPORT_GFX_DMG (1 << 2) |
187 | #define RADEON_PG_SUPPORT_UVD (1 << 3) | 187 | #define RADEON_PG_SUPPORT_UVD (1 << 3) |
@@ -1778,6 +1778,7 @@ struct radeon_asic { | |||
1778 | int (*force_performance_level)(struct radeon_device *rdev, enum radeon_dpm_forced_level level); | 1778 | int (*force_performance_level)(struct radeon_device *rdev, enum radeon_dpm_forced_level level); |
1779 | bool (*vblank_too_short)(struct radeon_device *rdev); | 1779 | bool (*vblank_too_short)(struct radeon_device *rdev); |
1780 | void (*powergate_uvd)(struct radeon_device *rdev, bool gate); | 1780 | void (*powergate_uvd)(struct radeon_device *rdev, bool gate); |
1781 | void (*enable_bapm)(struct radeon_device *rdev, bool enable); | ||
1781 | } dpm; | 1782 | } dpm; |
1782 | /* pageflipping */ | 1783 | /* pageflipping */ |
1783 | struct { | 1784 | struct { |
@@ -2110,6 +2111,28 @@ struct radeon_device { | |||
2110 | resource_size_t rmmio_size; | 2111 | resource_size_t rmmio_size; |
2111 | /* protects concurrent MM_INDEX/DATA based register access */ | 2112 | /* protects concurrent MM_INDEX/DATA based register access */ |
2112 | spinlock_t mmio_idx_lock; | 2113 | spinlock_t mmio_idx_lock; |
2114 | /* protects concurrent SMC based register access */ | ||
2115 | spinlock_t smc_idx_lock; | ||
2116 | /* protects concurrent PLL register access */ | ||
2117 | spinlock_t pll_idx_lock; | ||
2118 | /* protects concurrent MC register access */ | ||
2119 | spinlock_t mc_idx_lock; | ||
2120 | /* protects concurrent PCIE register access */ | ||
2121 | spinlock_t pcie_idx_lock; | ||
2122 | /* protects concurrent PCIE_PORT register access */ | ||
2123 | spinlock_t pciep_idx_lock; | ||
2124 | /* protects concurrent PIF register access */ | ||
2125 | spinlock_t pif_idx_lock; | ||
2126 | /* protects concurrent CG register access */ | ||
2127 | spinlock_t cg_idx_lock; | ||
2128 | /* protects concurrent UVD register access */ | ||
2129 | spinlock_t uvd_idx_lock; | ||
2130 | /* protects concurrent RCU register access */ | ||
2131 | spinlock_t rcu_idx_lock; | ||
2132 | /* protects concurrent DIDT register access */ | ||
2133 | spinlock_t didt_idx_lock; | ||
2134 | /* protects concurrent ENDPOINT (audio) register access */ | ||
2135 | spinlock_t end_idx_lock; | ||
2113 | void __iomem *rmmio; | 2136 | void __iomem *rmmio; |
2114 | radeon_rreg_t mc_rreg; | 2137 | radeon_rreg_t mc_rreg; |
2115 | radeon_wreg_t mc_wreg; | 2138 | radeon_wreg_t mc_wreg; |
@@ -2277,123 +2300,179 @@ void cik_mm_wdoorbell(struct radeon_device *rdev, u32 offset, u32 v); | |||
2277 | */ | 2300 | */ |
2278 | static inline uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg) | 2301 | static inline uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg) |
2279 | { | 2302 | { |
2303 | unsigned long flags; | ||
2280 | uint32_t r; | 2304 | uint32_t r; |
2281 | 2305 | ||
2306 | spin_lock_irqsave(&rdev->pcie_idx_lock, flags); | ||
2282 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); | 2307 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); |
2283 | r = RREG32(RADEON_PCIE_DATA); | 2308 | r = RREG32(RADEON_PCIE_DATA); |
2309 | spin_unlock_irqrestore(&rdev->pcie_idx_lock, flags); | ||
2284 | return r; | 2310 | return r; |
2285 | } | 2311 | } |
2286 | 2312 | ||
2287 | static inline void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 2313 | static inline void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
2288 | { | 2314 | { |
2315 | unsigned long flags; | ||
2316 | |||
2317 | spin_lock_irqsave(&rdev->pcie_idx_lock, flags); | ||
2289 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); | 2318 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); |
2290 | WREG32(RADEON_PCIE_DATA, (v)); | 2319 | WREG32(RADEON_PCIE_DATA, (v)); |
2320 | spin_unlock_irqrestore(&rdev->pcie_idx_lock, flags); | ||
2291 | } | 2321 | } |
2292 | 2322 | ||
2293 | static inline u32 tn_smc_rreg(struct radeon_device *rdev, u32 reg) | 2323 | static inline u32 tn_smc_rreg(struct radeon_device *rdev, u32 reg) |
2294 | { | 2324 | { |
2325 | unsigned long flags; | ||
2295 | u32 r; | 2326 | u32 r; |
2296 | 2327 | ||
2328 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
2297 | WREG32(TN_SMC_IND_INDEX_0, (reg)); | 2329 | WREG32(TN_SMC_IND_INDEX_0, (reg)); |
2298 | r = RREG32(TN_SMC_IND_DATA_0); | 2330 | r = RREG32(TN_SMC_IND_DATA_0); |
2331 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
2299 | return r; | 2332 | return r; |
2300 | } | 2333 | } |
2301 | 2334 | ||
2302 | static inline void tn_smc_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2335 | static inline void tn_smc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2303 | { | 2336 | { |
2337 | unsigned long flags; | ||
2338 | |||
2339 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
2304 | WREG32(TN_SMC_IND_INDEX_0, (reg)); | 2340 | WREG32(TN_SMC_IND_INDEX_0, (reg)); |
2305 | WREG32(TN_SMC_IND_DATA_0, (v)); | 2341 | WREG32(TN_SMC_IND_DATA_0, (v)); |
2342 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
2306 | } | 2343 | } |
2307 | 2344 | ||
2308 | static inline u32 r600_rcu_rreg(struct radeon_device *rdev, u32 reg) | 2345 | static inline u32 r600_rcu_rreg(struct radeon_device *rdev, u32 reg) |
2309 | { | 2346 | { |
2347 | unsigned long flags; | ||
2310 | u32 r; | 2348 | u32 r; |
2311 | 2349 | ||
2350 | spin_lock_irqsave(&rdev->rcu_idx_lock, flags); | ||
2312 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); | 2351 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); |
2313 | r = RREG32(R600_RCU_DATA); | 2352 | r = RREG32(R600_RCU_DATA); |
2353 | spin_unlock_irqrestore(&rdev->rcu_idx_lock, flags); | ||
2314 | return r; | 2354 | return r; |
2315 | } | 2355 | } |
2316 | 2356 | ||
2317 | static inline void r600_rcu_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2357 | static inline void r600_rcu_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2318 | { | 2358 | { |
2359 | unsigned long flags; | ||
2360 | |||
2361 | spin_lock_irqsave(&rdev->rcu_idx_lock, flags); | ||
2319 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); | 2362 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); |
2320 | WREG32(R600_RCU_DATA, (v)); | 2363 | WREG32(R600_RCU_DATA, (v)); |
2364 | spin_unlock_irqrestore(&rdev->rcu_idx_lock, flags); | ||
2321 | } | 2365 | } |
2322 | 2366 | ||
2323 | static inline u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg) | 2367 | static inline u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg) |
2324 | { | 2368 | { |
2369 | unsigned long flags; | ||
2325 | u32 r; | 2370 | u32 r; |
2326 | 2371 | ||
2372 | spin_lock_irqsave(&rdev->cg_idx_lock, flags); | ||
2327 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); | 2373 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); |
2328 | r = RREG32(EVERGREEN_CG_IND_DATA); | 2374 | r = RREG32(EVERGREEN_CG_IND_DATA); |
2375 | spin_unlock_irqrestore(&rdev->cg_idx_lock, flags); | ||
2329 | return r; | 2376 | return r; |
2330 | } | 2377 | } |
2331 | 2378 | ||
2332 | static inline void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2379 | static inline void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2333 | { | 2380 | { |
2381 | unsigned long flags; | ||
2382 | |||
2383 | spin_lock_irqsave(&rdev->cg_idx_lock, flags); | ||
2334 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); | 2384 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); |
2335 | WREG32(EVERGREEN_CG_IND_DATA, (v)); | 2385 | WREG32(EVERGREEN_CG_IND_DATA, (v)); |
2386 | spin_unlock_irqrestore(&rdev->cg_idx_lock, flags); | ||
2336 | } | 2387 | } |
2337 | 2388 | ||
2338 | static inline u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg) | 2389 | static inline u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg) |
2339 | { | 2390 | { |
2391 | unsigned long flags; | ||
2340 | u32 r; | 2392 | u32 r; |
2341 | 2393 | ||
2394 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
2342 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); | 2395 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); |
2343 | r = RREG32(EVERGREEN_PIF_PHY0_DATA); | 2396 | r = RREG32(EVERGREEN_PIF_PHY0_DATA); |
2397 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
2344 | return r; | 2398 | return r; |
2345 | } | 2399 | } |
2346 | 2400 | ||
2347 | static inline void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2401 | static inline void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2348 | { | 2402 | { |
2403 | unsigned long flags; | ||
2404 | |||
2405 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
2349 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); | 2406 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); |
2350 | WREG32(EVERGREEN_PIF_PHY0_DATA, (v)); | 2407 | WREG32(EVERGREEN_PIF_PHY0_DATA, (v)); |
2408 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
2351 | } | 2409 | } |
2352 | 2410 | ||
2353 | static inline u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg) | 2411 | static inline u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg) |
2354 | { | 2412 | { |
2413 | unsigned long flags; | ||
2355 | u32 r; | 2414 | u32 r; |
2356 | 2415 | ||
2416 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
2357 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); | 2417 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); |
2358 | r = RREG32(EVERGREEN_PIF_PHY1_DATA); | 2418 | r = RREG32(EVERGREEN_PIF_PHY1_DATA); |
2419 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
2359 | return r; | 2420 | return r; |
2360 | } | 2421 | } |
2361 | 2422 | ||
2362 | static inline void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2423 | static inline void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2363 | { | 2424 | { |
2425 | unsigned long flags; | ||
2426 | |||
2427 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
2364 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); | 2428 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); |
2365 | WREG32(EVERGREEN_PIF_PHY1_DATA, (v)); | 2429 | WREG32(EVERGREEN_PIF_PHY1_DATA, (v)); |
2430 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
2366 | } | 2431 | } |
2367 | 2432 | ||
2368 | static inline u32 r600_uvd_ctx_rreg(struct radeon_device *rdev, u32 reg) | 2433 | static inline u32 r600_uvd_ctx_rreg(struct radeon_device *rdev, u32 reg) |
2369 | { | 2434 | { |
2435 | unsigned long flags; | ||
2370 | u32 r; | 2436 | u32 r; |
2371 | 2437 | ||
2438 | spin_lock_irqsave(&rdev->uvd_idx_lock, flags); | ||
2372 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); | 2439 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); |
2373 | r = RREG32(R600_UVD_CTX_DATA); | 2440 | r = RREG32(R600_UVD_CTX_DATA); |
2441 | spin_unlock_irqrestore(&rdev->uvd_idx_lock, flags); | ||
2374 | return r; | 2442 | return r; |
2375 | } | 2443 | } |
2376 | 2444 | ||
2377 | static inline void r600_uvd_ctx_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2445 | static inline void r600_uvd_ctx_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2378 | { | 2446 | { |
2447 | unsigned long flags; | ||
2448 | |||
2449 | spin_lock_irqsave(&rdev->uvd_idx_lock, flags); | ||
2379 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); | 2450 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); |
2380 | WREG32(R600_UVD_CTX_DATA, (v)); | 2451 | WREG32(R600_UVD_CTX_DATA, (v)); |
2452 | spin_unlock_irqrestore(&rdev->uvd_idx_lock, flags); | ||
2381 | } | 2453 | } |
2382 | 2454 | ||
2383 | 2455 | ||
2384 | static inline u32 cik_didt_rreg(struct radeon_device *rdev, u32 reg) | 2456 | static inline u32 cik_didt_rreg(struct radeon_device *rdev, u32 reg) |
2385 | { | 2457 | { |
2458 | unsigned long flags; | ||
2386 | u32 r; | 2459 | u32 r; |
2387 | 2460 | ||
2461 | spin_lock_irqsave(&rdev->didt_idx_lock, flags); | ||
2388 | WREG32(CIK_DIDT_IND_INDEX, (reg)); | 2462 | WREG32(CIK_DIDT_IND_INDEX, (reg)); |
2389 | r = RREG32(CIK_DIDT_IND_DATA); | 2463 | r = RREG32(CIK_DIDT_IND_DATA); |
2464 | spin_unlock_irqrestore(&rdev->didt_idx_lock, flags); | ||
2390 | return r; | 2465 | return r; |
2391 | } | 2466 | } |
2392 | 2467 | ||
2393 | static inline void cik_didt_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2468 | static inline void cik_didt_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
2394 | { | 2469 | { |
2470 | unsigned long flags; | ||
2471 | |||
2472 | spin_lock_irqsave(&rdev->didt_idx_lock, flags); | ||
2395 | WREG32(CIK_DIDT_IND_INDEX, (reg)); | 2473 | WREG32(CIK_DIDT_IND_INDEX, (reg)); |
2396 | WREG32(CIK_DIDT_IND_DATA, (v)); | 2474 | WREG32(CIK_DIDT_IND_DATA, (v)); |
2475 | spin_unlock_irqrestore(&rdev->didt_idx_lock, flags); | ||
2397 | } | 2476 | } |
2398 | 2477 | ||
2399 | void r100_pll_errata_after_index(struct radeon_device *rdev); | 2478 | void r100_pll_errata_after_index(struct radeon_device *rdev); |
@@ -2569,6 +2648,7 @@ void radeon_ring_write(struct radeon_ring *ring, uint32_t v); | |||
2569 | #define radeon_dpm_force_performance_level(rdev, l) rdev->asic->dpm.force_performance_level((rdev), (l)) | 2648 | #define radeon_dpm_force_performance_level(rdev, l) rdev->asic->dpm.force_performance_level((rdev), (l)) |
2570 | #define radeon_dpm_vblank_too_short(rdev) rdev->asic->dpm.vblank_too_short((rdev)) | 2649 | #define radeon_dpm_vblank_too_short(rdev) rdev->asic->dpm.vblank_too_short((rdev)) |
2571 | #define radeon_dpm_powergate_uvd(rdev, g) rdev->asic->dpm.powergate_uvd((rdev), (g)) | 2650 | #define radeon_dpm_powergate_uvd(rdev, g) rdev->asic->dpm.powergate_uvd((rdev), (g)) |
2651 | #define radeon_dpm_enable_bapm(rdev, e) rdev->asic->dpm.enable_bapm((rdev), (e)) | ||
2572 | 2652 | ||
2573 | /* Common functions */ | 2653 | /* Common functions */ |
2574 | /* AGP */ | 2654 | /* AGP */ |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.c b/drivers/gpu/drm/radeon/radeon_asic.c index 630853b96841..8f7e04538fd6 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.c +++ b/drivers/gpu/drm/radeon/radeon_asic.c | |||
@@ -1004,6 +1004,8 @@ static struct radeon_asic rv6xx_asic = { | |||
1004 | .wait_for_vblank = &avivo_wait_for_vblank, | 1004 | .wait_for_vblank = &avivo_wait_for_vblank, |
1005 | .set_backlight_level = &atombios_set_backlight_level, | 1005 | .set_backlight_level = &atombios_set_backlight_level, |
1006 | .get_backlight_level = &atombios_get_backlight_level, | 1006 | .get_backlight_level = &atombios_get_backlight_level, |
1007 | .hdmi_enable = &r600_hdmi_enable, | ||
1008 | .hdmi_setmode = &r600_hdmi_setmode, | ||
1007 | }, | 1009 | }, |
1008 | .copy = { | 1010 | .copy = { |
1009 | .blit = &r600_copy_cpdma, | 1011 | .blit = &r600_copy_cpdma, |
@@ -1037,6 +1039,7 @@ static struct radeon_asic rv6xx_asic = { | |||
1037 | .set_pcie_lanes = &r600_set_pcie_lanes, | 1039 | .set_pcie_lanes = &r600_set_pcie_lanes, |
1038 | .set_clock_gating = NULL, | 1040 | .set_clock_gating = NULL, |
1039 | .get_temperature = &rv6xx_get_temp, | 1041 | .get_temperature = &rv6xx_get_temp, |
1042 | .set_uvd_clocks = &r600_set_uvd_clocks, | ||
1040 | }, | 1043 | }, |
1041 | .dpm = { | 1044 | .dpm = { |
1042 | .init = &rv6xx_dpm_init, | 1045 | .init = &rv6xx_dpm_init, |
@@ -1126,6 +1129,7 @@ static struct radeon_asic rs780_asic = { | |||
1126 | .set_pcie_lanes = NULL, | 1129 | .set_pcie_lanes = NULL, |
1127 | .set_clock_gating = NULL, | 1130 | .set_clock_gating = NULL, |
1128 | .get_temperature = &rv6xx_get_temp, | 1131 | .get_temperature = &rv6xx_get_temp, |
1132 | .set_uvd_clocks = &r600_set_uvd_clocks, | ||
1129 | }, | 1133 | }, |
1130 | .dpm = { | 1134 | .dpm = { |
1131 | .init = &rs780_dpm_init, | 1135 | .init = &rs780_dpm_init, |
@@ -1141,6 +1145,7 @@ static struct radeon_asic rs780_asic = { | |||
1141 | .get_mclk = &rs780_dpm_get_mclk, | 1145 | .get_mclk = &rs780_dpm_get_mclk, |
1142 | .print_power_state = &rs780_dpm_print_power_state, | 1146 | .print_power_state = &rs780_dpm_print_power_state, |
1143 | .debugfs_print_current_performance_level = &rs780_dpm_debugfs_print_current_performance_level, | 1147 | .debugfs_print_current_performance_level = &rs780_dpm_debugfs_print_current_performance_level, |
1148 | .force_performance_level = &rs780_dpm_force_performance_level, | ||
1144 | }, | 1149 | }, |
1145 | .pflip = { | 1150 | .pflip = { |
1146 | .pre_page_flip = &rs600_pre_page_flip, | 1151 | .pre_page_flip = &rs600_pre_page_flip, |
@@ -1791,6 +1796,7 @@ static struct radeon_asic trinity_asic = { | |||
1791 | .print_power_state = &trinity_dpm_print_power_state, | 1796 | .print_power_state = &trinity_dpm_print_power_state, |
1792 | .debugfs_print_current_performance_level = &trinity_dpm_debugfs_print_current_performance_level, | 1797 | .debugfs_print_current_performance_level = &trinity_dpm_debugfs_print_current_performance_level, |
1793 | .force_performance_level = &trinity_dpm_force_performance_level, | 1798 | .force_performance_level = &trinity_dpm_force_performance_level, |
1799 | .enable_bapm = &trinity_dpm_enable_bapm, | ||
1794 | }, | 1800 | }, |
1795 | .pflip = { | 1801 | .pflip = { |
1796 | .pre_page_flip = &evergreen_pre_page_flip, | 1802 | .pre_page_flip = &evergreen_pre_page_flip, |
@@ -2166,6 +2172,7 @@ static struct radeon_asic kv_asic = { | |||
2166 | .debugfs_print_current_performance_level = &kv_dpm_debugfs_print_current_performance_level, | 2172 | .debugfs_print_current_performance_level = &kv_dpm_debugfs_print_current_performance_level, |
2167 | .force_performance_level = &kv_dpm_force_performance_level, | 2173 | .force_performance_level = &kv_dpm_force_performance_level, |
2168 | .powergate_uvd = &kv_dpm_powergate_uvd, | 2174 | .powergate_uvd = &kv_dpm_powergate_uvd, |
2175 | .enable_bapm = &kv_dpm_enable_bapm, | ||
2169 | }, | 2176 | }, |
2170 | .pflip = { | 2177 | .pflip = { |
2171 | .pre_page_flip = &evergreen_pre_page_flip, | 2178 | .pre_page_flip = &evergreen_pre_page_flip, |
@@ -2390,7 +2397,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
2390 | RADEON_CG_SUPPORT_HDP_LS | | 2397 | RADEON_CG_SUPPORT_HDP_LS | |
2391 | RADEON_CG_SUPPORT_HDP_MGCG; | 2398 | RADEON_CG_SUPPORT_HDP_MGCG; |
2392 | rdev->pg_flags = 0 | | 2399 | rdev->pg_flags = 0 | |
2393 | /*RADEON_PG_SUPPORT_GFX_CG | */ | 2400 | /*RADEON_PG_SUPPORT_GFX_PG | */ |
2394 | RADEON_PG_SUPPORT_SDMA; | 2401 | RADEON_PG_SUPPORT_SDMA; |
2395 | break; | 2402 | break; |
2396 | case CHIP_OLAND: | 2403 | case CHIP_OLAND: |
@@ -2479,7 +2486,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
2479 | RADEON_CG_SUPPORT_HDP_LS | | 2486 | RADEON_CG_SUPPORT_HDP_LS | |
2480 | RADEON_CG_SUPPORT_HDP_MGCG; | 2487 | RADEON_CG_SUPPORT_HDP_MGCG; |
2481 | rdev->pg_flags = 0; | 2488 | rdev->pg_flags = 0; |
2482 | /*RADEON_PG_SUPPORT_GFX_CG | | 2489 | /*RADEON_PG_SUPPORT_GFX_PG | |
2483 | RADEON_PG_SUPPORT_GFX_SMG | | 2490 | RADEON_PG_SUPPORT_GFX_SMG | |
2484 | RADEON_PG_SUPPORT_GFX_DMG | | 2491 | RADEON_PG_SUPPORT_GFX_DMG | |
2485 | RADEON_PG_SUPPORT_UVD | | 2492 | RADEON_PG_SUPPORT_UVD | |
@@ -2507,7 +2514,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
2507 | RADEON_CG_SUPPORT_HDP_LS | | 2514 | RADEON_CG_SUPPORT_HDP_LS | |
2508 | RADEON_CG_SUPPORT_HDP_MGCG; | 2515 | RADEON_CG_SUPPORT_HDP_MGCG; |
2509 | rdev->pg_flags = 0; | 2516 | rdev->pg_flags = 0; |
2510 | /*RADEON_PG_SUPPORT_GFX_CG | | 2517 | /*RADEON_PG_SUPPORT_GFX_PG | |
2511 | RADEON_PG_SUPPORT_GFX_SMG | | 2518 | RADEON_PG_SUPPORT_GFX_SMG | |
2512 | RADEON_PG_SUPPORT_UVD | | 2519 | RADEON_PG_SUPPORT_UVD | |
2513 | RADEON_PG_SUPPORT_VCE | | 2520 | RADEON_PG_SUPPORT_VCE | |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.h b/drivers/gpu/drm/radeon/radeon_asic.h index 818bbe6b884b..70c29d5e080d 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.h +++ b/drivers/gpu/drm/radeon/radeon_asic.h | |||
@@ -389,6 +389,7 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev); | |||
389 | u32 r600_get_xclk(struct radeon_device *rdev); | 389 | u32 r600_get_xclk(struct radeon_device *rdev); |
390 | uint64_t r600_get_gpu_clock_counter(struct radeon_device *rdev); | 390 | uint64_t r600_get_gpu_clock_counter(struct radeon_device *rdev); |
391 | int rv6xx_get_temp(struct radeon_device *rdev); | 391 | int rv6xx_get_temp(struct radeon_device *rdev); |
392 | int r600_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk); | ||
392 | int r600_dpm_pre_set_power_state(struct radeon_device *rdev); | 393 | int r600_dpm_pre_set_power_state(struct radeon_device *rdev); |
393 | void r600_dpm_post_set_power_state(struct radeon_device *rdev); | 394 | void r600_dpm_post_set_power_state(struct radeon_device *rdev); |
394 | /* r600 dma */ | 395 | /* r600 dma */ |
@@ -428,6 +429,8 @@ void rs780_dpm_print_power_state(struct radeon_device *rdev, | |||
428 | struct radeon_ps *ps); | 429 | struct radeon_ps *ps); |
429 | void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, | 430 | void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, |
430 | struct seq_file *m); | 431 | struct seq_file *m); |
432 | int rs780_dpm_force_performance_level(struct radeon_device *rdev, | ||
433 | enum radeon_dpm_forced_level level); | ||
431 | 434 | ||
432 | /* | 435 | /* |
433 | * rv770,rv730,rv710,rv740 | 436 | * rv770,rv730,rv710,rv740 |
@@ -625,6 +628,7 @@ void trinity_dpm_debugfs_print_current_performance_level(struct radeon_device *r | |||
625 | struct seq_file *m); | 628 | struct seq_file *m); |
626 | int trinity_dpm_force_performance_level(struct radeon_device *rdev, | 629 | int trinity_dpm_force_performance_level(struct radeon_device *rdev, |
627 | enum radeon_dpm_forced_level level); | 630 | enum radeon_dpm_forced_level level); |
631 | void trinity_dpm_enable_bapm(struct radeon_device *rdev, bool enable); | ||
628 | 632 | ||
629 | /* DCE6 - SI */ | 633 | /* DCE6 - SI */ |
630 | void dce6_bandwidth_update(struct radeon_device *rdev); | 634 | void dce6_bandwidth_update(struct radeon_device *rdev); |
@@ -781,6 +785,7 @@ void kv_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, | |||
781 | int kv_dpm_force_performance_level(struct radeon_device *rdev, | 785 | int kv_dpm_force_performance_level(struct radeon_device *rdev, |
782 | enum radeon_dpm_forced_level level); | 786 | enum radeon_dpm_forced_level level); |
783 | void kv_dpm_powergate_uvd(struct radeon_device *rdev, bool gate); | 787 | void kv_dpm_powergate_uvd(struct radeon_device *rdev, bool gate); |
788 | void kv_dpm_enable_bapm(struct radeon_device *rdev, bool enable); | ||
784 | 789 | ||
785 | /* uvd v1.0 */ | 790 | /* uvd v1.0 */ |
786 | uint32_t uvd_v1_0_get_rptr(struct radeon_device *rdev, | 791 | uint32_t uvd_v1_0_get_rptr(struct radeon_device *rdev, |
diff --git a/drivers/gpu/drm/radeon/radeon_atombios.c b/drivers/gpu/drm/radeon/radeon_atombios.c index 404e25d285ba..f79ee184ffd5 100644 --- a/drivers/gpu/drm/radeon/radeon_atombios.c +++ b/drivers/gpu/drm/radeon/radeon_atombios.c | |||
@@ -1367,6 +1367,7 @@ bool radeon_atombios_get_ppll_ss_info(struct radeon_device *rdev, | |||
1367 | int index = GetIndexIntoMasterTable(DATA, PPLL_SS_Info); | 1367 | int index = GetIndexIntoMasterTable(DATA, PPLL_SS_Info); |
1368 | uint16_t data_offset, size; | 1368 | uint16_t data_offset, size; |
1369 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; | 1369 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; |
1370 | struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT *ss_assign; | ||
1370 | uint8_t frev, crev; | 1371 | uint8_t frev, crev; |
1371 | int i, num_indices; | 1372 | int i, num_indices; |
1372 | 1373 | ||
@@ -1378,18 +1379,21 @@ bool radeon_atombios_get_ppll_ss_info(struct radeon_device *rdev, | |||
1378 | 1379 | ||
1379 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1380 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
1380 | sizeof(ATOM_SPREAD_SPECTRUM_ASSIGNMENT); | 1381 | sizeof(ATOM_SPREAD_SPECTRUM_ASSIGNMENT); |
1381 | 1382 | ss_assign = (struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT*) | |
1383 | ((u8 *)&ss_info->asSS_Info[0]); | ||
1382 | for (i = 0; i < num_indices; i++) { | 1384 | for (i = 0; i < num_indices; i++) { |
1383 | if (ss_info->asSS_Info[i].ucSS_Id == id) { | 1385 | if (ss_assign->ucSS_Id == id) { |
1384 | ss->percentage = | 1386 | ss->percentage = |
1385 | le16_to_cpu(ss_info->asSS_Info[i].usSpreadSpectrumPercentage); | 1387 | le16_to_cpu(ss_assign->usSpreadSpectrumPercentage); |
1386 | ss->type = ss_info->asSS_Info[i].ucSpreadSpectrumType; | 1388 | ss->type = ss_assign->ucSpreadSpectrumType; |
1387 | ss->step = ss_info->asSS_Info[i].ucSS_Step; | 1389 | ss->step = ss_assign->ucSS_Step; |
1388 | ss->delay = ss_info->asSS_Info[i].ucSS_Delay; | 1390 | ss->delay = ss_assign->ucSS_Delay; |
1389 | ss->range = ss_info->asSS_Info[i].ucSS_Range; | 1391 | ss->range = ss_assign->ucSS_Range; |
1390 | ss->refdiv = ss_info->asSS_Info[i].ucRecommendedRef_Div; | 1392 | ss->refdiv = ss_assign->ucRecommendedRef_Div; |
1391 | return true; | 1393 | return true; |
1392 | } | 1394 | } |
1395 | ss_assign = (struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT*) | ||
1396 | ((u8 *)ss_assign + sizeof(struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT)); | ||
1393 | } | 1397 | } |
1394 | } | 1398 | } |
1395 | return false; | 1399 | return false; |
@@ -1477,6 +1481,12 @@ union asic_ss_info { | |||
1477 | struct _ATOM_ASIC_INTERNAL_SS_INFO_V3 info_3; | 1481 | struct _ATOM_ASIC_INTERNAL_SS_INFO_V3 info_3; |
1478 | }; | 1482 | }; |
1479 | 1483 | ||
1484 | union asic_ss_assignment { | ||
1485 | struct _ATOM_ASIC_SS_ASSIGNMENT v1; | ||
1486 | struct _ATOM_ASIC_SS_ASSIGNMENT_V2 v2; | ||
1487 | struct _ATOM_ASIC_SS_ASSIGNMENT_V3 v3; | ||
1488 | }; | ||
1489 | |||
1480 | bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | 1490 | bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, |
1481 | struct radeon_atom_ss *ss, | 1491 | struct radeon_atom_ss *ss, |
1482 | int id, u32 clock) | 1492 | int id, u32 clock) |
@@ -1485,6 +1495,7 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
1485 | int index = GetIndexIntoMasterTable(DATA, ASIC_InternalSS_Info); | 1495 | int index = GetIndexIntoMasterTable(DATA, ASIC_InternalSS_Info); |
1486 | uint16_t data_offset, size; | 1496 | uint16_t data_offset, size; |
1487 | union asic_ss_info *ss_info; | 1497 | union asic_ss_info *ss_info; |
1498 | union asic_ss_assignment *ss_assign; | ||
1488 | uint8_t frev, crev; | 1499 | uint8_t frev, crev; |
1489 | int i, num_indices; | 1500 | int i, num_indices; |
1490 | 1501 | ||
@@ -1509,45 +1520,52 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
1509 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1520 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
1510 | sizeof(ATOM_ASIC_SS_ASSIGNMENT); | 1521 | sizeof(ATOM_ASIC_SS_ASSIGNMENT); |
1511 | 1522 | ||
1523 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info.asSpreadSpectrum[0]); | ||
1512 | for (i = 0; i < num_indices; i++) { | 1524 | for (i = 0; i < num_indices; i++) { |
1513 | if ((ss_info->info.asSpreadSpectrum[i].ucClockIndication == id) && | 1525 | if ((ss_assign->v1.ucClockIndication == id) && |
1514 | (clock <= le32_to_cpu(ss_info->info.asSpreadSpectrum[i].ulTargetClockRange))) { | 1526 | (clock <= le32_to_cpu(ss_assign->v1.ulTargetClockRange))) { |
1515 | ss->percentage = | 1527 | ss->percentage = |
1516 | le16_to_cpu(ss_info->info.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1528 | le16_to_cpu(ss_assign->v1.usSpreadSpectrumPercentage); |
1517 | ss->type = ss_info->info.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1529 | ss->type = ss_assign->v1.ucSpreadSpectrumMode; |
1518 | ss->rate = le16_to_cpu(ss_info->info.asSpreadSpectrum[i].usSpreadRateInKhz); | 1530 | ss->rate = le16_to_cpu(ss_assign->v1.usSpreadRateInKhz); |
1519 | return true; | 1531 | return true; |
1520 | } | 1532 | } |
1533 | ss_assign = (union asic_ss_assignment *) | ||
1534 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT)); | ||
1521 | } | 1535 | } |
1522 | break; | 1536 | break; |
1523 | case 2: | 1537 | case 2: |
1524 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1538 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
1525 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2); | 1539 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2); |
1540 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info_2.asSpreadSpectrum[0]); | ||
1526 | for (i = 0; i < num_indices; i++) { | 1541 | for (i = 0; i < num_indices; i++) { |
1527 | if ((ss_info->info_2.asSpreadSpectrum[i].ucClockIndication == id) && | 1542 | if ((ss_assign->v2.ucClockIndication == id) && |
1528 | (clock <= le32_to_cpu(ss_info->info_2.asSpreadSpectrum[i].ulTargetClockRange))) { | 1543 | (clock <= le32_to_cpu(ss_assign->v2.ulTargetClockRange))) { |
1529 | ss->percentage = | 1544 | ss->percentage = |
1530 | le16_to_cpu(ss_info->info_2.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1545 | le16_to_cpu(ss_assign->v2.usSpreadSpectrumPercentage); |
1531 | ss->type = ss_info->info_2.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1546 | ss->type = ss_assign->v2.ucSpreadSpectrumMode; |
1532 | ss->rate = le16_to_cpu(ss_info->info_2.asSpreadSpectrum[i].usSpreadRateIn10Hz); | 1547 | ss->rate = le16_to_cpu(ss_assign->v2.usSpreadRateIn10Hz); |
1533 | if ((crev == 2) && | 1548 | if ((crev == 2) && |
1534 | ((id == ASIC_INTERNAL_ENGINE_SS) || | 1549 | ((id == ASIC_INTERNAL_ENGINE_SS) || |
1535 | (id == ASIC_INTERNAL_MEMORY_SS))) | 1550 | (id == ASIC_INTERNAL_MEMORY_SS))) |
1536 | ss->rate /= 100; | 1551 | ss->rate /= 100; |
1537 | return true; | 1552 | return true; |
1538 | } | 1553 | } |
1554 | ss_assign = (union asic_ss_assignment *) | ||
1555 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2)); | ||
1539 | } | 1556 | } |
1540 | break; | 1557 | break; |
1541 | case 3: | 1558 | case 3: |
1542 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1559 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
1543 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3); | 1560 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3); |
1561 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info_3.asSpreadSpectrum[0]); | ||
1544 | for (i = 0; i < num_indices; i++) { | 1562 | for (i = 0; i < num_indices; i++) { |
1545 | if ((ss_info->info_3.asSpreadSpectrum[i].ucClockIndication == id) && | 1563 | if ((ss_assign->v3.ucClockIndication == id) && |
1546 | (clock <= le32_to_cpu(ss_info->info_3.asSpreadSpectrum[i].ulTargetClockRange))) { | 1564 | (clock <= le32_to_cpu(ss_assign->v3.ulTargetClockRange))) { |
1547 | ss->percentage = | 1565 | ss->percentage = |
1548 | le16_to_cpu(ss_info->info_3.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1566 | le16_to_cpu(ss_assign->v3.usSpreadSpectrumPercentage); |
1549 | ss->type = ss_info->info_3.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1567 | ss->type = ss_assign->v3.ucSpreadSpectrumMode; |
1550 | ss->rate = le16_to_cpu(ss_info->info_3.asSpreadSpectrum[i].usSpreadRateIn10Hz); | 1568 | ss->rate = le16_to_cpu(ss_assign->v3.usSpreadRateIn10Hz); |
1551 | if ((id == ASIC_INTERNAL_ENGINE_SS) || | 1569 | if ((id == ASIC_INTERNAL_ENGINE_SS) || |
1552 | (id == ASIC_INTERNAL_MEMORY_SS)) | 1570 | (id == ASIC_INTERNAL_MEMORY_SS)) |
1553 | ss->rate /= 100; | 1571 | ss->rate /= 100; |
@@ -1555,6 +1573,8 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
1555 | radeon_atombios_get_igp_ss_overrides(rdev, ss, id); | 1573 | radeon_atombios_get_igp_ss_overrides(rdev, ss, id); |
1556 | return true; | 1574 | return true; |
1557 | } | 1575 | } |
1576 | ss_assign = (union asic_ss_assignment *) | ||
1577 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3)); | ||
1558 | } | 1578 | } |
1559 | break; | 1579 | break; |
1560 | default: | 1580 | default: |
diff --git a/drivers/gpu/drm/radeon/radeon_connectors.c b/drivers/gpu/drm/radeon/radeon_connectors.c index 2399f25ec037..79159b5da05b 100644 --- a/drivers/gpu/drm/radeon/radeon_connectors.c +++ b/drivers/gpu/drm/radeon/radeon_connectors.c | |||
@@ -396,6 +396,21 @@ static int radeon_connector_set_property(struct drm_connector *connector, struct | |||
396 | } | 396 | } |
397 | } | 397 | } |
398 | 398 | ||
399 | if (property == rdev->mode_info.audio_property) { | ||
400 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
401 | /* need to find digital encoder on connector */ | ||
402 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); | ||
403 | if (!encoder) | ||
404 | return 0; | ||
405 | |||
406 | radeon_encoder = to_radeon_encoder(encoder); | ||
407 | |||
408 | if (radeon_connector->audio != val) { | ||
409 | radeon_connector->audio = val; | ||
410 | radeon_property_change_mode(&radeon_encoder->base); | ||
411 | } | ||
412 | } | ||
413 | |||
399 | if (property == rdev->mode_info.underscan_property) { | 414 | if (property == rdev->mode_info.underscan_property) { |
400 | /* need to find digital encoder on connector */ | 415 | /* need to find digital encoder on connector */ |
401 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); | 416 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); |
@@ -1420,7 +1435,7 @@ radeon_dp_detect(struct drm_connector *connector, bool force) | |||
1420 | if (radeon_dp_getdpcd(radeon_connector)) | 1435 | if (radeon_dp_getdpcd(radeon_connector)) |
1421 | ret = connector_status_connected; | 1436 | ret = connector_status_connected; |
1422 | } else { | 1437 | } else { |
1423 | /* try non-aux ddc (DP to DVI/HMDI/etc. adapter) */ | 1438 | /* try non-aux ddc (DP to DVI/HDMI/etc. adapter) */ |
1424 | if (radeon_ddc_probe(radeon_connector, false)) | 1439 | if (radeon_ddc_probe(radeon_connector, false)) |
1425 | ret = connector_status_connected; | 1440 | ret = connector_status_connected; |
1426 | } | 1441 | } |
@@ -1489,6 +1504,24 @@ static const struct drm_connector_funcs radeon_dp_connector_funcs = { | |||
1489 | .force = radeon_dvi_force, | 1504 | .force = radeon_dvi_force, |
1490 | }; | 1505 | }; |
1491 | 1506 | ||
1507 | static const struct drm_connector_funcs radeon_edp_connector_funcs = { | ||
1508 | .dpms = drm_helper_connector_dpms, | ||
1509 | .detect = radeon_dp_detect, | ||
1510 | .fill_modes = drm_helper_probe_single_connector_modes, | ||
1511 | .set_property = radeon_lvds_set_property, | ||
1512 | .destroy = radeon_dp_connector_destroy, | ||
1513 | .force = radeon_dvi_force, | ||
1514 | }; | ||
1515 | |||
1516 | static const struct drm_connector_funcs radeon_lvds_bridge_connector_funcs = { | ||
1517 | .dpms = drm_helper_connector_dpms, | ||
1518 | .detect = radeon_dp_detect, | ||
1519 | .fill_modes = drm_helper_probe_single_connector_modes, | ||
1520 | .set_property = radeon_lvds_set_property, | ||
1521 | .destroy = radeon_dp_connector_destroy, | ||
1522 | .force = radeon_dvi_force, | ||
1523 | }; | ||
1524 | |||
1492 | void | 1525 | void |
1493 | radeon_add_atom_connector(struct drm_device *dev, | 1526 | radeon_add_atom_connector(struct drm_device *dev, |
1494 | uint32_t connector_id, | 1527 | uint32_t connector_id, |
@@ -1580,8 +1613,6 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1580 | goto failed; | 1613 | goto failed; |
1581 | radeon_dig_connector->igp_lane_info = igp_lane_info; | 1614 | radeon_dig_connector->igp_lane_info = igp_lane_info; |
1582 | radeon_connector->con_priv = radeon_dig_connector; | 1615 | radeon_connector->con_priv = radeon_dig_connector; |
1583 | drm_connector_init(dev, &radeon_connector->base, &radeon_dp_connector_funcs, connector_type); | ||
1584 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); | ||
1585 | if (i2c_bus->valid) { | 1616 | if (i2c_bus->valid) { |
1586 | /* add DP i2c bus */ | 1617 | /* add DP i2c bus */ |
1587 | if (connector_type == DRM_MODE_CONNECTOR_eDP) | 1618 | if (connector_type == DRM_MODE_CONNECTOR_eDP) |
@@ -1598,6 +1629,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1598 | case DRM_MODE_CONNECTOR_VGA: | 1629 | case DRM_MODE_CONNECTOR_VGA: |
1599 | case DRM_MODE_CONNECTOR_DVIA: | 1630 | case DRM_MODE_CONNECTOR_DVIA: |
1600 | default: | 1631 | default: |
1632 | drm_connector_init(dev, &radeon_connector->base, | ||
1633 | &radeon_dp_connector_funcs, connector_type); | ||
1634 | drm_connector_helper_add(&radeon_connector->base, | ||
1635 | &radeon_dp_connector_helper_funcs); | ||
1601 | connector->interlace_allowed = true; | 1636 | connector->interlace_allowed = true; |
1602 | connector->doublescan_allowed = true; | 1637 | connector->doublescan_allowed = true; |
1603 | radeon_connector->dac_load_detect = true; | 1638 | radeon_connector->dac_load_detect = true; |
@@ -1610,6 +1645,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1610 | case DRM_MODE_CONNECTOR_HDMIA: | 1645 | case DRM_MODE_CONNECTOR_HDMIA: |
1611 | case DRM_MODE_CONNECTOR_HDMIB: | 1646 | case DRM_MODE_CONNECTOR_HDMIB: |
1612 | case DRM_MODE_CONNECTOR_DisplayPort: | 1647 | case DRM_MODE_CONNECTOR_DisplayPort: |
1648 | drm_connector_init(dev, &radeon_connector->base, | ||
1649 | &radeon_dp_connector_funcs, connector_type); | ||
1650 | drm_connector_helper_add(&radeon_connector->base, | ||
1651 | &radeon_dp_connector_helper_funcs); | ||
1613 | drm_object_attach_property(&radeon_connector->base.base, | 1652 | drm_object_attach_property(&radeon_connector->base.base, |
1614 | rdev->mode_info.underscan_property, | 1653 | rdev->mode_info.underscan_property, |
1615 | UNDERSCAN_OFF); | 1654 | UNDERSCAN_OFF); |
@@ -1619,6 +1658,9 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1619 | drm_object_attach_property(&radeon_connector->base.base, | 1658 | drm_object_attach_property(&radeon_connector->base.base, |
1620 | rdev->mode_info.underscan_vborder_property, | 1659 | rdev->mode_info.underscan_vborder_property, |
1621 | 0); | 1660 | 0); |
1661 | drm_object_attach_property(&radeon_connector->base.base, | ||
1662 | rdev->mode_info.audio_property, | ||
1663 | RADEON_AUDIO_DISABLE); | ||
1622 | subpixel_order = SubPixelHorizontalRGB; | 1664 | subpixel_order = SubPixelHorizontalRGB; |
1623 | connector->interlace_allowed = true; | 1665 | connector->interlace_allowed = true; |
1624 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) | 1666 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) |
@@ -1634,6 +1676,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1634 | break; | 1676 | break; |
1635 | case DRM_MODE_CONNECTOR_LVDS: | 1677 | case DRM_MODE_CONNECTOR_LVDS: |
1636 | case DRM_MODE_CONNECTOR_eDP: | 1678 | case DRM_MODE_CONNECTOR_eDP: |
1679 | drm_connector_init(dev, &radeon_connector->base, | ||
1680 | &radeon_lvds_bridge_connector_funcs, connector_type); | ||
1681 | drm_connector_helper_add(&radeon_connector->base, | ||
1682 | &radeon_dp_connector_helper_funcs); | ||
1637 | drm_object_attach_property(&radeon_connector->base.base, | 1683 | drm_object_attach_property(&radeon_connector->base.base, |
1638 | dev->mode_config.scaling_mode_property, | 1684 | dev->mode_config.scaling_mode_property, |
1639 | DRM_MODE_SCALE_FULLSCREEN); | 1685 | DRM_MODE_SCALE_FULLSCREEN); |
@@ -1708,6 +1754,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1708 | rdev->mode_info.underscan_vborder_property, | 1754 | rdev->mode_info.underscan_vborder_property, |
1709 | 0); | 1755 | 0); |
1710 | } | 1756 | } |
1757 | if (ASIC_IS_DCE2(rdev)) { | ||
1758 | drm_object_attach_property(&radeon_connector->base.base, | ||
1759 | rdev->mode_info.audio_property, | ||
1760 | RADEON_AUDIO_DISABLE); | ||
1761 | } | ||
1711 | if (connector_type == DRM_MODE_CONNECTOR_DVII) { | 1762 | if (connector_type == DRM_MODE_CONNECTOR_DVII) { |
1712 | radeon_connector->dac_load_detect = true; | 1763 | radeon_connector->dac_load_detect = true; |
1713 | drm_object_attach_property(&radeon_connector->base.base, | 1764 | drm_object_attach_property(&radeon_connector->base.base, |
@@ -1748,6 +1799,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1748 | rdev->mode_info.underscan_vborder_property, | 1799 | rdev->mode_info.underscan_vborder_property, |
1749 | 0); | 1800 | 0); |
1750 | } | 1801 | } |
1802 | if (ASIC_IS_DCE2(rdev)) { | ||
1803 | drm_object_attach_property(&radeon_connector->base.base, | ||
1804 | rdev->mode_info.audio_property, | ||
1805 | RADEON_AUDIO_DISABLE); | ||
1806 | } | ||
1751 | subpixel_order = SubPixelHorizontalRGB; | 1807 | subpixel_order = SubPixelHorizontalRGB; |
1752 | connector->interlace_allowed = true; | 1808 | connector->interlace_allowed = true; |
1753 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) | 1809 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) |
@@ -1787,6 +1843,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1787 | rdev->mode_info.underscan_vborder_property, | 1843 | rdev->mode_info.underscan_vborder_property, |
1788 | 0); | 1844 | 0); |
1789 | } | 1845 | } |
1846 | if (ASIC_IS_DCE2(rdev)) { | ||
1847 | drm_object_attach_property(&radeon_connector->base.base, | ||
1848 | rdev->mode_info.audio_property, | ||
1849 | RADEON_AUDIO_DISABLE); | ||
1850 | } | ||
1790 | connector->interlace_allowed = true; | 1851 | connector->interlace_allowed = true; |
1791 | /* in theory with a DP to VGA converter... */ | 1852 | /* in theory with a DP to VGA converter... */ |
1792 | connector->doublescan_allowed = false; | 1853 | connector->doublescan_allowed = false; |
@@ -1797,7 +1858,7 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
1797 | goto failed; | 1858 | goto failed; |
1798 | radeon_dig_connector->igp_lane_info = igp_lane_info; | 1859 | radeon_dig_connector->igp_lane_info = igp_lane_info; |
1799 | radeon_connector->con_priv = radeon_dig_connector; | 1860 | radeon_connector->con_priv = radeon_dig_connector; |
1800 | drm_connector_init(dev, &radeon_connector->base, &radeon_dp_connector_funcs, connector_type); | 1861 | drm_connector_init(dev, &radeon_connector->base, &radeon_edp_connector_funcs, connector_type); |
1801 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); | 1862 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); |
1802 | if (i2c_bus->valid) { | 1863 | if (i2c_bus->valid) { |
1803 | /* add DP i2c bus */ | 1864 | /* add DP i2c bus */ |
diff --git a/drivers/gpu/drm/radeon/radeon_cs.c b/drivers/gpu/drm/radeon/radeon_cs.c index a56084410372..66c222836631 100644 --- a/drivers/gpu/drm/radeon/radeon_cs.c +++ b/drivers/gpu/drm/radeon/radeon_cs.c | |||
@@ -28,6 +28,7 @@ | |||
28 | #include <drm/radeon_drm.h> | 28 | #include <drm/radeon_drm.h> |
29 | #include "radeon_reg.h" | 29 | #include "radeon_reg.h" |
30 | #include "radeon.h" | 30 | #include "radeon.h" |
31 | #include "radeon_trace.h" | ||
31 | 32 | ||
32 | static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) | 33 | static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) |
33 | { | 34 | { |
@@ -80,10 +81,13 @@ static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) | |||
80 | p->relocs[i].lobj.bo = p->relocs[i].robj; | 81 | p->relocs[i].lobj.bo = p->relocs[i].robj; |
81 | p->relocs[i].lobj.written = !!r->write_domain; | 82 | p->relocs[i].lobj.written = !!r->write_domain; |
82 | 83 | ||
83 | /* the first reloc of an UVD job is the | 84 | /* the first reloc of an UVD job is the msg and that must be in |
84 | msg and that must be in VRAM */ | 85 | VRAM, also but everything into VRAM on AGP cards to avoid |
85 | if (p->ring == R600_RING_TYPE_UVD_INDEX && i == 0) { | 86 | image corruptions */ |
86 | /* TODO: is this still needed for NI+ ? */ | 87 | if (p->ring == R600_RING_TYPE_UVD_INDEX && |
88 | p->rdev->family < CHIP_PALM && | ||
89 | (i == 0 || drm_pci_device_is_agp(p->rdev->ddev))) { | ||
90 | |||
87 | p->relocs[i].lobj.domain = | 91 | p->relocs[i].lobj.domain = |
88 | RADEON_GEM_DOMAIN_VRAM; | 92 | RADEON_GEM_DOMAIN_VRAM; |
89 | 93 | ||
@@ -559,6 +563,8 @@ int radeon_cs_ioctl(struct drm_device *dev, void *data, struct drm_file *filp) | |||
559 | return r; | 563 | return r; |
560 | } | 564 | } |
561 | 565 | ||
566 | trace_radeon_cs(&parser); | ||
567 | |||
562 | r = radeon_cs_ib_chunk(rdev, &parser); | 568 | r = radeon_cs_ib_chunk(rdev, &parser); |
563 | if (r) { | 569 | if (r) { |
564 | goto out; | 570 | goto out; |
diff --git a/drivers/gpu/drm/radeon/radeon_device.c b/drivers/gpu/drm/radeon/radeon_device.c index 16cb8792b1e6..841d0e09be3e 100644 --- a/drivers/gpu/drm/radeon/radeon_device.c +++ b/drivers/gpu/drm/radeon/radeon_device.c | |||
@@ -1249,6 +1249,17 @@ int radeon_device_init(struct radeon_device *rdev, | |||
1249 | /* Registers mapping */ | 1249 | /* Registers mapping */ |
1250 | /* TODO: block userspace mapping of io register */ | 1250 | /* TODO: block userspace mapping of io register */ |
1251 | spin_lock_init(&rdev->mmio_idx_lock); | 1251 | spin_lock_init(&rdev->mmio_idx_lock); |
1252 | spin_lock_init(&rdev->smc_idx_lock); | ||
1253 | spin_lock_init(&rdev->pll_idx_lock); | ||
1254 | spin_lock_init(&rdev->mc_idx_lock); | ||
1255 | spin_lock_init(&rdev->pcie_idx_lock); | ||
1256 | spin_lock_init(&rdev->pciep_idx_lock); | ||
1257 | spin_lock_init(&rdev->pif_idx_lock); | ||
1258 | spin_lock_init(&rdev->cg_idx_lock); | ||
1259 | spin_lock_init(&rdev->uvd_idx_lock); | ||
1260 | spin_lock_init(&rdev->rcu_idx_lock); | ||
1261 | spin_lock_init(&rdev->didt_idx_lock); | ||
1262 | spin_lock_init(&rdev->end_idx_lock); | ||
1252 | if (rdev->family >= CHIP_BONAIRE) { | 1263 | if (rdev->family >= CHIP_BONAIRE) { |
1253 | rdev->rmmio_base = pci_resource_start(rdev->pdev, 5); | 1264 | rdev->rmmio_base = pci_resource_start(rdev->pdev, 5); |
1254 | rdev->rmmio_size = pci_resource_len(rdev->pdev, 5); | 1265 | rdev->rmmio_size = pci_resource_len(rdev->pdev, 5); |
@@ -1309,13 +1320,22 @@ int radeon_device_init(struct radeon_device *rdev, | |||
1309 | return r; | 1320 | return r; |
1310 | } | 1321 | } |
1311 | if ((radeon_testing & 1)) { | 1322 | if ((radeon_testing & 1)) { |
1312 | radeon_test_moves(rdev); | 1323 | if (rdev->accel_working) |
1324 | radeon_test_moves(rdev); | ||
1325 | else | ||
1326 | DRM_INFO("radeon: acceleration disabled, skipping move tests\n"); | ||
1313 | } | 1327 | } |
1314 | if ((radeon_testing & 2)) { | 1328 | if ((radeon_testing & 2)) { |
1315 | radeon_test_syncing(rdev); | 1329 | if (rdev->accel_working) |
1330 | radeon_test_syncing(rdev); | ||
1331 | else | ||
1332 | DRM_INFO("radeon: acceleration disabled, skipping sync tests\n"); | ||
1316 | } | 1333 | } |
1317 | if (radeon_benchmarking) { | 1334 | if (radeon_benchmarking) { |
1318 | radeon_benchmark(rdev, radeon_benchmarking); | 1335 | if (rdev->accel_working) |
1336 | radeon_benchmark(rdev, radeon_benchmarking); | ||
1337 | else | ||
1338 | DRM_INFO("radeon: acceleration disabled, skipping benchmarks\n"); | ||
1319 | } | 1339 | } |
1320 | return 0; | 1340 | return 0; |
1321 | } | 1341 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c index b055bddaa94c..0d1aa050d41d 100644 --- a/drivers/gpu/drm/radeon/radeon_display.c +++ b/drivers/gpu/drm/radeon/radeon_display.c | |||
@@ -1172,6 +1172,12 @@ static struct drm_prop_enum_list radeon_underscan_enum_list[] = | |||
1172 | { UNDERSCAN_AUTO, "auto" }, | 1172 | { UNDERSCAN_AUTO, "auto" }, |
1173 | }; | 1173 | }; |
1174 | 1174 | ||
1175 | static struct drm_prop_enum_list radeon_audio_enum_list[] = | ||
1176 | { { RADEON_AUDIO_DISABLE, "off" }, | ||
1177 | { RADEON_AUDIO_ENABLE, "on" }, | ||
1178 | { RADEON_AUDIO_AUTO, "auto" }, | ||
1179 | }; | ||
1180 | |||
1175 | static int radeon_modeset_create_props(struct radeon_device *rdev) | 1181 | static int radeon_modeset_create_props(struct radeon_device *rdev) |
1176 | { | 1182 | { |
1177 | int sz; | 1183 | int sz; |
@@ -1222,6 +1228,12 @@ static int radeon_modeset_create_props(struct radeon_device *rdev) | |||
1222 | if (!rdev->mode_info.underscan_vborder_property) | 1228 | if (!rdev->mode_info.underscan_vborder_property) |
1223 | return -ENOMEM; | 1229 | return -ENOMEM; |
1224 | 1230 | ||
1231 | sz = ARRAY_SIZE(radeon_audio_enum_list); | ||
1232 | rdev->mode_info.audio_property = | ||
1233 | drm_property_create_enum(rdev->ddev, 0, | ||
1234 | "audio", | ||
1235 | radeon_audio_enum_list, sz); | ||
1236 | |||
1225 | return 0; | 1237 | return 0; |
1226 | } | 1238 | } |
1227 | 1239 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_drv.c b/drivers/gpu/drm/radeon/radeon_drv.c index cb4445f55a96..cdd12dcd988b 100644 --- a/drivers/gpu/drm/radeon/radeon_drv.c +++ b/drivers/gpu/drm/radeon/radeon_drv.c | |||
@@ -153,7 +153,7 @@ int radeon_benchmarking = 0; | |||
153 | int radeon_testing = 0; | 153 | int radeon_testing = 0; |
154 | int radeon_connector_table = 0; | 154 | int radeon_connector_table = 0; |
155 | int radeon_tv = 1; | 155 | int radeon_tv = 1; |
156 | int radeon_audio = 0; | 156 | int radeon_audio = 1; |
157 | int radeon_disp_priority = 0; | 157 | int radeon_disp_priority = 0; |
158 | int radeon_hw_i2c = 0; | 158 | int radeon_hw_i2c = 0; |
159 | int radeon_pcie_gen2 = -1; | 159 | int radeon_pcie_gen2 = -1; |
diff --git a/drivers/gpu/drm/radeon/radeon_mode.h b/drivers/gpu/drm/radeon/radeon_mode.h index d908d8d68f6b..ef63d3f00b2f 100644 --- a/drivers/gpu/drm/radeon/radeon_mode.h +++ b/drivers/gpu/drm/radeon/radeon_mode.h | |||
@@ -247,6 +247,8 @@ struct radeon_mode_info { | |||
247 | struct drm_property *underscan_property; | 247 | struct drm_property *underscan_property; |
248 | struct drm_property *underscan_hborder_property; | 248 | struct drm_property *underscan_hborder_property; |
249 | struct drm_property *underscan_vborder_property; | 249 | struct drm_property *underscan_vborder_property; |
250 | /* audio */ | ||
251 | struct drm_property *audio_property; | ||
250 | /* hardcoded DFP edid from BIOS */ | 252 | /* hardcoded DFP edid from BIOS */ |
251 | struct edid *bios_hardcoded_edid; | 253 | struct edid *bios_hardcoded_edid; |
252 | int bios_hardcoded_edid_size; | 254 | int bios_hardcoded_edid_size; |
@@ -471,6 +473,12 @@ struct radeon_router { | |||
471 | u8 cd_mux_state; | 473 | u8 cd_mux_state; |
472 | }; | 474 | }; |
473 | 475 | ||
476 | enum radeon_connector_audio { | ||
477 | RADEON_AUDIO_DISABLE = 0, | ||
478 | RADEON_AUDIO_ENABLE = 1, | ||
479 | RADEON_AUDIO_AUTO = 2 | ||
480 | }; | ||
481 | |||
474 | struct radeon_connector { | 482 | struct radeon_connector { |
475 | struct drm_connector base; | 483 | struct drm_connector base; |
476 | uint32_t connector_id; | 484 | uint32_t connector_id; |
@@ -489,6 +497,7 @@ struct radeon_connector { | |||
489 | struct radeon_hpd hpd; | 497 | struct radeon_hpd hpd; |
490 | struct radeon_router router; | 498 | struct radeon_router router; |
491 | struct radeon_i2c_chan *router_bus; | 499 | struct radeon_i2c_chan *router_bus; |
500 | enum radeon_connector_audio audio; | ||
492 | }; | 501 | }; |
493 | 502 | ||
494 | struct radeon_framebuffer { | 503 | struct radeon_framebuffer { |
diff --git a/drivers/gpu/drm/radeon/radeon_pm.c b/drivers/gpu/drm/radeon/radeon_pm.c index d7555369a3e5..4f6b7fc7ad3c 100644 --- a/drivers/gpu/drm/radeon/radeon_pm.c +++ b/drivers/gpu/drm/radeon/radeon_pm.c | |||
@@ -67,7 +67,16 @@ int radeon_pm_get_type_index(struct radeon_device *rdev, | |||
67 | 67 | ||
68 | void radeon_pm_acpi_event_handler(struct radeon_device *rdev) | 68 | void radeon_pm_acpi_event_handler(struct radeon_device *rdev) |
69 | { | 69 | { |
70 | if (rdev->pm.pm_method == PM_METHOD_PROFILE) { | 70 | if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) { |
71 | mutex_lock(&rdev->pm.mutex); | ||
72 | if (power_supply_is_system_supplied() > 0) | ||
73 | rdev->pm.dpm.ac_power = true; | ||
74 | else | ||
75 | rdev->pm.dpm.ac_power = false; | ||
76 | if (rdev->asic->dpm.enable_bapm) | ||
77 | radeon_dpm_enable_bapm(rdev, rdev->pm.dpm.ac_power); | ||
78 | mutex_unlock(&rdev->pm.mutex); | ||
79 | } else if (rdev->pm.pm_method == PM_METHOD_PROFILE) { | ||
71 | if (rdev->pm.profile == PM_PROFILE_AUTO) { | 80 | if (rdev->pm.profile == PM_PROFILE_AUTO) { |
72 | mutex_lock(&rdev->pm.mutex); | 81 | mutex_lock(&rdev->pm.mutex); |
73 | radeon_pm_update_profile(rdev); | 82 | radeon_pm_update_profile(rdev); |
@@ -333,7 +342,7 @@ static ssize_t radeon_get_pm_profile(struct device *dev, | |||
333 | struct device_attribute *attr, | 342 | struct device_attribute *attr, |
334 | char *buf) | 343 | char *buf) |
335 | { | 344 | { |
336 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 345 | struct drm_device *ddev = dev_get_drvdata(dev); |
337 | struct radeon_device *rdev = ddev->dev_private; | 346 | struct radeon_device *rdev = ddev->dev_private; |
338 | int cp = rdev->pm.profile; | 347 | int cp = rdev->pm.profile; |
339 | 348 | ||
@@ -349,7 +358,7 @@ static ssize_t radeon_set_pm_profile(struct device *dev, | |||
349 | const char *buf, | 358 | const char *buf, |
350 | size_t count) | 359 | size_t count) |
351 | { | 360 | { |
352 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 361 | struct drm_device *ddev = dev_get_drvdata(dev); |
353 | struct radeon_device *rdev = ddev->dev_private; | 362 | struct radeon_device *rdev = ddev->dev_private; |
354 | 363 | ||
355 | mutex_lock(&rdev->pm.mutex); | 364 | mutex_lock(&rdev->pm.mutex); |
@@ -383,7 +392,7 @@ static ssize_t radeon_get_pm_method(struct device *dev, | |||
383 | struct device_attribute *attr, | 392 | struct device_attribute *attr, |
384 | char *buf) | 393 | char *buf) |
385 | { | 394 | { |
386 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 395 | struct drm_device *ddev = dev_get_drvdata(dev); |
387 | struct radeon_device *rdev = ddev->dev_private; | 396 | struct radeon_device *rdev = ddev->dev_private; |
388 | int pm = rdev->pm.pm_method; | 397 | int pm = rdev->pm.pm_method; |
389 | 398 | ||
@@ -397,7 +406,7 @@ static ssize_t radeon_set_pm_method(struct device *dev, | |||
397 | const char *buf, | 406 | const char *buf, |
398 | size_t count) | 407 | size_t count) |
399 | { | 408 | { |
400 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 409 | struct drm_device *ddev = dev_get_drvdata(dev); |
401 | struct radeon_device *rdev = ddev->dev_private; | 410 | struct radeon_device *rdev = ddev->dev_private; |
402 | 411 | ||
403 | /* we don't support the legacy modes with dpm */ | 412 | /* we don't support the legacy modes with dpm */ |
@@ -433,7 +442,7 @@ static ssize_t radeon_get_dpm_state(struct device *dev, | |||
433 | struct device_attribute *attr, | 442 | struct device_attribute *attr, |
434 | char *buf) | 443 | char *buf) |
435 | { | 444 | { |
436 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 445 | struct drm_device *ddev = dev_get_drvdata(dev); |
437 | struct radeon_device *rdev = ddev->dev_private; | 446 | struct radeon_device *rdev = ddev->dev_private; |
438 | enum radeon_pm_state_type pm = rdev->pm.dpm.user_state; | 447 | enum radeon_pm_state_type pm = rdev->pm.dpm.user_state; |
439 | 448 | ||
@@ -447,7 +456,7 @@ static ssize_t radeon_set_dpm_state(struct device *dev, | |||
447 | const char *buf, | 456 | const char *buf, |
448 | size_t count) | 457 | size_t count) |
449 | { | 458 | { |
450 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 459 | struct drm_device *ddev = dev_get_drvdata(dev); |
451 | struct radeon_device *rdev = ddev->dev_private; | 460 | struct radeon_device *rdev = ddev->dev_private; |
452 | 461 | ||
453 | mutex_lock(&rdev->pm.mutex); | 462 | mutex_lock(&rdev->pm.mutex); |
@@ -472,7 +481,7 @@ static ssize_t radeon_get_dpm_forced_performance_level(struct device *dev, | |||
472 | struct device_attribute *attr, | 481 | struct device_attribute *attr, |
473 | char *buf) | 482 | char *buf) |
474 | { | 483 | { |
475 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 484 | struct drm_device *ddev = dev_get_drvdata(dev); |
476 | struct radeon_device *rdev = ddev->dev_private; | 485 | struct radeon_device *rdev = ddev->dev_private; |
477 | enum radeon_dpm_forced_level level = rdev->pm.dpm.forced_level; | 486 | enum radeon_dpm_forced_level level = rdev->pm.dpm.forced_level; |
478 | 487 | ||
@@ -486,7 +495,7 @@ static ssize_t radeon_set_dpm_forced_performance_level(struct device *dev, | |||
486 | const char *buf, | 495 | const char *buf, |
487 | size_t count) | 496 | size_t count) |
488 | { | 497 | { |
489 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 498 | struct drm_device *ddev = dev_get_drvdata(dev); |
490 | struct radeon_device *rdev = ddev->dev_private; | 499 | struct radeon_device *rdev = ddev->dev_private; |
491 | enum radeon_dpm_forced_level level; | 500 | enum radeon_dpm_forced_level level; |
492 | int ret = 0; | 501 | int ret = 0; |
@@ -524,7 +533,7 @@ static ssize_t radeon_hwmon_show_temp(struct device *dev, | |||
524 | struct device_attribute *attr, | 533 | struct device_attribute *attr, |
525 | char *buf) | 534 | char *buf) |
526 | { | 535 | { |
527 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 536 | struct drm_device *ddev = dev_get_drvdata(dev); |
528 | struct radeon_device *rdev = ddev->dev_private; | 537 | struct radeon_device *rdev = ddev->dev_private; |
529 | int temp; | 538 | int temp; |
530 | 539 | ||
@@ -536,6 +545,23 @@ static ssize_t radeon_hwmon_show_temp(struct device *dev, | |||
536 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); | 545 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); |
537 | } | 546 | } |
538 | 547 | ||
548 | static ssize_t radeon_hwmon_show_temp_thresh(struct device *dev, | ||
549 | struct device_attribute *attr, | ||
550 | char *buf) | ||
551 | { | ||
552 | struct drm_device *ddev = dev_get_drvdata(dev); | ||
553 | struct radeon_device *rdev = ddev->dev_private; | ||
554 | int hyst = to_sensor_dev_attr(attr)->index; | ||
555 | int temp; | ||
556 | |||
557 | if (hyst) | ||
558 | temp = rdev->pm.dpm.thermal.min_temp; | ||
559 | else | ||
560 | temp = rdev->pm.dpm.thermal.max_temp; | ||
561 | |||
562 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); | ||
563 | } | ||
564 | |||
539 | static ssize_t radeon_hwmon_show_name(struct device *dev, | 565 | static ssize_t radeon_hwmon_show_name(struct device *dev, |
540 | struct device_attribute *attr, | 566 | struct device_attribute *attr, |
541 | char *buf) | 567 | char *buf) |
@@ -544,16 +570,37 @@ static ssize_t radeon_hwmon_show_name(struct device *dev, | |||
544 | } | 570 | } |
545 | 571 | ||
546 | static SENSOR_DEVICE_ATTR(temp1_input, S_IRUGO, radeon_hwmon_show_temp, NULL, 0); | 572 | static SENSOR_DEVICE_ATTR(temp1_input, S_IRUGO, radeon_hwmon_show_temp, NULL, 0); |
573 | static SENSOR_DEVICE_ATTR(temp1_crit, S_IRUGO, radeon_hwmon_show_temp_thresh, NULL, 0); | ||
574 | static SENSOR_DEVICE_ATTR(temp1_crit_hyst, S_IRUGO, radeon_hwmon_show_temp_thresh, NULL, 1); | ||
547 | static SENSOR_DEVICE_ATTR(name, S_IRUGO, radeon_hwmon_show_name, NULL, 0); | 575 | static SENSOR_DEVICE_ATTR(name, S_IRUGO, radeon_hwmon_show_name, NULL, 0); |
548 | 576 | ||
549 | static struct attribute *hwmon_attributes[] = { | 577 | static struct attribute *hwmon_attributes[] = { |
550 | &sensor_dev_attr_temp1_input.dev_attr.attr, | 578 | &sensor_dev_attr_temp1_input.dev_attr.attr, |
579 | &sensor_dev_attr_temp1_crit.dev_attr.attr, | ||
580 | &sensor_dev_attr_temp1_crit_hyst.dev_attr.attr, | ||
551 | &sensor_dev_attr_name.dev_attr.attr, | 581 | &sensor_dev_attr_name.dev_attr.attr, |
552 | NULL | 582 | NULL |
553 | }; | 583 | }; |
554 | 584 | ||
585 | static umode_t hwmon_attributes_visible(struct kobject *kobj, | ||
586 | struct attribute *attr, int index) | ||
587 | { | ||
588 | struct device *dev = container_of(kobj, struct device, kobj); | ||
589 | struct drm_device *ddev = dev_get_drvdata(dev); | ||
590 | struct radeon_device *rdev = ddev->dev_private; | ||
591 | |||
592 | /* Skip limit attributes if DPM is not enabled */ | ||
593 | if (rdev->pm.pm_method != PM_METHOD_DPM && | ||
594 | (attr == &sensor_dev_attr_temp1_crit.dev_attr.attr || | ||
595 | attr == &sensor_dev_attr_temp1_crit_hyst.dev_attr.attr)) | ||
596 | return 0; | ||
597 | |||
598 | return attr->mode; | ||
599 | } | ||
600 | |||
555 | static const struct attribute_group hwmon_attrgroup = { | 601 | static const struct attribute_group hwmon_attrgroup = { |
556 | .attrs = hwmon_attributes, | 602 | .attrs = hwmon_attributes, |
603 | .is_visible = hwmon_attributes_visible, | ||
557 | }; | 604 | }; |
558 | 605 | ||
559 | static int radeon_hwmon_init(struct radeon_device *rdev) | 606 | static int radeon_hwmon_init(struct radeon_device *rdev) |
@@ -870,10 +917,13 @@ static void radeon_dpm_change_power_state_locked(struct radeon_device *rdev) | |||
870 | 917 | ||
871 | radeon_dpm_post_set_power_state(rdev); | 918 | radeon_dpm_post_set_power_state(rdev); |
872 | 919 | ||
873 | /* force low perf level for thermal */ | 920 | if (rdev->asic->dpm.force_performance_level) { |
874 | if (rdev->pm.dpm.thermal_active && | 921 | if (rdev->pm.dpm.thermal_active) |
875 | rdev->asic->dpm.force_performance_level) { | 922 | /* force low perf level for thermal */ |
876 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_LOW); | 923 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_LOW); |
924 | else | ||
925 | /* otherwise, enable auto */ | ||
926 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
877 | } | 927 | } |
878 | 928 | ||
879 | done: | 929 | done: |
@@ -895,6 +945,8 @@ void radeon_dpm_enable_uvd(struct radeon_device *rdev, bool enable) | |||
895 | if (enable) { | 945 | if (enable) { |
896 | mutex_lock(&rdev->pm.mutex); | 946 | mutex_lock(&rdev->pm.mutex); |
897 | rdev->pm.dpm.uvd_active = true; | 947 | rdev->pm.dpm.uvd_active = true; |
948 | /* disable this for now */ | ||
949 | #if 0 | ||
898 | if ((rdev->pm.dpm.sd == 1) && (rdev->pm.dpm.hd == 0)) | 950 | if ((rdev->pm.dpm.sd == 1) && (rdev->pm.dpm.hd == 0)) |
899 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD_SD; | 951 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD_SD; |
900 | else if ((rdev->pm.dpm.sd == 2) && (rdev->pm.dpm.hd == 0)) | 952 | else if ((rdev->pm.dpm.sd == 2) && (rdev->pm.dpm.hd == 0)) |
@@ -904,6 +956,7 @@ void radeon_dpm_enable_uvd(struct radeon_device *rdev, bool enable) | |||
904 | else if ((rdev->pm.dpm.sd == 0) && (rdev->pm.dpm.hd == 2)) | 956 | else if ((rdev->pm.dpm.sd == 0) && (rdev->pm.dpm.hd == 2)) |
905 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD_HD2; | 957 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD_HD2; |
906 | else | 958 | else |
959 | #endif | ||
907 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD; | 960 | dpm_state = POWER_STATE_TYPE_INTERNAL_UVD; |
908 | rdev->pm.dpm.state = dpm_state; | 961 | rdev->pm.dpm.state = dpm_state; |
909 | mutex_unlock(&rdev->pm.mutex); | 962 | mutex_unlock(&rdev->pm.mutex); |
@@ -952,7 +1005,7 @@ static void radeon_pm_resume_old(struct radeon_device *rdev) | |||
952 | { | 1005 | { |
953 | /* set up the default clocks if the MC ucode is loaded */ | 1006 | /* set up the default clocks if the MC ucode is loaded */ |
954 | if ((rdev->family >= CHIP_BARTS) && | 1007 | if ((rdev->family >= CHIP_BARTS) && |
955 | (rdev->family <= CHIP_HAINAN) && | 1008 | (rdev->family <= CHIP_CAYMAN) && |
956 | rdev->mc_fw) { | 1009 | rdev->mc_fw) { |
957 | if (rdev->pm.default_vddc) | 1010 | if (rdev->pm.default_vddc) |
958 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1011 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
@@ -996,7 +1049,7 @@ static void radeon_pm_resume_dpm(struct radeon_device *rdev) | |||
996 | if (ret) { | 1049 | if (ret) { |
997 | DRM_ERROR("radeon: dpm resume failed\n"); | 1050 | DRM_ERROR("radeon: dpm resume failed\n"); |
998 | if ((rdev->family >= CHIP_BARTS) && | 1051 | if ((rdev->family >= CHIP_BARTS) && |
999 | (rdev->family <= CHIP_HAINAN) && | 1052 | (rdev->family <= CHIP_CAYMAN) && |
1000 | rdev->mc_fw) { | 1053 | rdev->mc_fw) { |
1001 | if (rdev->pm.default_vddc) | 1054 | if (rdev->pm.default_vddc) |
1002 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1055 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
@@ -1047,7 +1100,7 @@ static int radeon_pm_init_old(struct radeon_device *rdev) | |||
1047 | radeon_pm_init_profile(rdev); | 1100 | radeon_pm_init_profile(rdev); |
1048 | /* set up the default clocks if the MC ucode is loaded */ | 1101 | /* set up the default clocks if the MC ucode is loaded */ |
1049 | if ((rdev->family >= CHIP_BARTS) && | 1102 | if ((rdev->family >= CHIP_BARTS) && |
1050 | (rdev->family <= CHIP_HAINAN) && | 1103 | (rdev->family <= CHIP_CAYMAN) && |
1051 | rdev->mc_fw) { | 1104 | rdev->mc_fw) { |
1052 | if (rdev->pm.default_vddc) | 1105 | if (rdev->pm.default_vddc) |
1053 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1106 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
@@ -1102,9 +1155,10 @@ static int radeon_pm_init_dpm(struct radeon_device *rdev) | |||
1102 | { | 1155 | { |
1103 | int ret; | 1156 | int ret; |
1104 | 1157 | ||
1105 | /* default to performance state */ | 1158 | /* default to balanced state */ |
1106 | rdev->pm.dpm.state = POWER_STATE_TYPE_BALANCED; | 1159 | rdev->pm.dpm.state = POWER_STATE_TYPE_BALANCED; |
1107 | rdev->pm.dpm.user_state = POWER_STATE_TYPE_BALANCED; | 1160 | rdev->pm.dpm.user_state = POWER_STATE_TYPE_BALANCED; |
1161 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1108 | rdev->pm.default_sclk = rdev->clock.default_sclk; | 1162 | rdev->pm.default_sclk = rdev->clock.default_sclk; |
1109 | rdev->pm.default_mclk = rdev->clock.default_mclk; | 1163 | rdev->pm.default_mclk = rdev->clock.default_mclk; |
1110 | rdev->pm.current_sclk = rdev->clock.default_sclk; | 1164 | rdev->pm.current_sclk = rdev->clock.default_sclk; |
@@ -1132,7 +1186,7 @@ static int radeon_pm_init_dpm(struct radeon_device *rdev) | |||
1132 | if (ret) { | 1186 | if (ret) { |
1133 | rdev->pm.dpm_enabled = false; | 1187 | rdev->pm.dpm_enabled = false; |
1134 | if ((rdev->family >= CHIP_BARTS) && | 1188 | if ((rdev->family >= CHIP_BARTS) && |
1135 | (rdev->family <= CHIP_HAINAN) && | 1189 | (rdev->family <= CHIP_CAYMAN) && |
1136 | rdev->mc_fw) { | 1190 | rdev->mc_fw) { |
1137 | if (rdev->pm.default_vddc) | 1191 | if (rdev->pm.default_vddc) |
1138 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1192 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
diff --git a/drivers/gpu/drm/radeon/radeon_ring.c b/drivers/gpu/drm/radeon/radeon_ring.c index 46a25f037b84..18254e1c3e71 100644 --- a/drivers/gpu/drm/radeon/radeon_ring.c +++ b/drivers/gpu/drm/radeon/radeon_ring.c | |||
@@ -839,9 +839,11 @@ static int radeon_debugfs_ring_info(struct seq_file *m, void *data) | |||
839 | * packet that is the root issue | 839 | * packet that is the root issue |
840 | */ | 840 | */ |
841 | i = (ring->rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; | 841 | i = (ring->rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; |
842 | for (j = 0; j <= (count + 32); j++) { | 842 | if (ring->ready) { |
843 | seq_printf(m, "r[%5d]=0x%08x\n", i, ring->ring[i]); | 843 | for (j = 0; j <= (count + 32); j++) { |
844 | i = (i + 1) & ring->ptr_mask; | 844 | seq_printf(m, "r[%5d]=0x%08x\n", i, ring->ring[i]); |
845 | i = (i + 1) & ring->ptr_mask; | ||
846 | } | ||
845 | } | 847 | } |
846 | return 0; | 848 | return 0; |
847 | } | 849 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_test.c b/drivers/gpu/drm/radeon/radeon_test.c index f4d6bcee9006..12e8099a0823 100644 --- a/drivers/gpu/drm/radeon/radeon_test.c +++ b/drivers/gpu/drm/radeon/radeon_test.c | |||
@@ -36,8 +36,8 @@ static void radeon_do_test_moves(struct radeon_device *rdev, int flag) | |||
36 | struct radeon_bo *vram_obj = NULL; | 36 | struct radeon_bo *vram_obj = NULL; |
37 | struct radeon_bo **gtt_obj = NULL; | 37 | struct radeon_bo **gtt_obj = NULL; |
38 | uint64_t gtt_addr, vram_addr; | 38 | uint64_t gtt_addr, vram_addr; |
39 | unsigned i, n, size; | 39 | unsigned n, size; |
40 | int r, ring; | 40 | int i, r, ring; |
41 | 41 | ||
42 | switch (flag) { | 42 | switch (flag) { |
43 | case RADEON_TEST_COPY_DMA: | 43 | case RADEON_TEST_COPY_DMA: |
diff --git a/drivers/gpu/drm/radeon/radeon_trace.h b/drivers/gpu/drm/radeon/radeon_trace.h index eafd8160a155..f7e367815964 100644 --- a/drivers/gpu/drm/radeon/radeon_trace.h +++ b/drivers/gpu/drm/radeon/radeon_trace.h | |||
@@ -27,6 +27,26 @@ TRACE_EVENT(radeon_bo_create, | |||
27 | TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages) | 27 | TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages) |
28 | ); | 28 | ); |
29 | 29 | ||
30 | TRACE_EVENT(radeon_cs, | ||
31 | TP_PROTO(struct radeon_cs_parser *p), | ||
32 | TP_ARGS(p), | ||
33 | TP_STRUCT__entry( | ||
34 | __field(u32, ring) | ||
35 | __field(u32, dw) | ||
36 | __field(u32, fences) | ||
37 | ), | ||
38 | |||
39 | TP_fast_assign( | ||
40 | __entry->ring = p->ring; | ||
41 | __entry->dw = p->chunks[p->chunk_ib_idx].length_dw; | ||
42 | __entry->fences = radeon_fence_count_emitted( | ||
43 | p->rdev, p->ring); | ||
44 | ), | ||
45 | TP_printk("ring=%u, dw=%u, fences=%u", | ||
46 | __entry->ring, __entry->dw, | ||
47 | __entry->fences) | ||
48 | ); | ||
49 | |||
30 | DECLARE_EVENT_CLASS(radeon_fence_request, | 50 | DECLARE_EVENT_CLASS(radeon_fence_request, |
31 | 51 | ||
32 | TP_PROTO(struct drm_device *dev, u32 seqno), | 52 | TP_PROTO(struct drm_device *dev, u32 seqno), |
@@ -53,13 +73,6 @@ DEFINE_EVENT(radeon_fence_request, radeon_fence_emit, | |||
53 | TP_ARGS(dev, seqno) | 73 | TP_ARGS(dev, seqno) |
54 | ); | 74 | ); |
55 | 75 | ||
56 | DEFINE_EVENT(radeon_fence_request, radeon_fence_retire, | ||
57 | |||
58 | TP_PROTO(struct drm_device *dev, u32 seqno), | ||
59 | |||
60 | TP_ARGS(dev, seqno) | ||
61 | ); | ||
62 | |||
63 | DEFINE_EVENT(radeon_fence_request, radeon_fence_wait_begin, | 76 | DEFINE_EVENT(radeon_fence_request, radeon_fence_wait_begin, |
64 | 77 | ||
65 | TP_PROTO(struct drm_device *dev, u32 seqno), | 78 | TP_PROTO(struct drm_device *dev, u32 seqno), |
diff --git a/drivers/gpu/drm/radeon/radeon_uvd.c b/drivers/gpu/drm/radeon/radeon_uvd.c index 1a01bbff9bfa..4f2e73f79638 100644 --- a/drivers/gpu/drm/radeon/radeon_uvd.c +++ b/drivers/gpu/drm/radeon/radeon_uvd.c | |||
@@ -476,8 +476,7 @@ static int radeon_uvd_cs_reloc(struct radeon_cs_parser *p, | |||
476 | return -EINVAL; | 476 | return -EINVAL; |
477 | } | 477 | } |
478 | 478 | ||
479 | /* TODO: is this still necessary on NI+ ? */ | 479 | if (p->rdev->family < CHIP_PALM && (cmd == 0 || cmd == 0x3) && |
480 | if ((cmd == 0 || cmd == 0x3) && | ||
481 | (start >> 28) != (p->rdev->uvd.gpu_addr >> 28)) { | 480 | (start >> 28) != (p->rdev->uvd.gpu_addr >> 28)) { |
482 | DRM_ERROR("msg/fb buffer %LX-%LX out of 256MB segment!\n", | 481 | DRM_ERROR("msg/fb buffer %LX-%LX out of 256MB segment!\n", |
483 | start, end); | 482 | start, end); |
@@ -799,7 +798,8 @@ void radeon_uvd_note_usage(struct radeon_device *rdev) | |||
799 | (rdev->pm.dpm.hd != hd)) { | 798 | (rdev->pm.dpm.hd != hd)) { |
800 | rdev->pm.dpm.sd = sd; | 799 | rdev->pm.dpm.sd = sd; |
801 | rdev->pm.dpm.hd = hd; | 800 | rdev->pm.dpm.hd = hd; |
802 | streams_changed = true; | 801 | /* disable this for now */ |
802 | /*streams_changed = true;*/ | ||
803 | } | 803 | } |
804 | } | 804 | } |
805 | 805 | ||
diff --git a/drivers/gpu/drm/radeon/rs400.c b/drivers/gpu/drm/radeon/rs400.c index b8074a8ec75a..9566b5940a5a 100644 --- a/drivers/gpu/drm/radeon/rs400.c +++ b/drivers/gpu/drm/radeon/rs400.c | |||
@@ -274,19 +274,26 @@ static void rs400_mc_init(struct radeon_device *rdev) | |||
274 | 274 | ||
275 | uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 275 | uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
276 | { | 276 | { |
277 | unsigned long flags; | ||
277 | uint32_t r; | 278 | uint32_t r; |
278 | 279 | ||
280 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
279 | WREG32(RS480_NB_MC_INDEX, reg & 0xff); | 281 | WREG32(RS480_NB_MC_INDEX, reg & 0xff); |
280 | r = RREG32(RS480_NB_MC_DATA); | 282 | r = RREG32(RS480_NB_MC_DATA); |
281 | WREG32(RS480_NB_MC_INDEX, 0xff); | 283 | WREG32(RS480_NB_MC_INDEX, 0xff); |
284 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
282 | return r; | 285 | return r; |
283 | } | 286 | } |
284 | 287 | ||
285 | void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 288 | void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
286 | { | 289 | { |
290 | unsigned long flags; | ||
291 | |||
292 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
287 | WREG32(RS480_NB_MC_INDEX, ((reg) & 0xff) | RS480_NB_MC_IND_WR_EN); | 293 | WREG32(RS480_NB_MC_INDEX, ((reg) & 0xff) | RS480_NB_MC_IND_WR_EN); |
288 | WREG32(RS480_NB_MC_DATA, (v)); | 294 | WREG32(RS480_NB_MC_DATA, (v)); |
289 | WREG32(RS480_NB_MC_INDEX, 0xff); | 295 | WREG32(RS480_NB_MC_INDEX, 0xff); |
296 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
290 | } | 297 | } |
291 | 298 | ||
292 | #if defined(CONFIG_DEBUG_FS) | 299 | #if defined(CONFIG_DEBUG_FS) |
diff --git a/drivers/gpu/drm/radeon/rs600.c b/drivers/gpu/drm/radeon/rs600.c index 670b555d2ca2..6acba8017b9a 100644 --- a/drivers/gpu/drm/radeon/rs600.c +++ b/drivers/gpu/drm/radeon/rs600.c | |||
@@ -847,16 +847,26 @@ void rs600_bandwidth_update(struct radeon_device *rdev) | |||
847 | 847 | ||
848 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 848 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
849 | { | 849 | { |
850 | unsigned long flags; | ||
851 | u32 r; | ||
852 | |||
853 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
850 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | | 854 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | |
851 | S_000070_MC_IND_CITF_ARB0(1)); | 855 | S_000070_MC_IND_CITF_ARB0(1)); |
852 | return RREG32(R_000074_MC_IND_DATA); | 856 | r = RREG32(R_000074_MC_IND_DATA); |
857 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
858 | return r; | ||
853 | } | 859 | } |
854 | 860 | ||
855 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 861 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
856 | { | 862 | { |
863 | unsigned long flags; | ||
864 | |||
865 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
857 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | | 866 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | |
858 | S_000070_MC_IND_CITF_ARB0(1) | S_000070_MC_IND_WR_EN(1)); | 867 | S_000070_MC_IND_CITF_ARB0(1) | S_000070_MC_IND_WR_EN(1)); |
859 | WREG32(R_000074_MC_IND_DATA, v); | 868 | WREG32(R_000074_MC_IND_DATA, v); |
869 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
860 | } | 870 | } |
861 | 871 | ||
862 | static void rs600_debugfs(struct radeon_device *rdev) | 872 | static void rs600_debugfs(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/rs690.c b/drivers/gpu/drm/radeon/rs690.c index d8ddfb34545d..1447d794c22a 100644 --- a/drivers/gpu/drm/radeon/rs690.c +++ b/drivers/gpu/drm/radeon/rs690.c | |||
@@ -631,20 +631,27 @@ void rs690_bandwidth_update(struct radeon_device *rdev) | |||
631 | 631 | ||
632 | uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 632 | uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
633 | { | 633 | { |
634 | unsigned long flags; | ||
634 | uint32_t r; | 635 | uint32_t r; |
635 | 636 | ||
637 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
636 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg)); | 638 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg)); |
637 | r = RREG32(R_00007C_MC_DATA); | 639 | r = RREG32(R_00007C_MC_DATA); |
638 | WREG32(R_000078_MC_INDEX, ~C_000078_MC_IND_ADDR); | 640 | WREG32(R_000078_MC_INDEX, ~C_000078_MC_IND_ADDR); |
641 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
639 | return r; | 642 | return r; |
640 | } | 643 | } |
641 | 644 | ||
642 | void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 645 | void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
643 | { | 646 | { |
647 | unsigned long flags; | ||
648 | |||
649 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
644 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg) | | 650 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg) | |
645 | S_000078_MC_IND_WR_EN(1)); | 651 | S_000078_MC_IND_WR_EN(1)); |
646 | WREG32(R_00007C_MC_DATA, v); | 652 | WREG32(R_00007C_MC_DATA, v); |
647 | WREG32(R_000078_MC_INDEX, 0x7F); | 653 | WREG32(R_000078_MC_INDEX, 0x7F); |
654 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
648 | } | 655 | } |
649 | 656 | ||
650 | static void rs690_mc_program(struct radeon_device *rdev) | 657 | static void rs690_mc_program(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/rs780_dpm.c b/drivers/gpu/drm/radeon/rs780_dpm.c index d1a1ce73bd45..6af8505cf4d2 100644 --- a/drivers/gpu/drm/radeon/rs780_dpm.c +++ b/drivers/gpu/drm/radeon/rs780_dpm.c | |||
@@ -62,9 +62,7 @@ static void rs780_get_pm_mode_parameters(struct radeon_device *rdev) | |||
62 | radeon_crtc = to_radeon_crtc(crtc); | 62 | radeon_crtc = to_radeon_crtc(crtc); |
63 | pi->crtc_id = radeon_crtc->crtc_id; | 63 | pi->crtc_id = radeon_crtc->crtc_id; |
64 | if (crtc->mode.htotal && crtc->mode.vtotal) | 64 | if (crtc->mode.htotal && crtc->mode.vtotal) |
65 | pi->refresh_rate = | 65 | pi->refresh_rate = drm_mode_vrefresh(&crtc->mode); |
66 | (crtc->mode.clock * 1000) / | ||
67 | (crtc->mode.htotal * crtc->mode.vtotal); | ||
68 | break; | 66 | break; |
69 | } | 67 | } |
70 | } | 68 | } |
@@ -376,9 +374,8 @@ static void rs780_disable_vbios_powersaving(struct radeon_device *rdev) | |||
376 | WREG32_P(CG_INTGFX_MISC, 0, ~0xFFF00000); | 374 | WREG32_P(CG_INTGFX_MISC, 0, ~0xFFF00000); |
377 | } | 375 | } |
378 | 376 | ||
379 | static void rs780_force_voltage_to_high(struct radeon_device *rdev) | 377 | static void rs780_force_voltage(struct radeon_device *rdev, u16 voltage) |
380 | { | 378 | { |
381 | struct igp_power_info *pi = rs780_get_pi(rdev); | ||
382 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); | 379 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); |
383 | 380 | ||
384 | if ((current_state->max_voltage == RS780_VDDC_LEVEL_HIGH) && | 381 | if ((current_state->max_voltage == RS780_VDDC_LEVEL_HIGH) && |
@@ -390,7 +387,7 @@ static void rs780_force_voltage_to_high(struct radeon_device *rdev) | |||
390 | udelay(1); | 387 | udelay(1); |
391 | 388 | ||
392 | WREG32_P(FVTHROT_PWM_CTRL_REG0, | 389 | WREG32_P(FVTHROT_PWM_CTRL_REG0, |
393 | STARTING_PWM_HIGHTIME(pi->max_voltage), | 390 | STARTING_PWM_HIGHTIME(voltage), |
394 | ~STARTING_PWM_HIGHTIME_MASK); | 391 | ~STARTING_PWM_HIGHTIME_MASK); |
395 | 392 | ||
396 | WREG32_P(FVTHROT_PWM_CTRL_REG0, | 393 | WREG32_P(FVTHROT_PWM_CTRL_REG0, |
@@ -404,6 +401,26 @@ static void rs780_force_voltage_to_high(struct radeon_device *rdev) | |||
404 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | 401 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); |
405 | } | 402 | } |
406 | 403 | ||
404 | static void rs780_force_fbdiv(struct radeon_device *rdev, u32 fb_div) | ||
405 | { | ||
406 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); | ||
407 | |||
408 | if (current_state->sclk_low == current_state->sclk_high) | ||
409 | return; | ||
410 | |||
411 | WREG32_P(GFX_MACRO_BYPASS_CNTL, SPLL_BYPASS_CNTL, ~SPLL_BYPASS_CNTL); | ||
412 | |||
413 | WREG32_P(FVTHROT_FBDIV_REG2, FORCED_FEEDBACK_DIV(fb_div), | ||
414 | ~FORCED_FEEDBACK_DIV_MASK); | ||
415 | WREG32_P(FVTHROT_FBDIV_REG1, STARTING_FEEDBACK_DIV(fb_div), | ||
416 | ~STARTING_FEEDBACK_DIV_MASK); | ||
417 | WREG32_P(FVTHROT_FBDIV_REG1, FORCE_FEEDBACK_DIV, ~FORCE_FEEDBACK_DIV); | ||
418 | |||
419 | udelay(100); | ||
420 | |||
421 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | ||
422 | } | ||
423 | |||
407 | static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, | 424 | static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, |
408 | struct radeon_ps *new_ps, | 425 | struct radeon_ps *new_ps, |
409 | struct radeon_ps *old_ps) | 426 | struct radeon_ps *old_ps) |
@@ -432,17 +449,13 @@ static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, | |||
432 | if (ret) | 449 | if (ret) |
433 | return ret; | 450 | return ret; |
434 | 451 | ||
435 | WREG32_P(GFX_MACRO_BYPASS_CNTL, SPLL_BYPASS_CNTL, ~SPLL_BYPASS_CNTL); | 452 | if ((min_dividers.ref_div != max_dividers.ref_div) || |
436 | 453 | (min_dividers.post_div != max_dividers.post_div) || | |
437 | WREG32_P(FVTHROT_FBDIV_REG2, FORCED_FEEDBACK_DIV(max_dividers.fb_div), | 454 | (max_dividers.ref_div != current_max_dividers.ref_div) || |
438 | ~FORCED_FEEDBACK_DIV_MASK); | 455 | (max_dividers.post_div != current_max_dividers.post_div)) |
439 | WREG32_P(FVTHROT_FBDIV_REG1, STARTING_FEEDBACK_DIV(max_dividers.fb_div), | 456 | return -EINVAL; |
440 | ~STARTING_FEEDBACK_DIV_MASK); | ||
441 | WREG32_P(FVTHROT_FBDIV_REG1, FORCE_FEEDBACK_DIV, ~FORCE_FEEDBACK_DIV); | ||
442 | |||
443 | udelay(100); | ||
444 | 457 | ||
445 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | 458 | rs780_force_fbdiv(rdev, max_dividers.fb_div); |
446 | 459 | ||
447 | if (max_dividers.fb_div > min_dividers.fb_div) { | 460 | if (max_dividers.fb_div > min_dividers.fb_div) { |
448 | WREG32_P(FVTHROT_FBDIV_REG0, | 461 | WREG32_P(FVTHROT_FBDIV_REG0, |
@@ -486,6 +499,9 @@ static void rs780_activate_engine_clk_scaling(struct radeon_device *rdev, | |||
486 | (new_state->sclk_low == old_state->sclk_low)) | 499 | (new_state->sclk_low == old_state->sclk_low)) |
487 | return; | 500 | return; |
488 | 501 | ||
502 | if (new_state->sclk_high == new_state->sclk_low) | ||
503 | return; | ||
504 | |||
489 | rs780_clk_scaling_enable(rdev, true); | 505 | rs780_clk_scaling_enable(rdev, true); |
490 | } | 506 | } |
491 | 507 | ||
@@ -649,7 +665,7 @@ int rs780_dpm_set_power_state(struct radeon_device *rdev) | |||
649 | rs780_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); | 665 | rs780_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); |
650 | 666 | ||
651 | if (pi->voltage_control) { | 667 | if (pi->voltage_control) { |
652 | rs780_force_voltage_to_high(rdev); | 668 | rs780_force_voltage(rdev, pi->max_voltage); |
653 | mdelay(5); | 669 | mdelay(5); |
654 | } | 670 | } |
655 | 671 | ||
@@ -717,14 +733,18 @@ static void rs780_parse_pplib_non_clock_info(struct radeon_device *rdev, | |||
717 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { | 733 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { |
718 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); | 734 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); |
719 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); | 735 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); |
720 | } else if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
721 | rps->vclk = RS780_DEFAULT_VCLK_FREQ; | ||
722 | rps->dclk = RS780_DEFAULT_DCLK_FREQ; | ||
723 | } else { | 736 | } else { |
724 | rps->vclk = 0; | 737 | rps->vclk = 0; |
725 | rps->dclk = 0; | 738 | rps->dclk = 0; |
726 | } | 739 | } |
727 | 740 | ||
741 | if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
742 | if ((rps->vclk == 0) || (rps->dclk == 0)) { | ||
743 | rps->vclk = RS780_DEFAULT_VCLK_FREQ; | ||
744 | rps->dclk = RS780_DEFAULT_DCLK_FREQ; | ||
745 | } | ||
746 | } | ||
747 | |||
728 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) | 748 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) |
729 | rdev->pm.dpm.boot_ps = rps; | 749 | rdev->pm.dpm.boot_ps = rps; |
730 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) | 750 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) |
@@ -986,3 +1006,55 @@ void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rde | |||
986 | seq_printf(m, "power level 1 sclk: %u vddc_index: %d\n", | 1006 | seq_printf(m, "power level 1 sclk: %u vddc_index: %d\n", |
987 | ps->sclk_high, ps->max_voltage); | 1007 | ps->sclk_high, ps->max_voltage); |
988 | } | 1008 | } |
1009 | |||
1010 | int rs780_dpm_force_performance_level(struct radeon_device *rdev, | ||
1011 | enum radeon_dpm_forced_level level) | ||
1012 | { | ||
1013 | struct igp_power_info *pi = rs780_get_pi(rdev); | ||
1014 | struct radeon_ps *rps = rdev->pm.dpm.current_ps; | ||
1015 | struct igp_ps *ps = rs780_get_ps(rps); | ||
1016 | struct atom_clock_dividers dividers; | ||
1017 | int ret; | ||
1018 | |||
1019 | rs780_clk_scaling_enable(rdev, false); | ||
1020 | rs780_voltage_scaling_enable(rdev, false); | ||
1021 | |||
1022 | if (level == RADEON_DPM_FORCED_LEVEL_HIGH) { | ||
1023 | if (pi->voltage_control) | ||
1024 | rs780_force_voltage(rdev, pi->max_voltage); | ||
1025 | |||
1026 | ret = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM, | ||
1027 | ps->sclk_high, false, ÷rs); | ||
1028 | if (ret) | ||
1029 | return ret; | ||
1030 | |||
1031 | rs780_force_fbdiv(rdev, dividers.fb_div); | ||
1032 | } else if (level == RADEON_DPM_FORCED_LEVEL_LOW) { | ||
1033 | ret = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM, | ||
1034 | ps->sclk_low, false, ÷rs); | ||
1035 | if (ret) | ||
1036 | return ret; | ||
1037 | |||
1038 | rs780_force_fbdiv(rdev, dividers.fb_div); | ||
1039 | |||
1040 | if (pi->voltage_control) | ||
1041 | rs780_force_voltage(rdev, pi->min_voltage); | ||
1042 | } else { | ||
1043 | if (pi->voltage_control) | ||
1044 | rs780_force_voltage(rdev, pi->max_voltage); | ||
1045 | |||
1046 | if (ps->sclk_high != ps->sclk_low) { | ||
1047 | WREG32_P(FVTHROT_FBDIV_REG1, 0, ~FORCE_FEEDBACK_DIV); | ||
1048 | rs780_clk_scaling_enable(rdev, true); | ||
1049 | } | ||
1050 | |||
1051 | if (pi->voltage_control) { | ||
1052 | rs780_voltage_scaling_enable(rdev, true); | ||
1053 | rs780_enable_voltage_scaling(rdev, rps); | ||
1054 | } | ||
1055 | } | ||
1056 | |||
1057 | rdev->pm.dpm.forced_level = level; | ||
1058 | |||
1059 | return 0; | ||
1060 | } | ||
diff --git a/drivers/gpu/drm/radeon/rv515.c b/drivers/gpu/drm/radeon/rv515.c index 8ea1573ae820..873eb4b193b4 100644 --- a/drivers/gpu/drm/radeon/rv515.c +++ b/drivers/gpu/drm/radeon/rv515.c | |||
@@ -209,19 +209,27 @@ static void rv515_mc_init(struct radeon_device *rdev) | |||
209 | 209 | ||
210 | uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 210 | uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
211 | { | 211 | { |
212 | unsigned long flags; | ||
212 | uint32_t r; | 213 | uint32_t r; |
213 | 214 | ||
215 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
214 | WREG32(MC_IND_INDEX, 0x7f0000 | (reg & 0xffff)); | 216 | WREG32(MC_IND_INDEX, 0x7f0000 | (reg & 0xffff)); |
215 | r = RREG32(MC_IND_DATA); | 217 | r = RREG32(MC_IND_DATA); |
216 | WREG32(MC_IND_INDEX, 0); | 218 | WREG32(MC_IND_INDEX, 0); |
219 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
220 | |||
217 | return r; | 221 | return r; |
218 | } | 222 | } |
219 | 223 | ||
220 | void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 224 | void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
221 | { | 225 | { |
226 | unsigned long flags; | ||
227 | |||
228 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
222 | WREG32(MC_IND_INDEX, 0xff0000 | ((reg) & 0xffff)); | 229 | WREG32(MC_IND_INDEX, 0xff0000 | ((reg) & 0xffff)); |
223 | WREG32(MC_IND_DATA, (v)); | 230 | WREG32(MC_IND_DATA, (v)); |
224 | WREG32(MC_IND_INDEX, 0); | 231 | WREG32(MC_IND_INDEX, 0); |
232 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
225 | } | 233 | } |
226 | 234 | ||
227 | #if defined(CONFIG_DEBUG_FS) | 235 | #if defined(CONFIG_DEBUG_FS) |
diff --git a/drivers/gpu/drm/radeon/rv6xx_dpm.c b/drivers/gpu/drm/radeon/rv6xx_dpm.c index ab1f2016f21e..5811d277a36a 100644 --- a/drivers/gpu/drm/radeon/rv6xx_dpm.c +++ b/drivers/gpu/drm/radeon/rv6xx_dpm.c | |||
@@ -1758,8 +1758,6 @@ int rv6xx_dpm_set_power_state(struct radeon_device *rdev) | |||
1758 | 1758 | ||
1759 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1759 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
1760 | 1760 | ||
1761 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1762 | |||
1763 | return 0; | 1761 | return 0; |
1764 | } | 1762 | } |
1765 | 1763 | ||
diff --git a/drivers/gpu/drm/radeon/rv770_dpm.c b/drivers/gpu/drm/radeon/rv770_dpm.c index 8cbb85dae5aa..913b025ae9b3 100644 --- a/drivers/gpu/drm/radeon/rv770_dpm.c +++ b/drivers/gpu/drm/radeon/rv770_dpm.c | |||
@@ -2064,12 +2064,6 @@ int rv770_dpm_set_power_state(struct radeon_device *rdev) | |||
2064 | rv770_program_dcodt_after_state_switch(rdev, new_ps, old_ps); | 2064 | rv770_program_dcodt_after_state_switch(rdev, new_ps, old_ps); |
2065 | rv770_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 2065 | rv770_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
2066 | 2066 | ||
2067 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
2068 | if (ret) { | ||
2069 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
2070 | return ret; | ||
2071 | } | ||
2072 | |||
2073 | return 0; | 2067 | return 0; |
2074 | } | 2068 | } |
2075 | 2069 | ||
@@ -2147,14 +2141,18 @@ static void rv7xx_parse_pplib_non_clock_info(struct radeon_device *rdev, | |||
2147 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { | 2141 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { |
2148 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); | 2142 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); |
2149 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); | 2143 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); |
2150 | } else if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
2151 | rps->vclk = RV770_DEFAULT_VCLK_FREQ; | ||
2152 | rps->dclk = RV770_DEFAULT_DCLK_FREQ; | ||
2153 | } else { | 2144 | } else { |
2154 | rps->vclk = 0; | 2145 | rps->vclk = 0; |
2155 | rps->dclk = 0; | 2146 | rps->dclk = 0; |
2156 | } | 2147 | } |
2157 | 2148 | ||
2149 | if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
2150 | if ((rps->vclk == 0) || (rps->dclk == 0)) { | ||
2151 | rps->vclk = RV770_DEFAULT_VCLK_FREQ; | ||
2152 | rps->dclk = RV770_DEFAULT_DCLK_FREQ; | ||
2153 | } | ||
2154 | } | ||
2155 | |||
2158 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) | 2156 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) |
2159 | rdev->pm.dpm.boot_ps = rps; | 2157 | rdev->pm.dpm.boot_ps = rps; |
2160 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) | 2158 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) |
diff --git a/drivers/gpu/drm/radeon/rv770_smc.c b/drivers/gpu/drm/radeon/rv770_smc.c index ab95da570215..b2a224407365 100644 --- a/drivers/gpu/drm/radeon/rv770_smc.c +++ b/drivers/gpu/drm/radeon/rv770_smc.c | |||
@@ -274,8 +274,8 @@ static const u8 cayman_smc_int_vectors[] = | |||
274 | 0x08, 0x72, 0x08, 0x72 | 274 | 0x08, 0x72, 0x08, 0x72 |
275 | }; | 275 | }; |
276 | 276 | ||
277 | int rv770_set_smc_sram_address(struct radeon_device *rdev, | 277 | static int rv770_set_smc_sram_address(struct radeon_device *rdev, |
278 | u16 smc_address, u16 limit) | 278 | u16 smc_address, u16 limit) |
279 | { | 279 | { |
280 | u32 addr; | 280 | u32 addr; |
281 | 281 | ||
@@ -296,9 +296,10 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
296 | u16 smc_start_address, const u8 *src, | 296 | u16 smc_start_address, const u8 *src, |
297 | u16 byte_count, u16 limit) | 297 | u16 byte_count, u16 limit) |
298 | { | 298 | { |
299 | unsigned long flags; | ||
299 | u32 data, original_data, extra_shift; | 300 | u32 data, original_data, extra_shift; |
300 | u16 addr; | 301 | u16 addr; |
301 | int ret; | 302 | int ret = 0; |
302 | 303 | ||
303 | if (smc_start_address & 3) | 304 | if (smc_start_address & 3) |
304 | return -EINVAL; | 305 | return -EINVAL; |
@@ -307,13 +308,14 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
307 | 308 | ||
308 | addr = smc_start_address; | 309 | addr = smc_start_address; |
309 | 310 | ||
311 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
310 | while (byte_count >= 4) { | 312 | while (byte_count >= 4) { |
311 | /* SMC address space is BE */ | 313 | /* SMC address space is BE */ |
312 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 314 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
313 | 315 | ||
314 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 316 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
315 | if (ret) | 317 | if (ret) |
316 | return ret; | 318 | goto done; |
317 | 319 | ||
318 | WREG32(SMC_SRAM_DATA, data); | 320 | WREG32(SMC_SRAM_DATA, data); |
319 | 321 | ||
@@ -328,7 +330,7 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
328 | 330 | ||
329 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 331 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
330 | if (ret) | 332 | if (ret) |
331 | return ret; | 333 | goto done; |
332 | 334 | ||
333 | original_data = RREG32(SMC_SRAM_DATA); | 335 | original_data = RREG32(SMC_SRAM_DATA); |
334 | 336 | ||
@@ -346,12 +348,15 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
346 | 348 | ||
347 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 349 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
348 | if (ret) | 350 | if (ret) |
349 | return ret; | 351 | goto done; |
350 | 352 | ||
351 | WREG32(SMC_SRAM_DATA, data); | 353 | WREG32(SMC_SRAM_DATA, data); |
352 | } | 354 | } |
353 | 355 | ||
354 | return 0; | 356 | done: |
357 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
358 | |||
359 | return ret; | ||
355 | } | 360 | } |
356 | 361 | ||
357 | static int rv770_program_interrupt_vectors(struct radeon_device *rdev, | 362 | static int rv770_program_interrupt_vectors(struct radeon_device *rdev, |
@@ -461,12 +466,15 @@ PPSMC_Result rv770_wait_for_smc_inactive(struct radeon_device *rdev) | |||
461 | 466 | ||
462 | static void rv770_clear_smc_sram(struct radeon_device *rdev, u16 limit) | 467 | static void rv770_clear_smc_sram(struct radeon_device *rdev, u16 limit) |
463 | { | 468 | { |
469 | unsigned long flags; | ||
464 | u16 i; | 470 | u16 i; |
465 | 471 | ||
472 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
466 | for (i = 0; i < limit; i += 4) { | 473 | for (i = 0; i < limit; i += 4) { |
467 | rv770_set_smc_sram_address(rdev, i, limit); | 474 | rv770_set_smc_sram_address(rdev, i, limit); |
468 | WREG32(SMC_SRAM_DATA, 0); | 475 | WREG32(SMC_SRAM_DATA, 0); |
469 | } | 476 | } |
477 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
470 | } | 478 | } |
471 | 479 | ||
472 | int rv770_load_smc_ucode(struct radeon_device *rdev, | 480 | int rv770_load_smc_ucode(struct radeon_device *rdev, |
@@ -595,27 +603,29 @@ int rv770_load_smc_ucode(struct radeon_device *rdev, | |||
595 | int rv770_read_smc_sram_dword(struct radeon_device *rdev, | 603 | int rv770_read_smc_sram_dword(struct radeon_device *rdev, |
596 | u16 smc_address, u32 *value, u16 limit) | 604 | u16 smc_address, u32 *value, u16 limit) |
597 | { | 605 | { |
606 | unsigned long flags; | ||
598 | int ret; | 607 | int ret; |
599 | 608 | ||
609 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
600 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); | 610 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); |
601 | if (ret) | 611 | if (ret == 0) |
602 | return ret; | 612 | *value = RREG32(SMC_SRAM_DATA); |
603 | 613 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | |
604 | *value = RREG32(SMC_SRAM_DATA); | ||
605 | 614 | ||
606 | return 0; | 615 | return ret; |
607 | } | 616 | } |
608 | 617 | ||
609 | int rv770_write_smc_sram_dword(struct radeon_device *rdev, | 618 | int rv770_write_smc_sram_dword(struct radeon_device *rdev, |
610 | u16 smc_address, u32 value, u16 limit) | 619 | u16 smc_address, u32 value, u16 limit) |
611 | { | 620 | { |
621 | unsigned long flags; | ||
612 | int ret; | 622 | int ret; |
613 | 623 | ||
624 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
614 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); | 625 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); |
615 | if (ret) | 626 | if (ret == 0) |
616 | return ret; | 627 | WREG32(SMC_SRAM_DATA, value); |
628 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
617 | 629 | ||
618 | WREG32(SMC_SRAM_DATA, value); | 630 | return ret; |
619 | |||
620 | return 0; | ||
621 | } | 631 | } |
diff --git a/drivers/gpu/drm/radeon/rv770_smc.h b/drivers/gpu/drm/radeon/rv770_smc.h index f78d92a4b325..3b2c963c4880 100644 --- a/drivers/gpu/drm/radeon/rv770_smc.h +++ b/drivers/gpu/drm/radeon/rv770_smc.h | |||
@@ -187,8 +187,6 @@ typedef struct RV770_SMC_STATETABLE RV770_SMC_STATETABLE; | |||
187 | #define RV770_SMC_SOFT_REGISTER_uvd_enabled 0x9C | 187 | #define RV770_SMC_SOFT_REGISTER_uvd_enabled 0x9C |
188 | #define RV770_SMC_SOFT_REGISTER_is_asic_lombok 0xA0 | 188 | #define RV770_SMC_SOFT_REGISTER_is_asic_lombok 0xA0 |
189 | 189 | ||
190 | int rv770_set_smc_sram_address(struct radeon_device *rdev, | ||
191 | u16 smc_address, u16 limit); | ||
192 | int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | 190 | int rv770_copy_bytes_to_smc(struct radeon_device *rdev, |
193 | u16 smc_start_address, const u8 *src, | 191 | u16 smc_start_address, const u8 *src, |
194 | u16 byte_count, u16 limit); | 192 | u16 byte_count, u16 limit); |
diff --git a/drivers/gpu/drm/radeon/rv770d.h b/drivers/gpu/drm/radeon/rv770d.h index 9fe60e542922..1ae277152cc7 100644 --- a/drivers/gpu/drm/radeon/rv770d.h +++ b/drivers/gpu/drm/radeon/rv770d.h | |||
@@ -852,7 +852,7 @@ | |||
852 | #define AFMT_VBI_PACKET_CONTROL 0x7608 | 852 | #define AFMT_VBI_PACKET_CONTROL 0x7608 |
853 | # define AFMT_GENERIC0_UPDATE (1 << 2) | 853 | # define AFMT_GENERIC0_UPDATE (1 << 2) |
854 | #define AFMT_INFOFRAME_CONTROL0 0x760c | 854 | #define AFMT_INFOFRAME_CONTROL0 0x760c |
855 | # define AFMT_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hmdi regs */ | 855 | # define AFMT_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hdmi regs */ |
856 | # define AFMT_AUDIO_INFO_UPDATE (1 << 7) | 856 | # define AFMT_AUDIO_INFO_UPDATE (1 << 7) |
857 | # define AFMT_MPEG_INFO_UPDATE (1 << 10) | 857 | # define AFMT_MPEG_INFO_UPDATE (1 << 10) |
858 | #define AFMT_GENERIC0_7 0x7610 | 858 | #define AFMT_GENERIC0_7 0x7610 |
diff --git a/drivers/gpu/drm/radeon/si.c b/drivers/gpu/drm/radeon/si.c index 3e23b757dcfa..d4652af425b8 100644 --- a/drivers/gpu/drm/radeon/si.c +++ b/drivers/gpu/drm/radeon/si.c | |||
@@ -83,6 +83,11 @@ extern void si_dma_vm_set_page(struct radeon_device *rdev, | |||
83 | uint64_t pe, | 83 | uint64_t pe, |
84 | uint64_t addr, unsigned count, | 84 | uint64_t addr, unsigned count, |
85 | uint32_t incr, uint32_t flags); | 85 | uint32_t incr, uint32_t flags); |
86 | static void si_enable_gui_idle_interrupt(struct radeon_device *rdev, | ||
87 | bool enable); | ||
88 | static void si_fini_pg(struct radeon_device *rdev); | ||
89 | static void si_fini_cg(struct radeon_device *rdev); | ||
90 | static void si_rlc_stop(struct radeon_device *rdev); | ||
86 | 91 | ||
87 | static const u32 verde_rlc_save_restore_register_list[] = | 92 | static const u32 verde_rlc_save_restore_register_list[] = |
88 | { | 93 | { |
@@ -3386,6 +3391,8 @@ static int si_cp_resume(struct radeon_device *rdev) | |||
3386 | u32 rb_bufsz; | 3391 | u32 rb_bufsz; |
3387 | int r; | 3392 | int r; |
3388 | 3393 | ||
3394 | si_enable_gui_idle_interrupt(rdev, false); | ||
3395 | |||
3389 | WREG32(CP_SEM_WAIT_TIMER, 0x0); | 3396 | WREG32(CP_SEM_WAIT_TIMER, 0x0); |
3390 | WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0); | 3397 | WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0); |
3391 | 3398 | ||
@@ -3501,6 +3508,8 @@ static int si_cp_resume(struct radeon_device *rdev) | |||
3501 | rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false; | 3508 | rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false; |
3502 | } | 3509 | } |
3503 | 3510 | ||
3511 | si_enable_gui_idle_interrupt(rdev, true); | ||
3512 | |||
3504 | return 0; | 3513 | return 0; |
3505 | } | 3514 | } |
3506 | 3515 | ||
@@ -3602,6 +3611,13 @@ static void si_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask) | |||
3602 | dev_info(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n", | 3611 | dev_info(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n", |
3603 | RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS)); | 3612 | RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS)); |
3604 | 3613 | ||
3614 | /* disable PG/CG */ | ||
3615 | si_fini_pg(rdev); | ||
3616 | si_fini_cg(rdev); | ||
3617 | |||
3618 | /* stop the rlc */ | ||
3619 | si_rlc_stop(rdev); | ||
3620 | |||
3605 | /* Disable CP parsing/prefetching */ | 3621 | /* Disable CP parsing/prefetching */ |
3606 | WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT | CP_CE_HALT); | 3622 | WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT | CP_CE_HALT); |
3607 | 3623 | ||
@@ -4888,7 +4904,7 @@ static void si_enable_gfx_cgpg(struct radeon_device *rdev, | |||
4888 | { | 4904 | { |
4889 | u32 tmp; | 4905 | u32 tmp; |
4890 | 4906 | ||
4891 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG)) { | 4907 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG)) { |
4892 | tmp = RLC_PUD(0x10) | RLC_PDD(0x10) | RLC_TTPD(0x10) | RLC_MSD(0x10); | 4908 | tmp = RLC_PUD(0x10) | RLC_PDD(0x10) | RLC_TTPD(0x10) | RLC_MSD(0x10); |
4893 | WREG32(RLC_TTOP_D, tmp); | 4909 | WREG32(RLC_TTOP_D, tmp); |
4894 | 4910 | ||
@@ -5250,6 +5266,7 @@ void si_update_cg(struct radeon_device *rdev, | |||
5250 | u32 block, bool enable) | 5266 | u32 block, bool enable) |
5251 | { | 5267 | { |
5252 | if (block & RADEON_CG_BLOCK_GFX) { | 5268 | if (block & RADEON_CG_BLOCK_GFX) { |
5269 | si_enable_gui_idle_interrupt(rdev, false); | ||
5253 | /* order matters! */ | 5270 | /* order matters! */ |
5254 | if (enable) { | 5271 | if (enable) { |
5255 | si_enable_mgcg(rdev, true); | 5272 | si_enable_mgcg(rdev, true); |
@@ -5258,6 +5275,7 @@ void si_update_cg(struct radeon_device *rdev, | |||
5258 | si_enable_cgcg(rdev, false); | 5275 | si_enable_cgcg(rdev, false); |
5259 | si_enable_mgcg(rdev, false); | 5276 | si_enable_mgcg(rdev, false); |
5260 | } | 5277 | } |
5278 | si_enable_gui_idle_interrupt(rdev, true); | ||
5261 | } | 5279 | } |
5262 | 5280 | ||
5263 | if (block & RADEON_CG_BLOCK_MC) { | 5281 | if (block & RADEON_CG_BLOCK_MC) { |
@@ -5408,7 +5426,7 @@ static void si_init_pg(struct radeon_device *rdev) | |||
5408 | si_init_dma_pg(rdev); | 5426 | si_init_dma_pg(rdev); |
5409 | } | 5427 | } |
5410 | si_init_ao_cu_mask(rdev); | 5428 | si_init_ao_cu_mask(rdev); |
5411 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5429 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
5412 | si_init_gfx_cgpg(rdev); | 5430 | si_init_gfx_cgpg(rdev); |
5413 | } | 5431 | } |
5414 | si_enable_dma_pg(rdev, true); | 5432 | si_enable_dma_pg(rdev, true); |
@@ -5560,7 +5578,9 @@ static void si_disable_interrupt_state(struct radeon_device *rdev) | |||
5560 | { | 5578 | { |
5561 | u32 tmp; | 5579 | u32 tmp; |
5562 | 5580 | ||
5563 | WREG32(CP_INT_CNTL_RING0, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | 5581 | tmp = RREG32(CP_INT_CNTL_RING0) & |
5582 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
5583 | WREG32(CP_INT_CNTL_RING0, tmp); | ||
5564 | WREG32(CP_INT_CNTL_RING1, 0); | 5584 | WREG32(CP_INT_CNTL_RING1, 0); |
5565 | WREG32(CP_INT_CNTL_RING2, 0); | 5585 | WREG32(CP_INT_CNTL_RING2, 0); |
5566 | tmp = RREG32(DMA_CNTL + DMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; | 5586 | tmp = RREG32(DMA_CNTL + DMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; |
@@ -5685,7 +5705,7 @@ static int si_irq_init(struct radeon_device *rdev) | |||
5685 | 5705 | ||
5686 | int si_irq_set(struct radeon_device *rdev) | 5706 | int si_irq_set(struct radeon_device *rdev) |
5687 | { | 5707 | { |
5688 | u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE; | 5708 | u32 cp_int_cntl; |
5689 | u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0; | 5709 | u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0; |
5690 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; | 5710 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; |
5691 | u32 hpd1 = 0, hpd2 = 0, hpd3 = 0, hpd4 = 0, hpd5 = 0, hpd6 = 0; | 5711 | u32 hpd1 = 0, hpd2 = 0, hpd3 = 0, hpd4 = 0, hpd5 = 0, hpd6 = 0; |
@@ -5706,6 +5726,9 @@ int si_irq_set(struct radeon_device *rdev) | |||
5706 | return 0; | 5726 | return 0; |
5707 | } | 5727 | } |
5708 | 5728 | ||
5729 | cp_int_cntl = RREG32(CP_INT_CNTL_RING0) & | ||
5730 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
5731 | |||
5709 | if (!ASIC_IS_NODCE(rdev)) { | 5732 | if (!ASIC_IS_NODCE(rdev)) { |
5710 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; | 5733 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; |
5711 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; | 5734 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; |
diff --git a/drivers/gpu/drm/radeon/si_dpm.c b/drivers/gpu/drm/radeon/si_dpm.c index 5be9b4e72350..2332aa1bf93c 100644 --- a/drivers/gpu/drm/radeon/si_dpm.c +++ b/drivers/gpu/drm/radeon/si_dpm.c | |||
@@ -2910,6 +2910,7 @@ static void si_apply_state_adjust_rules(struct radeon_device *rdev, | |||
2910 | bool disable_sclk_switching = false; | 2910 | bool disable_sclk_switching = false; |
2911 | u32 mclk, sclk; | 2911 | u32 mclk, sclk; |
2912 | u16 vddc, vddci; | 2912 | u16 vddc, vddci; |
2913 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
2913 | int i; | 2914 | int i; |
2914 | 2915 | ||
2915 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 2916 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
@@ -2943,6 +2944,29 @@ static void si_apply_state_adjust_rules(struct radeon_device *rdev, | |||
2943 | } | 2944 | } |
2944 | } | 2945 | } |
2945 | 2946 | ||
2947 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
2948 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
2949 | &max_sclk_vddc); | ||
2950 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
2951 | &max_mclk_vddci); | ||
2952 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
2953 | &max_mclk_vddc); | ||
2954 | |||
2955 | for (i = 0; i < ps->performance_level_count; i++) { | ||
2956 | if (max_sclk_vddc) { | ||
2957 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
2958 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
2959 | } | ||
2960 | if (max_mclk_vddci) { | ||
2961 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
2962 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
2963 | } | ||
2964 | if (max_mclk_vddc) { | ||
2965 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
2966 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
2967 | } | ||
2968 | } | ||
2969 | |||
2946 | /* XXX validate the min clocks required for display */ | 2970 | /* XXX validate the min clocks required for display */ |
2947 | 2971 | ||
2948 | if (disable_mclk_switching) { | 2972 | if (disable_mclk_switching) { |
@@ -5184,7 +5208,7 @@ static int si_set_mc_special_registers(struct radeon_device *rdev, | |||
5184 | table->mc_reg_table_entry[k].mc_data[j] |= 0x100; | 5208 | table->mc_reg_table_entry[k].mc_data[j] |= 0x100; |
5185 | } | 5209 | } |
5186 | j++; | 5210 | j++; |
5187 | if (j > SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) | 5211 | if (j >= SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) |
5188 | return -EINVAL; | 5212 | return -EINVAL; |
5189 | 5213 | ||
5190 | if (!pi->mem_gddr5) { | 5214 | if (!pi->mem_gddr5) { |
@@ -5194,7 +5218,7 @@ static int si_set_mc_special_registers(struct radeon_device *rdev, | |||
5194 | table->mc_reg_table_entry[k].mc_data[j] = | 5218 | table->mc_reg_table_entry[k].mc_data[j] = |
5195 | (table->mc_reg_table_entry[k].mc_data[i] & 0xffff0000) >> 16; | 5219 | (table->mc_reg_table_entry[k].mc_data[i] & 0xffff0000) >> 16; |
5196 | j++; | 5220 | j++; |
5197 | if (j > SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) | 5221 | if (j >= SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) |
5198 | return -EINVAL; | 5222 | return -EINVAL; |
5199 | } | 5223 | } |
5200 | break; | 5224 | break; |
@@ -5207,7 +5231,7 @@ static int si_set_mc_special_registers(struct radeon_device *rdev, | |||
5207 | (temp_reg & 0xffff0000) | | 5231 | (temp_reg & 0xffff0000) | |
5208 | (table->mc_reg_table_entry[k].mc_data[i] & 0x0000ffff); | 5232 | (table->mc_reg_table_entry[k].mc_data[i] & 0x0000ffff); |
5209 | j++; | 5233 | j++; |
5210 | if (j > SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) | 5234 | if (j >= SMC_SISLANDS_MC_REGISTER_ARRAY_SIZE) |
5211 | return -EINVAL; | 5235 | return -EINVAL; |
5212 | break; | 5236 | break; |
5213 | default: | 5237 | default: |
@@ -6075,12 +6099,6 @@ int si_dpm_set_power_state(struct radeon_device *rdev) | |||
6075 | return ret; | 6099 | return ret; |
6076 | } | 6100 | } |
6077 | 6101 | ||
6078 | ret = si_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
6079 | if (ret) { | ||
6080 | DRM_ERROR("si_dpm_force_performance_level failed\n"); | ||
6081 | return ret; | ||
6082 | } | ||
6083 | |||
6084 | si_update_cg(rdev, (RADEON_CG_BLOCK_GFX | | 6102 | si_update_cg(rdev, (RADEON_CG_BLOCK_GFX | |
6085 | RADEON_CG_BLOCK_MC | | 6103 | RADEON_CG_BLOCK_MC | |
6086 | RADEON_CG_BLOCK_SDMA | | 6104 | RADEON_CG_BLOCK_SDMA | |
diff --git a/drivers/gpu/drm/radeon/si_smc.c b/drivers/gpu/drm/radeon/si_smc.c index 5f524c0a541e..d422a1cbf727 100644 --- a/drivers/gpu/drm/radeon/si_smc.c +++ b/drivers/gpu/drm/radeon/si_smc.c | |||
@@ -29,8 +29,8 @@ | |||
29 | #include "ppsmc.h" | 29 | #include "ppsmc.h" |
30 | #include "radeon_ucode.h" | 30 | #include "radeon_ucode.h" |
31 | 31 | ||
32 | int si_set_smc_sram_address(struct radeon_device *rdev, | 32 | static int si_set_smc_sram_address(struct radeon_device *rdev, |
33 | u32 smc_address, u32 limit) | 33 | u32 smc_address, u32 limit) |
34 | { | 34 | { |
35 | if (smc_address & 3) | 35 | if (smc_address & 3) |
36 | return -EINVAL; | 36 | return -EINVAL; |
@@ -47,7 +47,8 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
47 | u32 smc_start_address, | 47 | u32 smc_start_address, |
48 | const u8 *src, u32 byte_count, u32 limit) | 48 | const u8 *src, u32 byte_count, u32 limit) |
49 | { | 49 | { |
50 | int ret; | 50 | unsigned long flags; |
51 | int ret = 0; | ||
51 | u32 data, original_data, addr, extra_shift; | 52 | u32 data, original_data, addr, extra_shift; |
52 | 53 | ||
53 | if (smc_start_address & 3) | 54 | if (smc_start_address & 3) |
@@ -57,13 +58,14 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
57 | 58 | ||
58 | addr = smc_start_address; | 59 | addr = smc_start_address; |
59 | 60 | ||
61 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
60 | while (byte_count >= 4) { | 62 | while (byte_count >= 4) { |
61 | /* SMC address space is BE */ | 63 | /* SMC address space is BE */ |
62 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 64 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
63 | 65 | ||
64 | ret = si_set_smc_sram_address(rdev, addr, limit); | 66 | ret = si_set_smc_sram_address(rdev, addr, limit); |
65 | if (ret) | 67 | if (ret) |
66 | return ret; | 68 | goto done; |
67 | 69 | ||
68 | WREG32(SMC_IND_DATA_0, data); | 70 | WREG32(SMC_IND_DATA_0, data); |
69 | 71 | ||
@@ -78,7 +80,7 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
78 | 80 | ||
79 | ret = si_set_smc_sram_address(rdev, addr, limit); | 81 | ret = si_set_smc_sram_address(rdev, addr, limit); |
80 | if (ret) | 82 | if (ret) |
81 | return ret; | 83 | goto done; |
82 | 84 | ||
83 | original_data = RREG32(SMC_IND_DATA_0); | 85 | original_data = RREG32(SMC_IND_DATA_0); |
84 | 86 | ||
@@ -96,11 +98,15 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
96 | 98 | ||
97 | ret = si_set_smc_sram_address(rdev, addr, limit); | 99 | ret = si_set_smc_sram_address(rdev, addr, limit); |
98 | if (ret) | 100 | if (ret) |
99 | return ret; | 101 | goto done; |
100 | 102 | ||
101 | WREG32(SMC_IND_DATA_0, data); | 103 | WREG32(SMC_IND_DATA_0, data); |
102 | } | 104 | } |
103 | return 0; | 105 | |
106 | done: | ||
107 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
108 | |||
109 | return ret; | ||
104 | } | 110 | } |
105 | 111 | ||
106 | void si_start_smc(struct radeon_device *rdev) | 112 | void si_start_smc(struct radeon_device *rdev) |
@@ -203,6 +209,7 @@ PPSMC_Result si_wait_for_smc_inactive(struct radeon_device *rdev) | |||
203 | 209 | ||
204 | int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | 210 | int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) |
205 | { | 211 | { |
212 | unsigned long flags; | ||
206 | u32 ucode_start_address; | 213 | u32 ucode_start_address; |
207 | u32 ucode_size; | 214 | u32 ucode_size; |
208 | const u8 *src; | 215 | const u8 *src; |
@@ -241,6 +248,7 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
241 | return -EINVAL; | 248 | return -EINVAL; |
242 | 249 | ||
243 | src = (const u8 *)rdev->smc_fw->data; | 250 | src = (const u8 *)rdev->smc_fw->data; |
251 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
244 | WREG32(SMC_IND_INDEX_0, ucode_start_address); | 252 | WREG32(SMC_IND_INDEX_0, ucode_start_address); |
245 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); | 253 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); |
246 | while (ucode_size >= 4) { | 254 | while (ucode_size >= 4) { |
@@ -253,6 +261,7 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
253 | ucode_size -= 4; | 261 | ucode_size -= 4; |
254 | } | 262 | } |
255 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); | 263 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); |
264 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
256 | 265 | ||
257 | return 0; | 266 | return 0; |
258 | } | 267 | } |
@@ -260,25 +269,29 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
260 | int si_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 269 | int si_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
261 | u32 *value, u32 limit) | 270 | u32 *value, u32 limit) |
262 | { | 271 | { |
272 | unsigned long flags; | ||
263 | int ret; | 273 | int ret; |
264 | 274 | ||
275 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
265 | ret = si_set_smc_sram_address(rdev, smc_address, limit); | 276 | ret = si_set_smc_sram_address(rdev, smc_address, limit); |
266 | if (ret) | 277 | if (ret == 0) |
267 | return ret; | 278 | *value = RREG32(SMC_IND_DATA_0); |
279 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
268 | 280 | ||
269 | *value = RREG32(SMC_IND_DATA_0); | 281 | return ret; |
270 | return 0; | ||
271 | } | 282 | } |
272 | 283 | ||
273 | int si_write_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 284 | int si_write_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
274 | u32 value, u32 limit) | 285 | u32 value, u32 limit) |
275 | { | 286 | { |
287 | unsigned long flags; | ||
276 | int ret; | 288 | int ret; |
277 | 289 | ||
290 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
278 | ret = si_set_smc_sram_address(rdev, smc_address, limit); | 291 | ret = si_set_smc_sram_address(rdev, smc_address, limit); |
279 | if (ret) | 292 | if (ret == 0) |
280 | return ret; | 293 | WREG32(SMC_IND_DATA_0, value); |
294 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
281 | 295 | ||
282 | WREG32(SMC_IND_DATA_0, value); | 296 | return ret; |
283 | return 0; | ||
284 | } | 297 | } |
diff --git a/drivers/gpu/drm/radeon/sid.h b/drivers/gpu/drm/radeon/sid.h index 52d2ab6b67a0..7e2e0ea66a00 100644 --- a/drivers/gpu/drm/radeon/sid.h +++ b/drivers/gpu/drm/radeon/sid.h | |||
@@ -1553,7 +1553,7 @@ | |||
1553 | * 6. COMMAND [30:21] | BYTE_COUNT [20:0] | 1553 | * 6. COMMAND [30:21] | BYTE_COUNT [20:0] |
1554 | */ | 1554 | */ |
1555 | # define PACKET3_CP_DMA_DST_SEL(x) ((x) << 20) | 1555 | # define PACKET3_CP_DMA_DST_SEL(x) ((x) << 20) |
1556 | /* 0 - SRC_ADDR | 1556 | /* 0 - DST_ADDR |
1557 | * 1 - GDS | 1557 | * 1 - GDS |
1558 | */ | 1558 | */ |
1559 | # define PACKET3_CP_DMA_ENGINE(x) ((x) << 27) | 1559 | # define PACKET3_CP_DMA_ENGINE(x) ((x) << 27) |
@@ -1568,7 +1568,7 @@ | |||
1568 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) | 1568 | # define PACKET3_CP_DMA_CP_SYNC (1 << 31) |
1569 | /* COMMAND */ | 1569 | /* COMMAND */ |
1570 | # define PACKET3_CP_DMA_DIS_WC (1 << 21) | 1570 | # define PACKET3_CP_DMA_DIS_WC (1 << 21) |
1571 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 23) | 1571 | # define PACKET3_CP_DMA_CMD_SRC_SWAP(x) ((x) << 22) |
1572 | /* 0 - none | 1572 | /* 0 - none |
1573 | * 1 - 8 in 16 | 1573 | * 1 - 8 in 16 |
1574 | * 2 - 8 in 32 | 1574 | * 2 - 8 in 32 |
diff --git a/drivers/gpu/drm/radeon/sumo_dpm.c b/drivers/gpu/drm/radeon/sumo_dpm.c index 864761c0120e..96ea6db8bf57 100644 --- a/drivers/gpu/drm/radeon/sumo_dpm.c +++ b/drivers/gpu/drm/radeon/sumo_dpm.c | |||
@@ -1319,8 +1319,6 @@ int sumo_dpm_set_power_state(struct radeon_device *rdev) | |||
1319 | if (pi->enable_dpm) | 1319 | if (pi->enable_dpm) |
1320 | sumo_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1320 | sumo_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
1321 | 1321 | ||
1322 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1323 | |||
1324 | return 0; | 1322 | return 0; |
1325 | } | 1323 | } |
1326 | 1324 | ||
diff --git a/drivers/gpu/drm/radeon/trinity_dpm.c b/drivers/gpu/drm/radeon/trinity_dpm.c index b07b7b8f1aff..9364129ba292 100644 --- a/drivers/gpu/drm/radeon/trinity_dpm.c +++ b/drivers/gpu/drm/radeon/trinity_dpm.c | |||
@@ -1068,6 +1068,17 @@ static void trinity_update_requested_ps(struct radeon_device *rdev, | |||
1068 | pi->requested_rps.ps_priv = &pi->requested_ps; | 1068 | pi->requested_rps.ps_priv = &pi->requested_ps; |
1069 | } | 1069 | } |
1070 | 1070 | ||
1071 | void trinity_dpm_enable_bapm(struct radeon_device *rdev, bool enable) | ||
1072 | { | ||
1073 | struct trinity_power_info *pi = trinity_get_pi(rdev); | ||
1074 | |||
1075 | if (pi->enable_bapm) { | ||
1076 | trinity_acquire_mutex(rdev); | ||
1077 | trinity_dpm_bapm_enable(rdev, enable); | ||
1078 | trinity_release_mutex(rdev); | ||
1079 | } | ||
1080 | } | ||
1081 | |||
1071 | int trinity_dpm_enable(struct radeon_device *rdev) | 1082 | int trinity_dpm_enable(struct radeon_device *rdev) |
1072 | { | 1083 | { |
1073 | struct trinity_power_info *pi = trinity_get_pi(rdev); | 1084 | struct trinity_power_info *pi = trinity_get_pi(rdev); |
@@ -1091,6 +1102,7 @@ int trinity_dpm_enable(struct radeon_device *rdev) | |||
1091 | trinity_program_sclk_dpm(rdev); | 1102 | trinity_program_sclk_dpm(rdev); |
1092 | trinity_start_dpm(rdev); | 1103 | trinity_start_dpm(rdev); |
1093 | trinity_wait_for_dpm_enabled(rdev); | 1104 | trinity_wait_for_dpm_enabled(rdev); |
1105 | trinity_dpm_bapm_enable(rdev, false); | ||
1094 | trinity_release_mutex(rdev); | 1106 | trinity_release_mutex(rdev); |
1095 | 1107 | ||
1096 | if (rdev->irq.installed && | 1108 | if (rdev->irq.installed && |
@@ -1116,6 +1128,7 @@ void trinity_dpm_disable(struct radeon_device *rdev) | |||
1116 | trinity_release_mutex(rdev); | 1128 | trinity_release_mutex(rdev); |
1117 | return; | 1129 | return; |
1118 | } | 1130 | } |
1131 | trinity_dpm_bapm_enable(rdev, false); | ||
1119 | trinity_disable_clock_power_gating(rdev); | 1132 | trinity_disable_clock_power_gating(rdev); |
1120 | sumo_clear_vc(rdev); | 1133 | sumo_clear_vc(rdev); |
1121 | trinity_wait_for_level_0(rdev); | 1134 | trinity_wait_for_level_0(rdev); |
@@ -1212,6 +1225,8 @@ int trinity_dpm_set_power_state(struct radeon_device *rdev) | |||
1212 | 1225 | ||
1213 | trinity_acquire_mutex(rdev); | 1226 | trinity_acquire_mutex(rdev); |
1214 | if (pi->enable_dpm) { | 1227 | if (pi->enable_dpm) { |
1228 | if (pi->enable_bapm) | ||
1229 | trinity_dpm_bapm_enable(rdev, rdev->pm.dpm.ac_power); | ||
1215 | trinity_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); | 1230 | trinity_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); |
1216 | trinity_enable_power_level_0(rdev); | 1231 | trinity_enable_power_level_0(rdev); |
1217 | trinity_force_level_0(rdev); | 1232 | trinity_force_level_0(rdev); |
@@ -1221,7 +1236,6 @@ int trinity_dpm_set_power_state(struct radeon_device *rdev) | |||
1221 | trinity_force_level_0(rdev); | 1236 | trinity_force_level_0(rdev); |
1222 | trinity_unforce_levels(rdev); | 1237 | trinity_unforce_levels(rdev); |
1223 | trinity_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1238 | trinity_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
1224 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1225 | } | 1239 | } |
1226 | trinity_release_mutex(rdev); | 1240 | trinity_release_mutex(rdev); |
1227 | 1241 | ||
@@ -1854,6 +1868,7 @@ int trinity_dpm_init(struct radeon_device *rdev) | |||
1854 | for (i = 0; i < SUMO_MAX_HARDWARE_POWERLEVELS; i++) | 1868 | for (i = 0; i < SUMO_MAX_HARDWARE_POWERLEVELS; i++) |
1855 | pi->at[i] = TRINITY_AT_DFLT; | 1869 | pi->at[i] = TRINITY_AT_DFLT; |
1856 | 1870 | ||
1871 | pi->enable_bapm = false; | ||
1857 | pi->enable_nbps_policy = true; | 1872 | pi->enable_nbps_policy = true; |
1858 | pi->enable_sclk_ds = true; | 1873 | pi->enable_sclk_ds = true; |
1859 | pi->enable_gfx_power_gating = true; | 1874 | pi->enable_gfx_power_gating = true; |
diff --git a/drivers/gpu/drm/radeon/trinity_dpm.h b/drivers/gpu/drm/radeon/trinity_dpm.h index e82df071f8b3..c261657750ca 100644 --- a/drivers/gpu/drm/radeon/trinity_dpm.h +++ b/drivers/gpu/drm/radeon/trinity_dpm.h | |||
@@ -108,6 +108,7 @@ struct trinity_power_info { | |||
108 | bool enable_auto_thermal_throttling; | 108 | bool enable_auto_thermal_throttling; |
109 | bool enable_dpm; | 109 | bool enable_dpm; |
110 | bool enable_sclk_ds; | 110 | bool enable_sclk_ds; |
111 | bool enable_bapm; | ||
111 | bool uvd_dpm; | 112 | bool uvd_dpm; |
112 | struct radeon_ps current_rps; | 113 | struct radeon_ps current_rps; |
113 | struct trinity_ps current_ps; | 114 | struct trinity_ps current_ps; |
@@ -118,6 +119,7 @@ struct trinity_power_info { | |||
118 | #define TRINITY_AT_DFLT 30 | 119 | #define TRINITY_AT_DFLT 30 |
119 | 120 | ||
120 | /* trinity_smc.c */ | 121 | /* trinity_smc.c */ |
122 | int trinity_dpm_bapm_enable(struct radeon_device *rdev, bool enable); | ||
121 | int trinity_dpm_config(struct radeon_device *rdev, bool enable); | 123 | int trinity_dpm_config(struct radeon_device *rdev, bool enable); |
122 | int trinity_uvd_dpm_config(struct radeon_device *rdev); | 124 | int trinity_uvd_dpm_config(struct radeon_device *rdev); |
123 | int trinity_dpm_force_state(struct radeon_device *rdev, u32 n); | 125 | int trinity_dpm_force_state(struct radeon_device *rdev, u32 n); |
diff --git a/drivers/gpu/drm/radeon/trinity_smc.c b/drivers/gpu/drm/radeon/trinity_smc.c index a42d89f1830c..9672bcbc7312 100644 --- a/drivers/gpu/drm/radeon/trinity_smc.c +++ b/drivers/gpu/drm/radeon/trinity_smc.c | |||
@@ -56,6 +56,14 @@ static int trinity_notify_message_to_smu(struct radeon_device *rdev, u32 id) | |||
56 | return 0; | 56 | return 0; |
57 | } | 57 | } |
58 | 58 | ||
59 | int trinity_dpm_bapm_enable(struct radeon_device *rdev, bool enable) | ||
60 | { | ||
61 | if (enable) | ||
62 | return trinity_notify_message_to_smu(rdev, PPSMC_MSG_EnableBAPM); | ||
63 | else | ||
64 | return trinity_notify_message_to_smu(rdev, PPSMC_MSG_DisableBAPM); | ||
65 | } | ||
66 | |||
59 | int trinity_dpm_config(struct radeon_device *rdev, bool enable) | 67 | int trinity_dpm_config(struct radeon_device *rdev, bool enable) |
60 | { | 68 | { |
61 | if (enable) | 69 | if (enable) |
diff --git a/drivers/gpu/drm/radeon/uvd_v1_0.c b/drivers/gpu/drm/radeon/uvd_v1_0.c index 7266805d9786..3100fa9cb52f 100644 --- a/drivers/gpu/drm/radeon/uvd_v1_0.c +++ b/drivers/gpu/drm/radeon/uvd_v1_0.c | |||
@@ -212,8 +212,8 @@ int uvd_v1_0_start(struct radeon_device *rdev) | |||
212 | /* enable VCPU clock */ | 212 | /* enable VCPU clock */ |
213 | WREG32(UVD_VCPU_CNTL, 1 << 9); | 213 | WREG32(UVD_VCPU_CNTL, 1 << 9); |
214 | 214 | ||
215 | /* enable UMC */ | 215 | /* enable UMC and NC0 */ |
216 | WREG32_P(UVD_LMI_CTRL2, 0, ~(1 << 8)); | 216 | WREG32_P(UVD_LMI_CTRL2, 1 << 13, ~((1 << 8) | (1 << 13))); |
217 | 217 | ||
218 | /* boot up the VCPU */ | 218 | /* boot up the VCPU */ |
219 | WREG32(UVD_SOFT_RESET, 0); | 219 | WREG32(UVD_SOFT_RESET, 0); |
diff --git a/drivers/gpu/drm/ttm/ttm_object.c b/drivers/gpu/drm/ttm/ttm_object.c index 58a5f3261c0b..a868176c258a 100644 --- a/drivers/gpu/drm/ttm/ttm_object.c +++ b/drivers/gpu/drm/ttm/ttm_object.c | |||
@@ -218,7 +218,7 @@ struct ttm_base_object *ttm_base_object_lookup(struct ttm_object_file *tfile, | |||
218 | uint32_t key) | 218 | uint32_t key) |
219 | { | 219 | { |
220 | struct ttm_object_device *tdev = tfile->tdev; | 220 | struct ttm_object_device *tdev = tfile->tdev; |
221 | struct ttm_base_object *base; | 221 | struct ttm_base_object *uninitialized_var(base); |
222 | struct drm_hash_item *hash; | 222 | struct drm_hash_item *hash; |
223 | int ret; | 223 | int ret; |
224 | 224 | ||
diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c index 5e93a52d4f2c..210d50365162 100644 --- a/drivers/gpu/drm/ttm/ttm_tt.c +++ b/drivers/gpu/drm/ttm/ttm_tt.c | |||
@@ -170,7 +170,7 @@ void ttm_tt_destroy(struct ttm_tt *ttm) | |||
170 | ttm_tt_unbind(ttm); | 170 | ttm_tt_unbind(ttm); |
171 | } | 171 | } |
172 | 172 | ||
173 | if (likely(ttm->pages != NULL)) { | 173 | if (ttm->state == tt_unbound) { |
174 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); | 174 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); |
175 | } | 175 | } |
176 | 176 | ||
diff --git a/drivers/gpu/drm/udl/udl_gem.c b/drivers/gpu/drm/udl/udl_gem.c index 8dbe9d0ae9a7..8bf646183bac 100644 --- a/drivers/gpu/drm/udl/udl_gem.c +++ b/drivers/gpu/drm/udl/udl_gem.c | |||
@@ -97,7 +97,6 @@ int udl_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
97 | ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page); | 97 | ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page); |
98 | switch (ret) { | 98 | switch (ret) { |
99 | case -EAGAIN: | 99 | case -EAGAIN: |
100 | set_need_resched(); | ||
101 | case 0: | 100 | case 0: |
102 | case -ERESTARTSYS: | 101 | case -ERESTARTSYS: |
103 | return VM_FAULT_NOPAGE; | 102 | return VM_FAULT_NOPAGE; |