diff options
Diffstat (limited to 'drivers/gpu/drm')
87 files changed, 1369 insertions, 530 deletions
diff --git a/drivers/gpu/drm/ast/ast_drv.h b/drivers/gpu/drm/ast/ast_drv.h index 796dbb212a41..8492b68e873c 100644 --- a/drivers/gpu/drm/ast/ast_drv.h +++ b/drivers/gpu/drm/ast/ast_drv.h | |||
| @@ -177,7 +177,7 @@ uint8_t ast_get_index_reg_mask(struct ast_private *ast, | |||
| 177 | 177 | ||
| 178 | static inline void ast_open_key(struct ast_private *ast) | 178 | static inline void ast_open_key(struct ast_private *ast) |
| 179 | { | 179 | { |
| 180 | ast_set_index_reg_mask(ast, AST_IO_CRTC_PORT, 0xA1, 0xFF, 0x04); | 180 | ast_set_index_reg(ast, AST_IO_CRTC_PORT, 0x80, 0xA8); |
| 181 | } | 181 | } |
| 182 | 182 | ||
| 183 | #define AST_VIDMEM_SIZE_8M 0x00800000 | 183 | #define AST_VIDMEM_SIZE_8M 0x00800000 |
diff --git a/drivers/gpu/drm/drm_context.c b/drivers/gpu/drm/drm_context.c index b4fb86d89850..224ff965bcf7 100644 --- a/drivers/gpu/drm/drm_context.c +++ b/drivers/gpu/drm/drm_context.c | |||
| @@ -42,6 +42,10 @@ | |||
| 42 | 42 | ||
| 43 | #include <drm/drmP.h> | 43 | #include <drm/drmP.h> |
| 44 | 44 | ||
| 45 | /******************************************************************/ | ||
| 46 | /** \name Context bitmap support */ | ||
| 47 | /*@{*/ | ||
| 48 | |||
| 45 | /** | 49 | /** |
| 46 | * Free a handle from the context bitmap. | 50 | * Free a handle from the context bitmap. |
| 47 | * | 51 | * |
| @@ -52,48 +56,13 @@ | |||
| 52 | * in drm_device::ctx_idr, while holding the drm_device::struct_mutex | 56 | * in drm_device::ctx_idr, while holding the drm_device::struct_mutex |
| 53 | * lock. | 57 | * lock. |
| 54 | */ | 58 | */ |
| 55 | static void drm_ctxbitmap_free(struct drm_device * dev, int ctx_handle) | 59 | void drm_ctxbitmap_free(struct drm_device * dev, int ctx_handle) |
| 56 | { | 60 | { |
| 57 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 58 | return; | ||
| 59 | |||
| 60 | mutex_lock(&dev->struct_mutex); | 61 | mutex_lock(&dev->struct_mutex); |
| 61 | idr_remove(&dev->ctx_idr, ctx_handle); | 62 | idr_remove(&dev->ctx_idr, ctx_handle); |
| 62 | mutex_unlock(&dev->struct_mutex); | 63 | mutex_unlock(&dev->struct_mutex); |
| 63 | } | 64 | } |
| 64 | 65 | ||
| 65 | /******************************************************************/ | ||
| 66 | /** \name Context bitmap support */ | ||
| 67 | /*@{*/ | ||
| 68 | |||
| 69 | void drm_legacy_ctxbitmap_release(struct drm_device *dev, | ||
| 70 | struct drm_file *file_priv) | ||
| 71 | { | ||
| 72 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 73 | return; | ||
| 74 | |||
| 75 | mutex_lock(&dev->ctxlist_mutex); | ||
| 76 | if (!list_empty(&dev->ctxlist)) { | ||
| 77 | struct drm_ctx_list *pos, *n; | ||
| 78 | |||
| 79 | list_for_each_entry_safe(pos, n, &dev->ctxlist, head) { | ||
| 80 | if (pos->tag == file_priv && | ||
| 81 | pos->handle != DRM_KERNEL_CONTEXT) { | ||
| 82 | if (dev->driver->context_dtor) | ||
| 83 | dev->driver->context_dtor(dev, | ||
| 84 | pos->handle); | ||
| 85 | |||
| 86 | drm_ctxbitmap_free(dev, pos->handle); | ||
| 87 | |||
| 88 | list_del(&pos->head); | ||
| 89 | kfree(pos); | ||
| 90 | --dev->ctx_count; | ||
| 91 | } | ||
| 92 | } | ||
| 93 | } | ||
| 94 | mutex_unlock(&dev->ctxlist_mutex); | ||
| 95 | } | ||
| 96 | |||
| 97 | /** | 66 | /** |
| 98 | * Context bitmap allocation. | 67 | * Context bitmap allocation. |
| 99 | * | 68 | * |
| @@ -121,12 +90,10 @@ static int drm_ctxbitmap_next(struct drm_device * dev) | |||
| 121 | * | 90 | * |
| 122 | * Initialise the drm_device::ctx_idr | 91 | * Initialise the drm_device::ctx_idr |
| 123 | */ | 92 | */ |
| 124 | void drm_legacy_ctxbitmap_init(struct drm_device * dev) | 93 | int drm_ctxbitmap_init(struct drm_device * dev) |
| 125 | { | 94 | { |
| 126 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 127 | return; | ||
| 128 | |||
| 129 | idr_init(&dev->ctx_idr); | 95 | idr_init(&dev->ctx_idr); |
| 96 | return 0; | ||
| 130 | } | 97 | } |
| 131 | 98 | ||
| 132 | /** | 99 | /** |
| @@ -137,7 +104,7 @@ void drm_legacy_ctxbitmap_init(struct drm_device * dev) | |||
| 137 | * Free all idr members using drm_ctx_sarea_free helper function | 104 | * Free all idr members using drm_ctx_sarea_free helper function |
| 138 | * while holding the drm_device::struct_mutex lock. | 105 | * while holding the drm_device::struct_mutex lock. |
| 139 | */ | 106 | */ |
| 140 | void drm_legacy_ctxbitmap_cleanup(struct drm_device * dev) | 107 | void drm_ctxbitmap_cleanup(struct drm_device * dev) |
| 141 | { | 108 | { |
| 142 | mutex_lock(&dev->struct_mutex); | 109 | mutex_lock(&dev->struct_mutex); |
| 143 | idr_destroy(&dev->ctx_idr); | 110 | idr_destroy(&dev->ctx_idr); |
| @@ -169,9 +136,6 @@ int drm_getsareactx(struct drm_device *dev, void *data, | |||
| 169 | struct drm_local_map *map; | 136 | struct drm_local_map *map; |
| 170 | struct drm_map_list *_entry; | 137 | struct drm_map_list *_entry; |
| 171 | 138 | ||
| 172 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 173 | return -EINVAL; | ||
| 174 | |||
| 175 | mutex_lock(&dev->struct_mutex); | 139 | mutex_lock(&dev->struct_mutex); |
| 176 | 140 | ||
| 177 | map = idr_find(&dev->ctx_idr, request->ctx_id); | 141 | map = idr_find(&dev->ctx_idr, request->ctx_id); |
| @@ -216,9 +180,6 @@ int drm_setsareactx(struct drm_device *dev, void *data, | |||
| 216 | struct drm_local_map *map = NULL; | 180 | struct drm_local_map *map = NULL; |
| 217 | struct drm_map_list *r_list = NULL; | 181 | struct drm_map_list *r_list = NULL; |
| 218 | 182 | ||
| 219 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 220 | return -EINVAL; | ||
| 221 | |||
| 222 | mutex_lock(&dev->struct_mutex); | 183 | mutex_lock(&dev->struct_mutex); |
| 223 | list_for_each_entry(r_list, &dev->maplist, head) { | 184 | list_for_each_entry(r_list, &dev->maplist, head) { |
| 224 | if (r_list->map | 185 | if (r_list->map |
| @@ -319,9 +280,6 @@ int drm_resctx(struct drm_device *dev, void *data, | |||
| 319 | struct drm_ctx ctx; | 280 | struct drm_ctx ctx; |
| 320 | int i; | 281 | int i; |
| 321 | 282 | ||
| 322 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 323 | return -EINVAL; | ||
| 324 | |||
| 325 | if (res->count >= DRM_RESERVED_CONTEXTS) { | 283 | if (res->count >= DRM_RESERVED_CONTEXTS) { |
| 326 | memset(&ctx, 0, sizeof(ctx)); | 284 | memset(&ctx, 0, sizeof(ctx)); |
| 327 | for (i = 0; i < DRM_RESERVED_CONTEXTS; i++) { | 285 | for (i = 0; i < DRM_RESERVED_CONTEXTS; i++) { |
| @@ -352,9 +310,6 @@ int drm_addctx(struct drm_device *dev, void *data, | |||
| 352 | struct drm_ctx_list *ctx_entry; | 310 | struct drm_ctx_list *ctx_entry; |
| 353 | struct drm_ctx *ctx = data; | 311 | struct drm_ctx *ctx = data; |
| 354 | 312 | ||
| 355 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 356 | return -EINVAL; | ||
| 357 | |||
| 358 | ctx->handle = drm_ctxbitmap_next(dev); | 313 | ctx->handle = drm_ctxbitmap_next(dev); |
| 359 | if (ctx->handle == DRM_KERNEL_CONTEXT) { | 314 | if (ctx->handle == DRM_KERNEL_CONTEXT) { |
| 360 | /* Skip kernel's context and get a new one. */ | 315 | /* Skip kernel's context and get a new one. */ |
| @@ -398,9 +353,6 @@ int drm_getctx(struct drm_device *dev, void *data, struct drm_file *file_priv) | |||
| 398 | { | 353 | { |
| 399 | struct drm_ctx *ctx = data; | 354 | struct drm_ctx *ctx = data; |
| 400 | 355 | ||
| 401 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 402 | return -EINVAL; | ||
| 403 | |||
| 404 | /* This is 0, because we don't handle any context flags */ | 356 | /* This is 0, because we don't handle any context flags */ |
| 405 | ctx->flags = 0; | 357 | ctx->flags = 0; |
| 406 | 358 | ||
| @@ -423,9 +375,6 @@ int drm_switchctx(struct drm_device *dev, void *data, | |||
| 423 | { | 375 | { |
| 424 | struct drm_ctx *ctx = data; | 376 | struct drm_ctx *ctx = data; |
| 425 | 377 | ||
| 426 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 427 | return -EINVAL; | ||
| 428 | |||
| 429 | DRM_DEBUG("%d\n", ctx->handle); | 378 | DRM_DEBUG("%d\n", ctx->handle); |
| 430 | return drm_context_switch(dev, dev->last_context, ctx->handle); | 379 | return drm_context_switch(dev, dev->last_context, ctx->handle); |
| 431 | } | 380 | } |
| @@ -446,9 +395,6 @@ int drm_newctx(struct drm_device *dev, void *data, | |||
| 446 | { | 395 | { |
| 447 | struct drm_ctx *ctx = data; | 396 | struct drm_ctx *ctx = data; |
| 448 | 397 | ||
| 449 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 450 | return -EINVAL; | ||
| 451 | |||
| 452 | DRM_DEBUG("%d\n", ctx->handle); | 398 | DRM_DEBUG("%d\n", ctx->handle); |
| 453 | drm_context_switch_complete(dev, file_priv, ctx->handle); | 399 | drm_context_switch_complete(dev, file_priv, ctx->handle); |
| 454 | 400 | ||
| @@ -471,9 +417,6 @@ int drm_rmctx(struct drm_device *dev, void *data, | |||
| 471 | { | 417 | { |
| 472 | struct drm_ctx *ctx = data; | 418 | struct drm_ctx *ctx = data; |
| 473 | 419 | ||
| 474 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | ||
| 475 | return -EINVAL; | ||
| 476 | |||
| 477 | DRM_DEBUG("%d\n", ctx->handle); | 420 | DRM_DEBUG("%d\n", ctx->handle); |
| 478 | if (ctx->handle != DRM_KERNEL_CONTEXT) { | 421 | if (ctx->handle != DRM_KERNEL_CONTEXT) { |
| 479 | if (dev->driver->context_dtor) | 422 | if (dev->driver->context_dtor) |
diff --git a/drivers/gpu/drm/drm_fb_helper.c b/drivers/gpu/drm/drm_fb_helper.c index 3d13ca6e257f..f6f6cc7fc133 100644 --- a/drivers/gpu/drm/drm_fb_helper.c +++ b/drivers/gpu/drm/drm_fb_helper.c | |||
| @@ -416,6 +416,14 @@ static void drm_fb_helper_dpms(struct fb_info *info, int dpms_mode) | |||
| 416 | return; | 416 | return; |
| 417 | 417 | ||
| 418 | /* | 418 | /* |
| 419 | * fbdev->blank can be called from irq context in case of a panic. | ||
| 420 | * Since we already have our own special panic handler which will | ||
| 421 | * restore the fbdev console mode completely, just bail out early. | ||
| 422 | */ | ||
| 423 | if (oops_in_progress) | ||
| 424 | return; | ||
| 425 | |||
| 426 | /* | ||
| 419 | * For each CRTC in this fb, turn the connectors on/off. | 427 | * For each CRTC in this fb, turn the connectors on/off. |
| 420 | */ | 428 | */ |
| 421 | drm_modeset_lock_all(dev); | 429 | drm_modeset_lock_all(dev); |
diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c index 4be8e09a32ef..3f84277d7036 100644 --- a/drivers/gpu/drm/drm_fops.c +++ b/drivers/gpu/drm/drm_fops.c | |||
| @@ -439,7 +439,26 @@ int drm_release(struct inode *inode, struct file *filp) | |||
| 439 | if (dev->driver->driver_features & DRIVER_GEM) | 439 | if (dev->driver->driver_features & DRIVER_GEM) |
| 440 | drm_gem_release(dev, file_priv); | 440 | drm_gem_release(dev, file_priv); |
| 441 | 441 | ||
| 442 | drm_legacy_ctxbitmap_release(dev, file_priv); | 442 | mutex_lock(&dev->ctxlist_mutex); |
| 443 | if (!list_empty(&dev->ctxlist)) { | ||
| 444 | struct drm_ctx_list *pos, *n; | ||
| 445 | |||
| 446 | list_for_each_entry_safe(pos, n, &dev->ctxlist, head) { | ||
| 447 | if (pos->tag == file_priv && | ||
| 448 | pos->handle != DRM_KERNEL_CONTEXT) { | ||
| 449 | if (dev->driver->context_dtor) | ||
| 450 | dev->driver->context_dtor(dev, | ||
| 451 | pos->handle); | ||
| 452 | |||
| 453 | drm_ctxbitmap_free(dev, pos->handle); | ||
| 454 | |||
| 455 | list_del(&pos->head); | ||
| 456 | kfree(pos); | ||
| 457 | --dev->ctx_count; | ||
| 458 | } | ||
| 459 | } | ||
| 460 | } | ||
| 461 | mutex_unlock(&dev->ctxlist_mutex); | ||
| 443 | 462 | ||
| 444 | mutex_lock(&dev->struct_mutex); | 463 | mutex_lock(&dev->struct_mutex); |
| 445 | 464 | ||
diff --git a/drivers/gpu/drm/drm_stub.c b/drivers/gpu/drm/drm_stub.c index e7eb0276f7f1..39d864576be4 100644 --- a/drivers/gpu/drm/drm_stub.c +++ b/drivers/gpu/drm/drm_stub.c | |||
| @@ -292,7 +292,13 @@ int drm_fill_in_dev(struct drm_device *dev, | |||
| 292 | goto error_out_unreg; | 292 | goto error_out_unreg; |
| 293 | } | 293 | } |
| 294 | 294 | ||
| 295 | drm_legacy_ctxbitmap_init(dev); | 295 | |
| 296 | |||
| 297 | retcode = drm_ctxbitmap_init(dev); | ||
| 298 | if (retcode) { | ||
| 299 | DRM_ERROR("Cannot allocate memory for context bitmap.\n"); | ||
| 300 | goto error_out_unreg; | ||
| 301 | } | ||
| 296 | 302 | ||
| 297 | if (driver->driver_features & DRIVER_GEM) { | 303 | if (driver->driver_features & DRIVER_GEM) { |
| 298 | retcode = drm_gem_init(dev); | 304 | retcode = drm_gem_init(dev); |
| @@ -446,7 +452,7 @@ void drm_put_dev(struct drm_device *dev) | |||
| 446 | drm_rmmap(dev, r_list->map); | 452 | drm_rmmap(dev, r_list->map); |
| 447 | drm_ht_remove(&dev->map_hash); | 453 | drm_ht_remove(&dev->map_hash); |
| 448 | 454 | ||
| 449 | drm_legacy_ctxbitmap_cleanup(dev); | 455 | drm_ctxbitmap_cleanup(dev); |
| 450 | 456 | ||
| 451 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | 457 | if (drm_core_check_feature(dev, DRIVER_MODESET)) |
| 452 | drm_put_minor(&dev->control); | 458 | drm_put_minor(&dev->control); |
diff --git a/drivers/gpu/drm/exynos/Kconfig b/drivers/gpu/drm/exynos/Kconfig index 4752f223e5b2..45b6ef595965 100644 --- a/drivers/gpu/drm/exynos/Kconfig +++ b/drivers/gpu/drm/exynos/Kconfig | |||
| @@ -56,7 +56,7 @@ config DRM_EXYNOS_IPP | |||
| 56 | 56 | ||
| 57 | config DRM_EXYNOS_FIMC | 57 | config DRM_EXYNOS_FIMC |
| 58 | bool "Exynos DRM FIMC" | 58 | bool "Exynos DRM FIMC" |
| 59 | depends on DRM_EXYNOS_IPP && MFD_SYSCON && OF | 59 | depends on DRM_EXYNOS_IPP && MFD_SYSCON |
| 60 | help | 60 | help |
| 61 | Choose this option if you want to use Exynos FIMC for DRM. | 61 | Choose this option if you want to use Exynos FIMC for DRM. |
| 62 | 62 | ||
diff --git a/drivers/gpu/drm/exynos/exynos_drm_buf.c b/drivers/gpu/drm/exynos/exynos_drm_buf.c index 3445a0f3a6b2..9c8088462c26 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_buf.c +++ b/drivers/gpu/drm/exynos/exynos_drm_buf.c | |||
| @@ -63,7 +63,8 @@ static int lowlevel_buffer_allocate(struct drm_device *dev, | |||
| 63 | return -ENOMEM; | 63 | return -ENOMEM; |
| 64 | } | 64 | } |
| 65 | 65 | ||
| 66 | buf->kvaddr = dma_alloc_attrs(dev->dev, buf->size, | 66 | buf->kvaddr = (void __iomem *)dma_alloc_attrs(dev->dev, |
| 67 | buf->size, | ||
| 67 | &buf->dma_addr, GFP_KERNEL, | 68 | &buf->dma_addr, GFP_KERNEL, |
| 68 | &buf->dma_attrs); | 69 | &buf->dma_attrs); |
| 69 | if (!buf->kvaddr) { | 70 | if (!buf->kvaddr) { |
| @@ -90,9 +91,9 @@ static int lowlevel_buffer_allocate(struct drm_device *dev, | |||
| 90 | } | 91 | } |
| 91 | 92 | ||
| 92 | buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages); | 93 | buf->sgt = drm_prime_pages_to_sg(buf->pages, nr_pages); |
| 93 | if (!buf->sgt) { | 94 | if (IS_ERR(buf->sgt)) { |
| 94 | DRM_ERROR("failed to get sg table.\n"); | 95 | DRM_ERROR("failed to get sg table.\n"); |
| 95 | ret = -ENOMEM; | 96 | ret = PTR_ERR(buf->sgt); |
| 96 | goto err_free_attrs; | 97 | goto err_free_attrs; |
| 97 | } | 98 | } |
| 98 | 99 | ||
diff --git a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c index 78e868bcf1ec..e7c2f2d07f19 100644 --- a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c +++ b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c | |||
| @@ -99,12 +99,13 @@ static int exynos_drm_fbdev_update(struct drm_fb_helper *helper, | |||
| 99 | if (is_drm_iommu_supported(dev)) { | 99 | if (is_drm_iommu_supported(dev)) { |
| 100 | unsigned int nr_pages = buffer->size >> PAGE_SHIFT; | 100 | unsigned int nr_pages = buffer->size >> PAGE_SHIFT; |
| 101 | 101 | ||
| 102 | buffer->kvaddr = vmap(buffer->pages, nr_pages, VM_MAP, | 102 | buffer->kvaddr = (void __iomem *) vmap(buffer->pages, |
| 103 | nr_pages, VM_MAP, | ||
| 103 | pgprot_writecombine(PAGE_KERNEL)); | 104 | pgprot_writecombine(PAGE_KERNEL)); |
| 104 | } else { | 105 | } else { |
| 105 | phys_addr_t dma_addr = buffer->dma_addr; | 106 | phys_addr_t dma_addr = buffer->dma_addr; |
| 106 | if (dma_addr) | 107 | if (dma_addr) |
| 107 | buffer->kvaddr = phys_to_virt(dma_addr); | 108 | buffer->kvaddr = (void __iomem *)phys_to_virt(dma_addr); |
| 108 | else | 109 | else |
| 109 | buffer->kvaddr = (void __iomem *)NULL; | 110 | buffer->kvaddr = (void __iomem *)NULL; |
| 110 | } | 111 | } |
diff --git a/drivers/gpu/drm/i2c/tda998x_drv.c b/drivers/gpu/drm/i2c/tda998x_drv.c index b1f8fc69023f..60e84043aa34 100644 --- a/drivers/gpu/drm/i2c/tda998x_drv.c +++ b/drivers/gpu/drm/i2c/tda998x_drv.c | |||
| @@ -707,8 +707,7 @@ tda998x_encoder_dpms(struct drm_encoder *encoder, int mode) | |||
| 707 | reg_write(encoder, REG_VIP_CNTRL_2, priv->vip_cntrl_2); | 707 | reg_write(encoder, REG_VIP_CNTRL_2, priv->vip_cntrl_2); |
| 708 | break; | 708 | break; |
| 709 | case DRM_MODE_DPMS_OFF: | 709 | case DRM_MODE_DPMS_OFF: |
| 710 | /* disable audio and video ports */ | 710 | /* disable video ports */ |
| 711 | reg_write(encoder, REG_ENA_AP, 0x00); | ||
| 712 | reg_write(encoder, REG_ENA_VP_0, 0x00); | 711 | reg_write(encoder, REG_ENA_VP_0, 0x00); |
| 713 | reg_write(encoder, REG_ENA_VP_1, 0x00); | 712 | reg_write(encoder, REG_ENA_VP_1, 0x00); |
| 714 | reg_write(encoder, REG_ENA_VP_2, 0x00); | 713 | reg_write(encoder, REG_ENA_VP_2, 0x00); |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 8507c6d1e642..cdfb9da0e4ce 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
| @@ -1392,14 +1392,11 @@ out: | |||
| 1392 | if (i915_terminally_wedged(&dev_priv->gpu_error)) | 1392 | if (i915_terminally_wedged(&dev_priv->gpu_error)) |
| 1393 | return VM_FAULT_SIGBUS; | 1393 | return VM_FAULT_SIGBUS; |
| 1394 | case -EAGAIN: | 1394 | case -EAGAIN: |
| 1395 | /* Give the error handler a chance to run and move the | 1395 | /* |
| 1396 | * objects off the GPU active list. Next time we service the | 1396 | * EAGAIN means the gpu is hung and we'll wait for the error |
| 1397 | * fault, we should be able to transition the page into the | 1397 | * handler to reset everything when re-faulting in |
| 1398 | * GTT without touching the GPU (and so avoid further | 1398 | * i915_mutex_lock_interruptible. |
| 1399 | * EIO/EGAIN). If the GPU is wedged, then there is no issue | ||
| 1400 | * with coherency, just lost writes. | ||
| 1401 | */ | 1399 | */ |
| 1402 | set_need_resched(); | ||
| 1403 | case 0: | 1400 | case 0: |
| 1404 | case -ERESTARTSYS: | 1401 | case -ERESTARTSYS: |
| 1405 | case -EINTR: | 1402 | case -EINTR: |
| @@ -4803,10 +4800,10 @@ i915_gem_inactive_count(struct shrinker *shrinker, struct shrink_control *sc) | |||
| 4803 | 4800 | ||
| 4804 | if (!mutex_trylock(&dev->struct_mutex)) { | 4801 | if (!mutex_trylock(&dev->struct_mutex)) { |
| 4805 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) | 4802 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) |
| 4806 | return SHRINK_STOP; | 4803 | return 0; |
| 4807 | 4804 | ||
| 4808 | if (dev_priv->mm.shrinker_no_lock_stealing) | 4805 | if (dev_priv->mm.shrinker_no_lock_stealing) |
| 4809 | return SHRINK_STOP; | 4806 | return 0; |
| 4810 | 4807 | ||
| 4811 | unlock = false; | 4808 | unlock = false; |
| 4812 | } | 4809 | } |
| @@ -4904,10 +4901,10 @@ i915_gem_inactive_scan(struct shrinker *shrinker, struct shrink_control *sc) | |||
| 4904 | 4901 | ||
| 4905 | if (!mutex_trylock(&dev->struct_mutex)) { | 4902 | if (!mutex_trylock(&dev->struct_mutex)) { |
| 4906 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) | 4903 | if (!mutex_is_locked_by(&dev->struct_mutex, current)) |
| 4907 | return 0; | 4904 | return SHRINK_STOP; |
| 4908 | 4905 | ||
| 4909 | if (dev_priv->mm.shrinker_no_lock_stealing) | 4906 | if (dev_priv->mm.shrinker_no_lock_stealing) |
| 4910 | return 0; | 4907 | return SHRINK_STOP; |
| 4911 | 4908 | ||
| 4912 | unlock = false; | 4909 | unlock = false; |
| 4913 | } | 4910 | } |
diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c index aba9d7498996..dae364f0028c 100644 --- a/drivers/gpu/drm/i915/i915_gpu_error.c +++ b/drivers/gpu/drm/i915/i915_gpu_error.c | |||
| @@ -143,8 +143,10 @@ static void i915_error_vprintf(struct drm_i915_error_state_buf *e, | |||
| 143 | 143 | ||
| 144 | /* Seek the first printf which is hits start position */ | 144 | /* Seek the first printf which is hits start position */ |
| 145 | if (e->pos < e->start) { | 145 | if (e->pos < e->start) { |
| 146 | len = vsnprintf(NULL, 0, f, args); | 146 | va_list tmp; |
| 147 | if (!__i915_error_seek(e, len)) | 147 | |
| 148 | va_copy(tmp, args); | ||
| 149 | if (!__i915_error_seek(e, vsnprintf(NULL, 0, f, tmp))) | ||
| 148 | return; | 150 | return; |
| 149 | } | 151 | } |
| 150 | 152 | ||
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index 83cce0cdb769..4b91228fd9bd 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
| @@ -1469,6 +1469,34 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg) | |||
| 1469 | return ret; | 1469 | return ret; |
| 1470 | } | 1470 | } |
| 1471 | 1471 | ||
| 1472 | static void i915_error_wake_up(struct drm_i915_private *dev_priv, | ||
| 1473 | bool reset_completed) | ||
| 1474 | { | ||
| 1475 | struct intel_ring_buffer *ring; | ||
| 1476 | int i; | ||
| 1477 | |||
| 1478 | /* | ||
| 1479 | * Notify all waiters for GPU completion events that reset state has | ||
| 1480 | * been changed, and that they need to restart their wait after | ||
| 1481 | * checking for potential errors (and bail out to drop locks if there is | ||
| 1482 | * a gpu reset pending so that i915_error_work_func can acquire them). | ||
| 1483 | */ | ||
| 1484 | |||
| 1485 | /* Wake up __wait_seqno, potentially holding dev->struct_mutex. */ | ||
| 1486 | for_each_ring(ring, dev_priv, i) | ||
| 1487 | wake_up_all(&ring->irq_queue); | ||
| 1488 | |||
| 1489 | /* Wake up intel_crtc_wait_for_pending_flips, holding crtc->mutex. */ | ||
| 1490 | wake_up_all(&dev_priv->pending_flip_queue); | ||
| 1491 | |||
| 1492 | /* | ||
| 1493 | * Signal tasks blocked in i915_gem_wait_for_error that the pending | ||
| 1494 | * reset state is cleared. | ||
| 1495 | */ | ||
| 1496 | if (reset_completed) | ||
| 1497 | wake_up_all(&dev_priv->gpu_error.reset_queue); | ||
| 1498 | } | ||
| 1499 | |||
| 1472 | /** | 1500 | /** |
| 1473 | * i915_error_work_func - do process context error handling work | 1501 | * i915_error_work_func - do process context error handling work |
| 1474 | * @work: work struct | 1502 | * @work: work struct |
| @@ -1483,11 +1511,10 @@ static void i915_error_work_func(struct work_struct *work) | |||
| 1483 | drm_i915_private_t *dev_priv = container_of(error, drm_i915_private_t, | 1511 | drm_i915_private_t *dev_priv = container_of(error, drm_i915_private_t, |
| 1484 | gpu_error); | 1512 | gpu_error); |
| 1485 | struct drm_device *dev = dev_priv->dev; | 1513 | struct drm_device *dev = dev_priv->dev; |
| 1486 | struct intel_ring_buffer *ring; | ||
| 1487 | char *error_event[] = { I915_ERROR_UEVENT "=1", NULL }; | 1514 | char *error_event[] = { I915_ERROR_UEVENT "=1", NULL }; |
| 1488 | char *reset_event[] = { I915_RESET_UEVENT "=1", NULL }; | 1515 | char *reset_event[] = { I915_RESET_UEVENT "=1", NULL }; |
| 1489 | char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL }; | 1516 | char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL }; |
| 1490 | int i, ret; | 1517 | int ret; |
| 1491 | 1518 | ||
| 1492 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); | 1519 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); |
| 1493 | 1520 | ||
| @@ -1506,8 +1533,16 @@ static void i915_error_work_func(struct work_struct *work) | |||
| 1506 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, | 1533 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, |
| 1507 | reset_event); | 1534 | reset_event); |
| 1508 | 1535 | ||
| 1536 | /* | ||
| 1537 | * All state reset _must_ be completed before we update the | ||
| 1538 | * reset counter, for otherwise waiters might miss the reset | ||
| 1539 | * pending state and not properly drop locks, resulting in | ||
| 1540 | * deadlocks with the reset work. | ||
| 1541 | */ | ||
| 1509 | ret = i915_reset(dev); | 1542 | ret = i915_reset(dev); |
| 1510 | 1543 | ||
| 1544 | intel_display_handle_reset(dev); | ||
| 1545 | |||
| 1511 | if (ret == 0) { | 1546 | if (ret == 0) { |
| 1512 | /* | 1547 | /* |
| 1513 | * After all the gem state is reset, increment the reset | 1548 | * After all the gem state is reset, increment the reset |
| @@ -1528,12 +1563,11 @@ static void i915_error_work_func(struct work_struct *work) | |||
| 1528 | atomic_set(&error->reset_counter, I915_WEDGED); | 1563 | atomic_set(&error->reset_counter, I915_WEDGED); |
| 1529 | } | 1564 | } |
| 1530 | 1565 | ||
| 1531 | for_each_ring(ring, dev_priv, i) | 1566 | /* |
| 1532 | wake_up_all(&ring->irq_queue); | 1567 | * Note: The wake_up also serves as a memory barrier so that |
| 1533 | 1568 | * waiters see the update value of the reset counter atomic_t. | |
| 1534 | intel_display_handle_reset(dev); | 1569 | */ |
| 1535 | 1570 | i915_error_wake_up(dev_priv, true); | |
| 1536 | wake_up_all(&dev_priv->gpu_error.reset_queue); | ||
| 1537 | } | 1571 | } |
| 1538 | } | 1572 | } |
| 1539 | 1573 | ||
| @@ -1642,8 +1676,6 @@ static void i915_report_and_clear_eir(struct drm_device *dev) | |||
| 1642 | void i915_handle_error(struct drm_device *dev, bool wedged) | 1676 | void i915_handle_error(struct drm_device *dev, bool wedged) |
| 1643 | { | 1677 | { |
| 1644 | struct drm_i915_private *dev_priv = dev->dev_private; | 1678 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1645 | struct intel_ring_buffer *ring; | ||
| 1646 | int i; | ||
| 1647 | 1679 | ||
| 1648 | i915_capture_error_state(dev); | 1680 | i915_capture_error_state(dev); |
| 1649 | i915_report_and_clear_eir(dev); | 1681 | i915_report_and_clear_eir(dev); |
| @@ -1653,11 +1685,19 @@ void i915_handle_error(struct drm_device *dev, bool wedged) | |||
| 1653 | &dev_priv->gpu_error.reset_counter); | 1685 | &dev_priv->gpu_error.reset_counter); |
| 1654 | 1686 | ||
| 1655 | /* | 1687 | /* |
| 1656 | * Wakeup waiting processes so that the reset work item | 1688 | * Wakeup waiting processes so that the reset work function |
| 1657 | * doesn't deadlock trying to grab various locks. | 1689 | * i915_error_work_func doesn't deadlock trying to grab various |
| 1690 | * locks. By bumping the reset counter first, the woken | ||
| 1691 | * processes will see a reset in progress and back off, | ||
| 1692 | * releasing their locks and then wait for the reset completion. | ||
| 1693 | * We must do this for _all_ gpu waiters that might hold locks | ||
| 1694 | * that the reset work needs to acquire. | ||
| 1695 | * | ||
| 1696 | * Note: The wake_up serves as the required memory barrier to | ||
| 1697 | * ensure that the waiters see the updated value of the reset | ||
| 1698 | * counter atomic_t. | ||
| 1658 | */ | 1699 | */ |
| 1659 | for_each_ring(ring, dev_priv, i) | 1700 | i915_error_wake_up(dev_priv, false); |
| 1660 | wake_up_all(&ring->irq_queue); | ||
| 1661 | } | 1701 | } |
| 1662 | 1702 | ||
| 1663 | /* | 1703 | /* |
diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c index 63aca49d11a8..63de2701b974 100644 --- a/drivers/gpu/drm/i915/intel_ddi.c +++ b/drivers/gpu/drm/i915/intel_ddi.c | |||
| @@ -778,7 +778,7 @@ void intel_ddi_enable_transcoder_func(struct drm_crtc *crtc) | |||
| 778 | /* Can only use the always-on power well for eDP when | 778 | /* Can only use the always-on power well for eDP when |
| 779 | * not using the panel fitter, and when not using motion | 779 | * not using the panel fitter, and when not using motion |
| 780 | * blur mitigation (which we don't support). */ | 780 | * blur mitigation (which we don't support). */ |
| 781 | if (intel_crtc->config.pch_pfit.size) | 781 | if (intel_crtc->config.pch_pfit.enabled) |
| 782 | temp |= TRANS_DDI_EDP_INPUT_A_ONOFF; | 782 | temp |= TRANS_DDI_EDP_INPUT_A_ONOFF; |
| 783 | else | 783 | else |
| 784 | temp |= TRANS_DDI_EDP_INPUT_A_ON; | 784 | temp |= TRANS_DDI_EDP_INPUT_A_ON; |
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index 2489d0b4c7d2..e5822e79f912 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
| @@ -2249,7 +2249,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 2249 | I915_WRITE(PIPESRC(intel_crtc->pipe), | 2249 | I915_WRITE(PIPESRC(intel_crtc->pipe), |
| 2250 | ((crtc->mode.hdisplay - 1) << 16) | | 2250 | ((crtc->mode.hdisplay - 1) << 16) | |
| 2251 | (crtc->mode.vdisplay - 1)); | 2251 | (crtc->mode.vdisplay - 1)); |
| 2252 | if (!intel_crtc->config.pch_pfit.size && | 2252 | if (!intel_crtc->config.pch_pfit.enabled && |
| 2253 | (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS) || | 2253 | (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS) || |
| 2254 | intel_pipe_has_type(crtc, INTEL_OUTPUT_EDP))) { | 2254 | intel_pipe_has_type(crtc, INTEL_OUTPUT_EDP))) { |
| 2255 | I915_WRITE(PF_CTL(intel_crtc->pipe), 0); | 2255 | I915_WRITE(PF_CTL(intel_crtc->pipe), 0); |
| @@ -3203,7 +3203,7 @@ static void ironlake_pfit_enable(struct intel_crtc *crtc) | |||
| 3203 | struct drm_i915_private *dev_priv = dev->dev_private; | 3203 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 3204 | int pipe = crtc->pipe; | 3204 | int pipe = crtc->pipe; |
| 3205 | 3205 | ||
| 3206 | if (crtc->config.pch_pfit.size) { | 3206 | if (crtc->config.pch_pfit.enabled) { |
| 3207 | /* Force use of hard-coded filter coefficients | 3207 | /* Force use of hard-coded filter coefficients |
| 3208 | * as some pre-programmed values are broken, | 3208 | * as some pre-programmed values are broken, |
| 3209 | * e.g. x201. | 3209 | * e.g. x201. |
| @@ -3428,7 +3428,7 @@ static void ironlake_pfit_disable(struct intel_crtc *crtc) | |||
| 3428 | 3428 | ||
| 3429 | /* To avoid upsetting the power well on haswell only disable the pfit if | 3429 | /* To avoid upsetting the power well on haswell only disable the pfit if |
| 3430 | * it's in use. The hw state code will make sure we get this right. */ | 3430 | * it's in use. The hw state code will make sure we get this right. */ |
| 3431 | if (crtc->config.pch_pfit.size) { | 3431 | if (crtc->config.pch_pfit.enabled) { |
| 3432 | I915_WRITE(PF_CTL(pipe), 0); | 3432 | I915_WRITE(PF_CTL(pipe), 0); |
| 3433 | I915_WRITE(PF_WIN_POS(pipe), 0); | 3433 | I915_WRITE(PF_WIN_POS(pipe), 0); |
| 3434 | I915_WRITE(PF_WIN_SZ(pipe), 0); | 3434 | I915_WRITE(PF_WIN_SZ(pipe), 0); |
| @@ -4775,6 +4775,10 @@ static void i9xx_set_pipeconf(struct intel_crtc *intel_crtc) | |||
| 4775 | 4775 | ||
| 4776 | pipeconf = 0; | 4776 | pipeconf = 0; |
| 4777 | 4777 | ||
| 4778 | if (dev_priv->quirks & QUIRK_PIPEA_FORCE && | ||
| 4779 | I915_READ(PIPECONF(intel_crtc->pipe)) & PIPECONF_ENABLE) | ||
| 4780 | pipeconf |= PIPECONF_ENABLE; | ||
| 4781 | |||
| 4778 | if (intel_crtc->pipe == 0 && INTEL_INFO(dev)->gen < 4) { | 4782 | if (intel_crtc->pipe == 0 && INTEL_INFO(dev)->gen < 4) { |
| 4779 | /* Enable pixel doubling when the dot clock is > 90% of the (display) | 4783 | /* Enable pixel doubling when the dot clock is > 90% of the (display) |
| 4780 | * core speed. | 4784 | * core speed. |
| @@ -4877,9 +4881,6 @@ static int i9xx_crtc_mode_set(struct drm_crtc *crtc, | |||
| 4877 | return -EINVAL; | 4881 | return -EINVAL; |
| 4878 | } | 4882 | } |
| 4879 | 4883 | ||
| 4880 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
| 4881 | intel_crtc_update_cursor(crtc, true); | ||
| 4882 | |||
| 4883 | if (is_lvds && dev_priv->lvds_downclock_avail) { | 4884 | if (is_lvds && dev_priv->lvds_downclock_avail) { |
| 4884 | /* | 4885 | /* |
| 4885 | * Ensure we match the reduced clock's P to the target clock. | 4886 | * Ensure we match the reduced clock's P to the target clock. |
| @@ -5768,9 +5769,6 @@ static int ironlake_crtc_mode_set(struct drm_crtc *crtc, | |||
| 5768 | intel_crtc->config.dpll.p2 = clock.p2; | 5769 | intel_crtc->config.dpll.p2 = clock.p2; |
| 5769 | } | 5770 | } |
| 5770 | 5771 | ||
| 5771 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
| 5772 | intel_crtc_update_cursor(crtc, true); | ||
| 5773 | |||
| 5774 | /* CPU eDP is the only output that doesn't need a PCH PLL of its own. */ | 5772 | /* CPU eDP is the only output that doesn't need a PCH PLL of its own. */ |
| 5775 | if (intel_crtc->config.has_pch_encoder) { | 5773 | if (intel_crtc->config.has_pch_encoder) { |
| 5776 | fp = i9xx_dpll_compute_fp(&intel_crtc->config.dpll); | 5774 | fp = i9xx_dpll_compute_fp(&intel_crtc->config.dpll); |
| @@ -5859,6 +5857,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc, | |||
| 5859 | tmp = I915_READ(PF_CTL(crtc->pipe)); | 5857 | tmp = I915_READ(PF_CTL(crtc->pipe)); |
| 5860 | 5858 | ||
| 5861 | if (tmp & PF_ENABLE) { | 5859 | if (tmp & PF_ENABLE) { |
| 5860 | pipe_config->pch_pfit.enabled = true; | ||
| 5862 | pipe_config->pch_pfit.pos = I915_READ(PF_WIN_POS(crtc->pipe)); | 5861 | pipe_config->pch_pfit.pos = I915_READ(PF_WIN_POS(crtc->pipe)); |
| 5863 | pipe_config->pch_pfit.size = I915_READ(PF_WIN_SZ(crtc->pipe)); | 5862 | pipe_config->pch_pfit.size = I915_READ(PF_WIN_SZ(crtc->pipe)); |
| 5864 | 5863 | ||
| @@ -6236,7 +6235,7 @@ static void haswell_modeset_global_resources(struct drm_device *dev) | |||
| 6236 | if (!crtc->base.enabled) | 6235 | if (!crtc->base.enabled) |
| 6237 | continue; | 6236 | continue; |
| 6238 | 6237 | ||
| 6239 | if (crtc->pipe != PIPE_A || crtc->config.pch_pfit.size || | 6238 | if (crtc->pipe != PIPE_A || crtc->config.pch_pfit.enabled || |
| 6240 | crtc->config.cpu_transcoder != TRANSCODER_EDP) | 6239 | crtc->config.cpu_transcoder != TRANSCODER_EDP) |
| 6241 | enable = true; | 6240 | enable = true; |
| 6242 | } | 6241 | } |
| @@ -6259,9 +6258,6 @@ static int haswell_crtc_mode_set(struct drm_crtc *crtc, | |||
| 6259 | if (!intel_ddi_pll_mode_set(crtc)) | 6258 | if (!intel_ddi_pll_mode_set(crtc)) |
| 6260 | return -EINVAL; | 6259 | return -EINVAL; |
| 6261 | 6260 | ||
| 6262 | /* Ensure that the cursor is valid for the new mode before changing... */ | ||
| 6263 | intel_crtc_update_cursor(crtc, true); | ||
| 6264 | |||
| 6265 | if (intel_crtc->config.has_dp_encoder) | 6261 | if (intel_crtc->config.has_dp_encoder) |
| 6266 | intel_dp_set_m_n(intel_crtc); | 6262 | intel_dp_set_m_n(intel_crtc); |
| 6267 | 6263 | ||
| @@ -6494,15 +6490,15 @@ static void haswell_write_eld(struct drm_connector *connector, | |||
| 6494 | 6490 | ||
| 6495 | /* Set ELD valid state */ | 6491 | /* Set ELD valid state */ |
| 6496 | tmp = I915_READ(aud_cntrl_st2); | 6492 | tmp = I915_READ(aud_cntrl_st2); |
| 6497 | DRM_DEBUG_DRIVER("HDMI audio: pin eld vld status=0x%8x\n", tmp); | 6493 | DRM_DEBUG_DRIVER("HDMI audio: pin eld vld status=0x%08x\n", tmp); |
| 6498 | tmp |= (AUDIO_ELD_VALID_A << (pipe * 4)); | 6494 | tmp |= (AUDIO_ELD_VALID_A << (pipe * 4)); |
| 6499 | I915_WRITE(aud_cntrl_st2, tmp); | 6495 | I915_WRITE(aud_cntrl_st2, tmp); |
| 6500 | tmp = I915_READ(aud_cntrl_st2); | 6496 | tmp = I915_READ(aud_cntrl_st2); |
| 6501 | DRM_DEBUG_DRIVER("HDMI audio: eld vld status=0x%8x\n", tmp); | 6497 | DRM_DEBUG_DRIVER("HDMI audio: eld vld status=0x%08x\n", tmp); |
| 6502 | 6498 | ||
| 6503 | /* Enable HDMI mode */ | 6499 | /* Enable HDMI mode */ |
| 6504 | tmp = I915_READ(aud_config); | 6500 | tmp = I915_READ(aud_config); |
| 6505 | DRM_DEBUG_DRIVER("HDMI audio: audio conf: 0x%8x\n", tmp); | 6501 | DRM_DEBUG_DRIVER("HDMI audio: audio conf: 0x%08x\n", tmp); |
| 6506 | /* clear N_programing_enable and N_value_index */ | 6502 | /* clear N_programing_enable and N_value_index */ |
| 6507 | tmp &= ~(AUD_CONFIG_N_VALUE_INDEX | AUD_CONFIG_N_PROG_ENABLE); | 6503 | tmp &= ~(AUD_CONFIG_N_VALUE_INDEX | AUD_CONFIG_N_PROG_ENABLE); |
| 6508 | I915_WRITE(aud_config, tmp); | 6504 | I915_WRITE(aud_config, tmp); |
| @@ -6937,7 +6933,8 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc, | |||
| 6937 | intel_crtc->cursor_width = width; | 6933 | intel_crtc->cursor_width = width; |
| 6938 | intel_crtc->cursor_height = height; | 6934 | intel_crtc->cursor_height = height; |
| 6939 | 6935 | ||
| 6940 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | 6936 | if (intel_crtc->active) |
| 6937 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | ||
| 6941 | 6938 | ||
| 6942 | return 0; | 6939 | return 0; |
| 6943 | fail_unpin: | 6940 | fail_unpin: |
| @@ -6956,7 +6953,8 @@ static int intel_crtc_cursor_move(struct drm_crtc *crtc, int x, int y) | |||
| 6956 | intel_crtc->cursor_x = x; | 6953 | intel_crtc->cursor_x = x; |
| 6957 | intel_crtc->cursor_y = y; | 6954 | intel_crtc->cursor_y = y; |
| 6958 | 6955 | ||
| 6959 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | 6956 | if (intel_crtc->active) |
| 6957 | intel_crtc_update_cursor(crtc, intel_crtc->cursor_bo != NULL); | ||
| 6960 | 6958 | ||
| 6961 | return 0; | 6959 | return 0; |
| 6962 | } | 6960 | } |
| @@ -8205,9 +8203,10 @@ static void intel_dump_pipe_config(struct intel_crtc *crtc, | |||
| 8205 | pipe_config->gmch_pfit.control, | 8203 | pipe_config->gmch_pfit.control, |
| 8206 | pipe_config->gmch_pfit.pgm_ratios, | 8204 | pipe_config->gmch_pfit.pgm_ratios, |
| 8207 | pipe_config->gmch_pfit.lvds_border_bits); | 8205 | pipe_config->gmch_pfit.lvds_border_bits); |
| 8208 | DRM_DEBUG_KMS("pch pfit: pos: 0x%08x, size: 0x%08x\n", | 8206 | DRM_DEBUG_KMS("pch pfit: pos: 0x%08x, size: 0x%08x, %s\n", |
| 8209 | pipe_config->pch_pfit.pos, | 8207 | pipe_config->pch_pfit.pos, |
| 8210 | pipe_config->pch_pfit.size); | 8208 | pipe_config->pch_pfit.size, |
| 8209 | pipe_config->pch_pfit.enabled ? "enabled" : "disabled"); | ||
| 8211 | DRM_DEBUG_KMS("ips: %i\n", pipe_config->ips_enabled); | 8210 | DRM_DEBUG_KMS("ips: %i\n", pipe_config->ips_enabled); |
| 8212 | } | 8211 | } |
| 8213 | 8212 | ||
| @@ -8603,8 +8602,11 @@ intel_pipe_config_compare(struct drm_device *dev, | |||
| 8603 | if (INTEL_INFO(dev)->gen < 4) | 8602 | if (INTEL_INFO(dev)->gen < 4) |
| 8604 | PIPE_CONF_CHECK_I(gmch_pfit.pgm_ratios); | 8603 | PIPE_CONF_CHECK_I(gmch_pfit.pgm_ratios); |
| 8605 | PIPE_CONF_CHECK_I(gmch_pfit.lvds_border_bits); | 8604 | PIPE_CONF_CHECK_I(gmch_pfit.lvds_border_bits); |
| 8606 | PIPE_CONF_CHECK_I(pch_pfit.pos); | 8605 | PIPE_CONF_CHECK_I(pch_pfit.enabled); |
| 8607 | PIPE_CONF_CHECK_I(pch_pfit.size); | 8606 | if (current_config->pch_pfit.enabled) { |
| 8607 | PIPE_CONF_CHECK_I(pch_pfit.pos); | ||
| 8608 | PIPE_CONF_CHECK_I(pch_pfit.size); | ||
| 8609 | } | ||
| 8608 | 8610 | ||
| 8609 | PIPE_CONF_CHECK_I(ips_enabled); | 8611 | PIPE_CONF_CHECK_I(ips_enabled); |
| 8610 | 8612 | ||
diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c index 2151d13772b8..79c14e298ba6 100644 --- a/drivers/gpu/drm/i915/intel_dp.c +++ b/drivers/gpu/drm/i915/intel_dp.c | |||
| @@ -588,7 +588,18 @@ intel_dp_i2c_aux_ch(struct i2c_adapter *adapter, int mode, | |||
| 588 | DRM_DEBUG_KMS("aux_ch native nack\n"); | 588 | DRM_DEBUG_KMS("aux_ch native nack\n"); |
| 589 | return -EREMOTEIO; | 589 | return -EREMOTEIO; |
| 590 | case AUX_NATIVE_REPLY_DEFER: | 590 | case AUX_NATIVE_REPLY_DEFER: |
| 591 | udelay(100); | 591 | /* |
| 592 | * For now, just give more slack to branch devices. We | ||
| 593 | * could check the DPCD for I2C bit rate capabilities, | ||
| 594 | * and if available, adjust the interval. We could also | ||
| 595 | * be more careful with DP-to-Legacy adapters where a | ||
| 596 | * long legacy cable may force very low I2C bit rates. | ||
| 597 | */ | ||
| 598 | if (intel_dp->dpcd[DP_DOWNSTREAMPORT_PRESENT] & | ||
| 599 | DP_DWN_STRM_PORT_PRESENT) | ||
| 600 | usleep_range(500, 600); | ||
| 601 | else | ||
| 602 | usleep_range(300, 400); | ||
| 592 | continue; | 603 | continue; |
| 593 | default: | 604 | default: |
| 594 | DRM_ERROR("aux_ch invalid native reply 0x%02x\n", | 605 | DRM_ERROR("aux_ch invalid native reply 0x%02x\n", |
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h index a47799e832c6..28cae80495e2 100644 --- a/drivers/gpu/drm/i915/intel_drv.h +++ b/drivers/gpu/drm/i915/intel_drv.h | |||
| @@ -280,6 +280,7 @@ struct intel_crtc_config { | |||
| 280 | struct { | 280 | struct { |
| 281 | u32 pos; | 281 | u32 pos; |
| 282 | u32 size; | 282 | u32 size; |
| 283 | bool enabled; | ||
| 283 | } pch_pfit; | 284 | } pch_pfit; |
| 284 | 285 | ||
| 285 | /* FDI configuration, only valid if has_pch_encoder is set. */ | 286 | /* FDI configuration, only valid if has_pch_encoder is set. */ |
diff --git a/drivers/gpu/drm/i915/intel_dvo.c b/drivers/gpu/drm/i915/intel_dvo.c index 406303b509c1..7fa7df546c1e 100644 --- a/drivers/gpu/drm/i915/intel_dvo.c +++ b/drivers/gpu/drm/i915/intel_dvo.c | |||
| @@ -263,6 +263,8 @@ static bool intel_dvo_compute_config(struct intel_encoder *encoder, | |||
| 263 | C(vtotal); | 263 | C(vtotal); |
| 264 | C(clock); | 264 | C(clock); |
| 265 | #undef C | 265 | #undef C |
| 266 | |||
| 267 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
| 266 | } | 268 | } |
| 267 | 269 | ||
| 268 | if (intel_dvo->dev.dev_ops->mode_fixup) | 270 | if (intel_dvo->dev.dev_ops->mode_fixup) |
diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c index 42114ecbae0e..293564a2896a 100644 --- a/drivers/gpu/drm/i915/intel_panel.c +++ b/drivers/gpu/drm/i915/intel_panel.c | |||
| @@ -112,6 +112,7 @@ intel_pch_panel_fitting(struct intel_crtc *intel_crtc, | |||
| 112 | done: | 112 | done: |
| 113 | pipe_config->pch_pfit.pos = (x << 16) | y; | 113 | pipe_config->pch_pfit.pos = (x << 16) | y; |
| 114 | pipe_config->pch_pfit.size = (width << 16) | height; | 114 | pipe_config->pch_pfit.size = (width << 16) | height; |
| 115 | pipe_config->pch_pfit.enabled = pipe_config->pch_pfit.size != 0; | ||
| 115 | } | 116 | } |
| 116 | 117 | ||
| 117 | static void | 118 | static void |
diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c index 0c115cc4899f..dd176b7296c1 100644 --- a/drivers/gpu/drm/i915/intel_pm.c +++ b/drivers/gpu/drm/i915/intel_pm.c | |||
| @@ -2096,16 +2096,16 @@ static uint32_t ilk_pipe_pixel_rate(struct drm_device *dev, | |||
| 2096 | struct drm_crtc *crtc) | 2096 | struct drm_crtc *crtc) |
| 2097 | { | 2097 | { |
| 2098 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | 2098 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); |
| 2099 | uint32_t pixel_rate, pfit_size; | 2099 | uint32_t pixel_rate; |
| 2100 | 2100 | ||
| 2101 | pixel_rate = intel_crtc->config.adjusted_mode.clock; | 2101 | pixel_rate = intel_crtc->config.adjusted_mode.clock; |
| 2102 | 2102 | ||
| 2103 | /* We only use IF-ID interlacing. If we ever use PF-ID we'll need to | 2103 | /* We only use IF-ID interlacing. If we ever use PF-ID we'll need to |
| 2104 | * adjust the pixel_rate here. */ | 2104 | * adjust the pixel_rate here. */ |
| 2105 | 2105 | ||
| 2106 | pfit_size = intel_crtc->config.pch_pfit.size; | 2106 | if (intel_crtc->config.pch_pfit.enabled) { |
| 2107 | if (pfit_size) { | ||
| 2108 | uint64_t pipe_w, pipe_h, pfit_w, pfit_h; | 2107 | uint64_t pipe_w, pipe_h, pfit_w, pfit_h; |
| 2108 | uint32_t pfit_size = intel_crtc->config.pch_pfit.size; | ||
| 2109 | 2109 | ||
| 2110 | pipe_w = intel_crtc->config.requested_mode.hdisplay; | 2110 | pipe_w = intel_crtc->config.requested_mode.hdisplay; |
| 2111 | pipe_h = intel_crtc->config.requested_mode.vdisplay; | 2111 | pipe_h = intel_crtc->config.requested_mode.vdisplay; |
diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c index 85037b9d4934..49482fd5b76c 100644 --- a/drivers/gpu/drm/i915/intel_sdvo.c +++ b/drivers/gpu/drm/i915/intel_sdvo.c | |||
| @@ -788,6 +788,8 @@ static void intel_sdvo_get_dtd_from_mode(struct intel_sdvo_dtd *dtd, | |||
| 788 | uint16_t h_sync_offset, v_sync_offset; | 788 | uint16_t h_sync_offset, v_sync_offset; |
| 789 | int mode_clock; | 789 | int mode_clock; |
| 790 | 790 | ||
| 791 | memset(dtd, 0, sizeof(*dtd)); | ||
| 792 | |||
| 791 | width = mode->hdisplay; | 793 | width = mode->hdisplay; |
| 792 | height = mode->vdisplay; | 794 | height = mode->vdisplay; |
| 793 | 795 | ||
| @@ -830,44 +832,51 @@ static void intel_sdvo_get_dtd_from_mode(struct intel_sdvo_dtd *dtd, | |||
| 830 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) | 832 | if (mode->flags & DRM_MODE_FLAG_PVSYNC) |
| 831 | dtd->part2.dtd_flags |= DTD_FLAG_VSYNC_POSITIVE; | 833 | dtd->part2.dtd_flags |= DTD_FLAG_VSYNC_POSITIVE; |
| 832 | 834 | ||
| 833 | dtd->part2.sdvo_flags = 0; | ||
| 834 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; | 835 | dtd->part2.v_sync_off_high = v_sync_offset & 0xc0; |
| 835 | dtd->part2.reserved = 0; | ||
| 836 | } | 836 | } |
| 837 | 837 | ||
| 838 | static void intel_sdvo_get_mode_from_dtd(struct drm_display_mode * mode, | 838 | static void intel_sdvo_get_mode_from_dtd(struct drm_display_mode *pmode, |
| 839 | const struct intel_sdvo_dtd *dtd) | 839 | const struct intel_sdvo_dtd *dtd) |
| 840 | { | 840 | { |
| 841 | mode->hdisplay = dtd->part1.h_active; | 841 | struct drm_display_mode mode = {}; |
| 842 | mode->hdisplay += ((dtd->part1.h_high >> 4) & 0x0f) << 8; | 842 | |
| 843 | mode->hsync_start = mode->hdisplay + dtd->part2.h_sync_off; | 843 | mode.hdisplay = dtd->part1.h_active; |
| 844 | mode->hsync_start += (dtd->part2.sync_off_width_high & 0xc0) << 2; | 844 | mode.hdisplay += ((dtd->part1.h_high >> 4) & 0x0f) << 8; |
| 845 | mode->hsync_end = mode->hsync_start + dtd->part2.h_sync_width; | 845 | mode.hsync_start = mode.hdisplay + dtd->part2.h_sync_off; |
| 846 | mode->hsync_end += (dtd->part2.sync_off_width_high & 0x30) << 4; | 846 | mode.hsync_start += (dtd->part2.sync_off_width_high & 0xc0) << 2; |
| 847 | mode->htotal = mode->hdisplay + dtd->part1.h_blank; | 847 | mode.hsync_end = mode.hsync_start + dtd->part2.h_sync_width; |
| 848 | mode->htotal += (dtd->part1.h_high & 0xf) << 8; | 848 | mode.hsync_end += (dtd->part2.sync_off_width_high & 0x30) << 4; |
| 849 | 849 | mode.htotal = mode.hdisplay + dtd->part1.h_blank; | |
| 850 | mode->vdisplay = dtd->part1.v_active; | 850 | mode.htotal += (dtd->part1.h_high & 0xf) << 8; |
| 851 | mode->vdisplay += ((dtd->part1.v_high >> 4) & 0x0f) << 8; | 851 | |
| 852 | mode->vsync_start = mode->vdisplay; | 852 | mode.vdisplay = dtd->part1.v_active; |
| 853 | mode->vsync_start += (dtd->part2.v_sync_off_width >> 4) & 0xf; | 853 | mode.vdisplay += ((dtd->part1.v_high >> 4) & 0x0f) << 8; |
| 854 | mode->vsync_start += (dtd->part2.sync_off_width_high & 0x0c) << 2; | 854 | mode.vsync_start = mode.vdisplay; |
| 855 | mode->vsync_start += dtd->part2.v_sync_off_high & 0xc0; | 855 | mode.vsync_start += (dtd->part2.v_sync_off_width >> 4) & 0xf; |
| 856 | mode->vsync_end = mode->vsync_start + | 856 | mode.vsync_start += (dtd->part2.sync_off_width_high & 0x0c) << 2; |
| 857 | mode.vsync_start += dtd->part2.v_sync_off_high & 0xc0; | ||
| 858 | mode.vsync_end = mode.vsync_start + | ||
| 857 | (dtd->part2.v_sync_off_width & 0xf); | 859 | (dtd->part2.v_sync_off_width & 0xf); |
| 858 | mode->vsync_end += (dtd->part2.sync_off_width_high & 0x3) << 4; | 860 | mode.vsync_end += (dtd->part2.sync_off_width_high & 0x3) << 4; |
| 859 | mode->vtotal = mode->vdisplay + dtd->part1.v_blank; | 861 | mode.vtotal = mode.vdisplay + dtd->part1.v_blank; |
| 860 | mode->vtotal += (dtd->part1.v_high & 0xf) << 8; | 862 | mode.vtotal += (dtd->part1.v_high & 0xf) << 8; |
| 861 | 863 | ||
| 862 | mode->clock = dtd->part1.clock * 10; | 864 | mode.clock = dtd->part1.clock * 10; |
| 863 | 865 | ||
| 864 | mode->flags &= ~(DRM_MODE_FLAG_PHSYNC | DRM_MODE_FLAG_PVSYNC); | ||
| 865 | if (dtd->part2.dtd_flags & DTD_FLAG_INTERLACE) | 866 | if (dtd->part2.dtd_flags & DTD_FLAG_INTERLACE) |
| 866 | mode->flags |= DRM_MODE_FLAG_INTERLACE; | 867 | mode.flags |= DRM_MODE_FLAG_INTERLACE; |
| 867 | if (dtd->part2.dtd_flags & DTD_FLAG_HSYNC_POSITIVE) | 868 | if (dtd->part2.dtd_flags & DTD_FLAG_HSYNC_POSITIVE) |
| 868 | mode->flags |= DRM_MODE_FLAG_PHSYNC; | 869 | mode.flags |= DRM_MODE_FLAG_PHSYNC; |
| 870 | else | ||
| 871 | mode.flags |= DRM_MODE_FLAG_NHSYNC; | ||
| 869 | if (dtd->part2.dtd_flags & DTD_FLAG_VSYNC_POSITIVE) | 872 | if (dtd->part2.dtd_flags & DTD_FLAG_VSYNC_POSITIVE) |
| 870 | mode->flags |= DRM_MODE_FLAG_PVSYNC; | 873 | mode.flags |= DRM_MODE_FLAG_PVSYNC; |
| 874 | else | ||
| 875 | mode.flags |= DRM_MODE_FLAG_NVSYNC; | ||
| 876 | |||
| 877 | drm_mode_set_crtcinfo(&mode, 0); | ||
| 878 | |||
| 879 | drm_mode_copy(pmode, &mode); | ||
| 871 | } | 880 | } |
| 872 | 881 | ||
| 873 | static bool intel_sdvo_check_supp_encode(struct intel_sdvo *intel_sdvo) | 882 | static bool intel_sdvo_check_supp_encode(struct intel_sdvo *intel_sdvo) |
diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c index f2c6d7909ae2..dd6f84bf6c22 100644 --- a/drivers/gpu/drm/i915/intel_tv.c +++ b/drivers/gpu/drm/i915/intel_tv.c | |||
| @@ -916,6 +916,14 @@ intel_tv_compute_config(struct intel_encoder *encoder, | |||
| 916 | DRM_DEBUG_KMS("forcing bpc to 8 for TV\n"); | 916 | DRM_DEBUG_KMS("forcing bpc to 8 for TV\n"); |
| 917 | pipe_config->pipe_bpp = 8*3; | 917 | pipe_config->pipe_bpp = 8*3; |
| 918 | 918 | ||
| 919 | /* TV has it's own notion of sync and other mode flags, so clear them. */ | ||
| 920 | pipe_config->adjusted_mode.flags = 0; | ||
| 921 | |||
| 922 | /* | ||
| 923 | * FIXME: We don't check whether the input mode is actually what we want | ||
| 924 | * or whether userspace is doing something stupid. | ||
| 925 | */ | ||
| 926 | |||
| 919 | return true; | 927 | return true; |
| 920 | } | 928 | } |
| 921 | 929 | ||
diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c index a60584763b61..a0b9d8a95b16 100644 --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c | |||
| @@ -124,6 +124,8 @@ void adreno_recover(struct msm_gpu *gpu) | |||
| 124 | 124 | ||
| 125 | /* reset completed fence seqno, just discard anything pending: */ | 125 | /* reset completed fence seqno, just discard anything pending: */ |
| 126 | adreno_gpu->memptrs->fence = gpu->submitted_fence; | 126 | adreno_gpu->memptrs->fence = gpu->submitted_fence; |
| 127 | adreno_gpu->memptrs->rptr = 0; | ||
| 128 | adreno_gpu->memptrs->wptr = 0; | ||
| 127 | 129 | ||
| 128 | gpu->funcs->pm_resume(gpu); | 130 | gpu->funcs->pm_resume(gpu); |
| 129 | ret = gpu->funcs->hw_init(gpu); | 131 | ret = gpu->funcs->hw_init(gpu); |
| @@ -229,7 +231,7 @@ void adreno_idle(struct msm_gpu *gpu) | |||
| 229 | return; | 231 | return; |
| 230 | } while(time_before(jiffies, t)); | 232 | } while(time_before(jiffies, t)); |
| 231 | 233 | ||
| 232 | DRM_ERROR("timeout waiting for %s to drain ringbuffer!\n", gpu->name); | 234 | DRM_ERROR("%s: timeout waiting to drain ringbuffer!\n", gpu->name); |
| 233 | 235 | ||
| 234 | /* TODO maybe we need to reset GPU here to recover from hang? */ | 236 | /* TODO maybe we need to reset GPU here to recover from hang? */ |
| 235 | } | 237 | } |
| @@ -256,11 +258,17 @@ void adreno_wait_ring(struct msm_gpu *gpu, uint32_t ndwords) | |||
| 256 | { | 258 | { |
| 257 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); | 259 | struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); |
| 258 | uint32_t freedwords; | 260 | uint32_t freedwords; |
| 261 | unsigned long t = jiffies + ADRENO_IDLE_TIMEOUT; | ||
| 259 | do { | 262 | do { |
| 260 | uint32_t size = gpu->rb->size / 4; | 263 | uint32_t size = gpu->rb->size / 4; |
| 261 | uint32_t wptr = get_wptr(gpu->rb); | 264 | uint32_t wptr = get_wptr(gpu->rb); |
| 262 | uint32_t rptr = adreno_gpu->memptrs->rptr; | 265 | uint32_t rptr = adreno_gpu->memptrs->rptr; |
| 263 | freedwords = (rptr + (size - 1) - wptr) % size; | 266 | freedwords = (rptr + (size - 1) - wptr) % size; |
| 267 | |||
| 268 | if (time_after(jiffies, t)) { | ||
| 269 | DRM_ERROR("%s: timeout waiting for ringbuffer space\n", gpu->name); | ||
| 270 | break; | ||
| 271 | } | ||
| 264 | } while(freedwords < ndwords); | 272 | } while(freedwords < ndwords); |
| 265 | } | 273 | } |
| 266 | 274 | ||
diff --git a/drivers/gpu/drm/msm/mdp4/mdp4_kms.c b/drivers/gpu/drm/msm/mdp4/mdp4_kms.c index 5db5bbaedae2..bc7fd11ad8be 100644 --- a/drivers/gpu/drm/msm/mdp4/mdp4_kms.c +++ b/drivers/gpu/drm/msm/mdp4/mdp4_kms.c | |||
| @@ -19,8 +19,6 @@ | |||
| 19 | #include "msm_drv.h" | 19 | #include "msm_drv.h" |
| 20 | #include "mdp4_kms.h" | 20 | #include "mdp4_kms.h" |
| 21 | 21 | ||
| 22 | #include <mach/iommu.h> | ||
| 23 | |||
| 24 | static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev); | 22 | static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev); |
| 25 | 23 | ||
| 26 | static int mdp4_hw_init(struct msm_kms *kms) | 24 | static int mdp4_hw_init(struct msm_kms *kms) |
diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c index 864c9773636b..b3a2f1629041 100644 --- a/drivers/gpu/drm/msm/msm_drv.c +++ b/drivers/gpu/drm/msm/msm_drv.c | |||
| @@ -18,8 +18,6 @@ | |||
| 18 | #include "msm_drv.h" | 18 | #include "msm_drv.h" |
| 19 | #include "msm_gpu.h" | 19 | #include "msm_gpu.h" |
| 20 | 20 | ||
| 21 | #include <mach/iommu.h> | ||
| 22 | |||
| 23 | static void msm_fb_output_poll_changed(struct drm_device *dev) | 21 | static void msm_fb_output_poll_changed(struct drm_device *dev) |
| 24 | { | 22 | { |
| 25 | struct msm_drm_private *priv = dev->dev_private; | 23 | struct msm_drm_private *priv = dev->dev_private; |
| @@ -62,6 +60,8 @@ int msm_iommu_attach(struct drm_device *dev, struct iommu_domain *iommu, | |||
| 62 | int i, ret; | 60 | int i, ret; |
| 63 | 61 | ||
| 64 | for (i = 0; i < cnt; i++) { | 62 | for (i = 0; i < cnt; i++) { |
| 63 | /* TODO maybe some day msm iommu won't require this hack: */ | ||
| 64 | struct device *msm_iommu_get_ctx(const char *ctx_name); | ||
| 65 | struct device *ctx = msm_iommu_get_ctx(names[i]); | 65 | struct device *ctx = msm_iommu_get_ctx(names[i]); |
| 66 | if (!ctx) | 66 | if (!ctx) |
| 67 | continue; | 67 | continue; |
| @@ -199,7 +199,7 @@ static int msm_load(struct drm_device *dev, unsigned long flags) | |||
| 199 | * imx drm driver on iMX5 | 199 | * imx drm driver on iMX5 |
| 200 | */ | 200 | */ |
| 201 | dev_err(dev->dev, "failed to load kms\n"); | 201 | dev_err(dev->dev, "failed to load kms\n"); |
| 202 | ret = PTR_ERR(priv->kms); | 202 | ret = PTR_ERR(kms); |
| 203 | goto fail; | 203 | goto fail; |
| 204 | } | 204 | } |
| 205 | 205 | ||
| @@ -499,25 +499,41 @@ int msm_wait_fence_interruptable(struct drm_device *dev, uint32_t fence, | |||
| 499 | struct timespec *timeout) | 499 | struct timespec *timeout) |
| 500 | { | 500 | { |
| 501 | struct msm_drm_private *priv = dev->dev_private; | 501 | struct msm_drm_private *priv = dev->dev_private; |
| 502 | unsigned long timeout_jiffies = timespec_to_jiffies(timeout); | ||
| 503 | unsigned long start_jiffies = jiffies; | ||
| 504 | unsigned long remaining_jiffies; | ||
| 505 | int ret; | 502 | int ret; |
| 506 | 503 | ||
| 507 | if (time_after(start_jiffies, timeout_jiffies)) | 504 | if (!priv->gpu) |
| 508 | remaining_jiffies = 0; | 505 | return 0; |
| 509 | else | 506 | |
| 510 | remaining_jiffies = timeout_jiffies - start_jiffies; | 507 | if (fence > priv->gpu->submitted_fence) { |
| 511 | 508 | DRM_ERROR("waiting on invalid fence: %u (of %u)\n", | |
| 512 | ret = wait_event_interruptible_timeout(priv->fence_event, | 509 | fence, priv->gpu->submitted_fence); |
| 513 | priv->completed_fence >= fence, | 510 | return -EINVAL; |
| 514 | remaining_jiffies); | 511 | } |
| 515 | if (ret == 0) { | 512 | |
| 516 | DBG("timeout waiting for fence: %u (completed: %u)", | 513 | if (!timeout) { |
| 517 | fence, priv->completed_fence); | 514 | /* no-wait: */ |
| 518 | ret = -ETIMEDOUT; | 515 | ret = fence_completed(dev, fence) ? 0 : -EBUSY; |
| 519 | } else if (ret != -ERESTARTSYS) { | 516 | } else { |
| 520 | ret = 0; | 517 | unsigned long timeout_jiffies = timespec_to_jiffies(timeout); |
| 518 | unsigned long start_jiffies = jiffies; | ||
| 519 | unsigned long remaining_jiffies; | ||
| 520 | |||
| 521 | if (time_after(start_jiffies, timeout_jiffies)) | ||
| 522 | remaining_jiffies = 0; | ||
| 523 | else | ||
| 524 | remaining_jiffies = timeout_jiffies - start_jiffies; | ||
| 525 | |||
| 526 | ret = wait_event_interruptible_timeout(priv->fence_event, | ||
| 527 | fence_completed(dev, fence), | ||
| 528 | remaining_jiffies); | ||
| 529 | |||
| 530 | if (ret == 0) { | ||
| 531 | DBG("timeout waiting for fence: %u (completed: %u)", | ||
| 532 | fence, priv->completed_fence); | ||
| 533 | ret = -ETIMEDOUT; | ||
| 534 | } else if (ret != -ERESTARTSYS) { | ||
| 535 | ret = 0; | ||
| 536 | } | ||
| 521 | } | 537 | } |
| 522 | 538 | ||
| 523 | return ret; | 539 | return ret; |
| @@ -681,7 +697,7 @@ static struct drm_driver msm_driver = { | |||
| 681 | .gem_vm_ops = &vm_ops, | 697 | .gem_vm_ops = &vm_ops, |
| 682 | .dumb_create = msm_gem_dumb_create, | 698 | .dumb_create = msm_gem_dumb_create, |
| 683 | .dumb_map_offset = msm_gem_dumb_map_offset, | 699 | .dumb_map_offset = msm_gem_dumb_map_offset, |
| 684 | .dumb_destroy = msm_gem_dumb_destroy, | 700 | .dumb_destroy = drm_gem_dumb_destroy, |
| 685 | #ifdef CONFIG_DEBUG_FS | 701 | #ifdef CONFIG_DEBUG_FS |
| 686 | .debugfs_init = msm_debugfs_init, | 702 | .debugfs_init = msm_debugfs_init, |
| 687 | .debugfs_cleanup = msm_debugfs_cleanup, | 703 | .debugfs_cleanup = msm_debugfs_cleanup, |
diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h index 80d75094bf0a..df8f1d084bc1 100644 --- a/drivers/gpu/drm/msm/msm_drv.h +++ b/drivers/gpu/drm/msm/msm_drv.h | |||
| @@ -153,7 +153,7 @@ void *msm_gem_vaddr(struct drm_gem_object *obj); | |||
| 153 | int msm_gem_queue_inactive_work(struct drm_gem_object *obj, | 153 | int msm_gem_queue_inactive_work(struct drm_gem_object *obj, |
| 154 | struct work_struct *work); | 154 | struct work_struct *work); |
| 155 | void msm_gem_move_to_active(struct drm_gem_object *obj, | 155 | void msm_gem_move_to_active(struct drm_gem_object *obj, |
| 156 | struct msm_gpu *gpu, uint32_t fence); | 156 | struct msm_gpu *gpu, bool write, uint32_t fence); |
| 157 | void msm_gem_move_to_inactive(struct drm_gem_object *obj); | 157 | void msm_gem_move_to_inactive(struct drm_gem_object *obj); |
| 158 | int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, | 158 | int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, |
| 159 | struct timespec *timeout); | 159 | struct timespec *timeout); |
| @@ -191,6 +191,12 @@ u32 msm_readl(const void __iomem *addr); | |||
| 191 | #define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) | 191 | #define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) |
| 192 | #define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) | 192 | #define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__) |
| 193 | 193 | ||
| 194 | static inline bool fence_completed(struct drm_device *dev, uint32_t fence) | ||
| 195 | { | ||
| 196 | struct msm_drm_private *priv = dev->dev_private; | ||
| 197 | return priv->completed_fence >= fence; | ||
| 198 | } | ||
| 199 | |||
| 194 | static inline int align_pitch(int width, int bpp) | 200 | static inline int align_pitch(int width, int bpp) |
| 195 | { | 201 | { |
| 196 | int bytespp = (bpp + 7) / 8; | 202 | int bytespp = (bpp + 7) / 8; |
diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c index 6b5a6c8c7658..2bae46c66a30 100644 --- a/drivers/gpu/drm/msm/msm_gem.c +++ b/drivers/gpu/drm/msm/msm_gem.c | |||
| @@ -40,9 +40,9 @@ static struct page **get_pages(struct drm_gem_object *obj) | |||
| 40 | } | 40 | } |
| 41 | 41 | ||
| 42 | msm_obj->sgt = drm_prime_pages_to_sg(p, npages); | 42 | msm_obj->sgt = drm_prime_pages_to_sg(p, npages); |
| 43 | if (!msm_obj->sgt) { | 43 | if (IS_ERR(msm_obj->sgt)) { |
| 44 | dev_err(dev->dev, "failed to allocate sgt\n"); | 44 | dev_err(dev->dev, "failed to allocate sgt\n"); |
| 45 | return ERR_PTR(-ENOMEM); | 45 | return ERR_CAST(msm_obj->sgt); |
| 46 | } | 46 | } |
| 47 | 47 | ||
| 48 | msm_obj->pages = p; | 48 | msm_obj->pages = p; |
| @@ -159,7 +159,6 @@ out_unlock: | |||
| 159 | out: | 159 | out: |
| 160 | switch (ret) { | 160 | switch (ret) { |
| 161 | case -EAGAIN: | 161 | case -EAGAIN: |
| 162 | set_need_resched(); | ||
| 163 | case 0: | 162 | case 0: |
| 164 | case -ERESTARTSYS: | 163 | case -ERESTARTSYS: |
| 165 | case -EINTR: | 164 | case -EINTR: |
| @@ -320,13 +319,6 @@ int msm_gem_dumb_create(struct drm_file *file, struct drm_device *dev, | |||
| 320 | MSM_BO_SCANOUT | MSM_BO_WC, &args->handle); | 319 | MSM_BO_SCANOUT | MSM_BO_WC, &args->handle); |
| 321 | } | 320 | } |
| 322 | 321 | ||
| 323 | int msm_gem_dumb_destroy(struct drm_file *file, struct drm_device *dev, | ||
| 324 | uint32_t handle) | ||
| 325 | { | ||
| 326 | /* No special work needed, drop the reference and see what falls out */ | ||
| 327 | return drm_gem_handle_delete(file, handle); | ||
| 328 | } | ||
| 329 | |||
| 330 | int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, | 322 | int msm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, |
| 331 | uint32_t handle, uint64_t *offset) | 323 | uint32_t handle, uint64_t *offset) |
| 332 | { | 324 | { |
| @@ -393,11 +385,14 @@ int msm_gem_queue_inactive_work(struct drm_gem_object *obj, | |||
| 393 | } | 385 | } |
| 394 | 386 | ||
| 395 | void msm_gem_move_to_active(struct drm_gem_object *obj, | 387 | void msm_gem_move_to_active(struct drm_gem_object *obj, |
| 396 | struct msm_gpu *gpu, uint32_t fence) | 388 | struct msm_gpu *gpu, bool write, uint32_t fence) |
| 397 | { | 389 | { |
| 398 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 390 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
| 399 | msm_obj->gpu = gpu; | 391 | msm_obj->gpu = gpu; |
| 400 | msm_obj->fence = fence; | 392 | if (write) |
| 393 | msm_obj->write_fence = fence; | ||
| 394 | else | ||
| 395 | msm_obj->read_fence = fence; | ||
| 401 | list_del_init(&msm_obj->mm_list); | 396 | list_del_init(&msm_obj->mm_list); |
| 402 | list_add_tail(&msm_obj->mm_list, &gpu->active_list); | 397 | list_add_tail(&msm_obj->mm_list, &gpu->active_list); |
| 403 | } | 398 | } |
| @@ -411,7 +406,8 @@ void msm_gem_move_to_inactive(struct drm_gem_object *obj) | |||
| 411 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); | 406 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); |
| 412 | 407 | ||
| 413 | msm_obj->gpu = NULL; | 408 | msm_obj->gpu = NULL; |
| 414 | msm_obj->fence = 0; | 409 | msm_obj->read_fence = 0; |
| 410 | msm_obj->write_fence = 0; | ||
| 415 | list_del_init(&msm_obj->mm_list); | 411 | list_del_init(&msm_obj->mm_list); |
| 416 | list_add_tail(&msm_obj->mm_list, &priv->inactive_list); | 412 | list_add_tail(&msm_obj->mm_list, &priv->inactive_list); |
| 417 | 413 | ||
| @@ -433,8 +429,18 @@ int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, | |||
| 433 | struct msm_gem_object *msm_obj = to_msm_bo(obj); | 429 | struct msm_gem_object *msm_obj = to_msm_bo(obj); |
| 434 | int ret = 0; | 430 | int ret = 0; |
| 435 | 431 | ||
| 436 | if (is_active(msm_obj) && !(op & MSM_PREP_NOSYNC)) | 432 | if (is_active(msm_obj)) { |
| 437 | ret = msm_wait_fence_interruptable(dev, msm_obj->fence, timeout); | 433 | uint32_t fence = 0; |
| 434 | |||
| 435 | if (op & MSM_PREP_READ) | ||
| 436 | fence = msm_obj->write_fence; | ||
| 437 | if (op & MSM_PREP_WRITE) | ||
| 438 | fence = max(fence, msm_obj->read_fence); | ||
| 439 | if (op & MSM_PREP_NOSYNC) | ||
| 440 | timeout = NULL; | ||
| 441 | |||
| 442 | ret = msm_wait_fence_interruptable(dev, fence, timeout); | ||
| 443 | } | ||
| 438 | 444 | ||
| 439 | /* TODO cache maintenance */ | 445 | /* TODO cache maintenance */ |
| 440 | 446 | ||
| @@ -455,9 +461,10 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m) | |||
| 455 | uint64_t off = drm_vma_node_start(&obj->vma_node); | 461 | uint64_t off = drm_vma_node_start(&obj->vma_node); |
| 456 | 462 | ||
| 457 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); | 463 | WARN_ON(!mutex_is_locked(&dev->struct_mutex)); |
| 458 | seq_printf(m, "%08x: %c(%d) %2d (%2d) %08llx %p %d\n", | 464 | seq_printf(m, "%08x: %c(r=%u,w=%u) %2d (%2d) %08llx %p %d\n", |
| 459 | msm_obj->flags, is_active(msm_obj) ? 'A' : 'I', | 465 | msm_obj->flags, is_active(msm_obj) ? 'A' : 'I', |
| 460 | msm_obj->fence, obj->name, obj->refcount.refcount.counter, | 466 | msm_obj->read_fence, msm_obj->write_fence, |
| 467 | obj->name, obj->refcount.refcount.counter, | ||
| 461 | off, msm_obj->vaddr, obj->size); | 468 | off, msm_obj->vaddr, obj->size); |
| 462 | } | 469 | } |
| 463 | 470 | ||
diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h index d746f13d283c..0676f32e2c6a 100644 --- a/drivers/gpu/drm/msm/msm_gem.h +++ b/drivers/gpu/drm/msm/msm_gem.h | |||
| @@ -36,7 +36,7 @@ struct msm_gem_object { | |||
| 36 | */ | 36 | */ |
| 37 | struct list_head mm_list; | 37 | struct list_head mm_list; |
| 38 | struct msm_gpu *gpu; /* non-null if active */ | 38 | struct msm_gpu *gpu; /* non-null if active */ |
| 39 | uint32_t fence; | 39 | uint32_t read_fence, write_fence; |
| 40 | 40 | ||
| 41 | /* Transiently in the process of submit ioctl, objects associated | 41 | /* Transiently in the process of submit ioctl, objects associated |
| 42 | * with the submit are on submit->bo_list.. this only lasts for | 42 | * with the submit are on submit->bo_list.. this only lasts for |
diff --git a/drivers/gpu/drm/msm/msm_gem_submit.c b/drivers/gpu/drm/msm/msm_gem_submit.c index 3e1ef3a00f60..5281d4bc37f7 100644 --- a/drivers/gpu/drm/msm/msm_gem_submit.c +++ b/drivers/gpu/drm/msm/msm_gem_submit.c | |||
| @@ -78,7 +78,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
| 78 | } | 78 | } |
| 79 | 79 | ||
| 80 | if (submit_bo.flags & BO_INVALID_FLAGS) { | 80 | if (submit_bo.flags & BO_INVALID_FLAGS) { |
| 81 | DBG("invalid flags: %x", submit_bo.flags); | 81 | DRM_ERROR("invalid flags: %x\n", submit_bo.flags); |
| 82 | ret = -EINVAL; | 82 | ret = -EINVAL; |
| 83 | goto out_unlock; | 83 | goto out_unlock; |
| 84 | } | 84 | } |
| @@ -92,7 +92,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
| 92 | */ | 92 | */ |
| 93 | obj = idr_find(&file->object_idr, submit_bo.handle); | 93 | obj = idr_find(&file->object_idr, submit_bo.handle); |
| 94 | if (!obj) { | 94 | if (!obj) { |
| 95 | DBG("invalid handle %u at index %u", submit_bo.handle, i); | 95 | DRM_ERROR("invalid handle %u at index %u\n", submit_bo.handle, i); |
| 96 | ret = -EINVAL; | 96 | ret = -EINVAL; |
| 97 | goto out_unlock; | 97 | goto out_unlock; |
| 98 | } | 98 | } |
| @@ -100,7 +100,7 @@ static int submit_lookup_objects(struct msm_gem_submit *submit, | |||
| 100 | msm_obj = to_msm_bo(obj); | 100 | msm_obj = to_msm_bo(obj); |
| 101 | 101 | ||
| 102 | if (!list_empty(&msm_obj->submit_entry)) { | 102 | if (!list_empty(&msm_obj->submit_entry)) { |
| 103 | DBG("handle %u at index %u already on submit list", | 103 | DRM_ERROR("handle %u at index %u already on submit list\n", |
| 104 | submit_bo.handle, i); | 104 | submit_bo.handle, i); |
| 105 | ret = -EINVAL; | 105 | ret = -EINVAL; |
| 106 | goto out_unlock; | 106 | goto out_unlock; |
| @@ -216,8 +216,9 @@ static int submit_bo(struct msm_gem_submit *submit, uint32_t idx, | |||
| 216 | struct msm_gem_object **obj, uint32_t *iova, bool *valid) | 216 | struct msm_gem_object **obj, uint32_t *iova, bool *valid) |
| 217 | { | 217 | { |
| 218 | if (idx >= submit->nr_bos) { | 218 | if (idx >= submit->nr_bos) { |
| 219 | DBG("invalid buffer index: %u (out of %u)", idx, submit->nr_bos); | 219 | DRM_ERROR("invalid buffer index: %u (out of %u)\n", |
| 220 | return EINVAL; | 220 | idx, submit->nr_bos); |
| 221 | return -EINVAL; | ||
| 221 | } | 222 | } |
| 222 | 223 | ||
| 223 | if (obj) | 224 | if (obj) |
| @@ -239,7 +240,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
| 239 | int ret; | 240 | int ret; |
| 240 | 241 | ||
| 241 | if (offset % 4) { | 242 | if (offset % 4) { |
| 242 | DBG("non-aligned cmdstream buffer: %u", offset); | 243 | DRM_ERROR("non-aligned cmdstream buffer: %u\n", offset); |
| 243 | return -EINVAL; | 244 | return -EINVAL; |
| 244 | } | 245 | } |
| 245 | 246 | ||
| @@ -266,7 +267,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
| 266 | return -EFAULT; | 267 | return -EFAULT; |
| 267 | 268 | ||
| 268 | if (submit_reloc.submit_offset % 4) { | 269 | if (submit_reloc.submit_offset % 4) { |
| 269 | DBG("non-aligned reloc offset: %u", | 270 | DRM_ERROR("non-aligned reloc offset: %u\n", |
| 270 | submit_reloc.submit_offset); | 271 | submit_reloc.submit_offset); |
| 271 | return -EINVAL; | 272 | return -EINVAL; |
| 272 | } | 273 | } |
| @@ -276,7 +277,7 @@ static int submit_reloc(struct msm_gem_submit *submit, struct msm_gem_object *ob | |||
| 276 | 277 | ||
| 277 | if ((off >= (obj->base.size / 4)) || | 278 | if ((off >= (obj->base.size / 4)) || |
| 278 | (off < last_offset)) { | 279 | (off < last_offset)) { |
| 279 | DBG("invalid offset %u at reloc %u", off, i); | 280 | DRM_ERROR("invalid offset %u at reloc %u\n", off, i); |
| 280 | return -EINVAL; | 281 | return -EINVAL; |
| 281 | } | 282 | } |
| 282 | 283 | ||
| @@ -374,14 +375,15 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data, | |||
| 374 | goto out; | 375 | goto out; |
| 375 | 376 | ||
| 376 | if (submit_cmd.size % 4) { | 377 | if (submit_cmd.size % 4) { |
| 377 | DBG("non-aligned cmdstream buffer size: %u", | 378 | DRM_ERROR("non-aligned cmdstream buffer size: %u\n", |
| 378 | submit_cmd.size); | 379 | submit_cmd.size); |
| 379 | ret = -EINVAL; | 380 | ret = -EINVAL; |
| 380 | goto out; | 381 | goto out; |
| 381 | } | 382 | } |
| 382 | 383 | ||
| 383 | if (submit_cmd.size >= msm_obj->base.size) { | 384 | if ((submit_cmd.size + submit_cmd.submit_offset) >= |
| 384 | DBG("invalid cmdstream size: %u", submit_cmd.size); | 385 | msm_obj->base.size) { |
| 386 | DRM_ERROR("invalid cmdstream size: %u\n", submit_cmd.size); | ||
| 385 | ret = -EINVAL; | 387 | ret = -EINVAL; |
| 386 | goto out; | 388 | goto out; |
| 387 | } | 389 | } |
diff --git a/drivers/gpu/drm/msm/msm_gpu.c b/drivers/gpu/drm/msm/msm_gpu.c index e1e1ec9321ff..3bab937965d1 100644 --- a/drivers/gpu/drm/msm/msm_gpu.c +++ b/drivers/gpu/drm/msm/msm_gpu.c | |||
| @@ -29,13 +29,14 @@ | |||
| 29 | static void bs_init(struct msm_gpu *gpu, struct platform_device *pdev) | 29 | static void bs_init(struct msm_gpu *gpu, struct platform_device *pdev) |
| 30 | { | 30 | { |
| 31 | struct drm_device *dev = gpu->dev; | 31 | struct drm_device *dev = gpu->dev; |
| 32 | struct kgsl_device_platform_data *pdata = pdev->dev.platform_data; | 32 | struct kgsl_device_platform_data *pdata; |
| 33 | 33 | ||
| 34 | if (!pdev) { | 34 | if (!pdev) { |
| 35 | dev_err(dev->dev, "could not find dtv pdata\n"); | 35 | dev_err(dev->dev, "could not find dtv pdata\n"); |
| 36 | return; | 36 | return; |
| 37 | } | 37 | } |
| 38 | 38 | ||
| 39 | pdata = pdev->dev.platform_data; | ||
| 39 | if (pdata->bus_scale_table) { | 40 | if (pdata->bus_scale_table) { |
| 40 | gpu->bsc = msm_bus_scale_register_client(pdata->bus_scale_table); | 41 | gpu->bsc = msm_bus_scale_register_client(pdata->bus_scale_table); |
| 41 | DBG("bus scale client: %08x", gpu->bsc); | 42 | DBG("bus scale client: %08x", gpu->bsc); |
| @@ -230,6 +231,8 @@ static void hangcheck_timer_reset(struct msm_gpu *gpu) | |||
| 230 | static void hangcheck_handler(unsigned long data) | 231 | static void hangcheck_handler(unsigned long data) |
| 231 | { | 232 | { |
| 232 | struct msm_gpu *gpu = (struct msm_gpu *)data; | 233 | struct msm_gpu *gpu = (struct msm_gpu *)data; |
| 234 | struct drm_device *dev = gpu->dev; | ||
| 235 | struct msm_drm_private *priv = dev->dev_private; | ||
| 233 | uint32_t fence = gpu->funcs->last_fence(gpu); | 236 | uint32_t fence = gpu->funcs->last_fence(gpu); |
| 234 | 237 | ||
| 235 | if (fence != gpu->hangcheck_fence) { | 238 | if (fence != gpu->hangcheck_fence) { |
| @@ -237,14 +240,22 @@ static void hangcheck_handler(unsigned long data) | |||
| 237 | gpu->hangcheck_fence = fence; | 240 | gpu->hangcheck_fence = fence; |
| 238 | } else if (fence < gpu->submitted_fence) { | 241 | } else if (fence < gpu->submitted_fence) { |
| 239 | /* no progress and not done.. hung! */ | 242 | /* no progress and not done.. hung! */ |
| 240 | struct msm_drm_private *priv = gpu->dev->dev_private; | ||
| 241 | gpu->hangcheck_fence = fence; | 243 | gpu->hangcheck_fence = fence; |
| 244 | dev_err(dev->dev, "%s: hangcheck detected gpu lockup!\n", | ||
| 245 | gpu->name); | ||
| 246 | dev_err(dev->dev, "%s: completed fence: %u\n", | ||
| 247 | gpu->name, fence); | ||
| 248 | dev_err(dev->dev, "%s: submitted fence: %u\n", | ||
| 249 | gpu->name, gpu->submitted_fence); | ||
| 242 | queue_work(priv->wq, &gpu->recover_work); | 250 | queue_work(priv->wq, &gpu->recover_work); |
| 243 | } | 251 | } |
| 244 | 252 | ||
| 245 | /* if still more pending work, reset the hangcheck timer: */ | 253 | /* if still more pending work, reset the hangcheck timer: */ |
| 246 | if (gpu->submitted_fence > gpu->hangcheck_fence) | 254 | if (gpu->submitted_fence > gpu->hangcheck_fence) |
| 247 | hangcheck_timer_reset(gpu); | 255 | hangcheck_timer_reset(gpu); |
| 256 | |||
| 257 | /* workaround for missing irq: */ | ||
| 258 | queue_work(priv->wq, &gpu->retire_work); | ||
| 248 | } | 259 | } |
| 249 | 260 | ||
| 250 | /* | 261 | /* |
| @@ -265,7 +276,8 @@ static void retire_worker(struct work_struct *work) | |||
| 265 | obj = list_first_entry(&gpu->active_list, | 276 | obj = list_first_entry(&gpu->active_list, |
| 266 | struct msm_gem_object, mm_list); | 277 | struct msm_gem_object, mm_list); |
| 267 | 278 | ||
| 268 | if (obj->fence <= fence) { | 279 | if ((obj->read_fence <= fence) && |
| 280 | (obj->write_fence <= fence)) { | ||
| 269 | /* move to inactive: */ | 281 | /* move to inactive: */ |
| 270 | msm_gem_move_to_inactive(&obj->base); | 282 | msm_gem_move_to_inactive(&obj->base); |
| 271 | msm_gem_put_iova(&obj->base, gpu->id); | 283 | msm_gem_put_iova(&obj->base, gpu->id); |
| @@ -321,7 +333,11 @@ int msm_gpu_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, | |||
| 321 | submit->gpu->id, &iova); | 333 | submit->gpu->id, &iova); |
| 322 | } | 334 | } |
| 323 | 335 | ||
| 324 | msm_gem_move_to_active(&msm_obj->base, gpu, submit->fence); | 336 | if (submit->bos[i].flags & MSM_SUBMIT_BO_READ) |
| 337 | msm_gem_move_to_active(&msm_obj->base, gpu, false, submit->fence); | ||
| 338 | |||
| 339 | if (submit->bos[i].flags & MSM_SUBMIT_BO_WRITE) | ||
| 340 | msm_gem_move_to_active(&msm_obj->base, gpu, true, submit->fence); | ||
| 325 | } | 341 | } |
| 326 | hangcheck_timer_reset(gpu); | 342 | hangcheck_timer_reset(gpu); |
| 327 | mutex_unlock(&dev->struct_mutex); | 343 | mutex_unlock(&dev->struct_mutex); |
diff --git a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c index 2e11ea02cf87..57cda2a1437b 100644 --- a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c +++ b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c | |||
| @@ -579,8 +579,22 @@ static void | |||
| 579 | init_reserved(struct nvbios_init *init) | 579 | init_reserved(struct nvbios_init *init) |
| 580 | { | 580 | { |
| 581 | u8 opcode = nv_ro08(init->bios, init->offset); | 581 | u8 opcode = nv_ro08(init->bios, init->offset); |
| 582 | trace("RESERVED\t0x%02x\n", opcode); | 582 | u8 length, i; |
| 583 | init->offset += 1; | 583 | |
| 584 | switch (opcode) { | ||
| 585 | case 0xaa: | ||
| 586 | length = 4; | ||
| 587 | break; | ||
| 588 | default: | ||
| 589 | length = 1; | ||
| 590 | break; | ||
| 591 | } | ||
| 592 | |||
| 593 | trace("RESERVED 0x%02x\t", opcode); | ||
| 594 | for (i = 1; i < length; i++) | ||
| 595 | cont(" 0x%02x", nv_ro08(init->bios, init->offset + i)); | ||
| 596 | cont("\n"); | ||
| 597 | init->offset += length; | ||
| 584 | } | 598 | } |
| 585 | 599 | ||
| 586 | /** | 600 | /** |
| @@ -1437,7 +1451,7 @@ init_configure_mem(struct nvbios_init *init) | |||
| 1437 | data = init_rdvgai(init, 0x03c4, 0x01); | 1451 | data = init_rdvgai(init, 0x03c4, 0x01); |
| 1438 | init_wrvgai(init, 0x03c4, 0x01, data | 0x20); | 1452 | init_wrvgai(init, 0x03c4, 0x01, data | 0x20); |
| 1439 | 1453 | ||
| 1440 | while ((addr = nv_ro32(bios, sdata)) != 0xffffffff) { | 1454 | for (; (addr = nv_ro32(bios, sdata)) != 0xffffffff; sdata += 4) { |
| 1441 | switch (addr) { | 1455 | switch (addr) { |
| 1442 | case 0x10021c: /* CKE_NORMAL */ | 1456 | case 0x10021c: /* CKE_NORMAL */ |
| 1443 | case 0x1002d0: /* CMD_REFRESH */ | 1457 | case 0x1002d0: /* CMD_REFRESH */ |
| @@ -2135,6 +2149,7 @@ static struct nvbios_init_opcode { | |||
| 2135 | [0x99] = { init_zm_auxch }, | 2149 | [0x99] = { init_zm_auxch }, |
| 2136 | [0x9a] = { init_i2c_long_if }, | 2150 | [0x9a] = { init_i2c_long_if }, |
| 2137 | [0xa9] = { init_gpio_ne }, | 2151 | [0xa9] = { init_gpio_ne }, |
| 2152 | [0xaa] = { init_reserved }, | ||
| 2138 | }; | 2153 | }; |
| 2139 | 2154 | ||
| 2140 | #define init_opcode_nr (sizeof(init_opcode) / sizeof(init_opcode[0])) | 2155 | #define init_opcode_nr (sizeof(init_opcode) / sizeof(init_opcode[0])) |
diff --git a/drivers/gpu/drm/nouveau/nouveau_display.c b/drivers/gpu/drm/nouveau/nouveau_display.c index d2712e6e5d31..7848590f5568 100644 --- a/drivers/gpu/drm/nouveau/nouveau_display.c +++ b/drivers/gpu/drm/nouveau/nouveau_display.c | |||
| @@ -278,7 +278,6 @@ nouveau_display_create(struct drm_device *dev) | |||
| 278 | { | 278 | { |
| 279 | struct nouveau_drm *drm = nouveau_drm(dev); | 279 | struct nouveau_drm *drm = nouveau_drm(dev); |
| 280 | struct nouveau_display *disp; | 280 | struct nouveau_display *disp; |
| 281 | u32 pclass = dev->pdev->class >> 8; | ||
| 282 | int ret, gen; | 281 | int ret, gen; |
| 283 | 282 | ||
| 284 | disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL); | 283 | disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL); |
| @@ -340,29 +339,25 @@ nouveau_display_create(struct drm_device *dev) | |||
| 340 | drm_kms_helper_poll_init(dev); | 339 | drm_kms_helper_poll_init(dev); |
| 341 | drm_kms_helper_poll_disable(dev); | 340 | drm_kms_helper_poll_disable(dev); |
| 342 | 341 | ||
| 343 | if (nouveau_modeset == 1 || | 342 | if (drm->vbios.dcb.entries) { |
| 344 | (nouveau_modeset < 0 && pclass == PCI_CLASS_DISPLAY_VGA)) { | 343 | if (nv_device(drm->device)->card_type < NV_50) |
| 345 | if (drm->vbios.dcb.entries) { | 344 | ret = nv04_display_create(dev); |
| 346 | if (nv_device(drm->device)->card_type < NV_50) | 345 | else |
| 347 | ret = nv04_display_create(dev); | 346 | ret = nv50_display_create(dev); |
| 348 | else | 347 | } else { |
| 349 | ret = nv50_display_create(dev); | 348 | ret = 0; |
| 350 | } else { | 349 | } |
| 351 | ret = 0; | ||
| 352 | } | ||
| 353 | |||
| 354 | if (ret) | ||
| 355 | goto disp_create_err; | ||
| 356 | 350 | ||
| 357 | if (dev->mode_config.num_crtc) { | 351 | if (ret) |
| 358 | ret = drm_vblank_init(dev, dev->mode_config.num_crtc); | 352 | goto disp_create_err; |
| 359 | if (ret) | ||
| 360 | goto vblank_err; | ||
| 361 | } | ||
| 362 | 353 | ||
| 363 | nouveau_backlight_init(dev); | 354 | if (dev->mode_config.num_crtc) { |
| 355 | ret = drm_vblank_init(dev, dev->mode_config.num_crtc); | ||
| 356 | if (ret) | ||
| 357 | goto vblank_err; | ||
| 364 | } | 358 | } |
| 365 | 359 | ||
| 360 | nouveau_backlight_init(dev); | ||
| 366 | return 0; | 361 | return 0; |
| 367 | 362 | ||
| 368 | vblank_err: | 363 | vblank_err: |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fbcon.c b/drivers/gpu/drm/nouveau/nouveau_fbcon.c index 8f6d63d7edd3..a86ecf65c164 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fbcon.c +++ b/drivers/gpu/drm/nouveau/nouveau_fbcon.c | |||
| @@ -454,7 +454,8 @@ nouveau_fbcon_init(struct drm_device *dev) | |||
| 454 | int preferred_bpp; | 454 | int preferred_bpp; |
| 455 | int ret; | 455 | int ret; |
| 456 | 456 | ||
| 457 | if (!dev->mode_config.num_crtc) | 457 | if (!dev->mode_config.num_crtc || |
| 458 | (dev->pdev->class >> 8) != PCI_CLASS_DISPLAY_VGA) | ||
| 458 | return 0; | 459 | return 0; |
| 459 | 460 | ||
| 460 | fbcon = kzalloc(sizeof(struct nouveau_fbdev), GFP_KERNEL); | 461 | fbcon = kzalloc(sizeof(struct nouveau_fbdev), GFP_KERNEL); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/drivers/gpu/drm/nouveau/nouveau_sgdma.c index ca5492ac2da5..0843ebc910d4 100644 --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c | |||
| @@ -104,9 +104,7 @@ nouveau_sgdma_create_ttm(struct ttm_bo_device *bdev, | |||
| 104 | else | 104 | else |
| 105 | nvbe->ttm.ttm.func = &nv50_sgdma_backend; | 105 | nvbe->ttm.ttm.func = &nv50_sgdma_backend; |
| 106 | 106 | ||
| 107 | if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) { | 107 | if (ttm_dma_tt_init(&nvbe->ttm, bdev, size, page_flags, dummy_read_page)) |
| 108 | kfree(nvbe); | ||
| 109 | return NULL; | 108 | return NULL; |
| 110 | } | ||
| 111 | return &nvbe->ttm.ttm; | 109 | return &nvbe->ttm.ttm; |
| 112 | } | 110 | } |
diff --git a/drivers/gpu/drm/radeon/atombios_encoders.c b/drivers/gpu/drm/radeon/atombios_encoders.c index dfac7965ea28..32923d2f6002 100644 --- a/drivers/gpu/drm/radeon/atombios_encoders.c +++ b/drivers/gpu/drm/radeon/atombios_encoders.c | |||
| @@ -707,8 +707,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 707 | switch (connector->connector_type) { | 707 | switch (connector->connector_type) { |
| 708 | case DRM_MODE_CONNECTOR_DVII: | 708 | case DRM_MODE_CONNECTOR_DVII: |
| 709 | case DRM_MODE_CONNECTOR_HDMIB: /* HDMI-B is basically DL-DVI; analog works fine */ | 709 | case DRM_MODE_CONNECTOR_HDMIB: /* HDMI-B is basically DL-DVI; analog works fine */ |
| 710 | if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 710 | if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
| 711 | radeon_audio) | 711 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
| 712 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
| 712 | return ATOM_ENCODER_MODE_HDMI; | 713 | return ATOM_ENCODER_MODE_HDMI; |
| 713 | else if (radeon_connector->use_digital) | 714 | else if (radeon_connector->use_digital) |
| 714 | return ATOM_ENCODER_MODE_DVI; | 715 | return ATOM_ENCODER_MODE_DVI; |
| @@ -718,8 +719,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 718 | case DRM_MODE_CONNECTOR_DVID: | 719 | case DRM_MODE_CONNECTOR_DVID: |
| 719 | case DRM_MODE_CONNECTOR_HDMIA: | 720 | case DRM_MODE_CONNECTOR_HDMIA: |
| 720 | default: | 721 | default: |
| 721 | if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 722 | if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
| 722 | radeon_audio) | 723 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
| 724 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
| 723 | return ATOM_ENCODER_MODE_HDMI; | 725 | return ATOM_ENCODER_MODE_HDMI; |
| 724 | else | 726 | else |
| 725 | return ATOM_ENCODER_MODE_DVI; | 727 | return ATOM_ENCODER_MODE_DVI; |
| @@ -732,8 +734,9 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 732 | if ((dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) || | 734 | if ((dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) || |
| 733 | (dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_eDP)) | 735 | (dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_eDP)) |
| 734 | return ATOM_ENCODER_MODE_DP; | 736 | return ATOM_ENCODER_MODE_DP; |
| 735 | else if (drm_detect_hdmi_monitor(radeon_connector->edid) && | 737 | else if ((radeon_connector->audio == RADEON_AUDIO_ENABLE) || |
| 736 | radeon_audio) | 738 | (drm_detect_hdmi_monitor(radeon_connector->edid) && |
| 739 | (radeon_connector->audio == RADEON_AUDIO_AUTO))) | ||
| 737 | return ATOM_ENCODER_MODE_HDMI; | 740 | return ATOM_ENCODER_MODE_HDMI; |
| 738 | else | 741 | else |
| 739 | return ATOM_ENCODER_MODE_DVI; | 742 | return ATOM_ENCODER_MODE_DVI; |
| @@ -1647,8 +1650,12 @@ radeon_atom_encoder_dpms_dig(struct drm_encoder *encoder, int mode) | |||
| 1647 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE, 0); | 1650 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE, 0); |
| 1648 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP, 0, 0); | 1651 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP, 0, 0); |
| 1649 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE, 0, 0); | 1652 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE, 0, 0); |
| 1650 | /* some early dce3.2 boards have a bug in their transmitter control table */ | 1653 | /* some dce3.x boards have a bug in their transmitter control table. |
| 1651 | if ((rdev->family != CHIP_RV710) && (rdev->family != CHIP_RV730)) | 1654 | * ACTION_ENABLE_OUTPUT can probably be dropped since ACTION_ENABLE |
| 1655 | * does the same thing and more. | ||
| 1656 | */ | ||
| 1657 | if ((rdev->family != CHIP_RV710) && (rdev->family != CHIP_RV730) && | ||
| 1658 | (rdev->family != CHIP_RS880)) | ||
| 1652 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE_OUTPUT, 0, 0); | 1659 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE_OUTPUT, 0, 0); |
| 1653 | } | 1660 | } |
| 1654 | if (ENCODER_MODE_IS_DP(atombios_get_encoder_mode(encoder)) && connector) { | 1661 | if (ENCODER_MODE_IS_DP(atombios_get_encoder_mode(encoder)) && connector) { |
diff --git a/drivers/gpu/drm/radeon/btc_dpm.c b/drivers/gpu/drm/radeon/btc_dpm.c index 084e69414fd1..b162e98a2953 100644 --- a/drivers/gpu/drm/radeon/btc_dpm.c +++ b/drivers/gpu/drm/radeon/btc_dpm.c | |||
| @@ -1168,6 +1168,23 @@ static const struct radeon_blacklist_clocks btc_blacklist_clocks[] = | |||
| 1168 | { 25000, 30000, RADEON_SCLK_UP } | 1168 | { 25000, 30000, RADEON_SCLK_UP } |
| 1169 | }; | 1169 | }; |
| 1170 | 1170 | ||
| 1171 | void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
| 1172 | u32 *max_clock) | ||
| 1173 | { | ||
| 1174 | u32 i, clock = 0; | ||
| 1175 | |||
| 1176 | if ((table == NULL) || (table->count == 0)) { | ||
| 1177 | *max_clock = clock; | ||
| 1178 | return; | ||
| 1179 | } | ||
| 1180 | |||
| 1181 | for (i = 0; i < table->count; i++) { | ||
| 1182 | if (clock < table->entries[i].clk) | ||
| 1183 | clock = table->entries[i].clk; | ||
| 1184 | } | ||
| 1185 | *max_clock = clock; | ||
| 1186 | } | ||
| 1187 | |||
| 1171 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, | 1188 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, |
| 1172 | u32 clock, u16 max_voltage, u16 *voltage) | 1189 | u32 clock, u16 max_voltage, u16 *voltage) |
| 1173 | { | 1190 | { |
| @@ -2080,6 +2097,7 @@ static void btc_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 2080 | bool disable_mclk_switching; | 2097 | bool disable_mclk_switching; |
| 2081 | u32 mclk, sclk; | 2098 | u32 mclk, sclk; |
| 2082 | u16 vddc, vddci; | 2099 | u16 vddc, vddci; |
| 2100 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
| 2083 | 2101 | ||
| 2084 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 2102 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
| 2085 | btc_dpm_vblank_too_short(rdev)) | 2103 | btc_dpm_vblank_too_short(rdev)) |
| @@ -2121,6 +2139,39 @@ static void btc_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 2121 | ps->low.vddci = max_limits->vddci; | 2139 | ps->low.vddci = max_limits->vddci; |
| 2122 | } | 2140 | } |
| 2123 | 2141 | ||
| 2142 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
| 2143 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
| 2144 | &max_sclk_vddc); | ||
| 2145 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
| 2146 | &max_mclk_vddci); | ||
| 2147 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
| 2148 | &max_mclk_vddc); | ||
| 2149 | |||
| 2150 | if (max_sclk_vddc) { | ||
| 2151 | if (ps->low.sclk > max_sclk_vddc) | ||
| 2152 | ps->low.sclk = max_sclk_vddc; | ||
| 2153 | if (ps->medium.sclk > max_sclk_vddc) | ||
| 2154 | ps->medium.sclk = max_sclk_vddc; | ||
| 2155 | if (ps->high.sclk > max_sclk_vddc) | ||
| 2156 | ps->high.sclk = max_sclk_vddc; | ||
| 2157 | } | ||
| 2158 | if (max_mclk_vddci) { | ||
| 2159 | if (ps->low.mclk > max_mclk_vddci) | ||
| 2160 | ps->low.mclk = max_mclk_vddci; | ||
| 2161 | if (ps->medium.mclk > max_mclk_vddci) | ||
| 2162 | ps->medium.mclk = max_mclk_vddci; | ||
| 2163 | if (ps->high.mclk > max_mclk_vddci) | ||
| 2164 | ps->high.mclk = max_mclk_vddci; | ||
| 2165 | } | ||
| 2166 | if (max_mclk_vddc) { | ||
| 2167 | if (ps->low.mclk > max_mclk_vddc) | ||
| 2168 | ps->low.mclk = max_mclk_vddc; | ||
| 2169 | if (ps->medium.mclk > max_mclk_vddc) | ||
| 2170 | ps->medium.mclk = max_mclk_vddc; | ||
| 2171 | if (ps->high.mclk > max_mclk_vddc) | ||
| 2172 | ps->high.mclk = max_mclk_vddc; | ||
| 2173 | } | ||
| 2174 | |||
| 2124 | /* XXX validate the min clocks required for display */ | 2175 | /* XXX validate the min clocks required for display */ |
| 2125 | 2176 | ||
| 2126 | if (disable_mclk_switching) { | 2177 | if (disable_mclk_switching) { |
| @@ -2340,12 +2391,6 @@ int btc_dpm_set_power_state(struct radeon_device *rdev) | |||
| 2340 | return ret; | 2391 | return ret; |
| 2341 | } | 2392 | } |
| 2342 | 2393 | ||
| 2343 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 2344 | if (ret) { | ||
| 2345 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
| 2346 | return ret; | ||
| 2347 | } | ||
| 2348 | |||
| 2349 | return 0; | 2394 | return 0; |
| 2350 | } | 2395 | } |
| 2351 | 2396 | ||
diff --git a/drivers/gpu/drm/radeon/btc_dpm.h b/drivers/gpu/drm/radeon/btc_dpm.h index 1a15e0e41950..3b6f12b7760b 100644 --- a/drivers/gpu/drm/radeon/btc_dpm.h +++ b/drivers/gpu/drm/radeon/btc_dpm.h | |||
| @@ -46,6 +46,8 @@ void btc_adjust_clock_combinations(struct radeon_device *rdev, | |||
| 46 | struct rv7xx_pl *pl); | 46 | struct rv7xx_pl *pl); |
| 47 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, | 47 | void btc_apply_voltage_dependency_rules(struct radeon_clock_voltage_dependency_table *table, |
| 48 | u32 clock, u16 max_voltage, u16 *voltage); | 48 | u32 clock, u16 max_voltage, u16 *voltage); |
| 49 | void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
| 50 | u32 *max_clock); | ||
| 49 | void btc_apply_voltage_delta_rules(struct radeon_device *rdev, | 51 | void btc_apply_voltage_delta_rules(struct radeon_device *rdev, |
| 50 | u16 max_vddc, u16 max_vddci, | 52 | u16 max_vddc, u16 max_vddci, |
| 51 | u16 *vddc, u16 *vddci); | 53 | u16 *vddc, u16 *vddci); |
diff --git a/drivers/gpu/drm/radeon/ci_dpm.c b/drivers/gpu/drm/radeon/ci_dpm.c index 3cce533397c6..51e947a97edf 100644 --- a/drivers/gpu/drm/radeon/ci_dpm.c +++ b/drivers/gpu/drm/radeon/ci_dpm.c | |||
| @@ -146,6 +146,8 @@ static const struct ci_pt_config_reg didt_config_ci[] = | |||
| 146 | }; | 146 | }; |
| 147 | 147 | ||
| 148 | extern u8 rv770_get_memory_module_index(struct radeon_device *rdev); | 148 | extern u8 rv770_get_memory_module_index(struct radeon_device *rdev); |
| 149 | extern void btc_get_max_clock_from_voltage_dependency_table(struct radeon_clock_voltage_dependency_table *table, | ||
| 150 | u32 *max_clock); | ||
| 149 | extern int ni_copy_and_switch_arb_sets(struct radeon_device *rdev, | 151 | extern int ni_copy_and_switch_arb_sets(struct radeon_device *rdev, |
| 150 | u32 arb_freq_src, u32 arb_freq_dest); | 152 | u32 arb_freq_src, u32 arb_freq_dest); |
| 151 | extern u8 si_get_ddr3_mclk_frequency_ratio(u32 memory_clock); | 153 | extern u8 si_get_ddr3_mclk_frequency_ratio(u32 memory_clock); |
| @@ -712,6 +714,7 @@ static void ci_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 712 | struct radeon_clock_and_voltage_limits *max_limits; | 714 | struct radeon_clock_and_voltage_limits *max_limits; |
| 713 | bool disable_mclk_switching; | 715 | bool disable_mclk_switching; |
| 714 | u32 sclk, mclk; | 716 | u32 sclk, mclk; |
| 717 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
| 715 | int i; | 718 | int i; |
| 716 | 719 | ||
| 717 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 720 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
| @@ -739,6 +742,29 @@ static void ci_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 739 | } | 742 | } |
| 740 | } | 743 | } |
| 741 | 744 | ||
| 745 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
| 746 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
| 747 | &max_sclk_vddc); | ||
| 748 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
| 749 | &max_mclk_vddci); | ||
| 750 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
| 751 | &max_mclk_vddc); | ||
| 752 | |||
| 753 | for (i = 0; i < ps->performance_level_count; i++) { | ||
| 754 | if (max_sclk_vddc) { | ||
| 755 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
| 756 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
| 757 | } | ||
| 758 | if (max_mclk_vddci) { | ||
| 759 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
| 760 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
| 761 | } | ||
| 762 | if (max_mclk_vddc) { | ||
| 763 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
| 764 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
| 765 | } | ||
| 766 | } | ||
| 767 | |||
| 742 | /* XXX validate the min clocks required for display */ | 768 | /* XXX validate the min clocks required for display */ |
| 743 | 769 | ||
| 744 | if (disable_mclk_switching) { | 770 | if (disable_mclk_switching) { |
| @@ -4748,12 +4774,6 @@ int ci_dpm_set_power_state(struct radeon_device *rdev) | |||
| 4748 | if (pi->pcie_performance_request) | 4774 | if (pi->pcie_performance_request) |
| 4749 | ci_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); | 4775 | ci_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); |
| 4750 | 4776 | ||
| 4751 | ret = ci_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 4752 | if (ret) { | ||
| 4753 | DRM_ERROR("ci_dpm_force_performance_level failed\n"); | ||
| 4754 | return ret; | ||
| 4755 | } | ||
| 4756 | |||
| 4757 | cik_update_cg(rdev, (RADEON_CG_BLOCK_GFX | | 4777 | cik_update_cg(rdev, (RADEON_CG_BLOCK_GFX | |
| 4758 | RADEON_CG_BLOCK_MC | | 4778 | RADEON_CG_BLOCK_MC | |
| 4759 | RADEON_CG_BLOCK_SDMA | | 4779 | RADEON_CG_BLOCK_SDMA | |
diff --git a/drivers/gpu/drm/radeon/ci_smc.c b/drivers/gpu/drm/radeon/ci_smc.c index 53b43dd3cf1e..252e10a41cf5 100644 --- a/drivers/gpu/drm/radeon/ci_smc.c +++ b/drivers/gpu/drm/radeon/ci_smc.c | |||
| @@ -47,10 +47,11 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 47 | u32 smc_start_address, | 47 | u32 smc_start_address, |
| 48 | const u8 *src, u32 byte_count, u32 limit) | 48 | const u8 *src, u32 byte_count, u32 limit) |
| 49 | { | 49 | { |
| 50 | unsigned long flags; | ||
| 50 | u32 data, original_data; | 51 | u32 data, original_data; |
| 51 | u32 addr; | 52 | u32 addr; |
| 52 | u32 extra_shift; | 53 | u32 extra_shift; |
| 53 | int ret; | 54 | int ret = 0; |
| 54 | 55 | ||
| 55 | if (smc_start_address & 3) | 56 | if (smc_start_address & 3) |
| 56 | return -EINVAL; | 57 | return -EINVAL; |
| @@ -59,13 +60,14 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 59 | 60 | ||
| 60 | addr = smc_start_address; | 61 | addr = smc_start_address; |
| 61 | 62 | ||
| 63 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 62 | while (byte_count >= 4) { | 64 | while (byte_count >= 4) { |
| 63 | /* SMC address space is BE */ | 65 | /* SMC address space is BE */ |
| 64 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 66 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
| 65 | 67 | ||
| 66 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 68 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
| 67 | if (ret) | 69 | if (ret) |
| 68 | return ret; | 70 | goto done; |
| 69 | 71 | ||
| 70 | WREG32(SMC_IND_DATA_0, data); | 72 | WREG32(SMC_IND_DATA_0, data); |
| 71 | 73 | ||
| @@ -80,7 +82,7 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 80 | 82 | ||
| 81 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 83 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
| 82 | if (ret) | 84 | if (ret) |
| 83 | return ret; | 85 | goto done; |
| 84 | 86 | ||
| 85 | original_data = RREG32(SMC_IND_DATA_0); | 87 | original_data = RREG32(SMC_IND_DATA_0); |
| 86 | 88 | ||
| @@ -97,11 +99,15 @@ int ci_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 97 | 99 | ||
| 98 | ret = ci_set_smc_sram_address(rdev, addr, limit); | 100 | ret = ci_set_smc_sram_address(rdev, addr, limit); |
| 99 | if (ret) | 101 | if (ret) |
| 100 | return ret; | 102 | goto done; |
| 101 | 103 | ||
| 102 | WREG32(SMC_IND_DATA_0, data); | 104 | WREG32(SMC_IND_DATA_0, data); |
| 103 | } | 105 | } |
| 104 | return 0; | 106 | |
| 107 | done: | ||
| 108 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 109 | |||
| 110 | return ret; | ||
| 105 | } | 111 | } |
| 106 | 112 | ||
| 107 | void ci_start_smc(struct radeon_device *rdev) | 113 | void ci_start_smc(struct radeon_device *rdev) |
| @@ -197,6 +203,7 @@ PPSMC_Result ci_wait_for_smc_inactive(struct radeon_device *rdev) | |||
| 197 | 203 | ||
| 198 | int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | 204 | int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) |
| 199 | { | 205 | { |
| 206 | unsigned long flags; | ||
| 200 | u32 ucode_start_address; | 207 | u32 ucode_start_address; |
| 201 | u32 ucode_size; | 208 | u32 ucode_size; |
| 202 | const u8 *src; | 209 | const u8 *src; |
| @@ -219,6 +226,7 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 219 | return -EINVAL; | 226 | return -EINVAL; |
| 220 | 227 | ||
| 221 | src = (const u8 *)rdev->smc_fw->data; | 228 | src = (const u8 *)rdev->smc_fw->data; |
| 229 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 222 | WREG32(SMC_IND_INDEX_0, ucode_start_address); | 230 | WREG32(SMC_IND_INDEX_0, ucode_start_address); |
| 223 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); | 231 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); |
| 224 | while (ucode_size >= 4) { | 232 | while (ucode_size >= 4) { |
| @@ -231,6 +239,7 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 231 | ucode_size -= 4; | 239 | ucode_size -= 4; |
| 232 | } | 240 | } |
| 233 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); | 241 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); |
| 242 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 234 | 243 | ||
| 235 | return 0; | 244 | return 0; |
| 236 | } | 245 | } |
| @@ -238,25 +247,29 @@ int ci_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 238 | int ci_read_smc_sram_dword(struct radeon_device *rdev, | 247 | int ci_read_smc_sram_dword(struct radeon_device *rdev, |
| 239 | u32 smc_address, u32 *value, u32 limit) | 248 | u32 smc_address, u32 *value, u32 limit) |
| 240 | { | 249 | { |
| 250 | unsigned long flags; | ||
| 241 | int ret; | 251 | int ret; |
| 242 | 252 | ||
| 253 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 243 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); | 254 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); |
| 244 | if (ret) | 255 | if (ret == 0) |
| 245 | return ret; | 256 | *value = RREG32(SMC_IND_DATA_0); |
| 257 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 246 | 258 | ||
| 247 | *value = RREG32(SMC_IND_DATA_0); | 259 | return ret; |
| 248 | return 0; | ||
| 249 | } | 260 | } |
| 250 | 261 | ||
| 251 | int ci_write_smc_sram_dword(struct radeon_device *rdev, | 262 | int ci_write_smc_sram_dword(struct radeon_device *rdev, |
| 252 | u32 smc_address, u32 value, u32 limit) | 263 | u32 smc_address, u32 value, u32 limit) |
| 253 | { | 264 | { |
| 265 | unsigned long flags; | ||
| 254 | int ret; | 266 | int ret; |
| 255 | 267 | ||
| 268 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 256 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); | 269 | ret = ci_set_smc_sram_address(rdev, smc_address, limit); |
| 257 | if (ret) | 270 | if (ret == 0) |
| 258 | return ret; | 271 | WREG32(SMC_IND_DATA_0, value); |
| 272 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 259 | 273 | ||
| 260 | WREG32(SMC_IND_DATA_0, value); | 274 | return ret; |
| 261 | return 0; | ||
| 262 | } | 275 | } |
diff --git a/drivers/gpu/drm/radeon/cik.c b/drivers/gpu/drm/radeon/cik.c index a3bba0587276..d02fd1c045d5 100644 --- a/drivers/gpu/drm/radeon/cik.c +++ b/drivers/gpu/drm/radeon/cik.c | |||
| @@ -77,6 +77,8 @@ static void cik_pcie_gen3_enable(struct radeon_device *rdev); | |||
| 77 | static void cik_program_aspm(struct radeon_device *rdev); | 77 | static void cik_program_aspm(struct radeon_device *rdev); |
| 78 | static void cik_init_pg(struct radeon_device *rdev); | 78 | static void cik_init_pg(struct radeon_device *rdev); |
| 79 | static void cik_init_cg(struct radeon_device *rdev); | 79 | static void cik_init_cg(struct radeon_device *rdev); |
| 80 | static void cik_enable_gui_idle_interrupt(struct radeon_device *rdev, | ||
| 81 | bool enable); | ||
| 80 | 82 | ||
| 81 | /* get temperature in millidegrees */ | 83 | /* get temperature in millidegrees */ |
| 82 | int ci_get_temp(struct radeon_device *rdev) | 84 | int ci_get_temp(struct radeon_device *rdev) |
| @@ -120,20 +122,27 @@ int kv_get_temp(struct radeon_device *rdev) | |||
| 120 | */ | 122 | */ |
| 121 | u32 cik_pciep_rreg(struct radeon_device *rdev, u32 reg) | 123 | u32 cik_pciep_rreg(struct radeon_device *rdev, u32 reg) |
| 122 | { | 124 | { |
| 125 | unsigned long flags; | ||
| 123 | u32 r; | 126 | u32 r; |
| 124 | 127 | ||
| 128 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
| 125 | WREG32(PCIE_INDEX, reg); | 129 | WREG32(PCIE_INDEX, reg); |
| 126 | (void)RREG32(PCIE_INDEX); | 130 | (void)RREG32(PCIE_INDEX); |
| 127 | r = RREG32(PCIE_DATA); | 131 | r = RREG32(PCIE_DATA); |
| 132 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
| 128 | return r; | 133 | return r; |
| 129 | } | 134 | } |
| 130 | 135 | ||
| 131 | void cik_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 136 | void cik_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 132 | { | 137 | { |
| 138 | unsigned long flags; | ||
| 139 | |||
| 140 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
| 133 | WREG32(PCIE_INDEX, reg); | 141 | WREG32(PCIE_INDEX, reg); |
| 134 | (void)RREG32(PCIE_INDEX); | 142 | (void)RREG32(PCIE_INDEX); |
| 135 | WREG32(PCIE_DATA, v); | 143 | WREG32(PCIE_DATA, v); |
| 136 | (void)RREG32(PCIE_DATA); | 144 | (void)RREG32(PCIE_DATA); |
| 145 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
| 137 | } | 146 | } |
| 138 | 147 | ||
| 139 | static const u32 spectre_rlc_save_restore_register_list[] = | 148 | static const u32 spectre_rlc_save_restore_register_list[] = |
| @@ -2722,7 +2731,8 @@ static void cik_gpu_init(struct radeon_device *rdev) | |||
| 2722 | } else if ((rdev->pdev->device == 0x1309) || | 2731 | } else if ((rdev->pdev->device == 0x1309) || |
| 2723 | (rdev->pdev->device == 0x130A) || | 2732 | (rdev->pdev->device == 0x130A) || |
| 2724 | (rdev->pdev->device == 0x130D) || | 2733 | (rdev->pdev->device == 0x130D) || |
| 2725 | (rdev->pdev->device == 0x1313)) { | 2734 | (rdev->pdev->device == 0x1313) || |
| 2735 | (rdev->pdev->device == 0x131D)) { | ||
| 2726 | rdev->config.cik.max_cu_per_sh = 6; | 2736 | rdev->config.cik.max_cu_per_sh = 6; |
| 2727 | rdev->config.cik.max_backends_per_se = 2; | 2737 | rdev->config.cik.max_backends_per_se = 2; |
| 2728 | } else if ((rdev->pdev->device == 0x1306) || | 2738 | } else if ((rdev->pdev->device == 0x1306) || |
| @@ -2835,10 +2845,8 @@ static void cik_gpu_init(struct radeon_device *rdev) | |||
| 2835 | rdev->config.cik.tile_config |= (3 << 0); | 2845 | rdev->config.cik.tile_config |= (3 << 0); |
| 2836 | break; | 2846 | break; |
| 2837 | } | 2847 | } |
| 2838 | if ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) | 2848 | rdev->config.cik.tile_config |= |
| 2839 | rdev->config.cik.tile_config |= 1 << 4; | 2849 | ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) << 4; |
| 2840 | else | ||
| 2841 | rdev->config.cik.tile_config |= 0 << 4; | ||
| 2842 | rdev->config.cik.tile_config |= | 2850 | rdev->config.cik.tile_config |= |
| 2843 | ((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8; | 2851 | ((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8; |
| 2844 | rdev->config.cik.tile_config |= | 2852 | rdev->config.cik.tile_config |= |
| @@ -4013,6 +4021,8 @@ static int cik_cp_resume(struct radeon_device *rdev) | |||
| 4013 | { | 4021 | { |
| 4014 | int r; | 4022 | int r; |
| 4015 | 4023 | ||
| 4024 | cik_enable_gui_idle_interrupt(rdev, false); | ||
| 4025 | |||
| 4016 | r = cik_cp_load_microcode(rdev); | 4026 | r = cik_cp_load_microcode(rdev); |
| 4017 | if (r) | 4027 | if (r) |
| 4018 | return r; | 4028 | return r; |
| @@ -4024,6 +4034,8 @@ static int cik_cp_resume(struct radeon_device *rdev) | |||
| 4024 | if (r) | 4034 | if (r) |
| 4025 | return r; | 4035 | return r; |
| 4026 | 4036 | ||
| 4037 | cik_enable_gui_idle_interrupt(rdev, true); | ||
| 4038 | |||
| 4027 | return 0; | 4039 | return 0; |
| 4028 | } | 4040 | } |
| 4029 | 4041 | ||
| @@ -4442,8 +4454,8 @@ static int cik_mc_init(struct radeon_device *rdev) | |||
| 4442 | rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0); | 4454 | rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0); |
| 4443 | rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0); | 4455 | rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0); |
| 4444 | /* size in MB on si */ | 4456 | /* size in MB on si */ |
| 4445 | rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024 * 1024; | 4457 | rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL; |
| 4446 | rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024 * 1024; | 4458 | rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL; |
| 4447 | rdev->mc.visible_vram_size = rdev->mc.aper_size; | 4459 | rdev->mc.visible_vram_size = rdev->mc.aper_size; |
| 4448 | si_vram_gtt_location(rdev, &rdev->mc); | 4460 | si_vram_gtt_location(rdev, &rdev->mc); |
| 4449 | radeon_update_bandwidth_info(rdev); | 4461 | radeon_update_bandwidth_info(rdev); |
| @@ -4721,12 +4733,13 @@ static void cik_vm_decode_fault(struct radeon_device *rdev, | |||
| 4721 | u32 mc_id = (status & MEMORY_CLIENT_ID_MASK) >> MEMORY_CLIENT_ID_SHIFT; | 4733 | u32 mc_id = (status & MEMORY_CLIENT_ID_MASK) >> MEMORY_CLIENT_ID_SHIFT; |
| 4722 | u32 vmid = (status & FAULT_VMID_MASK) >> FAULT_VMID_SHIFT; | 4734 | u32 vmid = (status & FAULT_VMID_MASK) >> FAULT_VMID_SHIFT; |
| 4723 | u32 protections = (status & PROTECTIONS_MASK) >> PROTECTIONS_SHIFT; | 4735 | u32 protections = (status & PROTECTIONS_MASK) >> PROTECTIONS_SHIFT; |
| 4724 | char *block = (char *)&mc_client; | 4736 | char block[5] = { mc_client >> 24, (mc_client >> 16) & 0xff, |
| 4737 | (mc_client >> 8) & 0xff, mc_client & 0xff, 0 }; | ||
| 4725 | 4738 | ||
| 4726 | printk("VM fault (0x%02x, vmid %d) at page %u, %s from %s (%d)\n", | 4739 | printk("VM fault (0x%02x, vmid %d) at page %u, %s from '%s' (0x%08x) (%d)\n", |
| 4727 | protections, vmid, addr, | 4740 | protections, vmid, addr, |
| 4728 | (status & MEMORY_CLIENT_RW_MASK) ? "write" : "read", | 4741 | (status & MEMORY_CLIENT_RW_MASK) ? "write" : "read", |
| 4729 | block, mc_id); | 4742 | block, mc_client, mc_id); |
| 4730 | } | 4743 | } |
| 4731 | 4744 | ||
| 4732 | /** | 4745 | /** |
| @@ -5376,7 +5389,9 @@ static void cik_enable_hdp_ls(struct radeon_device *rdev, | |||
| 5376 | void cik_update_cg(struct radeon_device *rdev, | 5389 | void cik_update_cg(struct radeon_device *rdev, |
| 5377 | u32 block, bool enable) | 5390 | u32 block, bool enable) |
| 5378 | { | 5391 | { |
| 5392 | |||
| 5379 | if (block & RADEON_CG_BLOCK_GFX) { | 5393 | if (block & RADEON_CG_BLOCK_GFX) { |
| 5394 | cik_enable_gui_idle_interrupt(rdev, false); | ||
| 5380 | /* order matters! */ | 5395 | /* order matters! */ |
| 5381 | if (enable) { | 5396 | if (enable) { |
| 5382 | cik_enable_mgcg(rdev, true); | 5397 | cik_enable_mgcg(rdev, true); |
| @@ -5385,6 +5400,7 @@ void cik_update_cg(struct radeon_device *rdev, | |||
| 5385 | cik_enable_cgcg(rdev, false); | 5400 | cik_enable_cgcg(rdev, false); |
| 5386 | cik_enable_mgcg(rdev, false); | 5401 | cik_enable_mgcg(rdev, false); |
| 5387 | } | 5402 | } |
| 5403 | cik_enable_gui_idle_interrupt(rdev, true); | ||
| 5388 | } | 5404 | } |
| 5389 | 5405 | ||
| 5390 | if (block & RADEON_CG_BLOCK_MC) { | 5406 | if (block & RADEON_CG_BLOCK_MC) { |
| @@ -5541,7 +5557,7 @@ static void cik_enable_gfx_cgpg(struct radeon_device *rdev, | |||
| 5541 | { | 5557 | { |
| 5542 | u32 data, orig; | 5558 | u32 data, orig; |
| 5543 | 5559 | ||
| 5544 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG)) { | 5560 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG)) { |
| 5545 | orig = data = RREG32(RLC_PG_CNTL); | 5561 | orig = data = RREG32(RLC_PG_CNTL); |
| 5546 | data |= GFX_PG_ENABLE; | 5562 | data |= GFX_PG_ENABLE; |
| 5547 | if (orig != data) | 5563 | if (orig != data) |
| @@ -5805,7 +5821,7 @@ static void cik_init_pg(struct radeon_device *rdev) | |||
| 5805 | if (rdev->pg_flags) { | 5821 | if (rdev->pg_flags) { |
| 5806 | cik_enable_sck_slowdown_on_pu(rdev, true); | 5822 | cik_enable_sck_slowdown_on_pu(rdev, true); |
| 5807 | cik_enable_sck_slowdown_on_pd(rdev, true); | 5823 | cik_enable_sck_slowdown_on_pd(rdev, true); |
| 5808 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5824 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
| 5809 | cik_init_gfx_cgpg(rdev); | 5825 | cik_init_gfx_cgpg(rdev); |
| 5810 | cik_enable_cp_pg(rdev, true); | 5826 | cik_enable_cp_pg(rdev, true); |
| 5811 | cik_enable_gds_pg(rdev, true); | 5827 | cik_enable_gds_pg(rdev, true); |
| @@ -5819,7 +5835,7 @@ static void cik_fini_pg(struct radeon_device *rdev) | |||
| 5819 | { | 5835 | { |
| 5820 | if (rdev->pg_flags) { | 5836 | if (rdev->pg_flags) { |
| 5821 | cik_update_gfx_pg(rdev, false); | 5837 | cik_update_gfx_pg(rdev, false); |
| 5822 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5838 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
| 5823 | cik_enable_cp_pg(rdev, false); | 5839 | cik_enable_cp_pg(rdev, false); |
| 5824 | cik_enable_gds_pg(rdev, false); | 5840 | cik_enable_gds_pg(rdev, false); |
| 5825 | } | 5841 | } |
| @@ -5895,7 +5911,9 @@ static void cik_disable_interrupt_state(struct radeon_device *rdev) | |||
| 5895 | u32 tmp; | 5911 | u32 tmp; |
| 5896 | 5912 | ||
| 5897 | /* gfx ring */ | 5913 | /* gfx ring */ |
| 5898 | WREG32(CP_INT_CNTL_RING0, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | 5914 | tmp = RREG32(CP_INT_CNTL_RING0) & |
| 5915 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
| 5916 | WREG32(CP_INT_CNTL_RING0, tmp); | ||
| 5899 | /* sdma */ | 5917 | /* sdma */ |
| 5900 | tmp = RREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; | 5918 | tmp = RREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; |
| 5901 | WREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET, tmp); | 5919 | WREG32(SDMA0_CNTL + SDMA0_REGISTER_OFFSET, tmp); |
| @@ -6036,8 +6054,7 @@ static int cik_irq_init(struct radeon_device *rdev) | |||
| 6036 | */ | 6054 | */ |
| 6037 | int cik_irq_set(struct radeon_device *rdev) | 6055 | int cik_irq_set(struct radeon_device *rdev) |
| 6038 | { | 6056 | { |
| 6039 | u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE | | 6057 | u32 cp_int_cntl; |
| 6040 | PRIV_INSTR_INT_ENABLE | PRIV_REG_INT_ENABLE; | ||
| 6041 | u32 cp_m1p0, cp_m1p1, cp_m1p2, cp_m1p3; | 6058 | u32 cp_m1p0, cp_m1p1, cp_m1p2, cp_m1p3; |
| 6042 | u32 cp_m2p0, cp_m2p1, cp_m2p2, cp_m2p3; | 6059 | u32 cp_m2p0, cp_m2p1, cp_m2p2, cp_m2p3; |
| 6043 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; | 6060 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; |
| @@ -6058,6 +6075,10 @@ int cik_irq_set(struct radeon_device *rdev) | |||
| 6058 | return 0; | 6075 | return 0; |
| 6059 | } | 6076 | } |
| 6060 | 6077 | ||
| 6078 | cp_int_cntl = RREG32(CP_INT_CNTL_RING0) & | ||
| 6079 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
| 6080 | cp_int_cntl |= PRIV_INSTR_INT_ENABLE | PRIV_REG_INT_ENABLE; | ||
| 6081 | |||
| 6061 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6082 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; |
| 6062 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6083 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; |
| 6063 | hpd3 = RREG32(DC_HPD3_INT_CONTROL) & ~DC_HPDx_INT_EN; | 6084 | hpd3 = RREG32(DC_HPD3_INT_CONTROL) & ~DC_HPDx_INT_EN; |
diff --git a/drivers/gpu/drm/radeon/cypress_dpm.c b/drivers/gpu/drm/radeon/cypress_dpm.c index 95a66db08d9b..91bb470de0a3 100644 --- a/drivers/gpu/drm/radeon/cypress_dpm.c +++ b/drivers/gpu/drm/radeon/cypress_dpm.c | |||
| @@ -2014,12 +2014,6 @@ int cypress_dpm_set_power_state(struct radeon_device *rdev) | |||
| 2014 | if (eg_pi->pcie_performance_request) | 2014 | if (eg_pi->pcie_performance_request) |
| 2015 | cypress_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); | 2015 | cypress_notify_link_speed_change_after_state_change(rdev, new_ps, old_ps); |
| 2016 | 2016 | ||
| 2017 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 2018 | if (ret) { | ||
| 2019 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
| 2020 | return ret; | ||
| 2021 | } | ||
| 2022 | |||
| 2023 | return 0; | 2017 | return 0; |
| 2024 | } | 2018 | } |
| 2025 | 2019 | ||
diff --git a/drivers/gpu/drm/radeon/dce6_afmt.c b/drivers/gpu/drm/radeon/dce6_afmt.c index 8953255e894b..85a69d2ea3d2 100644 --- a/drivers/gpu/drm/radeon/dce6_afmt.c +++ b/drivers/gpu/drm/radeon/dce6_afmt.c | |||
| @@ -28,22 +28,30 @@ | |||
| 28 | static u32 dce6_endpoint_rreg(struct radeon_device *rdev, | 28 | static u32 dce6_endpoint_rreg(struct radeon_device *rdev, |
| 29 | u32 block_offset, u32 reg) | 29 | u32 block_offset, u32 reg) |
| 30 | { | 30 | { |
| 31 | unsigned long flags; | ||
| 31 | u32 r; | 32 | u32 r; |
| 32 | 33 | ||
| 34 | spin_lock_irqsave(&rdev->end_idx_lock, flags); | ||
| 33 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); | 35 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); |
| 34 | r = RREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset); | 36 | r = RREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset); |
| 37 | spin_unlock_irqrestore(&rdev->end_idx_lock, flags); | ||
| 38 | |||
| 35 | return r; | 39 | return r; |
| 36 | } | 40 | } |
| 37 | 41 | ||
| 38 | static void dce6_endpoint_wreg(struct radeon_device *rdev, | 42 | static void dce6_endpoint_wreg(struct radeon_device *rdev, |
| 39 | u32 block_offset, u32 reg, u32 v) | 43 | u32 block_offset, u32 reg, u32 v) |
| 40 | { | 44 | { |
| 45 | unsigned long flags; | ||
| 46 | |||
| 47 | spin_lock_irqsave(&rdev->end_idx_lock, flags); | ||
| 41 | if (ASIC_IS_DCE8(rdev)) | 48 | if (ASIC_IS_DCE8(rdev)) |
| 42 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); | 49 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, reg); |
| 43 | else | 50 | else |
| 44 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, | 51 | WREG32(AZ_F0_CODEC_ENDPOINT_INDEX + block_offset, |
| 45 | AZ_ENDPOINT_REG_WRITE_EN | AZ_ENDPOINT_REG_INDEX(reg)); | 52 | AZ_ENDPOINT_REG_WRITE_EN | AZ_ENDPOINT_REG_INDEX(reg)); |
| 46 | WREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset, v); | 53 | WREG32(AZ_F0_CODEC_ENDPOINT_DATA + block_offset, v); |
| 54 | spin_unlock_irqrestore(&rdev->end_idx_lock, flags); | ||
| 47 | } | 55 | } |
| 48 | 56 | ||
| 49 | #define RREG32_ENDPOINT(block, reg) dce6_endpoint_rreg(rdev, (block), (reg)) | 57 | #define RREG32_ENDPOINT(block, reg) dce6_endpoint_rreg(rdev, (block), (reg)) |
| @@ -86,12 +94,12 @@ void dce6_afmt_select_pin(struct drm_encoder *encoder) | |||
| 86 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | 94 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); |
| 87 | struct radeon_encoder_atom_dig *dig = radeon_encoder->enc_priv; | 95 | struct radeon_encoder_atom_dig *dig = radeon_encoder->enc_priv; |
| 88 | u32 offset = dig->afmt->offset; | 96 | u32 offset = dig->afmt->offset; |
| 89 | u32 id = dig->afmt->pin->id; | ||
| 90 | 97 | ||
| 91 | if (!dig->afmt->pin) | 98 | if (!dig->afmt->pin) |
| 92 | return; | 99 | return; |
| 93 | 100 | ||
| 94 | WREG32(AFMT_AUDIO_SRC_CONTROL + offset, AFMT_AUDIO_SRC_SELECT(id)); | 101 | WREG32(AFMT_AUDIO_SRC_CONTROL + offset, |
| 102 | AFMT_AUDIO_SRC_SELECT(dig->afmt->pin->id)); | ||
| 95 | } | 103 | } |
| 96 | 104 | ||
| 97 | void dce6_afmt_write_speaker_allocation(struct drm_encoder *encoder) | 105 | void dce6_afmt_write_speaker_allocation(struct drm_encoder *encoder) |
diff --git a/drivers/gpu/drm/radeon/kv_dpm.c b/drivers/gpu/drm/radeon/kv_dpm.c index ecd60809db4e..71399065db04 100644 --- a/drivers/gpu/drm/radeon/kv_dpm.c +++ b/drivers/gpu/drm/radeon/kv_dpm.c | |||
| @@ -40,6 +40,7 @@ static int kv_calculate_dpm_settings(struct radeon_device *rdev); | |||
| 40 | static void kv_enable_new_levels(struct radeon_device *rdev); | 40 | static void kv_enable_new_levels(struct radeon_device *rdev); |
| 41 | static void kv_program_nbps_index_settings(struct radeon_device *rdev, | 41 | static void kv_program_nbps_index_settings(struct radeon_device *rdev, |
| 42 | struct radeon_ps *new_rps); | 42 | struct radeon_ps *new_rps); |
| 43 | static int kv_set_enabled_level(struct radeon_device *rdev, u32 level); | ||
| 43 | static int kv_set_enabled_levels(struct radeon_device *rdev); | 44 | static int kv_set_enabled_levels(struct radeon_device *rdev); |
| 44 | static int kv_force_dpm_highest(struct radeon_device *rdev); | 45 | static int kv_force_dpm_highest(struct radeon_device *rdev); |
| 45 | static int kv_force_dpm_lowest(struct radeon_device *rdev); | 46 | static int kv_force_dpm_lowest(struct radeon_device *rdev); |
| @@ -519,7 +520,7 @@ static int kv_set_dpm_boot_state(struct radeon_device *rdev) | |||
| 519 | 520 | ||
| 520 | static void kv_program_vc(struct radeon_device *rdev) | 521 | static void kv_program_vc(struct radeon_device *rdev) |
| 521 | { | 522 | { |
| 522 | WREG32_SMC(CG_FTV_0, 0x3FFFC000); | 523 | WREG32_SMC(CG_FTV_0, 0x3FFFC100); |
| 523 | } | 524 | } |
| 524 | 525 | ||
| 525 | static void kv_clear_vc(struct radeon_device *rdev) | 526 | static void kv_clear_vc(struct radeon_device *rdev) |
| @@ -638,7 +639,10 @@ static int kv_force_lowest_valid(struct radeon_device *rdev) | |||
| 638 | 639 | ||
| 639 | static int kv_unforce_levels(struct radeon_device *rdev) | 640 | static int kv_unforce_levels(struct radeon_device *rdev) |
| 640 | { | 641 | { |
| 641 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_NoForcedLevel); | 642 | if (rdev->family == CHIP_KABINI) |
| 643 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_NoForcedLevel); | ||
| 644 | else | ||
| 645 | return kv_set_enabled_levels(rdev); | ||
| 642 | } | 646 | } |
| 643 | 647 | ||
| 644 | static int kv_update_sclk_t(struct radeon_device *rdev) | 648 | static int kv_update_sclk_t(struct radeon_device *rdev) |
| @@ -667,9 +671,8 @@ static int kv_program_bootup_state(struct radeon_device *rdev) | |||
| 667 | &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk; | 671 | &rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk; |
| 668 | 672 | ||
| 669 | if (table && table->count) { | 673 | if (table && table->count) { |
| 670 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 674 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
| 671 | if ((table->entries[i].clk == pi->boot_pl.sclk) || | 675 | if (table->entries[i].clk == pi->boot_pl.sclk) |
| 672 | (i == 0)) | ||
| 673 | break; | 676 | break; |
| 674 | } | 677 | } |
| 675 | 678 | ||
| @@ -682,9 +685,8 @@ static int kv_program_bootup_state(struct radeon_device *rdev) | |||
| 682 | if (table->num_max_dpm_entries == 0) | 685 | if (table->num_max_dpm_entries == 0) |
| 683 | return -EINVAL; | 686 | return -EINVAL; |
| 684 | 687 | ||
| 685 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 688 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
| 686 | if ((table->entries[i].sclk_frequency == pi->boot_pl.sclk) || | 689 | if (table->entries[i].sclk_frequency == pi->boot_pl.sclk) |
| 687 | (i == 0)) | ||
| 688 | break; | 690 | break; |
| 689 | } | 691 | } |
| 690 | 692 | ||
| @@ -1078,6 +1080,13 @@ static int kv_enable_ulv(struct radeon_device *rdev, bool enable) | |||
| 1078 | PPSMC_MSG_EnableULV : PPSMC_MSG_DisableULV); | 1080 | PPSMC_MSG_EnableULV : PPSMC_MSG_DisableULV); |
| 1079 | } | 1081 | } |
| 1080 | 1082 | ||
| 1083 | static void kv_reset_acp_boot_level(struct radeon_device *rdev) | ||
| 1084 | { | ||
| 1085 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
| 1086 | |||
| 1087 | pi->acp_boot_level = 0xff; | ||
| 1088 | } | ||
| 1089 | |||
| 1081 | static void kv_update_current_ps(struct radeon_device *rdev, | 1090 | static void kv_update_current_ps(struct radeon_device *rdev, |
| 1082 | struct radeon_ps *rps) | 1091 | struct radeon_ps *rps) |
| 1083 | { | 1092 | { |
| @@ -1100,6 +1109,18 @@ static void kv_update_requested_ps(struct radeon_device *rdev, | |||
| 1100 | pi->requested_rps.ps_priv = &pi->requested_ps; | 1109 | pi->requested_rps.ps_priv = &pi->requested_ps; |
| 1101 | } | 1110 | } |
| 1102 | 1111 | ||
| 1112 | void kv_dpm_enable_bapm(struct radeon_device *rdev, bool enable) | ||
| 1113 | { | ||
| 1114 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
| 1115 | int ret; | ||
| 1116 | |||
| 1117 | if (pi->bapm_enable) { | ||
| 1118 | ret = kv_smc_bapm_enable(rdev, enable); | ||
| 1119 | if (ret) | ||
| 1120 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
| 1121 | } | ||
| 1122 | } | ||
| 1123 | |||
| 1103 | int kv_dpm_enable(struct radeon_device *rdev) | 1124 | int kv_dpm_enable(struct radeon_device *rdev) |
| 1104 | { | 1125 | { |
| 1105 | struct kv_power_info *pi = kv_get_pi(rdev); | 1126 | struct kv_power_info *pi = kv_get_pi(rdev); |
| @@ -1192,6 +1213,8 @@ int kv_dpm_enable(struct radeon_device *rdev) | |||
| 1192 | return ret; | 1213 | return ret; |
| 1193 | } | 1214 | } |
| 1194 | 1215 | ||
| 1216 | kv_reset_acp_boot_level(rdev); | ||
| 1217 | |||
| 1195 | if (rdev->irq.installed && | 1218 | if (rdev->irq.installed && |
| 1196 | r600_is_internal_thermal_sensor(rdev->pm.int_thermal_type)) { | 1219 | r600_is_internal_thermal_sensor(rdev->pm.int_thermal_type)) { |
| 1197 | ret = kv_set_thermal_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX); | 1220 | ret = kv_set_thermal_temperature_range(rdev, R600_TEMP_RANGE_MIN, R600_TEMP_RANGE_MAX); |
| @@ -1203,6 +1226,12 @@ int kv_dpm_enable(struct radeon_device *rdev) | |||
| 1203 | radeon_irq_set(rdev); | 1226 | radeon_irq_set(rdev); |
| 1204 | } | 1227 | } |
| 1205 | 1228 | ||
| 1229 | ret = kv_smc_bapm_enable(rdev, false); | ||
| 1230 | if (ret) { | ||
| 1231 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
| 1232 | return ret; | ||
| 1233 | } | ||
| 1234 | |||
| 1206 | /* powerdown unused blocks for now */ | 1235 | /* powerdown unused blocks for now */ |
| 1207 | kv_dpm_powergate_acp(rdev, true); | 1236 | kv_dpm_powergate_acp(rdev, true); |
| 1208 | kv_dpm_powergate_samu(rdev, true); | 1237 | kv_dpm_powergate_samu(rdev, true); |
| @@ -1226,6 +1255,8 @@ void kv_dpm_disable(struct radeon_device *rdev) | |||
| 1226 | RADEON_CG_BLOCK_BIF | | 1255 | RADEON_CG_BLOCK_BIF | |
| 1227 | RADEON_CG_BLOCK_HDP), false); | 1256 | RADEON_CG_BLOCK_HDP), false); |
| 1228 | 1257 | ||
| 1258 | kv_smc_bapm_enable(rdev, false); | ||
| 1259 | |||
| 1229 | /* powerup blocks */ | 1260 | /* powerup blocks */ |
| 1230 | kv_dpm_powergate_acp(rdev, false); | 1261 | kv_dpm_powergate_acp(rdev, false); |
| 1231 | kv_dpm_powergate_samu(rdev, false); | 1262 | kv_dpm_powergate_samu(rdev, false); |
| @@ -1450,6 +1481,39 @@ static int kv_update_samu_dpm(struct radeon_device *rdev, bool gate) | |||
| 1450 | return kv_enable_samu_dpm(rdev, !gate); | 1481 | return kv_enable_samu_dpm(rdev, !gate); |
| 1451 | } | 1482 | } |
| 1452 | 1483 | ||
| 1484 | static u8 kv_get_acp_boot_level(struct radeon_device *rdev) | ||
| 1485 | { | ||
| 1486 | u8 i; | ||
| 1487 | struct radeon_clock_voltage_dependency_table *table = | ||
| 1488 | &rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table; | ||
| 1489 | |||
| 1490 | for (i = 0; i < table->count; i++) { | ||
| 1491 | if (table->entries[i].clk >= 0) /* XXX */ | ||
| 1492 | break; | ||
| 1493 | } | ||
| 1494 | |||
| 1495 | if (i >= table->count) | ||
| 1496 | i = table->count - 1; | ||
| 1497 | |||
| 1498 | return i; | ||
| 1499 | } | ||
| 1500 | |||
| 1501 | static void kv_update_acp_boot_level(struct radeon_device *rdev) | ||
| 1502 | { | ||
| 1503 | struct kv_power_info *pi = kv_get_pi(rdev); | ||
| 1504 | u8 acp_boot_level; | ||
| 1505 | |||
| 1506 | if (!pi->caps_stable_p_state) { | ||
| 1507 | acp_boot_level = kv_get_acp_boot_level(rdev); | ||
| 1508 | if (acp_boot_level != pi->acp_boot_level) { | ||
| 1509 | pi->acp_boot_level = acp_boot_level; | ||
| 1510 | kv_send_msg_to_smc_with_parameter(rdev, | ||
| 1511 | PPSMC_MSG_ACPDPM_SetEnabledMask, | ||
| 1512 | (1 << pi->acp_boot_level)); | ||
| 1513 | } | ||
| 1514 | } | ||
| 1515 | } | ||
| 1516 | |||
| 1453 | static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) | 1517 | static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) |
| 1454 | { | 1518 | { |
| 1455 | struct kv_power_info *pi = kv_get_pi(rdev); | 1519 | struct kv_power_info *pi = kv_get_pi(rdev); |
| @@ -1461,7 +1525,7 @@ static int kv_update_acp_dpm(struct radeon_device *rdev, bool gate) | |||
| 1461 | if (pi->caps_stable_p_state) | 1525 | if (pi->caps_stable_p_state) |
| 1462 | pi->acp_boot_level = table->count - 1; | 1526 | pi->acp_boot_level = table->count - 1; |
| 1463 | else | 1527 | else |
| 1464 | pi->acp_boot_level = 0; | 1528 | pi->acp_boot_level = kv_get_acp_boot_level(rdev); |
| 1465 | 1529 | ||
| 1466 | ret = kv_copy_bytes_to_smc(rdev, | 1530 | ret = kv_copy_bytes_to_smc(rdev, |
| 1467 | pi->dpm_table_start + | 1531 | pi->dpm_table_start + |
| @@ -1588,13 +1652,11 @@ static void kv_set_valid_clock_range(struct radeon_device *rdev, | |||
| 1588 | } | 1652 | } |
| 1589 | } | 1653 | } |
| 1590 | 1654 | ||
| 1591 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 1655 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
| 1592 | if ((table->entries[i].clk <= new_ps->levels[new_ps->num_levels -1].sclk) || | 1656 | if (table->entries[i].clk <= new_ps->levels[new_ps->num_levels - 1].sclk) |
| 1593 | (i == 0)) { | ||
| 1594 | pi->highest_valid = i; | ||
| 1595 | break; | 1657 | break; |
| 1596 | } | ||
| 1597 | } | 1658 | } |
| 1659 | pi->highest_valid = i; | ||
| 1598 | 1660 | ||
| 1599 | if (pi->lowest_valid > pi->highest_valid) { | 1661 | if (pi->lowest_valid > pi->highest_valid) { |
| 1600 | if ((new_ps->levels[0].sclk - table->entries[pi->highest_valid].clk) > | 1662 | if ((new_ps->levels[0].sclk - table->entries[pi->highest_valid].clk) > |
| @@ -1615,14 +1677,12 @@ static void kv_set_valid_clock_range(struct radeon_device *rdev, | |||
| 1615 | } | 1677 | } |
| 1616 | } | 1678 | } |
| 1617 | 1679 | ||
| 1618 | for (i = pi->graphics_dpm_level_count - 1; i >= 0; i--) { | 1680 | for (i = pi->graphics_dpm_level_count - 1; i > 0; i--) { |
| 1619 | if (table->entries[i].sclk_frequency <= | 1681 | if (table->entries[i].sclk_frequency <= |
| 1620 | new_ps->levels[new_ps->num_levels - 1].sclk || | 1682 | new_ps->levels[new_ps->num_levels - 1].sclk) |
| 1621 | i == 0) { | ||
| 1622 | pi->highest_valid = i; | ||
| 1623 | break; | 1683 | break; |
| 1624 | } | ||
| 1625 | } | 1684 | } |
| 1685 | pi->highest_valid = i; | ||
| 1626 | 1686 | ||
| 1627 | if (pi->lowest_valid > pi->highest_valid) { | 1687 | if (pi->lowest_valid > pi->highest_valid) { |
| 1628 | if ((new_ps->levels[0].sclk - | 1688 | if ((new_ps->levels[0].sclk - |
| @@ -1724,6 +1784,14 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1724 | RADEON_CG_BLOCK_BIF | | 1784 | RADEON_CG_BLOCK_BIF | |
| 1725 | RADEON_CG_BLOCK_HDP), false); | 1785 | RADEON_CG_BLOCK_HDP), false); |
| 1726 | 1786 | ||
| 1787 | if (pi->bapm_enable) { | ||
| 1788 | ret = kv_smc_bapm_enable(rdev, rdev->pm.dpm.ac_power); | ||
| 1789 | if (ret) { | ||
| 1790 | DRM_ERROR("kv_smc_bapm_enable failed\n"); | ||
| 1791 | return ret; | ||
| 1792 | } | ||
| 1793 | } | ||
| 1794 | |||
| 1727 | if (rdev->family == CHIP_KABINI) { | 1795 | if (rdev->family == CHIP_KABINI) { |
| 1728 | if (pi->enable_dpm) { | 1796 | if (pi->enable_dpm) { |
| 1729 | kv_set_valid_clock_range(rdev, new_ps); | 1797 | kv_set_valid_clock_range(rdev, new_ps); |
| @@ -1775,6 +1843,7 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1775 | return ret; | 1843 | return ret; |
| 1776 | } | 1844 | } |
| 1777 | #endif | 1845 | #endif |
| 1846 | kv_update_acp_boot_level(rdev); | ||
| 1778 | kv_update_sclk_t(rdev); | 1847 | kv_update_sclk_t(rdev); |
| 1779 | kv_enable_nb_dpm(rdev); | 1848 | kv_enable_nb_dpm(rdev); |
| 1780 | } | 1849 | } |
| @@ -1785,7 +1854,6 @@ int kv_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1785 | RADEON_CG_BLOCK_BIF | | 1854 | RADEON_CG_BLOCK_BIF | |
| 1786 | RADEON_CG_BLOCK_HDP), true); | 1855 | RADEON_CG_BLOCK_HDP), true); |
| 1787 | 1856 | ||
| 1788 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
| 1789 | return 0; | 1857 | return 0; |
| 1790 | } | 1858 | } |
| 1791 | 1859 | ||
| @@ -1806,12 +1874,23 @@ void kv_dpm_setup_asic(struct radeon_device *rdev) | |||
| 1806 | 1874 | ||
| 1807 | void kv_dpm_reset_asic(struct radeon_device *rdev) | 1875 | void kv_dpm_reset_asic(struct radeon_device *rdev) |
| 1808 | { | 1876 | { |
| 1809 | kv_force_lowest_valid(rdev); | 1877 | struct kv_power_info *pi = kv_get_pi(rdev); |
| 1810 | kv_init_graphics_levels(rdev); | 1878 | |
| 1811 | kv_program_bootup_state(rdev); | 1879 | if (rdev->family == CHIP_KABINI) { |
| 1812 | kv_upload_dpm_settings(rdev); | 1880 | kv_force_lowest_valid(rdev); |
| 1813 | kv_force_lowest_valid(rdev); | 1881 | kv_init_graphics_levels(rdev); |
| 1814 | kv_unforce_levels(rdev); | 1882 | kv_program_bootup_state(rdev); |
| 1883 | kv_upload_dpm_settings(rdev); | ||
| 1884 | kv_force_lowest_valid(rdev); | ||
| 1885 | kv_unforce_levels(rdev); | ||
| 1886 | } else { | ||
| 1887 | kv_init_graphics_levels(rdev); | ||
| 1888 | kv_program_bootup_state(rdev); | ||
| 1889 | kv_freeze_sclk_dpm(rdev, true); | ||
| 1890 | kv_upload_dpm_settings(rdev); | ||
| 1891 | kv_freeze_sclk_dpm(rdev, false); | ||
| 1892 | kv_set_enabled_level(rdev, pi->graphics_boot_level); | ||
| 1893 | } | ||
| 1815 | } | 1894 | } |
| 1816 | 1895 | ||
| 1817 | //XXX use sumo_dpm_display_configuration_changed | 1896 | //XXX use sumo_dpm_display_configuration_changed |
| @@ -1871,12 +1950,15 @@ static int kv_force_dpm_highest(struct radeon_device *rdev) | |||
| 1871 | if (ret) | 1950 | if (ret) |
| 1872 | return ret; | 1951 | return ret; |
| 1873 | 1952 | ||
| 1874 | for (i = SMU7_MAX_LEVELS_GRAPHICS - 1; i >= 0; i--) { | 1953 | for (i = SMU7_MAX_LEVELS_GRAPHICS - 1; i > 0; i--) { |
| 1875 | if (enable_mask & (1 << i)) | 1954 | if (enable_mask & (1 << i)) |
| 1876 | break; | 1955 | break; |
| 1877 | } | 1956 | } |
| 1878 | 1957 | ||
| 1879 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | 1958 | if (rdev->family == CHIP_KABINI) |
| 1959 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | ||
| 1960 | else | ||
| 1961 | return kv_set_enabled_level(rdev, i); | ||
| 1880 | } | 1962 | } |
| 1881 | 1963 | ||
| 1882 | static int kv_force_dpm_lowest(struct radeon_device *rdev) | 1964 | static int kv_force_dpm_lowest(struct radeon_device *rdev) |
| @@ -1893,7 +1975,10 @@ static int kv_force_dpm_lowest(struct radeon_device *rdev) | |||
| 1893 | break; | 1975 | break; |
| 1894 | } | 1976 | } |
| 1895 | 1977 | ||
| 1896 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | 1978 | if (rdev->family == CHIP_KABINI) |
| 1979 | return kv_send_msg_to_smc_with_parameter(rdev, PPSMC_MSG_DPM_ForceState, i); | ||
| 1980 | else | ||
| 1981 | return kv_set_enabled_level(rdev, i); | ||
| 1897 | } | 1982 | } |
| 1898 | 1983 | ||
| 1899 | static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, | 1984 | static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, |
| @@ -1911,9 +1996,9 @@ static u8 kv_get_sleep_divider_id_from_clock(struct radeon_device *rdev, | |||
| 1911 | if (!pi->caps_sclk_ds) | 1996 | if (!pi->caps_sclk_ds) |
| 1912 | return 0; | 1997 | return 0; |
| 1913 | 1998 | ||
| 1914 | for (i = KV_MAX_DEEPSLEEP_DIVIDER_ID; i <= 0; i--) { | 1999 | for (i = KV_MAX_DEEPSLEEP_DIVIDER_ID; i > 0; i--) { |
| 1915 | temp = sclk / sumo_get_sleep_divider_from_id(i); | 2000 | temp = sclk / sumo_get_sleep_divider_from_id(i); |
| 1916 | if ((temp >= min) || (i == 0)) | 2001 | if (temp >= min) |
| 1917 | break; | 2002 | break; |
| 1918 | } | 2003 | } |
| 1919 | 2004 | ||
| @@ -2039,12 +2124,12 @@ static void kv_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 2039 | ps->dpmx_nb_ps_lo = 0x1; | 2124 | ps->dpmx_nb_ps_lo = 0x1; |
| 2040 | ps->dpmx_nb_ps_hi = 0x0; | 2125 | ps->dpmx_nb_ps_hi = 0x0; |
| 2041 | } else { | 2126 | } else { |
| 2042 | ps->dpm0_pg_nb_ps_lo = 0x1; | 2127 | ps->dpm0_pg_nb_ps_lo = 0x3; |
| 2043 | ps->dpm0_pg_nb_ps_hi = 0x0; | 2128 | ps->dpm0_pg_nb_ps_hi = 0x0; |
| 2044 | ps->dpmx_nb_ps_lo = 0x2; | 2129 | ps->dpmx_nb_ps_lo = 0x3; |
| 2045 | ps->dpmx_nb_ps_hi = 0x1; | 2130 | ps->dpmx_nb_ps_hi = 0x0; |
| 2046 | 2131 | ||
| 2047 | if (pi->sys_info.nb_dpm_enable && pi->battery_state) { | 2132 | if (pi->sys_info.nb_dpm_enable) { |
| 2048 | force_high = (mclk >= pi->sys_info.nbp_memory_clock[3]) || | 2133 | force_high = (mclk >= pi->sys_info.nbp_memory_clock[3]) || |
| 2049 | pi->video_start || (rdev->pm.dpm.new_active_crtc_count >= 3) || | 2134 | pi->video_start || (rdev->pm.dpm.new_active_crtc_count >= 3) || |
| 2050 | pi->disable_nb_ps3_in_battery; | 2135 | pi->disable_nb_ps3_in_battery; |
| @@ -2210,6 +2295,15 @@ static void kv_enable_new_levels(struct radeon_device *rdev) | |||
| 2210 | } | 2295 | } |
| 2211 | } | 2296 | } |
| 2212 | 2297 | ||
| 2298 | static int kv_set_enabled_level(struct radeon_device *rdev, u32 level) | ||
| 2299 | { | ||
| 2300 | u32 new_mask = (1 << level); | ||
| 2301 | |||
| 2302 | return kv_send_msg_to_smc_with_parameter(rdev, | ||
| 2303 | PPSMC_MSG_SCLKDPM_SetEnabledMask, | ||
| 2304 | new_mask); | ||
| 2305 | } | ||
| 2306 | |||
| 2213 | static int kv_set_enabled_levels(struct radeon_device *rdev) | 2307 | static int kv_set_enabled_levels(struct radeon_device *rdev) |
| 2214 | { | 2308 | { |
| 2215 | struct kv_power_info *pi = kv_get_pi(rdev); | 2309 | struct kv_power_info *pi = kv_get_pi(rdev); |
diff --git a/drivers/gpu/drm/radeon/kv_dpm.h b/drivers/gpu/drm/radeon/kv_dpm.h index 32bb079572d7..8cef7525d7a8 100644 --- a/drivers/gpu/drm/radeon/kv_dpm.h +++ b/drivers/gpu/drm/radeon/kv_dpm.h | |||
| @@ -192,6 +192,7 @@ int kv_send_msg_to_smc_with_parameter(struct radeon_device *rdev, | |||
| 192 | int kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 192 | int kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
| 193 | u32 *value, u32 limit); | 193 | u32 *value, u32 limit); |
| 194 | int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable); | 194 | int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable); |
| 195 | int kv_smc_bapm_enable(struct radeon_device *rdev, bool enable); | ||
| 195 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, | 196 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, |
| 196 | u32 smc_start_address, | 197 | u32 smc_start_address, |
| 197 | const u8 *src, u32 byte_count, u32 limit); | 198 | const u8 *src, u32 byte_count, u32 limit); |
diff --git a/drivers/gpu/drm/radeon/kv_smc.c b/drivers/gpu/drm/radeon/kv_smc.c index 34a226d7e34a..0000b59a6d05 100644 --- a/drivers/gpu/drm/radeon/kv_smc.c +++ b/drivers/gpu/drm/radeon/kv_smc.c | |||
| @@ -107,6 +107,14 @@ int kv_smc_dpm_enable(struct radeon_device *rdev, bool enable) | |||
| 107 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DPM_Disable); | 107 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DPM_Disable); |
| 108 | } | 108 | } |
| 109 | 109 | ||
| 110 | int kv_smc_bapm_enable(struct radeon_device *rdev, bool enable) | ||
| 111 | { | ||
| 112 | if (enable) | ||
| 113 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_EnableBAPM); | ||
| 114 | else | ||
| 115 | return kv_notify_message_to_smu(rdev, PPSMC_MSG_DisableBAPM); | ||
| 116 | } | ||
| 117 | |||
| 110 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, | 118 | int kv_copy_bytes_to_smc(struct radeon_device *rdev, |
| 111 | u32 smc_start_address, | 119 | u32 smc_start_address, |
| 112 | const u8 *src, u32 byte_count, u32 limit) | 120 | const u8 *src, u32 byte_count, u32 limit) |
diff --git a/drivers/gpu/drm/radeon/ni_dpm.c b/drivers/gpu/drm/radeon/ni_dpm.c index f7b625c9e0e9..f26339028154 100644 --- a/drivers/gpu/drm/radeon/ni_dpm.c +++ b/drivers/gpu/drm/radeon/ni_dpm.c | |||
| @@ -787,6 +787,7 @@ static void ni_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 787 | bool disable_mclk_switching; | 787 | bool disable_mclk_switching; |
| 788 | u32 mclk, sclk; | 788 | u32 mclk, sclk; |
| 789 | u16 vddc, vddci; | 789 | u16 vddc, vddci; |
| 790 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
| 790 | int i; | 791 | int i; |
| 791 | 792 | ||
| 792 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 793 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
| @@ -813,6 +814,29 @@ static void ni_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 813 | } | 814 | } |
| 814 | } | 815 | } |
| 815 | 816 | ||
| 817 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
| 818 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
| 819 | &max_sclk_vddc); | ||
| 820 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
| 821 | &max_mclk_vddci); | ||
| 822 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
| 823 | &max_mclk_vddc); | ||
| 824 | |||
| 825 | for (i = 0; i < ps->performance_level_count; i++) { | ||
| 826 | if (max_sclk_vddc) { | ||
| 827 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
| 828 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
| 829 | } | ||
| 830 | if (max_mclk_vddci) { | ||
| 831 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
| 832 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
| 833 | } | ||
| 834 | if (max_mclk_vddc) { | ||
| 835 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
| 836 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
| 837 | } | ||
| 838 | } | ||
| 839 | |||
| 816 | /* XXX validate the min clocks required for display */ | 840 | /* XXX validate the min clocks required for display */ |
| 817 | 841 | ||
| 818 | if (disable_mclk_switching) { | 842 | if (disable_mclk_switching) { |
| @@ -3865,12 +3889,6 @@ int ni_dpm_set_power_state(struct radeon_device *rdev) | |||
| 3865 | return ret; | 3889 | return ret; |
| 3866 | } | 3890 | } |
| 3867 | 3891 | ||
| 3868 | ret = ni_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 3869 | if (ret) { | ||
| 3870 | DRM_ERROR("ni_dpm_force_performance_level failed\n"); | ||
| 3871 | return ret; | ||
| 3872 | } | ||
| 3873 | |||
| 3874 | return 0; | 3892 | return 0; |
| 3875 | } | 3893 | } |
| 3876 | 3894 | ||
diff --git a/drivers/gpu/drm/radeon/ppsmc.h b/drivers/gpu/drm/radeon/ppsmc.h index 682842804bce..5670b8291285 100644 --- a/drivers/gpu/drm/radeon/ppsmc.h +++ b/drivers/gpu/drm/radeon/ppsmc.h | |||
| @@ -163,6 +163,8 @@ typedef uint8_t PPSMC_Result; | |||
| 163 | #define PPSMC_MSG_VCEPowerON ((uint32_t) 0x10f) | 163 | #define PPSMC_MSG_VCEPowerON ((uint32_t) 0x10f) |
| 164 | #define PPSMC_MSG_DCE_RemoveVoltageAdjustment ((uint32_t) 0x11d) | 164 | #define PPSMC_MSG_DCE_RemoveVoltageAdjustment ((uint32_t) 0x11d) |
| 165 | #define PPSMC_MSG_DCE_AllowVoltageAdjustment ((uint32_t) 0x11e) | 165 | #define PPSMC_MSG_DCE_AllowVoltageAdjustment ((uint32_t) 0x11e) |
| 166 | #define PPSMC_MSG_EnableBAPM ((uint32_t) 0x120) | ||
| 167 | #define PPSMC_MSG_DisableBAPM ((uint32_t) 0x121) | ||
| 166 | #define PPSMC_MSG_UVD_DPM_Config ((uint32_t) 0x124) | 168 | #define PPSMC_MSG_UVD_DPM_Config ((uint32_t) 0x124) |
| 167 | 169 | ||
| 168 | 170 | ||
diff --git a/drivers/gpu/drm/radeon/r100.c b/drivers/gpu/drm/radeon/r100.c index 9fc61dd68bc0..d71333033b2b 100644 --- a/drivers/gpu/drm/radeon/r100.c +++ b/drivers/gpu/drm/radeon/r100.c | |||
| @@ -2853,21 +2853,28 @@ static void r100_pll_errata_after_data(struct radeon_device *rdev) | |||
| 2853 | 2853 | ||
| 2854 | uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg) | 2854 | uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg) |
| 2855 | { | 2855 | { |
| 2856 | unsigned long flags; | ||
| 2856 | uint32_t data; | 2857 | uint32_t data; |
| 2857 | 2858 | ||
| 2859 | spin_lock_irqsave(&rdev->pll_idx_lock, flags); | ||
| 2858 | WREG8(RADEON_CLOCK_CNTL_INDEX, reg & 0x3f); | 2860 | WREG8(RADEON_CLOCK_CNTL_INDEX, reg & 0x3f); |
| 2859 | r100_pll_errata_after_index(rdev); | 2861 | r100_pll_errata_after_index(rdev); |
| 2860 | data = RREG32(RADEON_CLOCK_CNTL_DATA); | 2862 | data = RREG32(RADEON_CLOCK_CNTL_DATA); |
| 2861 | r100_pll_errata_after_data(rdev); | 2863 | r100_pll_errata_after_data(rdev); |
| 2864 | spin_unlock_irqrestore(&rdev->pll_idx_lock, flags); | ||
| 2862 | return data; | 2865 | return data; |
| 2863 | } | 2866 | } |
| 2864 | 2867 | ||
| 2865 | void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 2868 | void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 2866 | { | 2869 | { |
| 2870 | unsigned long flags; | ||
| 2871 | |||
| 2872 | spin_lock_irqsave(&rdev->pll_idx_lock, flags); | ||
| 2867 | WREG8(RADEON_CLOCK_CNTL_INDEX, ((reg & 0x3f) | RADEON_PLL_WR_EN)); | 2873 | WREG8(RADEON_CLOCK_CNTL_INDEX, ((reg & 0x3f) | RADEON_PLL_WR_EN)); |
| 2868 | r100_pll_errata_after_index(rdev); | 2874 | r100_pll_errata_after_index(rdev); |
| 2869 | WREG32(RADEON_CLOCK_CNTL_DATA, v); | 2875 | WREG32(RADEON_CLOCK_CNTL_DATA, v); |
| 2870 | r100_pll_errata_after_data(rdev); | 2876 | r100_pll_errata_after_data(rdev); |
| 2877 | spin_unlock_irqrestore(&rdev->pll_idx_lock, flags); | ||
| 2871 | } | 2878 | } |
| 2872 | 2879 | ||
| 2873 | static void r100_set_safe_registers(struct radeon_device *rdev) | 2880 | static void r100_set_safe_registers(struct radeon_device *rdev) |
| @@ -2926,9 +2933,11 @@ static int r100_debugfs_cp_ring_info(struct seq_file *m, void *data) | |||
| 2926 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", rdp); | 2933 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", rdp); |
| 2927 | seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw); | 2934 | seq_printf(m, "%u free dwords in ring\n", ring->ring_free_dw); |
| 2928 | seq_printf(m, "%u dwords in ring\n", count); | 2935 | seq_printf(m, "%u dwords in ring\n", count); |
| 2929 | for (j = 0; j <= count; j++) { | 2936 | if (ring->ready) { |
| 2930 | i = (rdp + j) & ring->ptr_mask; | 2937 | for (j = 0; j <= count; j++) { |
| 2931 | seq_printf(m, "r[%04d]=0x%08x\n", i, ring->ring[i]); | 2938 | i = (rdp + j) & ring->ptr_mask; |
| 2939 | seq_printf(m, "r[%04d]=0x%08x\n", i, ring->ring[i]); | ||
| 2940 | } | ||
| 2932 | } | 2941 | } |
| 2933 | return 0; | 2942 | return 0; |
| 2934 | } | 2943 | } |
diff --git a/drivers/gpu/drm/radeon/r420.c b/drivers/gpu/drm/radeon/r420.c index 4e796ecf9ea4..6edf2b3a52b4 100644 --- a/drivers/gpu/drm/radeon/r420.c +++ b/drivers/gpu/drm/radeon/r420.c | |||
| @@ -160,18 +160,25 @@ void r420_pipes_init(struct radeon_device *rdev) | |||
| 160 | 160 | ||
| 161 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) | 161 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) |
| 162 | { | 162 | { |
| 163 | unsigned long flags; | ||
| 163 | u32 r; | 164 | u32 r; |
| 164 | 165 | ||
| 166 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 165 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); | 167 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); |
| 166 | r = RREG32(R_0001FC_MC_IND_DATA); | 168 | r = RREG32(R_0001FC_MC_IND_DATA); |
| 169 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 167 | return r; | 170 | return r; |
| 168 | } | 171 | } |
| 169 | 172 | ||
| 170 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 173 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 171 | { | 174 | { |
| 175 | unsigned long flags; | ||
| 176 | |||
| 177 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 172 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | | 178 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | |
| 173 | S_0001F8_MC_IND_WR_EN(1)); | 179 | S_0001F8_MC_IND_WR_EN(1)); |
| 174 | WREG32(R_0001FC_MC_IND_DATA, v); | 180 | WREG32(R_0001FC_MC_IND_DATA, v); |
| 181 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 175 | } | 182 | } |
| 176 | 183 | ||
| 177 | static void r420_debugfs(struct radeon_device *rdev) | 184 | static void r420_debugfs(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index ea4d3734e6d9..2a1b1876b431 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c | |||
| @@ -119,6 +119,11 @@ u32 r600_get_xclk(struct radeon_device *rdev) | |||
| 119 | return rdev->clock.spll.reference_freq; | 119 | return rdev->clock.spll.reference_freq; |
| 120 | } | 120 | } |
| 121 | 121 | ||
| 122 | int r600_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk) | ||
| 123 | { | ||
| 124 | return 0; | ||
| 125 | } | ||
| 126 | |||
| 122 | /* get temperature in millidegrees */ | 127 | /* get temperature in millidegrees */ |
| 123 | int rv6xx_get_temp(struct radeon_device *rdev) | 128 | int rv6xx_get_temp(struct radeon_device *rdev) |
| 124 | { | 129 | { |
| @@ -1045,20 +1050,27 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev) | |||
| 1045 | 1050 | ||
| 1046 | uint32_t rs780_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 1051 | uint32_t rs780_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
| 1047 | { | 1052 | { |
| 1053 | unsigned long flags; | ||
| 1048 | uint32_t r; | 1054 | uint32_t r; |
| 1049 | 1055 | ||
| 1056 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 1050 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg)); | 1057 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg)); |
| 1051 | r = RREG32(R_0028FC_MC_DATA); | 1058 | r = RREG32(R_0028FC_MC_DATA); |
| 1052 | WREG32(R_0028F8_MC_INDEX, ~C_0028F8_MC_IND_ADDR); | 1059 | WREG32(R_0028F8_MC_INDEX, ~C_0028F8_MC_IND_ADDR); |
| 1060 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 1053 | return r; | 1061 | return r; |
| 1054 | } | 1062 | } |
| 1055 | 1063 | ||
| 1056 | void rs780_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 1064 | void rs780_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 1057 | { | 1065 | { |
| 1066 | unsigned long flags; | ||
| 1067 | |||
| 1068 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 1058 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg) | | 1069 | WREG32(R_0028F8_MC_INDEX, S_0028F8_MC_IND_ADDR(reg) | |
| 1059 | S_0028F8_MC_IND_WR_EN(1)); | 1070 | S_0028F8_MC_IND_WR_EN(1)); |
| 1060 | WREG32(R_0028FC_MC_DATA, v); | 1071 | WREG32(R_0028FC_MC_DATA, v); |
| 1061 | WREG32(R_0028F8_MC_INDEX, 0x7F); | 1072 | WREG32(R_0028F8_MC_INDEX, 0x7F); |
| 1073 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 1062 | } | 1074 | } |
| 1063 | 1075 | ||
| 1064 | static void r600_mc_program(struct radeon_device *rdev) | 1076 | static void r600_mc_program(struct radeon_device *rdev) |
| @@ -2092,20 +2104,27 @@ static void r600_gpu_init(struct radeon_device *rdev) | |||
| 2092 | */ | 2104 | */ |
| 2093 | u32 r600_pciep_rreg(struct radeon_device *rdev, u32 reg) | 2105 | u32 r600_pciep_rreg(struct radeon_device *rdev, u32 reg) |
| 2094 | { | 2106 | { |
| 2107 | unsigned long flags; | ||
| 2095 | u32 r; | 2108 | u32 r; |
| 2096 | 2109 | ||
| 2110 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
| 2097 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); | 2111 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); |
| 2098 | (void)RREG32(PCIE_PORT_INDEX); | 2112 | (void)RREG32(PCIE_PORT_INDEX); |
| 2099 | r = RREG32(PCIE_PORT_DATA); | 2113 | r = RREG32(PCIE_PORT_DATA); |
| 2114 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
| 2100 | return r; | 2115 | return r; |
| 2101 | } | 2116 | } |
| 2102 | 2117 | ||
| 2103 | void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2118 | void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2104 | { | 2119 | { |
| 2120 | unsigned long flags; | ||
| 2121 | |||
| 2122 | spin_lock_irqsave(&rdev->pciep_idx_lock, flags); | ||
| 2105 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); | 2123 | WREG32(PCIE_PORT_INDEX, ((reg) & 0xff)); |
| 2106 | (void)RREG32(PCIE_PORT_INDEX); | 2124 | (void)RREG32(PCIE_PORT_INDEX); |
| 2107 | WREG32(PCIE_PORT_DATA, (v)); | 2125 | WREG32(PCIE_PORT_DATA, (v)); |
| 2108 | (void)RREG32(PCIE_PORT_DATA); | 2126 | (void)RREG32(PCIE_PORT_DATA); |
| 2127 | spin_unlock_irqrestore(&rdev->pciep_idx_lock, flags); | ||
| 2109 | } | 2128 | } |
| 2110 | 2129 | ||
| 2111 | /* | 2130 | /* |
diff --git a/drivers/gpu/drm/radeon/r600_dpm.c b/drivers/gpu/drm/radeon/r600_dpm.c index fa0de46fcc0d..5513d8f06252 100644 --- a/drivers/gpu/drm/radeon/r600_dpm.c +++ b/drivers/gpu/drm/radeon/r600_dpm.c | |||
| @@ -1084,7 +1084,7 @@ int r600_parse_extended_power_table(struct radeon_device *rdev) | |||
| 1084 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].dclk = | 1084 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].dclk = |
| 1085 | le16_to_cpu(uvd_clk->usDClkLow) | (uvd_clk->ucDClkHigh << 16); | 1085 | le16_to_cpu(uvd_clk->usDClkLow) | (uvd_clk->ucDClkHigh << 16); |
| 1086 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].v = | 1086 | rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries[i].v = |
| 1087 | le16_to_cpu(limits->entries[i].usVoltage); | 1087 | le16_to_cpu(entry->usVoltage); |
| 1088 | entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) | 1088 | entry = (ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record *) |
| 1089 | ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); | 1089 | ((u8 *)entry + sizeof(ATOM_PPLIB_UVD_Clock_Voltage_Limit_Record)); |
| 1090 | } | 1090 | } |
| @@ -1219,30 +1219,20 @@ int r600_parse_extended_power_table(struct radeon_device *rdev) | |||
| 1219 | 1219 | ||
| 1220 | void r600_free_extended_power_table(struct radeon_device *rdev) | 1220 | void r600_free_extended_power_table(struct radeon_device *rdev) |
| 1221 | { | 1221 | { |
| 1222 | if (rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries) | 1222 | struct radeon_dpm_dynamic_state *dyn_state = &rdev->pm.dpm.dyn_state; |
| 1223 | kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); | 1223 | |
| 1224 | if (rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries) | 1224 | kfree(dyn_state->vddc_dependency_on_sclk.entries); |
| 1225 | kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); | 1225 | kfree(dyn_state->vddci_dependency_on_mclk.entries); |
| 1226 | if (rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries) | 1226 | kfree(dyn_state->vddc_dependency_on_mclk.entries); |
| 1227 | kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk.entries); | 1227 | kfree(dyn_state->mvdd_dependency_on_mclk.entries); |
| 1228 | if (rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries) | 1228 | kfree(dyn_state->cac_leakage_table.entries); |
| 1229 | kfree(rdev->pm.dpm.dyn_state.mvdd_dependency_on_mclk.entries); | 1229 | kfree(dyn_state->phase_shedding_limits_table.entries); |
| 1230 | if (rdev->pm.dpm.dyn_state.cac_leakage_table.entries) | 1230 | kfree(dyn_state->ppm_table); |
| 1231 | kfree(rdev->pm.dpm.dyn_state.cac_leakage_table.entries); | 1231 | kfree(dyn_state->cac_tdp_table); |
| 1232 | if (rdev->pm.dpm.dyn_state.phase_shedding_limits_table.entries) | 1232 | kfree(dyn_state->vce_clock_voltage_dependency_table.entries); |
| 1233 | kfree(rdev->pm.dpm.dyn_state.phase_shedding_limits_table.entries); | 1233 | kfree(dyn_state->uvd_clock_voltage_dependency_table.entries); |
| 1234 | if (rdev->pm.dpm.dyn_state.ppm_table) | 1234 | kfree(dyn_state->samu_clock_voltage_dependency_table.entries); |
| 1235 | kfree(rdev->pm.dpm.dyn_state.ppm_table); | 1235 | kfree(dyn_state->acp_clock_voltage_dependency_table.entries); |
| 1236 | if (rdev->pm.dpm.dyn_state.cac_tdp_table) | ||
| 1237 | kfree(rdev->pm.dpm.dyn_state.cac_tdp_table); | ||
| 1238 | if (rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries) | ||
| 1239 | kfree(rdev->pm.dpm.dyn_state.vce_clock_voltage_dependency_table.entries); | ||
| 1240 | if (rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries) | ||
| 1241 | kfree(rdev->pm.dpm.dyn_state.uvd_clock_voltage_dependency_table.entries); | ||
| 1242 | if (rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries) | ||
| 1243 | kfree(rdev->pm.dpm.dyn_state.samu_clock_voltage_dependency_table.entries); | ||
| 1244 | if (rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries) | ||
| 1245 | kfree(rdev->pm.dpm.dyn_state.acp_clock_voltage_dependency_table.entries); | ||
| 1246 | } | 1236 | } |
| 1247 | 1237 | ||
| 1248 | enum radeon_pcie_gen r600_get_pcie_gen_support(struct radeon_device *rdev, | 1238 | enum radeon_pcie_gen r600_get_pcie_gen_support(struct radeon_device *rdev, |
diff --git a/drivers/gpu/drm/radeon/r600_hdmi.c b/drivers/gpu/drm/radeon/r600_hdmi.c index f443010ce90b..b0fa6002af3e 100644 --- a/drivers/gpu/drm/radeon/r600_hdmi.c +++ b/drivers/gpu/drm/radeon/r600_hdmi.c | |||
| @@ -257,10 +257,7 @@ void r600_audio_set_dto(struct drm_encoder *encoder, u32 clock) | |||
| 257 | * number (coefficient of two integer numbers. DCCG_AUDIO_DTOx_PHASE | 257 | * number (coefficient of two integer numbers. DCCG_AUDIO_DTOx_PHASE |
| 258 | * is the numerator, DCCG_AUDIO_DTOx_MODULE is the denominator | 258 | * is the numerator, DCCG_AUDIO_DTOx_MODULE is the denominator |
| 259 | */ | 259 | */ |
| 260 | if (ASIC_IS_DCE3(rdev)) { | 260 | if (ASIC_IS_DCE32(rdev)) { |
| 261 | /* according to the reg specs, this should DCE3.2 only, but in | ||
| 262 | * practice it seems to cover DCE3.0 as well. | ||
| 263 | */ | ||
| 264 | if (dig->dig_encoder == 0) { | 261 | if (dig->dig_encoder == 0) { |
| 265 | dto_cntl = RREG32(DCCG_AUDIO_DTO0_CNTL) & ~DCCG_AUDIO_DTO_WALLCLOCK_RATIO_MASK; | 262 | dto_cntl = RREG32(DCCG_AUDIO_DTO0_CNTL) & ~DCCG_AUDIO_DTO_WALLCLOCK_RATIO_MASK; |
| 266 | dto_cntl |= DCCG_AUDIO_DTO_WALLCLOCK_RATIO(wallclock_ratio); | 263 | dto_cntl |= DCCG_AUDIO_DTO_WALLCLOCK_RATIO(wallclock_ratio); |
| @@ -276,8 +273,21 @@ void r600_audio_set_dto(struct drm_encoder *encoder, u32 clock) | |||
| 276 | WREG32(DCCG_AUDIO_DTO1_MODULE, dto_modulo); | 273 | WREG32(DCCG_AUDIO_DTO1_MODULE, dto_modulo); |
| 277 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ | 274 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ |
| 278 | } | 275 | } |
| 276 | } else if (ASIC_IS_DCE3(rdev)) { | ||
| 277 | /* according to the reg specs, this should DCE3.2 only, but in | ||
| 278 | * practice it seems to cover DCE3.0/3.1 as well. | ||
| 279 | */ | ||
| 280 | if (dig->dig_encoder == 0) { | ||
| 281 | WREG32(DCCG_AUDIO_DTO0_PHASE, base_rate * 100); | ||
| 282 | WREG32(DCCG_AUDIO_DTO0_MODULE, clock * 100); | ||
| 283 | WREG32(DCCG_AUDIO_DTO_SELECT, 0); /* select DTO0 */ | ||
| 284 | } else { | ||
| 285 | WREG32(DCCG_AUDIO_DTO1_PHASE, base_rate * 100); | ||
| 286 | WREG32(DCCG_AUDIO_DTO1_MODULE, clock * 100); | ||
| 287 | WREG32(DCCG_AUDIO_DTO_SELECT, 1); /* select DTO1 */ | ||
| 288 | } | ||
| 279 | } else { | 289 | } else { |
| 280 | /* according to the reg specs, this should be DCE2.0 and DCE3.0 */ | 290 | /* according to the reg specs, this should be DCE2.0 and DCE3.0/3.1 */ |
| 281 | WREG32(AUDIO_DTO, AUDIO_DTO_PHASE(base_rate / 10) | | 291 | WREG32(AUDIO_DTO, AUDIO_DTO_PHASE(base_rate / 10) | |
| 282 | AUDIO_DTO_MODULE(clock / 10)); | 292 | AUDIO_DTO_MODULE(clock / 10)); |
| 283 | } | 293 | } |
diff --git a/drivers/gpu/drm/radeon/r600d.h b/drivers/gpu/drm/radeon/r600d.h index 454f90a849e4..e673fe26ea84 100644 --- a/drivers/gpu/drm/radeon/r600d.h +++ b/drivers/gpu/drm/radeon/r600d.h | |||
| @@ -1040,7 +1040,7 @@ | |||
| 1040 | # define HDMI0_AVI_INFO_CONT (1 << 1) | 1040 | # define HDMI0_AVI_INFO_CONT (1 << 1) |
| 1041 | # define HDMI0_AUDIO_INFO_SEND (1 << 4) | 1041 | # define HDMI0_AUDIO_INFO_SEND (1 << 4) |
| 1042 | # define HDMI0_AUDIO_INFO_CONT (1 << 5) | 1042 | # define HDMI0_AUDIO_INFO_CONT (1 << 5) |
| 1043 | # define HDMI0_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hmdi regs */ | 1043 | # define HDMI0_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hdmi regs */ |
| 1044 | # define HDMI0_AUDIO_INFO_UPDATE (1 << 7) | 1044 | # define HDMI0_AUDIO_INFO_UPDATE (1 << 7) |
| 1045 | # define HDMI0_MPEG_INFO_SEND (1 << 8) | 1045 | # define HDMI0_MPEG_INFO_SEND (1 << 8) |
| 1046 | # define HDMI0_MPEG_INFO_CONT (1 << 9) | 1046 | # define HDMI0_MPEG_INFO_CONT (1 << 9) |
diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h index ff8b564ce2b2..a400ac1c4147 100644 --- a/drivers/gpu/drm/radeon/radeon.h +++ b/drivers/gpu/drm/radeon/radeon.h | |||
| @@ -181,7 +181,7 @@ extern int radeon_aspm; | |||
| 181 | #define RADEON_CG_SUPPORT_HDP_MGCG (1 << 16) | 181 | #define RADEON_CG_SUPPORT_HDP_MGCG (1 << 16) |
| 182 | 182 | ||
| 183 | /* PG flags */ | 183 | /* PG flags */ |
| 184 | #define RADEON_PG_SUPPORT_GFX_CG (1 << 0) | 184 | #define RADEON_PG_SUPPORT_GFX_PG (1 << 0) |
| 185 | #define RADEON_PG_SUPPORT_GFX_SMG (1 << 1) | 185 | #define RADEON_PG_SUPPORT_GFX_SMG (1 << 1) |
| 186 | #define RADEON_PG_SUPPORT_GFX_DMG (1 << 2) | 186 | #define RADEON_PG_SUPPORT_GFX_DMG (1 << 2) |
| 187 | #define RADEON_PG_SUPPORT_UVD (1 << 3) | 187 | #define RADEON_PG_SUPPORT_UVD (1 << 3) |
| @@ -1778,6 +1778,7 @@ struct radeon_asic { | |||
| 1778 | int (*force_performance_level)(struct radeon_device *rdev, enum radeon_dpm_forced_level level); | 1778 | int (*force_performance_level)(struct radeon_device *rdev, enum radeon_dpm_forced_level level); |
| 1779 | bool (*vblank_too_short)(struct radeon_device *rdev); | 1779 | bool (*vblank_too_short)(struct radeon_device *rdev); |
| 1780 | void (*powergate_uvd)(struct radeon_device *rdev, bool gate); | 1780 | void (*powergate_uvd)(struct radeon_device *rdev, bool gate); |
| 1781 | void (*enable_bapm)(struct radeon_device *rdev, bool enable); | ||
| 1781 | } dpm; | 1782 | } dpm; |
| 1782 | /* pageflipping */ | 1783 | /* pageflipping */ |
| 1783 | struct { | 1784 | struct { |
| @@ -2110,6 +2111,28 @@ struct radeon_device { | |||
| 2110 | resource_size_t rmmio_size; | 2111 | resource_size_t rmmio_size; |
| 2111 | /* protects concurrent MM_INDEX/DATA based register access */ | 2112 | /* protects concurrent MM_INDEX/DATA based register access */ |
| 2112 | spinlock_t mmio_idx_lock; | 2113 | spinlock_t mmio_idx_lock; |
| 2114 | /* protects concurrent SMC based register access */ | ||
| 2115 | spinlock_t smc_idx_lock; | ||
| 2116 | /* protects concurrent PLL register access */ | ||
| 2117 | spinlock_t pll_idx_lock; | ||
| 2118 | /* protects concurrent MC register access */ | ||
| 2119 | spinlock_t mc_idx_lock; | ||
| 2120 | /* protects concurrent PCIE register access */ | ||
| 2121 | spinlock_t pcie_idx_lock; | ||
| 2122 | /* protects concurrent PCIE_PORT register access */ | ||
| 2123 | spinlock_t pciep_idx_lock; | ||
| 2124 | /* protects concurrent PIF register access */ | ||
| 2125 | spinlock_t pif_idx_lock; | ||
| 2126 | /* protects concurrent CG register access */ | ||
| 2127 | spinlock_t cg_idx_lock; | ||
| 2128 | /* protects concurrent UVD register access */ | ||
| 2129 | spinlock_t uvd_idx_lock; | ||
| 2130 | /* protects concurrent RCU register access */ | ||
| 2131 | spinlock_t rcu_idx_lock; | ||
| 2132 | /* protects concurrent DIDT register access */ | ||
| 2133 | spinlock_t didt_idx_lock; | ||
| 2134 | /* protects concurrent ENDPOINT (audio) register access */ | ||
| 2135 | spinlock_t end_idx_lock; | ||
| 2113 | void __iomem *rmmio; | 2136 | void __iomem *rmmio; |
| 2114 | radeon_rreg_t mc_rreg; | 2137 | radeon_rreg_t mc_rreg; |
| 2115 | radeon_wreg_t mc_wreg; | 2138 | radeon_wreg_t mc_wreg; |
| @@ -2277,123 +2300,179 @@ void cik_mm_wdoorbell(struct radeon_device *rdev, u32 offset, u32 v); | |||
| 2277 | */ | 2300 | */ |
| 2278 | static inline uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg) | 2301 | static inline uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg) |
| 2279 | { | 2302 | { |
| 2303 | unsigned long flags; | ||
| 2280 | uint32_t r; | 2304 | uint32_t r; |
| 2281 | 2305 | ||
| 2306 | spin_lock_irqsave(&rdev->pcie_idx_lock, flags); | ||
| 2282 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); | 2307 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); |
| 2283 | r = RREG32(RADEON_PCIE_DATA); | 2308 | r = RREG32(RADEON_PCIE_DATA); |
| 2309 | spin_unlock_irqrestore(&rdev->pcie_idx_lock, flags); | ||
| 2284 | return r; | 2310 | return r; |
| 2285 | } | 2311 | } |
| 2286 | 2312 | ||
| 2287 | static inline void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 2313 | static inline void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 2288 | { | 2314 | { |
| 2315 | unsigned long flags; | ||
| 2316 | |||
| 2317 | spin_lock_irqsave(&rdev->pcie_idx_lock, flags); | ||
| 2289 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); | 2318 | WREG32(RADEON_PCIE_INDEX, ((reg) & rdev->pcie_reg_mask)); |
| 2290 | WREG32(RADEON_PCIE_DATA, (v)); | 2319 | WREG32(RADEON_PCIE_DATA, (v)); |
| 2320 | spin_unlock_irqrestore(&rdev->pcie_idx_lock, flags); | ||
| 2291 | } | 2321 | } |
| 2292 | 2322 | ||
| 2293 | static inline u32 tn_smc_rreg(struct radeon_device *rdev, u32 reg) | 2323 | static inline u32 tn_smc_rreg(struct radeon_device *rdev, u32 reg) |
| 2294 | { | 2324 | { |
| 2325 | unsigned long flags; | ||
| 2295 | u32 r; | 2326 | u32 r; |
| 2296 | 2327 | ||
| 2328 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 2297 | WREG32(TN_SMC_IND_INDEX_0, (reg)); | 2329 | WREG32(TN_SMC_IND_INDEX_0, (reg)); |
| 2298 | r = RREG32(TN_SMC_IND_DATA_0); | 2330 | r = RREG32(TN_SMC_IND_DATA_0); |
| 2331 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 2299 | return r; | 2332 | return r; |
| 2300 | } | 2333 | } |
| 2301 | 2334 | ||
| 2302 | static inline void tn_smc_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2335 | static inline void tn_smc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2303 | { | 2336 | { |
| 2337 | unsigned long flags; | ||
| 2338 | |||
| 2339 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 2304 | WREG32(TN_SMC_IND_INDEX_0, (reg)); | 2340 | WREG32(TN_SMC_IND_INDEX_0, (reg)); |
| 2305 | WREG32(TN_SMC_IND_DATA_0, (v)); | 2341 | WREG32(TN_SMC_IND_DATA_0, (v)); |
| 2342 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 2306 | } | 2343 | } |
| 2307 | 2344 | ||
| 2308 | static inline u32 r600_rcu_rreg(struct radeon_device *rdev, u32 reg) | 2345 | static inline u32 r600_rcu_rreg(struct radeon_device *rdev, u32 reg) |
| 2309 | { | 2346 | { |
| 2347 | unsigned long flags; | ||
| 2310 | u32 r; | 2348 | u32 r; |
| 2311 | 2349 | ||
| 2350 | spin_lock_irqsave(&rdev->rcu_idx_lock, flags); | ||
| 2312 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); | 2351 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); |
| 2313 | r = RREG32(R600_RCU_DATA); | 2352 | r = RREG32(R600_RCU_DATA); |
| 2353 | spin_unlock_irqrestore(&rdev->rcu_idx_lock, flags); | ||
| 2314 | return r; | 2354 | return r; |
| 2315 | } | 2355 | } |
| 2316 | 2356 | ||
| 2317 | static inline void r600_rcu_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2357 | static inline void r600_rcu_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2318 | { | 2358 | { |
| 2359 | unsigned long flags; | ||
| 2360 | |||
| 2361 | spin_lock_irqsave(&rdev->rcu_idx_lock, flags); | ||
| 2319 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); | 2362 | WREG32(R600_RCU_INDEX, ((reg) & 0x1fff)); |
| 2320 | WREG32(R600_RCU_DATA, (v)); | 2363 | WREG32(R600_RCU_DATA, (v)); |
| 2364 | spin_unlock_irqrestore(&rdev->rcu_idx_lock, flags); | ||
| 2321 | } | 2365 | } |
| 2322 | 2366 | ||
| 2323 | static inline u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg) | 2367 | static inline u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg) |
| 2324 | { | 2368 | { |
| 2369 | unsigned long flags; | ||
| 2325 | u32 r; | 2370 | u32 r; |
| 2326 | 2371 | ||
| 2372 | spin_lock_irqsave(&rdev->cg_idx_lock, flags); | ||
| 2327 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); | 2373 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); |
| 2328 | r = RREG32(EVERGREEN_CG_IND_DATA); | 2374 | r = RREG32(EVERGREEN_CG_IND_DATA); |
| 2375 | spin_unlock_irqrestore(&rdev->cg_idx_lock, flags); | ||
| 2329 | return r; | 2376 | return r; |
| 2330 | } | 2377 | } |
| 2331 | 2378 | ||
| 2332 | static inline void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2379 | static inline void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2333 | { | 2380 | { |
| 2381 | unsigned long flags; | ||
| 2382 | |||
| 2383 | spin_lock_irqsave(&rdev->cg_idx_lock, flags); | ||
| 2334 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); | 2384 | WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff)); |
| 2335 | WREG32(EVERGREEN_CG_IND_DATA, (v)); | 2385 | WREG32(EVERGREEN_CG_IND_DATA, (v)); |
| 2386 | spin_unlock_irqrestore(&rdev->cg_idx_lock, flags); | ||
| 2336 | } | 2387 | } |
| 2337 | 2388 | ||
| 2338 | static inline u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg) | 2389 | static inline u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg) |
| 2339 | { | 2390 | { |
| 2391 | unsigned long flags; | ||
| 2340 | u32 r; | 2392 | u32 r; |
| 2341 | 2393 | ||
| 2394 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
| 2342 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); | 2395 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); |
| 2343 | r = RREG32(EVERGREEN_PIF_PHY0_DATA); | 2396 | r = RREG32(EVERGREEN_PIF_PHY0_DATA); |
| 2397 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
| 2344 | return r; | 2398 | return r; |
| 2345 | } | 2399 | } |
| 2346 | 2400 | ||
| 2347 | static inline void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2401 | static inline void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2348 | { | 2402 | { |
| 2403 | unsigned long flags; | ||
| 2404 | |||
| 2405 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
| 2349 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); | 2406 | WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff)); |
| 2350 | WREG32(EVERGREEN_PIF_PHY0_DATA, (v)); | 2407 | WREG32(EVERGREEN_PIF_PHY0_DATA, (v)); |
| 2408 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
| 2351 | } | 2409 | } |
| 2352 | 2410 | ||
| 2353 | static inline u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg) | 2411 | static inline u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg) |
| 2354 | { | 2412 | { |
| 2413 | unsigned long flags; | ||
| 2355 | u32 r; | 2414 | u32 r; |
| 2356 | 2415 | ||
| 2416 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
| 2357 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); | 2417 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); |
| 2358 | r = RREG32(EVERGREEN_PIF_PHY1_DATA); | 2418 | r = RREG32(EVERGREEN_PIF_PHY1_DATA); |
| 2419 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
| 2359 | return r; | 2420 | return r; |
| 2360 | } | 2421 | } |
| 2361 | 2422 | ||
| 2362 | static inline void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2423 | static inline void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2363 | { | 2424 | { |
| 2425 | unsigned long flags; | ||
| 2426 | |||
| 2427 | spin_lock_irqsave(&rdev->pif_idx_lock, flags); | ||
| 2364 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); | 2428 | WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff)); |
| 2365 | WREG32(EVERGREEN_PIF_PHY1_DATA, (v)); | 2429 | WREG32(EVERGREEN_PIF_PHY1_DATA, (v)); |
| 2430 | spin_unlock_irqrestore(&rdev->pif_idx_lock, flags); | ||
| 2366 | } | 2431 | } |
| 2367 | 2432 | ||
| 2368 | static inline u32 r600_uvd_ctx_rreg(struct radeon_device *rdev, u32 reg) | 2433 | static inline u32 r600_uvd_ctx_rreg(struct radeon_device *rdev, u32 reg) |
| 2369 | { | 2434 | { |
| 2435 | unsigned long flags; | ||
| 2370 | u32 r; | 2436 | u32 r; |
| 2371 | 2437 | ||
| 2438 | spin_lock_irqsave(&rdev->uvd_idx_lock, flags); | ||
| 2372 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); | 2439 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); |
| 2373 | r = RREG32(R600_UVD_CTX_DATA); | 2440 | r = RREG32(R600_UVD_CTX_DATA); |
| 2441 | spin_unlock_irqrestore(&rdev->uvd_idx_lock, flags); | ||
| 2374 | return r; | 2442 | return r; |
| 2375 | } | 2443 | } |
| 2376 | 2444 | ||
| 2377 | static inline void r600_uvd_ctx_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2445 | static inline void r600_uvd_ctx_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2378 | { | 2446 | { |
| 2447 | unsigned long flags; | ||
| 2448 | |||
| 2449 | spin_lock_irqsave(&rdev->uvd_idx_lock, flags); | ||
| 2379 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); | 2450 | WREG32(R600_UVD_CTX_INDEX, ((reg) & 0x1ff)); |
| 2380 | WREG32(R600_UVD_CTX_DATA, (v)); | 2451 | WREG32(R600_UVD_CTX_DATA, (v)); |
| 2452 | spin_unlock_irqrestore(&rdev->uvd_idx_lock, flags); | ||
| 2381 | } | 2453 | } |
| 2382 | 2454 | ||
| 2383 | 2455 | ||
| 2384 | static inline u32 cik_didt_rreg(struct radeon_device *rdev, u32 reg) | 2456 | static inline u32 cik_didt_rreg(struct radeon_device *rdev, u32 reg) |
| 2385 | { | 2457 | { |
| 2458 | unsigned long flags; | ||
| 2386 | u32 r; | 2459 | u32 r; |
| 2387 | 2460 | ||
| 2461 | spin_lock_irqsave(&rdev->didt_idx_lock, flags); | ||
| 2388 | WREG32(CIK_DIDT_IND_INDEX, (reg)); | 2462 | WREG32(CIK_DIDT_IND_INDEX, (reg)); |
| 2389 | r = RREG32(CIK_DIDT_IND_DATA); | 2463 | r = RREG32(CIK_DIDT_IND_DATA); |
| 2464 | spin_unlock_irqrestore(&rdev->didt_idx_lock, flags); | ||
| 2390 | return r; | 2465 | return r; |
| 2391 | } | 2466 | } |
| 2392 | 2467 | ||
| 2393 | static inline void cik_didt_wreg(struct radeon_device *rdev, u32 reg, u32 v) | 2468 | static inline void cik_didt_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
| 2394 | { | 2469 | { |
| 2470 | unsigned long flags; | ||
| 2471 | |||
| 2472 | spin_lock_irqsave(&rdev->didt_idx_lock, flags); | ||
| 2395 | WREG32(CIK_DIDT_IND_INDEX, (reg)); | 2473 | WREG32(CIK_DIDT_IND_INDEX, (reg)); |
| 2396 | WREG32(CIK_DIDT_IND_DATA, (v)); | 2474 | WREG32(CIK_DIDT_IND_DATA, (v)); |
| 2475 | spin_unlock_irqrestore(&rdev->didt_idx_lock, flags); | ||
| 2397 | } | 2476 | } |
| 2398 | 2477 | ||
| 2399 | void r100_pll_errata_after_index(struct radeon_device *rdev); | 2478 | void r100_pll_errata_after_index(struct radeon_device *rdev); |
| @@ -2569,6 +2648,7 @@ void radeon_ring_write(struct radeon_ring *ring, uint32_t v); | |||
| 2569 | #define radeon_dpm_force_performance_level(rdev, l) rdev->asic->dpm.force_performance_level((rdev), (l)) | 2648 | #define radeon_dpm_force_performance_level(rdev, l) rdev->asic->dpm.force_performance_level((rdev), (l)) |
| 2570 | #define radeon_dpm_vblank_too_short(rdev) rdev->asic->dpm.vblank_too_short((rdev)) | 2649 | #define radeon_dpm_vblank_too_short(rdev) rdev->asic->dpm.vblank_too_short((rdev)) |
| 2571 | #define radeon_dpm_powergate_uvd(rdev, g) rdev->asic->dpm.powergate_uvd((rdev), (g)) | 2650 | #define radeon_dpm_powergate_uvd(rdev, g) rdev->asic->dpm.powergate_uvd((rdev), (g)) |
| 2651 | #define radeon_dpm_enable_bapm(rdev, e) rdev->asic->dpm.enable_bapm((rdev), (e)) | ||
| 2572 | 2652 | ||
| 2573 | /* Common functions */ | 2653 | /* Common functions */ |
| 2574 | /* AGP */ | 2654 | /* AGP */ |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.c b/drivers/gpu/drm/radeon/radeon_asic.c index 630853b96841..8f7e04538fd6 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.c +++ b/drivers/gpu/drm/radeon/radeon_asic.c | |||
| @@ -1004,6 +1004,8 @@ static struct radeon_asic rv6xx_asic = { | |||
| 1004 | .wait_for_vblank = &avivo_wait_for_vblank, | 1004 | .wait_for_vblank = &avivo_wait_for_vblank, |
| 1005 | .set_backlight_level = &atombios_set_backlight_level, | 1005 | .set_backlight_level = &atombios_set_backlight_level, |
| 1006 | .get_backlight_level = &atombios_get_backlight_level, | 1006 | .get_backlight_level = &atombios_get_backlight_level, |
| 1007 | .hdmi_enable = &r600_hdmi_enable, | ||
| 1008 | .hdmi_setmode = &r600_hdmi_setmode, | ||
| 1007 | }, | 1009 | }, |
| 1008 | .copy = { | 1010 | .copy = { |
| 1009 | .blit = &r600_copy_cpdma, | 1011 | .blit = &r600_copy_cpdma, |
| @@ -1037,6 +1039,7 @@ static struct radeon_asic rv6xx_asic = { | |||
| 1037 | .set_pcie_lanes = &r600_set_pcie_lanes, | 1039 | .set_pcie_lanes = &r600_set_pcie_lanes, |
| 1038 | .set_clock_gating = NULL, | 1040 | .set_clock_gating = NULL, |
| 1039 | .get_temperature = &rv6xx_get_temp, | 1041 | .get_temperature = &rv6xx_get_temp, |
| 1042 | .set_uvd_clocks = &r600_set_uvd_clocks, | ||
| 1040 | }, | 1043 | }, |
| 1041 | .dpm = { | 1044 | .dpm = { |
| 1042 | .init = &rv6xx_dpm_init, | 1045 | .init = &rv6xx_dpm_init, |
| @@ -1126,6 +1129,7 @@ static struct radeon_asic rs780_asic = { | |||
| 1126 | .set_pcie_lanes = NULL, | 1129 | .set_pcie_lanes = NULL, |
| 1127 | .set_clock_gating = NULL, | 1130 | .set_clock_gating = NULL, |
| 1128 | .get_temperature = &rv6xx_get_temp, | 1131 | .get_temperature = &rv6xx_get_temp, |
| 1132 | .set_uvd_clocks = &r600_set_uvd_clocks, | ||
| 1129 | }, | 1133 | }, |
| 1130 | .dpm = { | 1134 | .dpm = { |
| 1131 | .init = &rs780_dpm_init, | 1135 | .init = &rs780_dpm_init, |
| @@ -1141,6 +1145,7 @@ static struct radeon_asic rs780_asic = { | |||
| 1141 | .get_mclk = &rs780_dpm_get_mclk, | 1145 | .get_mclk = &rs780_dpm_get_mclk, |
| 1142 | .print_power_state = &rs780_dpm_print_power_state, | 1146 | .print_power_state = &rs780_dpm_print_power_state, |
| 1143 | .debugfs_print_current_performance_level = &rs780_dpm_debugfs_print_current_performance_level, | 1147 | .debugfs_print_current_performance_level = &rs780_dpm_debugfs_print_current_performance_level, |
| 1148 | .force_performance_level = &rs780_dpm_force_performance_level, | ||
| 1144 | }, | 1149 | }, |
| 1145 | .pflip = { | 1150 | .pflip = { |
| 1146 | .pre_page_flip = &rs600_pre_page_flip, | 1151 | .pre_page_flip = &rs600_pre_page_flip, |
| @@ -1791,6 +1796,7 @@ static struct radeon_asic trinity_asic = { | |||
| 1791 | .print_power_state = &trinity_dpm_print_power_state, | 1796 | .print_power_state = &trinity_dpm_print_power_state, |
| 1792 | .debugfs_print_current_performance_level = &trinity_dpm_debugfs_print_current_performance_level, | 1797 | .debugfs_print_current_performance_level = &trinity_dpm_debugfs_print_current_performance_level, |
| 1793 | .force_performance_level = &trinity_dpm_force_performance_level, | 1798 | .force_performance_level = &trinity_dpm_force_performance_level, |
| 1799 | .enable_bapm = &trinity_dpm_enable_bapm, | ||
| 1794 | }, | 1800 | }, |
| 1795 | .pflip = { | 1801 | .pflip = { |
| 1796 | .pre_page_flip = &evergreen_pre_page_flip, | 1802 | .pre_page_flip = &evergreen_pre_page_flip, |
| @@ -2166,6 +2172,7 @@ static struct radeon_asic kv_asic = { | |||
| 2166 | .debugfs_print_current_performance_level = &kv_dpm_debugfs_print_current_performance_level, | 2172 | .debugfs_print_current_performance_level = &kv_dpm_debugfs_print_current_performance_level, |
| 2167 | .force_performance_level = &kv_dpm_force_performance_level, | 2173 | .force_performance_level = &kv_dpm_force_performance_level, |
| 2168 | .powergate_uvd = &kv_dpm_powergate_uvd, | 2174 | .powergate_uvd = &kv_dpm_powergate_uvd, |
| 2175 | .enable_bapm = &kv_dpm_enable_bapm, | ||
| 2169 | }, | 2176 | }, |
| 2170 | .pflip = { | 2177 | .pflip = { |
| 2171 | .pre_page_flip = &evergreen_pre_page_flip, | 2178 | .pre_page_flip = &evergreen_pre_page_flip, |
| @@ -2390,7 +2397,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
| 2390 | RADEON_CG_SUPPORT_HDP_LS | | 2397 | RADEON_CG_SUPPORT_HDP_LS | |
| 2391 | RADEON_CG_SUPPORT_HDP_MGCG; | 2398 | RADEON_CG_SUPPORT_HDP_MGCG; |
| 2392 | rdev->pg_flags = 0 | | 2399 | rdev->pg_flags = 0 | |
| 2393 | /*RADEON_PG_SUPPORT_GFX_CG | */ | 2400 | /*RADEON_PG_SUPPORT_GFX_PG | */ |
| 2394 | RADEON_PG_SUPPORT_SDMA; | 2401 | RADEON_PG_SUPPORT_SDMA; |
| 2395 | break; | 2402 | break; |
| 2396 | case CHIP_OLAND: | 2403 | case CHIP_OLAND: |
| @@ -2479,7 +2486,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
| 2479 | RADEON_CG_SUPPORT_HDP_LS | | 2486 | RADEON_CG_SUPPORT_HDP_LS | |
| 2480 | RADEON_CG_SUPPORT_HDP_MGCG; | 2487 | RADEON_CG_SUPPORT_HDP_MGCG; |
| 2481 | rdev->pg_flags = 0; | 2488 | rdev->pg_flags = 0; |
| 2482 | /*RADEON_PG_SUPPORT_GFX_CG | | 2489 | /*RADEON_PG_SUPPORT_GFX_PG | |
| 2483 | RADEON_PG_SUPPORT_GFX_SMG | | 2490 | RADEON_PG_SUPPORT_GFX_SMG | |
| 2484 | RADEON_PG_SUPPORT_GFX_DMG | | 2491 | RADEON_PG_SUPPORT_GFX_DMG | |
| 2485 | RADEON_PG_SUPPORT_UVD | | 2492 | RADEON_PG_SUPPORT_UVD | |
| @@ -2507,7 +2514,7 @@ int radeon_asic_init(struct radeon_device *rdev) | |||
| 2507 | RADEON_CG_SUPPORT_HDP_LS | | 2514 | RADEON_CG_SUPPORT_HDP_LS | |
| 2508 | RADEON_CG_SUPPORT_HDP_MGCG; | 2515 | RADEON_CG_SUPPORT_HDP_MGCG; |
| 2509 | rdev->pg_flags = 0; | 2516 | rdev->pg_flags = 0; |
| 2510 | /*RADEON_PG_SUPPORT_GFX_CG | | 2517 | /*RADEON_PG_SUPPORT_GFX_PG | |
| 2511 | RADEON_PG_SUPPORT_GFX_SMG | | 2518 | RADEON_PG_SUPPORT_GFX_SMG | |
| 2512 | RADEON_PG_SUPPORT_UVD | | 2519 | RADEON_PG_SUPPORT_UVD | |
| 2513 | RADEON_PG_SUPPORT_VCE | | 2520 | RADEON_PG_SUPPORT_VCE | |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.h b/drivers/gpu/drm/radeon/radeon_asic.h index 818bbe6b884b..70c29d5e080d 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.h +++ b/drivers/gpu/drm/radeon/radeon_asic.h | |||
| @@ -389,6 +389,7 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev); | |||
| 389 | u32 r600_get_xclk(struct radeon_device *rdev); | 389 | u32 r600_get_xclk(struct radeon_device *rdev); |
| 390 | uint64_t r600_get_gpu_clock_counter(struct radeon_device *rdev); | 390 | uint64_t r600_get_gpu_clock_counter(struct radeon_device *rdev); |
| 391 | int rv6xx_get_temp(struct radeon_device *rdev); | 391 | int rv6xx_get_temp(struct radeon_device *rdev); |
| 392 | int r600_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk); | ||
| 392 | int r600_dpm_pre_set_power_state(struct radeon_device *rdev); | 393 | int r600_dpm_pre_set_power_state(struct radeon_device *rdev); |
| 393 | void r600_dpm_post_set_power_state(struct radeon_device *rdev); | 394 | void r600_dpm_post_set_power_state(struct radeon_device *rdev); |
| 394 | /* r600 dma */ | 395 | /* r600 dma */ |
| @@ -428,6 +429,8 @@ void rs780_dpm_print_power_state(struct radeon_device *rdev, | |||
| 428 | struct radeon_ps *ps); | 429 | struct radeon_ps *ps); |
| 429 | void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, | 430 | void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, |
| 430 | struct seq_file *m); | 431 | struct seq_file *m); |
| 432 | int rs780_dpm_force_performance_level(struct radeon_device *rdev, | ||
| 433 | enum radeon_dpm_forced_level level); | ||
| 431 | 434 | ||
| 432 | /* | 435 | /* |
| 433 | * rv770,rv730,rv710,rv740 | 436 | * rv770,rv730,rv710,rv740 |
| @@ -625,6 +628,7 @@ void trinity_dpm_debugfs_print_current_performance_level(struct radeon_device *r | |||
| 625 | struct seq_file *m); | 628 | struct seq_file *m); |
| 626 | int trinity_dpm_force_performance_level(struct radeon_device *rdev, | 629 | int trinity_dpm_force_performance_level(struct radeon_device *rdev, |
| 627 | enum radeon_dpm_forced_level level); | 630 | enum radeon_dpm_forced_level level); |
| 631 | void trinity_dpm_enable_bapm(struct radeon_device *rdev, bool enable); | ||
| 628 | 632 | ||
| 629 | /* DCE6 - SI */ | 633 | /* DCE6 - SI */ |
| 630 | void dce6_bandwidth_update(struct radeon_device *rdev); | 634 | void dce6_bandwidth_update(struct radeon_device *rdev); |
| @@ -781,6 +785,7 @@ void kv_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, | |||
| 781 | int kv_dpm_force_performance_level(struct radeon_device *rdev, | 785 | int kv_dpm_force_performance_level(struct radeon_device *rdev, |
| 782 | enum radeon_dpm_forced_level level); | 786 | enum radeon_dpm_forced_level level); |
| 783 | void kv_dpm_powergate_uvd(struct radeon_device *rdev, bool gate); | 787 | void kv_dpm_powergate_uvd(struct radeon_device *rdev, bool gate); |
| 788 | void kv_dpm_enable_bapm(struct radeon_device *rdev, bool enable); | ||
| 784 | 789 | ||
| 785 | /* uvd v1.0 */ | 790 | /* uvd v1.0 */ |
| 786 | uint32_t uvd_v1_0_get_rptr(struct radeon_device *rdev, | 791 | uint32_t uvd_v1_0_get_rptr(struct radeon_device *rdev, |
diff --git a/drivers/gpu/drm/radeon/radeon_atombios.c b/drivers/gpu/drm/radeon/radeon_atombios.c index 404e25d285ba..f79ee184ffd5 100644 --- a/drivers/gpu/drm/radeon/radeon_atombios.c +++ b/drivers/gpu/drm/radeon/radeon_atombios.c | |||
| @@ -1367,6 +1367,7 @@ bool radeon_atombios_get_ppll_ss_info(struct radeon_device *rdev, | |||
| 1367 | int index = GetIndexIntoMasterTable(DATA, PPLL_SS_Info); | 1367 | int index = GetIndexIntoMasterTable(DATA, PPLL_SS_Info); |
| 1368 | uint16_t data_offset, size; | 1368 | uint16_t data_offset, size; |
| 1369 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; | 1369 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; |
| 1370 | struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT *ss_assign; | ||
| 1370 | uint8_t frev, crev; | 1371 | uint8_t frev, crev; |
| 1371 | int i, num_indices; | 1372 | int i, num_indices; |
| 1372 | 1373 | ||
| @@ -1378,18 +1379,21 @@ bool radeon_atombios_get_ppll_ss_info(struct radeon_device *rdev, | |||
| 1378 | 1379 | ||
| 1379 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1380 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
| 1380 | sizeof(ATOM_SPREAD_SPECTRUM_ASSIGNMENT); | 1381 | sizeof(ATOM_SPREAD_SPECTRUM_ASSIGNMENT); |
| 1381 | 1382 | ss_assign = (struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT*) | |
| 1383 | ((u8 *)&ss_info->asSS_Info[0]); | ||
| 1382 | for (i = 0; i < num_indices; i++) { | 1384 | for (i = 0; i < num_indices; i++) { |
| 1383 | if (ss_info->asSS_Info[i].ucSS_Id == id) { | 1385 | if (ss_assign->ucSS_Id == id) { |
| 1384 | ss->percentage = | 1386 | ss->percentage = |
| 1385 | le16_to_cpu(ss_info->asSS_Info[i].usSpreadSpectrumPercentage); | 1387 | le16_to_cpu(ss_assign->usSpreadSpectrumPercentage); |
| 1386 | ss->type = ss_info->asSS_Info[i].ucSpreadSpectrumType; | 1388 | ss->type = ss_assign->ucSpreadSpectrumType; |
| 1387 | ss->step = ss_info->asSS_Info[i].ucSS_Step; | 1389 | ss->step = ss_assign->ucSS_Step; |
| 1388 | ss->delay = ss_info->asSS_Info[i].ucSS_Delay; | 1390 | ss->delay = ss_assign->ucSS_Delay; |
| 1389 | ss->range = ss_info->asSS_Info[i].ucSS_Range; | 1391 | ss->range = ss_assign->ucSS_Range; |
| 1390 | ss->refdiv = ss_info->asSS_Info[i].ucRecommendedRef_Div; | 1392 | ss->refdiv = ss_assign->ucRecommendedRef_Div; |
| 1391 | return true; | 1393 | return true; |
| 1392 | } | 1394 | } |
| 1395 | ss_assign = (struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT*) | ||
| 1396 | ((u8 *)ss_assign + sizeof(struct _ATOM_SPREAD_SPECTRUM_ASSIGNMENT)); | ||
| 1393 | } | 1397 | } |
| 1394 | } | 1398 | } |
| 1395 | return false; | 1399 | return false; |
| @@ -1477,6 +1481,12 @@ union asic_ss_info { | |||
| 1477 | struct _ATOM_ASIC_INTERNAL_SS_INFO_V3 info_3; | 1481 | struct _ATOM_ASIC_INTERNAL_SS_INFO_V3 info_3; |
| 1478 | }; | 1482 | }; |
| 1479 | 1483 | ||
| 1484 | union asic_ss_assignment { | ||
| 1485 | struct _ATOM_ASIC_SS_ASSIGNMENT v1; | ||
| 1486 | struct _ATOM_ASIC_SS_ASSIGNMENT_V2 v2; | ||
| 1487 | struct _ATOM_ASIC_SS_ASSIGNMENT_V3 v3; | ||
| 1488 | }; | ||
| 1489 | |||
| 1480 | bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | 1490 | bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, |
| 1481 | struct radeon_atom_ss *ss, | 1491 | struct radeon_atom_ss *ss, |
| 1482 | int id, u32 clock) | 1492 | int id, u32 clock) |
| @@ -1485,6 +1495,7 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
| 1485 | int index = GetIndexIntoMasterTable(DATA, ASIC_InternalSS_Info); | 1495 | int index = GetIndexIntoMasterTable(DATA, ASIC_InternalSS_Info); |
| 1486 | uint16_t data_offset, size; | 1496 | uint16_t data_offset, size; |
| 1487 | union asic_ss_info *ss_info; | 1497 | union asic_ss_info *ss_info; |
| 1498 | union asic_ss_assignment *ss_assign; | ||
| 1488 | uint8_t frev, crev; | 1499 | uint8_t frev, crev; |
| 1489 | int i, num_indices; | 1500 | int i, num_indices; |
| 1490 | 1501 | ||
| @@ -1509,45 +1520,52 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
| 1509 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1520 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
| 1510 | sizeof(ATOM_ASIC_SS_ASSIGNMENT); | 1521 | sizeof(ATOM_ASIC_SS_ASSIGNMENT); |
| 1511 | 1522 | ||
| 1523 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info.asSpreadSpectrum[0]); | ||
| 1512 | for (i = 0; i < num_indices; i++) { | 1524 | for (i = 0; i < num_indices; i++) { |
| 1513 | if ((ss_info->info.asSpreadSpectrum[i].ucClockIndication == id) && | 1525 | if ((ss_assign->v1.ucClockIndication == id) && |
| 1514 | (clock <= le32_to_cpu(ss_info->info.asSpreadSpectrum[i].ulTargetClockRange))) { | 1526 | (clock <= le32_to_cpu(ss_assign->v1.ulTargetClockRange))) { |
| 1515 | ss->percentage = | 1527 | ss->percentage = |
| 1516 | le16_to_cpu(ss_info->info.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1528 | le16_to_cpu(ss_assign->v1.usSpreadSpectrumPercentage); |
| 1517 | ss->type = ss_info->info.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1529 | ss->type = ss_assign->v1.ucSpreadSpectrumMode; |
| 1518 | ss->rate = le16_to_cpu(ss_info->info.asSpreadSpectrum[i].usSpreadRateInKhz); | 1530 | ss->rate = le16_to_cpu(ss_assign->v1.usSpreadRateInKhz); |
| 1519 | return true; | 1531 | return true; |
| 1520 | } | 1532 | } |
| 1533 | ss_assign = (union asic_ss_assignment *) | ||
| 1534 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT)); | ||
| 1521 | } | 1535 | } |
| 1522 | break; | 1536 | break; |
| 1523 | case 2: | 1537 | case 2: |
| 1524 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1538 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
| 1525 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2); | 1539 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2); |
| 1540 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info_2.asSpreadSpectrum[0]); | ||
| 1526 | for (i = 0; i < num_indices; i++) { | 1541 | for (i = 0; i < num_indices; i++) { |
| 1527 | if ((ss_info->info_2.asSpreadSpectrum[i].ucClockIndication == id) && | 1542 | if ((ss_assign->v2.ucClockIndication == id) && |
| 1528 | (clock <= le32_to_cpu(ss_info->info_2.asSpreadSpectrum[i].ulTargetClockRange))) { | 1543 | (clock <= le32_to_cpu(ss_assign->v2.ulTargetClockRange))) { |
| 1529 | ss->percentage = | 1544 | ss->percentage = |
| 1530 | le16_to_cpu(ss_info->info_2.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1545 | le16_to_cpu(ss_assign->v2.usSpreadSpectrumPercentage); |
| 1531 | ss->type = ss_info->info_2.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1546 | ss->type = ss_assign->v2.ucSpreadSpectrumMode; |
| 1532 | ss->rate = le16_to_cpu(ss_info->info_2.asSpreadSpectrum[i].usSpreadRateIn10Hz); | 1547 | ss->rate = le16_to_cpu(ss_assign->v2.usSpreadRateIn10Hz); |
| 1533 | if ((crev == 2) && | 1548 | if ((crev == 2) && |
| 1534 | ((id == ASIC_INTERNAL_ENGINE_SS) || | 1549 | ((id == ASIC_INTERNAL_ENGINE_SS) || |
| 1535 | (id == ASIC_INTERNAL_MEMORY_SS))) | 1550 | (id == ASIC_INTERNAL_MEMORY_SS))) |
| 1536 | ss->rate /= 100; | 1551 | ss->rate /= 100; |
| 1537 | return true; | 1552 | return true; |
| 1538 | } | 1553 | } |
| 1554 | ss_assign = (union asic_ss_assignment *) | ||
| 1555 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT_V2)); | ||
| 1539 | } | 1556 | } |
| 1540 | break; | 1557 | break; |
| 1541 | case 3: | 1558 | case 3: |
| 1542 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / | 1559 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / |
| 1543 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3); | 1560 | sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3); |
| 1561 | ss_assign = (union asic_ss_assignment *)((u8 *)&ss_info->info_3.asSpreadSpectrum[0]); | ||
| 1544 | for (i = 0; i < num_indices; i++) { | 1562 | for (i = 0; i < num_indices; i++) { |
| 1545 | if ((ss_info->info_3.asSpreadSpectrum[i].ucClockIndication == id) && | 1563 | if ((ss_assign->v3.ucClockIndication == id) && |
| 1546 | (clock <= le32_to_cpu(ss_info->info_3.asSpreadSpectrum[i].ulTargetClockRange))) { | 1564 | (clock <= le32_to_cpu(ss_assign->v3.ulTargetClockRange))) { |
| 1547 | ss->percentage = | 1565 | ss->percentage = |
| 1548 | le16_to_cpu(ss_info->info_3.asSpreadSpectrum[i].usSpreadSpectrumPercentage); | 1566 | le16_to_cpu(ss_assign->v3.usSpreadSpectrumPercentage); |
| 1549 | ss->type = ss_info->info_3.asSpreadSpectrum[i].ucSpreadSpectrumMode; | 1567 | ss->type = ss_assign->v3.ucSpreadSpectrumMode; |
| 1550 | ss->rate = le16_to_cpu(ss_info->info_3.asSpreadSpectrum[i].usSpreadRateIn10Hz); | 1568 | ss->rate = le16_to_cpu(ss_assign->v3.usSpreadRateIn10Hz); |
| 1551 | if ((id == ASIC_INTERNAL_ENGINE_SS) || | 1569 | if ((id == ASIC_INTERNAL_ENGINE_SS) || |
| 1552 | (id == ASIC_INTERNAL_MEMORY_SS)) | 1570 | (id == ASIC_INTERNAL_MEMORY_SS)) |
| 1553 | ss->rate /= 100; | 1571 | ss->rate /= 100; |
| @@ -1555,6 +1573,8 @@ bool radeon_atombios_get_asic_ss_info(struct radeon_device *rdev, | |||
| 1555 | radeon_atombios_get_igp_ss_overrides(rdev, ss, id); | 1573 | radeon_atombios_get_igp_ss_overrides(rdev, ss, id); |
| 1556 | return true; | 1574 | return true; |
| 1557 | } | 1575 | } |
| 1576 | ss_assign = (union asic_ss_assignment *) | ||
| 1577 | ((u8 *)ss_assign + sizeof(ATOM_ASIC_SS_ASSIGNMENT_V3)); | ||
| 1558 | } | 1578 | } |
| 1559 | break; | 1579 | break; |
| 1560 | default: | 1580 | default: |
diff --git a/drivers/gpu/drm/radeon/radeon_connectors.c b/drivers/gpu/drm/radeon/radeon_connectors.c index 2399f25ec037..79159b5da05b 100644 --- a/drivers/gpu/drm/radeon/radeon_connectors.c +++ b/drivers/gpu/drm/radeon/radeon_connectors.c | |||
| @@ -396,6 +396,21 @@ static int radeon_connector_set_property(struct drm_connector *connector, struct | |||
| 396 | } | 396 | } |
| 397 | } | 397 | } |
| 398 | 398 | ||
| 399 | if (property == rdev->mode_info.audio_property) { | ||
| 400 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 401 | /* need to find digital encoder on connector */ | ||
| 402 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); | ||
| 403 | if (!encoder) | ||
| 404 | return 0; | ||
| 405 | |||
| 406 | radeon_encoder = to_radeon_encoder(encoder); | ||
| 407 | |||
| 408 | if (radeon_connector->audio != val) { | ||
| 409 | radeon_connector->audio = val; | ||
| 410 | radeon_property_change_mode(&radeon_encoder->base); | ||
| 411 | } | ||
| 412 | } | ||
| 413 | |||
| 399 | if (property == rdev->mode_info.underscan_property) { | 414 | if (property == rdev->mode_info.underscan_property) { |
| 400 | /* need to find digital encoder on connector */ | 415 | /* need to find digital encoder on connector */ |
| 401 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); | 416 | encoder = radeon_find_encoder(connector, DRM_MODE_ENCODER_TMDS); |
| @@ -1420,7 +1435,7 @@ radeon_dp_detect(struct drm_connector *connector, bool force) | |||
| 1420 | if (radeon_dp_getdpcd(radeon_connector)) | 1435 | if (radeon_dp_getdpcd(radeon_connector)) |
| 1421 | ret = connector_status_connected; | 1436 | ret = connector_status_connected; |
| 1422 | } else { | 1437 | } else { |
| 1423 | /* try non-aux ddc (DP to DVI/HMDI/etc. adapter) */ | 1438 | /* try non-aux ddc (DP to DVI/HDMI/etc. adapter) */ |
| 1424 | if (radeon_ddc_probe(radeon_connector, false)) | 1439 | if (radeon_ddc_probe(radeon_connector, false)) |
| 1425 | ret = connector_status_connected; | 1440 | ret = connector_status_connected; |
| 1426 | } | 1441 | } |
| @@ -1489,6 +1504,24 @@ static const struct drm_connector_funcs radeon_dp_connector_funcs = { | |||
| 1489 | .force = radeon_dvi_force, | 1504 | .force = radeon_dvi_force, |
| 1490 | }; | 1505 | }; |
| 1491 | 1506 | ||
| 1507 | static const struct drm_connector_funcs radeon_edp_connector_funcs = { | ||
| 1508 | .dpms = drm_helper_connector_dpms, | ||
| 1509 | .detect = radeon_dp_detect, | ||
| 1510 | .fill_modes = drm_helper_probe_single_connector_modes, | ||
| 1511 | .set_property = radeon_lvds_set_property, | ||
| 1512 | .destroy = radeon_dp_connector_destroy, | ||
| 1513 | .force = radeon_dvi_force, | ||
| 1514 | }; | ||
| 1515 | |||
| 1516 | static const struct drm_connector_funcs radeon_lvds_bridge_connector_funcs = { | ||
| 1517 | .dpms = drm_helper_connector_dpms, | ||
| 1518 | .detect = radeon_dp_detect, | ||
| 1519 | .fill_modes = drm_helper_probe_single_connector_modes, | ||
| 1520 | .set_property = radeon_lvds_set_property, | ||
| 1521 | .destroy = radeon_dp_connector_destroy, | ||
| 1522 | .force = radeon_dvi_force, | ||
| 1523 | }; | ||
| 1524 | |||
| 1492 | void | 1525 | void |
| 1493 | radeon_add_atom_connector(struct drm_device *dev, | 1526 | radeon_add_atom_connector(struct drm_device *dev, |
| 1494 | uint32_t connector_id, | 1527 | uint32_t connector_id, |
| @@ -1580,8 +1613,6 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1580 | goto failed; | 1613 | goto failed; |
| 1581 | radeon_dig_connector->igp_lane_info = igp_lane_info; | 1614 | radeon_dig_connector->igp_lane_info = igp_lane_info; |
| 1582 | radeon_connector->con_priv = radeon_dig_connector; | 1615 | radeon_connector->con_priv = radeon_dig_connector; |
| 1583 | drm_connector_init(dev, &radeon_connector->base, &radeon_dp_connector_funcs, connector_type); | ||
| 1584 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); | ||
| 1585 | if (i2c_bus->valid) { | 1616 | if (i2c_bus->valid) { |
| 1586 | /* add DP i2c bus */ | 1617 | /* add DP i2c bus */ |
| 1587 | if (connector_type == DRM_MODE_CONNECTOR_eDP) | 1618 | if (connector_type == DRM_MODE_CONNECTOR_eDP) |
| @@ -1598,6 +1629,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1598 | case DRM_MODE_CONNECTOR_VGA: | 1629 | case DRM_MODE_CONNECTOR_VGA: |
| 1599 | case DRM_MODE_CONNECTOR_DVIA: | 1630 | case DRM_MODE_CONNECTOR_DVIA: |
| 1600 | default: | 1631 | default: |
| 1632 | drm_connector_init(dev, &radeon_connector->base, | ||
| 1633 | &radeon_dp_connector_funcs, connector_type); | ||
| 1634 | drm_connector_helper_add(&radeon_connector->base, | ||
| 1635 | &radeon_dp_connector_helper_funcs); | ||
| 1601 | connector->interlace_allowed = true; | 1636 | connector->interlace_allowed = true; |
| 1602 | connector->doublescan_allowed = true; | 1637 | connector->doublescan_allowed = true; |
| 1603 | radeon_connector->dac_load_detect = true; | 1638 | radeon_connector->dac_load_detect = true; |
| @@ -1610,6 +1645,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1610 | case DRM_MODE_CONNECTOR_HDMIA: | 1645 | case DRM_MODE_CONNECTOR_HDMIA: |
| 1611 | case DRM_MODE_CONNECTOR_HDMIB: | 1646 | case DRM_MODE_CONNECTOR_HDMIB: |
| 1612 | case DRM_MODE_CONNECTOR_DisplayPort: | 1647 | case DRM_MODE_CONNECTOR_DisplayPort: |
| 1648 | drm_connector_init(dev, &radeon_connector->base, | ||
| 1649 | &radeon_dp_connector_funcs, connector_type); | ||
| 1650 | drm_connector_helper_add(&radeon_connector->base, | ||
| 1651 | &radeon_dp_connector_helper_funcs); | ||
| 1613 | drm_object_attach_property(&radeon_connector->base.base, | 1652 | drm_object_attach_property(&radeon_connector->base.base, |
| 1614 | rdev->mode_info.underscan_property, | 1653 | rdev->mode_info.underscan_property, |
| 1615 | UNDERSCAN_OFF); | 1654 | UNDERSCAN_OFF); |
| @@ -1619,6 +1658,9 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1619 | drm_object_attach_property(&radeon_connector->base.base, | 1658 | drm_object_attach_property(&radeon_connector->base.base, |
| 1620 | rdev->mode_info.underscan_vborder_property, | 1659 | rdev->mode_info.underscan_vborder_property, |
| 1621 | 0); | 1660 | 0); |
| 1661 | drm_object_attach_property(&radeon_connector->base.base, | ||
| 1662 | rdev->mode_info.audio_property, | ||
| 1663 | RADEON_AUDIO_DISABLE); | ||
| 1622 | subpixel_order = SubPixelHorizontalRGB; | 1664 | subpixel_order = SubPixelHorizontalRGB; |
| 1623 | connector->interlace_allowed = true; | 1665 | connector->interlace_allowed = true; |
| 1624 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) | 1666 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) |
| @@ -1634,6 +1676,10 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1634 | break; | 1676 | break; |
| 1635 | case DRM_MODE_CONNECTOR_LVDS: | 1677 | case DRM_MODE_CONNECTOR_LVDS: |
| 1636 | case DRM_MODE_CONNECTOR_eDP: | 1678 | case DRM_MODE_CONNECTOR_eDP: |
| 1679 | drm_connector_init(dev, &radeon_connector->base, | ||
| 1680 | &radeon_lvds_bridge_connector_funcs, connector_type); | ||
| 1681 | drm_connector_helper_add(&radeon_connector->base, | ||
| 1682 | &radeon_dp_connector_helper_funcs); | ||
| 1637 | drm_object_attach_property(&radeon_connector->base.base, | 1683 | drm_object_attach_property(&radeon_connector->base.base, |
| 1638 | dev->mode_config.scaling_mode_property, | 1684 | dev->mode_config.scaling_mode_property, |
| 1639 | DRM_MODE_SCALE_FULLSCREEN); | 1685 | DRM_MODE_SCALE_FULLSCREEN); |
| @@ -1708,6 +1754,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1708 | rdev->mode_info.underscan_vborder_property, | 1754 | rdev->mode_info.underscan_vborder_property, |
| 1709 | 0); | 1755 | 0); |
| 1710 | } | 1756 | } |
| 1757 | if (ASIC_IS_DCE2(rdev)) { | ||
| 1758 | drm_object_attach_property(&radeon_connector->base.base, | ||
| 1759 | rdev->mode_info.audio_property, | ||
| 1760 | RADEON_AUDIO_DISABLE); | ||
| 1761 | } | ||
| 1711 | if (connector_type == DRM_MODE_CONNECTOR_DVII) { | 1762 | if (connector_type == DRM_MODE_CONNECTOR_DVII) { |
| 1712 | radeon_connector->dac_load_detect = true; | 1763 | radeon_connector->dac_load_detect = true; |
| 1713 | drm_object_attach_property(&radeon_connector->base.base, | 1764 | drm_object_attach_property(&radeon_connector->base.base, |
| @@ -1748,6 +1799,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1748 | rdev->mode_info.underscan_vborder_property, | 1799 | rdev->mode_info.underscan_vborder_property, |
| 1749 | 0); | 1800 | 0); |
| 1750 | } | 1801 | } |
| 1802 | if (ASIC_IS_DCE2(rdev)) { | ||
| 1803 | drm_object_attach_property(&radeon_connector->base.base, | ||
| 1804 | rdev->mode_info.audio_property, | ||
| 1805 | RADEON_AUDIO_DISABLE); | ||
| 1806 | } | ||
| 1751 | subpixel_order = SubPixelHorizontalRGB; | 1807 | subpixel_order = SubPixelHorizontalRGB; |
| 1752 | connector->interlace_allowed = true; | 1808 | connector->interlace_allowed = true; |
| 1753 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) | 1809 | if (connector_type == DRM_MODE_CONNECTOR_HDMIB) |
| @@ -1787,6 +1843,11 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1787 | rdev->mode_info.underscan_vborder_property, | 1843 | rdev->mode_info.underscan_vborder_property, |
| 1788 | 0); | 1844 | 0); |
| 1789 | } | 1845 | } |
| 1846 | if (ASIC_IS_DCE2(rdev)) { | ||
| 1847 | drm_object_attach_property(&radeon_connector->base.base, | ||
| 1848 | rdev->mode_info.audio_property, | ||
| 1849 | RADEON_AUDIO_DISABLE); | ||
| 1850 | } | ||
| 1790 | connector->interlace_allowed = true; | 1851 | connector->interlace_allowed = true; |
| 1791 | /* in theory with a DP to VGA converter... */ | 1852 | /* in theory with a DP to VGA converter... */ |
| 1792 | connector->doublescan_allowed = false; | 1853 | connector->doublescan_allowed = false; |
| @@ -1797,7 +1858,7 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1797 | goto failed; | 1858 | goto failed; |
| 1798 | radeon_dig_connector->igp_lane_info = igp_lane_info; | 1859 | radeon_dig_connector->igp_lane_info = igp_lane_info; |
| 1799 | radeon_connector->con_priv = radeon_dig_connector; | 1860 | radeon_connector->con_priv = radeon_dig_connector; |
| 1800 | drm_connector_init(dev, &radeon_connector->base, &radeon_dp_connector_funcs, connector_type); | 1861 | drm_connector_init(dev, &radeon_connector->base, &radeon_edp_connector_funcs, connector_type); |
| 1801 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); | 1862 | drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); |
| 1802 | if (i2c_bus->valid) { | 1863 | if (i2c_bus->valid) { |
| 1803 | /* add DP i2c bus */ | 1864 | /* add DP i2c bus */ |
diff --git a/drivers/gpu/drm/radeon/radeon_cs.c b/drivers/gpu/drm/radeon/radeon_cs.c index a56084410372..66c222836631 100644 --- a/drivers/gpu/drm/radeon/radeon_cs.c +++ b/drivers/gpu/drm/radeon/radeon_cs.c | |||
| @@ -28,6 +28,7 @@ | |||
| 28 | #include <drm/radeon_drm.h> | 28 | #include <drm/radeon_drm.h> |
| 29 | #include "radeon_reg.h" | 29 | #include "radeon_reg.h" |
| 30 | #include "radeon.h" | 30 | #include "radeon.h" |
| 31 | #include "radeon_trace.h" | ||
| 31 | 32 | ||
| 32 | static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) | 33 | static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) |
| 33 | { | 34 | { |
| @@ -80,10 +81,13 @@ static int radeon_cs_parser_relocs(struct radeon_cs_parser *p) | |||
| 80 | p->relocs[i].lobj.bo = p->relocs[i].robj; | 81 | p->relocs[i].lobj.bo = p->relocs[i].robj; |
| 81 | p->relocs[i].lobj.written = !!r->write_domain; | 82 | p->relocs[i].lobj.written = !!r->write_domain; |
| 82 | 83 | ||
| 83 | /* the first reloc of an UVD job is the | 84 | /* the first reloc of an UVD job is the msg and that must be in |
| 84 | msg and that must be in VRAM */ | 85 | VRAM, also but everything into VRAM on AGP cards to avoid |
| 85 | if (p->ring == R600_RING_TYPE_UVD_INDEX && i == 0) { | 86 | image corruptions */ |
| 86 | /* TODO: is this still needed for NI+ ? */ | 87 | if (p->ring == R600_RING_TYPE_UVD_INDEX && |
| 88 | p->rdev->family < CHIP_PALM && | ||
| 89 | (i == 0 || drm_pci_device_is_agp(p->rdev->ddev))) { | ||
| 90 | |||
| 87 | p->relocs[i].lobj.domain = | 91 | p->relocs[i].lobj.domain = |
| 88 | RADEON_GEM_DOMAIN_VRAM; | 92 | RADEON_GEM_DOMAIN_VRAM; |
| 89 | 93 | ||
| @@ -559,6 +563,8 @@ int radeon_cs_ioctl(struct drm_device *dev, void *data, struct drm_file *filp) | |||
| 559 | return r; | 563 | return r; |
| 560 | } | 564 | } |
| 561 | 565 | ||
| 566 | trace_radeon_cs(&parser); | ||
| 567 | |||
| 562 | r = radeon_cs_ib_chunk(rdev, &parser); | 568 | r = radeon_cs_ib_chunk(rdev, &parser); |
| 563 | if (r) { | 569 | if (r) { |
| 564 | goto out; | 570 | goto out; |
diff --git a/drivers/gpu/drm/radeon/radeon_device.c b/drivers/gpu/drm/radeon/radeon_device.c index 16cb8792b1e6..841d0e09be3e 100644 --- a/drivers/gpu/drm/radeon/radeon_device.c +++ b/drivers/gpu/drm/radeon/radeon_device.c | |||
| @@ -1249,6 +1249,17 @@ int radeon_device_init(struct radeon_device *rdev, | |||
| 1249 | /* Registers mapping */ | 1249 | /* Registers mapping */ |
| 1250 | /* TODO: block userspace mapping of io register */ | 1250 | /* TODO: block userspace mapping of io register */ |
| 1251 | spin_lock_init(&rdev->mmio_idx_lock); | 1251 | spin_lock_init(&rdev->mmio_idx_lock); |
| 1252 | spin_lock_init(&rdev->smc_idx_lock); | ||
| 1253 | spin_lock_init(&rdev->pll_idx_lock); | ||
| 1254 | spin_lock_init(&rdev->mc_idx_lock); | ||
| 1255 | spin_lock_init(&rdev->pcie_idx_lock); | ||
| 1256 | spin_lock_init(&rdev->pciep_idx_lock); | ||
| 1257 | spin_lock_init(&rdev->pif_idx_lock); | ||
| 1258 | spin_lock_init(&rdev->cg_idx_lock); | ||
| 1259 | spin_lock_init(&rdev->uvd_idx_lock); | ||
| 1260 | spin_lock_init(&rdev->rcu_idx_lock); | ||
| 1261 | spin_lock_init(&rdev->didt_idx_lock); | ||
| 1262 | spin_lock_init(&rdev->end_idx_lock); | ||
| 1252 | if (rdev->family >= CHIP_BONAIRE) { | 1263 | if (rdev->family >= CHIP_BONAIRE) { |
| 1253 | rdev->rmmio_base = pci_resource_start(rdev->pdev, 5); | 1264 | rdev->rmmio_base = pci_resource_start(rdev->pdev, 5); |
| 1254 | rdev->rmmio_size = pci_resource_len(rdev->pdev, 5); | 1265 | rdev->rmmio_size = pci_resource_len(rdev->pdev, 5); |
| @@ -1309,13 +1320,22 @@ int radeon_device_init(struct radeon_device *rdev, | |||
| 1309 | return r; | 1320 | return r; |
| 1310 | } | 1321 | } |
| 1311 | if ((radeon_testing & 1)) { | 1322 | if ((radeon_testing & 1)) { |
| 1312 | radeon_test_moves(rdev); | 1323 | if (rdev->accel_working) |
| 1324 | radeon_test_moves(rdev); | ||
| 1325 | else | ||
| 1326 | DRM_INFO("radeon: acceleration disabled, skipping move tests\n"); | ||
| 1313 | } | 1327 | } |
| 1314 | if ((radeon_testing & 2)) { | 1328 | if ((radeon_testing & 2)) { |
| 1315 | radeon_test_syncing(rdev); | 1329 | if (rdev->accel_working) |
| 1330 | radeon_test_syncing(rdev); | ||
| 1331 | else | ||
| 1332 | DRM_INFO("radeon: acceleration disabled, skipping sync tests\n"); | ||
| 1316 | } | 1333 | } |
| 1317 | if (radeon_benchmarking) { | 1334 | if (radeon_benchmarking) { |
| 1318 | radeon_benchmark(rdev, radeon_benchmarking); | 1335 | if (rdev->accel_working) |
| 1336 | radeon_benchmark(rdev, radeon_benchmarking); | ||
| 1337 | else | ||
| 1338 | DRM_INFO("radeon: acceleration disabled, skipping benchmarks\n"); | ||
| 1319 | } | 1339 | } |
| 1320 | return 0; | 1340 | return 0; |
| 1321 | } | 1341 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c index b055bddaa94c..0d1aa050d41d 100644 --- a/drivers/gpu/drm/radeon/radeon_display.c +++ b/drivers/gpu/drm/radeon/radeon_display.c | |||
| @@ -1172,6 +1172,12 @@ static struct drm_prop_enum_list radeon_underscan_enum_list[] = | |||
| 1172 | { UNDERSCAN_AUTO, "auto" }, | 1172 | { UNDERSCAN_AUTO, "auto" }, |
| 1173 | }; | 1173 | }; |
| 1174 | 1174 | ||
| 1175 | static struct drm_prop_enum_list radeon_audio_enum_list[] = | ||
| 1176 | { { RADEON_AUDIO_DISABLE, "off" }, | ||
| 1177 | { RADEON_AUDIO_ENABLE, "on" }, | ||
| 1178 | { RADEON_AUDIO_AUTO, "auto" }, | ||
| 1179 | }; | ||
| 1180 | |||
| 1175 | static int radeon_modeset_create_props(struct radeon_device *rdev) | 1181 | static int radeon_modeset_create_props(struct radeon_device *rdev) |
| 1176 | { | 1182 | { |
| 1177 | int sz; | 1183 | int sz; |
| @@ -1222,6 +1228,12 @@ static int radeon_modeset_create_props(struct radeon_device *rdev) | |||
| 1222 | if (!rdev->mode_info.underscan_vborder_property) | 1228 | if (!rdev->mode_info.underscan_vborder_property) |
| 1223 | return -ENOMEM; | 1229 | return -ENOMEM; |
| 1224 | 1230 | ||
| 1231 | sz = ARRAY_SIZE(radeon_audio_enum_list); | ||
| 1232 | rdev->mode_info.audio_property = | ||
| 1233 | drm_property_create_enum(rdev->ddev, 0, | ||
| 1234 | "audio", | ||
| 1235 | radeon_audio_enum_list, sz); | ||
| 1236 | |||
| 1225 | return 0; | 1237 | return 0; |
| 1226 | } | 1238 | } |
| 1227 | 1239 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_drv.c b/drivers/gpu/drm/radeon/radeon_drv.c index cb4445f55a96..cdd12dcd988b 100644 --- a/drivers/gpu/drm/radeon/radeon_drv.c +++ b/drivers/gpu/drm/radeon/radeon_drv.c | |||
| @@ -153,7 +153,7 @@ int radeon_benchmarking = 0; | |||
| 153 | int radeon_testing = 0; | 153 | int radeon_testing = 0; |
| 154 | int radeon_connector_table = 0; | 154 | int radeon_connector_table = 0; |
| 155 | int radeon_tv = 1; | 155 | int radeon_tv = 1; |
| 156 | int radeon_audio = 0; | 156 | int radeon_audio = 1; |
| 157 | int radeon_disp_priority = 0; | 157 | int radeon_disp_priority = 0; |
| 158 | int radeon_hw_i2c = 0; | 158 | int radeon_hw_i2c = 0; |
| 159 | int radeon_pcie_gen2 = -1; | 159 | int radeon_pcie_gen2 = -1; |
diff --git a/drivers/gpu/drm/radeon/radeon_mode.h b/drivers/gpu/drm/radeon/radeon_mode.h index d908d8d68f6b..ef63d3f00b2f 100644 --- a/drivers/gpu/drm/radeon/radeon_mode.h +++ b/drivers/gpu/drm/radeon/radeon_mode.h | |||
| @@ -247,6 +247,8 @@ struct radeon_mode_info { | |||
| 247 | struct drm_property *underscan_property; | 247 | struct drm_property *underscan_property; |
| 248 | struct drm_property *underscan_hborder_property; | 248 | struct drm_property *underscan_hborder_property; |
| 249 | struct drm_property *underscan_vborder_property; | 249 | struct drm_property *underscan_vborder_property; |
| 250 | /* audio */ | ||
| 251 | struct drm_property *audio_property; | ||
| 250 | /* hardcoded DFP edid from BIOS */ | 252 | /* hardcoded DFP edid from BIOS */ |
| 251 | struct edid *bios_hardcoded_edid; | 253 | struct edid *bios_hardcoded_edid; |
| 252 | int bios_hardcoded_edid_size; | 254 | int bios_hardcoded_edid_size; |
| @@ -471,6 +473,12 @@ struct radeon_router { | |||
| 471 | u8 cd_mux_state; | 473 | u8 cd_mux_state; |
| 472 | }; | 474 | }; |
| 473 | 475 | ||
| 476 | enum radeon_connector_audio { | ||
| 477 | RADEON_AUDIO_DISABLE = 0, | ||
| 478 | RADEON_AUDIO_ENABLE = 1, | ||
| 479 | RADEON_AUDIO_AUTO = 2 | ||
| 480 | }; | ||
| 481 | |||
| 474 | struct radeon_connector { | 482 | struct radeon_connector { |
| 475 | struct drm_connector base; | 483 | struct drm_connector base; |
| 476 | uint32_t connector_id; | 484 | uint32_t connector_id; |
| @@ -489,6 +497,7 @@ struct radeon_connector { | |||
| 489 | struct radeon_hpd hpd; | 497 | struct radeon_hpd hpd; |
| 490 | struct radeon_router router; | 498 | struct radeon_router router; |
| 491 | struct radeon_i2c_chan *router_bus; | 499 | struct radeon_i2c_chan *router_bus; |
| 500 | enum radeon_connector_audio audio; | ||
| 492 | }; | 501 | }; |
| 493 | 502 | ||
| 494 | struct radeon_framebuffer { | 503 | struct radeon_framebuffer { |
diff --git a/drivers/gpu/drm/radeon/radeon_pm.c b/drivers/gpu/drm/radeon/radeon_pm.c index d7555369a3e5..ac07ad1d4f8c 100644 --- a/drivers/gpu/drm/radeon/radeon_pm.c +++ b/drivers/gpu/drm/radeon/radeon_pm.c | |||
| @@ -67,7 +67,16 @@ int radeon_pm_get_type_index(struct radeon_device *rdev, | |||
| 67 | 67 | ||
| 68 | void radeon_pm_acpi_event_handler(struct radeon_device *rdev) | 68 | void radeon_pm_acpi_event_handler(struct radeon_device *rdev) |
| 69 | { | 69 | { |
| 70 | if (rdev->pm.pm_method == PM_METHOD_PROFILE) { | 70 | if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) { |
| 71 | mutex_lock(&rdev->pm.mutex); | ||
| 72 | if (power_supply_is_system_supplied() > 0) | ||
| 73 | rdev->pm.dpm.ac_power = true; | ||
| 74 | else | ||
| 75 | rdev->pm.dpm.ac_power = false; | ||
| 76 | if (rdev->asic->dpm.enable_bapm) | ||
| 77 | radeon_dpm_enable_bapm(rdev, rdev->pm.dpm.ac_power); | ||
| 78 | mutex_unlock(&rdev->pm.mutex); | ||
| 79 | } else if (rdev->pm.pm_method == PM_METHOD_PROFILE) { | ||
| 71 | if (rdev->pm.profile == PM_PROFILE_AUTO) { | 80 | if (rdev->pm.profile == PM_PROFILE_AUTO) { |
| 72 | mutex_lock(&rdev->pm.mutex); | 81 | mutex_lock(&rdev->pm.mutex); |
| 73 | radeon_pm_update_profile(rdev); | 82 | radeon_pm_update_profile(rdev); |
| @@ -333,7 +342,7 @@ static ssize_t radeon_get_pm_profile(struct device *dev, | |||
| 333 | struct device_attribute *attr, | 342 | struct device_attribute *attr, |
| 334 | char *buf) | 343 | char *buf) |
| 335 | { | 344 | { |
| 336 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 345 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 337 | struct radeon_device *rdev = ddev->dev_private; | 346 | struct radeon_device *rdev = ddev->dev_private; |
| 338 | int cp = rdev->pm.profile; | 347 | int cp = rdev->pm.profile; |
| 339 | 348 | ||
| @@ -349,7 +358,7 @@ static ssize_t radeon_set_pm_profile(struct device *dev, | |||
| 349 | const char *buf, | 358 | const char *buf, |
| 350 | size_t count) | 359 | size_t count) |
| 351 | { | 360 | { |
| 352 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 361 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 353 | struct radeon_device *rdev = ddev->dev_private; | 362 | struct radeon_device *rdev = ddev->dev_private; |
| 354 | 363 | ||
| 355 | mutex_lock(&rdev->pm.mutex); | 364 | mutex_lock(&rdev->pm.mutex); |
| @@ -383,7 +392,7 @@ static ssize_t radeon_get_pm_method(struct device *dev, | |||
| 383 | struct device_attribute *attr, | 392 | struct device_attribute *attr, |
| 384 | char *buf) | 393 | char *buf) |
| 385 | { | 394 | { |
| 386 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 395 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 387 | struct radeon_device *rdev = ddev->dev_private; | 396 | struct radeon_device *rdev = ddev->dev_private; |
| 388 | int pm = rdev->pm.pm_method; | 397 | int pm = rdev->pm.pm_method; |
| 389 | 398 | ||
| @@ -397,7 +406,7 @@ static ssize_t radeon_set_pm_method(struct device *dev, | |||
| 397 | const char *buf, | 406 | const char *buf, |
| 398 | size_t count) | 407 | size_t count) |
| 399 | { | 408 | { |
| 400 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 409 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 401 | struct radeon_device *rdev = ddev->dev_private; | 410 | struct radeon_device *rdev = ddev->dev_private; |
| 402 | 411 | ||
| 403 | /* we don't support the legacy modes with dpm */ | 412 | /* we don't support the legacy modes with dpm */ |
| @@ -433,7 +442,7 @@ static ssize_t radeon_get_dpm_state(struct device *dev, | |||
| 433 | struct device_attribute *attr, | 442 | struct device_attribute *attr, |
| 434 | char *buf) | 443 | char *buf) |
| 435 | { | 444 | { |
| 436 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 445 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 437 | struct radeon_device *rdev = ddev->dev_private; | 446 | struct radeon_device *rdev = ddev->dev_private; |
| 438 | enum radeon_pm_state_type pm = rdev->pm.dpm.user_state; | 447 | enum radeon_pm_state_type pm = rdev->pm.dpm.user_state; |
| 439 | 448 | ||
| @@ -447,7 +456,7 @@ static ssize_t radeon_set_dpm_state(struct device *dev, | |||
| 447 | const char *buf, | 456 | const char *buf, |
| 448 | size_t count) | 457 | size_t count) |
| 449 | { | 458 | { |
| 450 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 459 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 451 | struct radeon_device *rdev = ddev->dev_private; | 460 | struct radeon_device *rdev = ddev->dev_private; |
| 452 | 461 | ||
| 453 | mutex_lock(&rdev->pm.mutex); | 462 | mutex_lock(&rdev->pm.mutex); |
| @@ -472,7 +481,7 @@ static ssize_t radeon_get_dpm_forced_performance_level(struct device *dev, | |||
| 472 | struct device_attribute *attr, | 481 | struct device_attribute *attr, |
| 473 | char *buf) | 482 | char *buf) |
| 474 | { | 483 | { |
| 475 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 484 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 476 | struct radeon_device *rdev = ddev->dev_private; | 485 | struct radeon_device *rdev = ddev->dev_private; |
| 477 | enum radeon_dpm_forced_level level = rdev->pm.dpm.forced_level; | 486 | enum radeon_dpm_forced_level level = rdev->pm.dpm.forced_level; |
| 478 | 487 | ||
| @@ -486,7 +495,7 @@ static ssize_t radeon_set_dpm_forced_performance_level(struct device *dev, | |||
| 486 | const char *buf, | 495 | const char *buf, |
| 487 | size_t count) | 496 | size_t count) |
| 488 | { | 497 | { |
| 489 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 498 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 490 | struct radeon_device *rdev = ddev->dev_private; | 499 | struct radeon_device *rdev = ddev->dev_private; |
| 491 | enum radeon_dpm_forced_level level; | 500 | enum radeon_dpm_forced_level level; |
| 492 | int ret = 0; | 501 | int ret = 0; |
| @@ -524,7 +533,7 @@ static ssize_t radeon_hwmon_show_temp(struct device *dev, | |||
| 524 | struct device_attribute *attr, | 533 | struct device_attribute *attr, |
| 525 | char *buf) | 534 | char *buf) |
| 526 | { | 535 | { |
| 527 | struct drm_device *ddev = pci_get_drvdata(to_pci_dev(dev)); | 536 | struct drm_device *ddev = dev_get_drvdata(dev); |
| 528 | struct radeon_device *rdev = ddev->dev_private; | 537 | struct radeon_device *rdev = ddev->dev_private; |
| 529 | int temp; | 538 | int temp; |
| 530 | 539 | ||
| @@ -536,6 +545,23 @@ static ssize_t radeon_hwmon_show_temp(struct device *dev, | |||
| 536 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); | 545 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); |
| 537 | } | 546 | } |
| 538 | 547 | ||
| 548 | static ssize_t radeon_hwmon_show_temp_thresh(struct device *dev, | ||
| 549 | struct device_attribute *attr, | ||
| 550 | char *buf) | ||
| 551 | { | ||
| 552 | struct drm_device *ddev = dev_get_drvdata(dev); | ||
| 553 | struct radeon_device *rdev = ddev->dev_private; | ||
| 554 | int hyst = to_sensor_dev_attr(attr)->index; | ||
| 555 | int temp; | ||
| 556 | |||
| 557 | if (hyst) | ||
| 558 | temp = rdev->pm.dpm.thermal.min_temp; | ||
| 559 | else | ||
| 560 | temp = rdev->pm.dpm.thermal.max_temp; | ||
| 561 | |||
| 562 | return snprintf(buf, PAGE_SIZE, "%d\n", temp); | ||
| 563 | } | ||
| 564 | |||
| 539 | static ssize_t radeon_hwmon_show_name(struct device *dev, | 565 | static ssize_t radeon_hwmon_show_name(struct device *dev, |
| 540 | struct device_attribute *attr, | 566 | struct device_attribute *attr, |
| 541 | char *buf) | 567 | char *buf) |
| @@ -544,16 +570,37 @@ static ssize_t radeon_hwmon_show_name(struct device *dev, | |||
| 544 | } | 570 | } |
| 545 | 571 | ||
| 546 | static SENSOR_DEVICE_ATTR(temp1_input, S_IRUGO, radeon_hwmon_show_temp, NULL, 0); | 572 | static SENSOR_DEVICE_ATTR(temp1_input, S_IRUGO, radeon_hwmon_show_temp, NULL, 0); |
| 573 | static SENSOR_DEVICE_ATTR(temp1_crit, S_IRUGO, radeon_hwmon_show_temp_thresh, NULL, 0); | ||
| 574 | static SENSOR_DEVICE_ATTR(temp1_crit_hyst, S_IRUGO, radeon_hwmon_show_temp_thresh, NULL, 1); | ||
| 547 | static SENSOR_DEVICE_ATTR(name, S_IRUGO, radeon_hwmon_show_name, NULL, 0); | 575 | static SENSOR_DEVICE_ATTR(name, S_IRUGO, radeon_hwmon_show_name, NULL, 0); |
| 548 | 576 | ||
| 549 | static struct attribute *hwmon_attributes[] = { | 577 | static struct attribute *hwmon_attributes[] = { |
| 550 | &sensor_dev_attr_temp1_input.dev_attr.attr, | 578 | &sensor_dev_attr_temp1_input.dev_attr.attr, |
| 579 | &sensor_dev_attr_temp1_crit.dev_attr.attr, | ||
| 580 | &sensor_dev_attr_temp1_crit_hyst.dev_attr.attr, | ||
| 551 | &sensor_dev_attr_name.dev_attr.attr, | 581 | &sensor_dev_attr_name.dev_attr.attr, |
| 552 | NULL | 582 | NULL |
| 553 | }; | 583 | }; |
| 554 | 584 | ||
| 585 | static umode_t hwmon_attributes_visible(struct kobject *kobj, | ||
| 586 | struct attribute *attr, int index) | ||
| 587 | { | ||
| 588 | struct device *dev = container_of(kobj, struct device, kobj); | ||
| 589 | struct drm_device *ddev = dev_get_drvdata(dev); | ||
| 590 | struct radeon_device *rdev = ddev->dev_private; | ||
| 591 | |||
| 592 | /* Skip limit attributes if DPM is not enabled */ | ||
| 593 | if (rdev->pm.pm_method != PM_METHOD_DPM && | ||
| 594 | (attr == &sensor_dev_attr_temp1_crit.dev_attr.attr || | ||
| 595 | attr == &sensor_dev_attr_temp1_crit_hyst.dev_attr.attr)) | ||
| 596 | return 0; | ||
| 597 | |||
| 598 | return attr->mode; | ||
| 599 | } | ||
| 600 | |||
| 555 | static const struct attribute_group hwmon_attrgroup = { | 601 | static const struct attribute_group hwmon_attrgroup = { |
| 556 | .attrs = hwmon_attributes, | 602 | .attrs = hwmon_attributes, |
| 603 | .is_visible = hwmon_attributes_visible, | ||
| 557 | }; | 604 | }; |
| 558 | 605 | ||
| 559 | static int radeon_hwmon_init(struct radeon_device *rdev) | 606 | static int radeon_hwmon_init(struct radeon_device *rdev) |
| @@ -870,10 +917,13 @@ static void radeon_dpm_change_power_state_locked(struct radeon_device *rdev) | |||
| 870 | 917 | ||
| 871 | radeon_dpm_post_set_power_state(rdev); | 918 | radeon_dpm_post_set_power_state(rdev); |
| 872 | 919 | ||
| 873 | /* force low perf level for thermal */ | 920 | if (rdev->asic->dpm.force_performance_level) { |
| 874 | if (rdev->pm.dpm.thermal_active && | 921 | if (rdev->pm.dpm.thermal_active) |
| 875 | rdev->asic->dpm.force_performance_level) { | 922 | /* force low perf level for thermal */ |
| 876 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_LOW); | 923 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_LOW); |
| 924 | else | ||
| 925 | /* otherwise, enable auto */ | ||
| 926 | radeon_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 877 | } | 927 | } |
| 878 | 928 | ||
| 879 | done: | 929 | done: |
| @@ -952,7 +1002,7 @@ static void radeon_pm_resume_old(struct radeon_device *rdev) | |||
| 952 | { | 1002 | { |
| 953 | /* set up the default clocks if the MC ucode is loaded */ | 1003 | /* set up the default clocks if the MC ucode is loaded */ |
| 954 | if ((rdev->family >= CHIP_BARTS) && | 1004 | if ((rdev->family >= CHIP_BARTS) && |
| 955 | (rdev->family <= CHIP_HAINAN) && | 1005 | (rdev->family <= CHIP_CAYMAN) && |
| 956 | rdev->mc_fw) { | 1006 | rdev->mc_fw) { |
| 957 | if (rdev->pm.default_vddc) | 1007 | if (rdev->pm.default_vddc) |
| 958 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1008 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
| @@ -996,7 +1046,7 @@ static void radeon_pm_resume_dpm(struct radeon_device *rdev) | |||
| 996 | if (ret) { | 1046 | if (ret) { |
| 997 | DRM_ERROR("radeon: dpm resume failed\n"); | 1047 | DRM_ERROR("radeon: dpm resume failed\n"); |
| 998 | if ((rdev->family >= CHIP_BARTS) && | 1048 | if ((rdev->family >= CHIP_BARTS) && |
| 999 | (rdev->family <= CHIP_HAINAN) && | 1049 | (rdev->family <= CHIP_CAYMAN) && |
| 1000 | rdev->mc_fw) { | 1050 | rdev->mc_fw) { |
| 1001 | if (rdev->pm.default_vddc) | 1051 | if (rdev->pm.default_vddc) |
| 1002 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1052 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
| @@ -1047,7 +1097,7 @@ static int radeon_pm_init_old(struct radeon_device *rdev) | |||
| 1047 | radeon_pm_init_profile(rdev); | 1097 | radeon_pm_init_profile(rdev); |
| 1048 | /* set up the default clocks if the MC ucode is loaded */ | 1098 | /* set up the default clocks if the MC ucode is loaded */ |
| 1049 | if ((rdev->family >= CHIP_BARTS) && | 1099 | if ((rdev->family >= CHIP_BARTS) && |
| 1050 | (rdev->family <= CHIP_HAINAN) && | 1100 | (rdev->family <= CHIP_CAYMAN) && |
| 1051 | rdev->mc_fw) { | 1101 | rdev->mc_fw) { |
| 1052 | if (rdev->pm.default_vddc) | 1102 | if (rdev->pm.default_vddc) |
| 1053 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1103 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
| @@ -1102,9 +1152,10 @@ static int radeon_pm_init_dpm(struct radeon_device *rdev) | |||
| 1102 | { | 1152 | { |
| 1103 | int ret; | 1153 | int ret; |
| 1104 | 1154 | ||
| 1105 | /* default to performance state */ | 1155 | /* default to balanced state */ |
| 1106 | rdev->pm.dpm.state = POWER_STATE_TYPE_BALANCED; | 1156 | rdev->pm.dpm.state = POWER_STATE_TYPE_BALANCED; |
| 1107 | rdev->pm.dpm.user_state = POWER_STATE_TYPE_BALANCED; | 1157 | rdev->pm.dpm.user_state = POWER_STATE_TYPE_BALANCED; |
| 1158 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
| 1108 | rdev->pm.default_sclk = rdev->clock.default_sclk; | 1159 | rdev->pm.default_sclk = rdev->clock.default_sclk; |
| 1109 | rdev->pm.default_mclk = rdev->clock.default_mclk; | 1160 | rdev->pm.default_mclk = rdev->clock.default_mclk; |
| 1110 | rdev->pm.current_sclk = rdev->clock.default_sclk; | 1161 | rdev->pm.current_sclk = rdev->clock.default_sclk; |
| @@ -1132,7 +1183,7 @@ static int radeon_pm_init_dpm(struct radeon_device *rdev) | |||
| 1132 | if (ret) { | 1183 | if (ret) { |
| 1133 | rdev->pm.dpm_enabled = false; | 1184 | rdev->pm.dpm_enabled = false; |
| 1134 | if ((rdev->family >= CHIP_BARTS) && | 1185 | if ((rdev->family >= CHIP_BARTS) && |
| 1135 | (rdev->family <= CHIP_HAINAN) && | 1186 | (rdev->family <= CHIP_CAYMAN) && |
| 1136 | rdev->mc_fw) { | 1187 | rdev->mc_fw) { |
| 1137 | if (rdev->pm.default_vddc) | 1188 | if (rdev->pm.default_vddc) |
| 1138 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, | 1189 | radeon_atom_set_voltage(rdev, rdev->pm.default_vddc, |
diff --git a/drivers/gpu/drm/radeon/radeon_ring.c b/drivers/gpu/drm/radeon/radeon_ring.c index 46a25f037b84..18254e1c3e71 100644 --- a/drivers/gpu/drm/radeon/radeon_ring.c +++ b/drivers/gpu/drm/radeon/radeon_ring.c | |||
| @@ -839,9 +839,11 @@ static int radeon_debugfs_ring_info(struct seq_file *m, void *data) | |||
| 839 | * packet that is the root issue | 839 | * packet that is the root issue |
| 840 | */ | 840 | */ |
| 841 | i = (ring->rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; | 841 | i = (ring->rptr + ring->ptr_mask + 1 - 32) & ring->ptr_mask; |
| 842 | for (j = 0; j <= (count + 32); j++) { | 842 | if (ring->ready) { |
| 843 | seq_printf(m, "r[%5d]=0x%08x\n", i, ring->ring[i]); | 843 | for (j = 0; j <= (count + 32); j++) { |
| 844 | i = (i + 1) & ring->ptr_mask; | 844 | seq_printf(m, "r[%5d]=0x%08x\n", i, ring->ring[i]); |
| 845 | i = (i + 1) & ring->ptr_mask; | ||
| 846 | } | ||
| 845 | } | 847 | } |
| 846 | return 0; | 848 | return 0; |
| 847 | } | 849 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_trace.h b/drivers/gpu/drm/radeon/radeon_trace.h index eafd8160a155..f7e367815964 100644 --- a/drivers/gpu/drm/radeon/radeon_trace.h +++ b/drivers/gpu/drm/radeon/radeon_trace.h | |||
| @@ -27,6 +27,26 @@ TRACE_EVENT(radeon_bo_create, | |||
| 27 | TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages) | 27 | TP_printk("bo=%p, pages=%u", __entry->bo, __entry->pages) |
| 28 | ); | 28 | ); |
| 29 | 29 | ||
| 30 | TRACE_EVENT(radeon_cs, | ||
| 31 | TP_PROTO(struct radeon_cs_parser *p), | ||
| 32 | TP_ARGS(p), | ||
| 33 | TP_STRUCT__entry( | ||
| 34 | __field(u32, ring) | ||
| 35 | __field(u32, dw) | ||
| 36 | __field(u32, fences) | ||
| 37 | ), | ||
| 38 | |||
| 39 | TP_fast_assign( | ||
| 40 | __entry->ring = p->ring; | ||
| 41 | __entry->dw = p->chunks[p->chunk_ib_idx].length_dw; | ||
| 42 | __entry->fences = radeon_fence_count_emitted( | ||
| 43 | p->rdev, p->ring); | ||
| 44 | ), | ||
| 45 | TP_printk("ring=%u, dw=%u, fences=%u", | ||
| 46 | __entry->ring, __entry->dw, | ||
| 47 | __entry->fences) | ||
| 48 | ); | ||
| 49 | |||
| 30 | DECLARE_EVENT_CLASS(radeon_fence_request, | 50 | DECLARE_EVENT_CLASS(radeon_fence_request, |
| 31 | 51 | ||
| 32 | TP_PROTO(struct drm_device *dev, u32 seqno), | 52 | TP_PROTO(struct drm_device *dev, u32 seqno), |
| @@ -53,13 +73,6 @@ DEFINE_EVENT(radeon_fence_request, radeon_fence_emit, | |||
| 53 | TP_ARGS(dev, seqno) | 73 | TP_ARGS(dev, seqno) |
| 54 | ); | 74 | ); |
| 55 | 75 | ||
| 56 | DEFINE_EVENT(radeon_fence_request, radeon_fence_retire, | ||
| 57 | |||
| 58 | TP_PROTO(struct drm_device *dev, u32 seqno), | ||
| 59 | |||
| 60 | TP_ARGS(dev, seqno) | ||
| 61 | ); | ||
| 62 | |||
| 63 | DEFINE_EVENT(radeon_fence_request, radeon_fence_wait_begin, | 76 | DEFINE_EVENT(radeon_fence_request, radeon_fence_wait_begin, |
| 64 | 77 | ||
| 65 | TP_PROTO(struct drm_device *dev, u32 seqno), | 78 | TP_PROTO(struct drm_device *dev, u32 seqno), |
diff --git a/drivers/gpu/drm/radeon/radeon_uvd.c b/drivers/gpu/drm/radeon/radeon_uvd.c index 1a01bbff9bfa..a0f11856ddde 100644 --- a/drivers/gpu/drm/radeon/radeon_uvd.c +++ b/drivers/gpu/drm/radeon/radeon_uvd.c | |||
| @@ -476,8 +476,7 @@ static int radeon_uvd_cs_reloc(struct radeon_cs_parser *p, | |||
| 476 | return -EINVAL; | 476 | return -EINVAL; |
| 477 | } | 477 | } |
| 478 | 478 | ||
| 479 | /* TODO: is this still necessary on NI+ ? */ | 479 | if (p->rdev->family < CHIP_PALM && (cmd == 0 || cmd == 0x3) && |
| 480 | if ((cmd == 0 || cmd == 0x3) && | ||
| 481 | (start >> 28) != (p->rdev->uvd.gpu_addr >> 28)) { | 480 | (start >> 28) != (p->rdev->uvd.gpu_addr >> 28)) { |
| 482 | DRM_ERROR("msg/fb buffer %LX-%LX out of 256MB segment!\n", | 481 | DRM_ERROR("msg/fb buffer %LX-%LX out of 256MB segment!\n", |
| 483 | start, end); | 482 | start, end); |
diff --git a/drivers/gpu/drm/radeon/rs400.c b/drivers/gpu/drm/radeon/rs400.c index b8074a8ec75a..9566b5940a5a 100644 --- a/drivers/gpu/drm/radeon/rs400.c +++ b/drivers/gpu/drm/radeon/rs400.c | |||
| @@ -274,19 +274,26 @@ static void rs400_mc_init(struct radeon_device *rdev) | |||
| 274 | 274 | ||
| 275 | uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 275 | uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
| 276 | { | 276 | { |
| 277 | unsigned long flags; | ||
| 277 | uint32_t r; | 278 | uint32_t r; |
| 278 | 279 | ||
| 280 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 279 | WREG32(RS480_NB_MC_INDEX, reg & 0xff); | 281 | WREG32(RS480_NB_MC_INDEX, reg & 0xff); |
| 280 | r = RREG32(RS480_NB_MC_DATA); | 282 | r = RREG32(RS480_NB_MC_DATA); |
| 281 | WREG32(RS480_NB_MC_INDEX, 0xff); | 283 | WREG32(RS480_NB_MC_INDEX, 0xff); |
| 284 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 282 | return r; | 285 | return r; |
| 283 | } | 286 | } |
| 284 | 287 | ||
| 285 | void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 288 | void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 286 | { | 289 | { |
| 290 | unsigned long flags; | ||
| 291 | |||
| 292 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 287 | WREG32(RS480_NB_MC_INDEX, ((reg) & 0xff) | RS480_NB_MC_IND_WR_EN); | 293 | WREG32(RS480_NB_MC_INDEX, ((reg) & 0xff) | RS480_NB_MC_IND_WR_EN); |
| 288 | WREG32(RS480_NB_MC_DATA, (v)); | 294 | WREG32(RS480_NB_MC_DATA, (v)); |
| 289 | WREG32(RS480_NB_MC_INDEX, 0xff); | 295 | WREG32(RS480_NB_MC_INDEX, 0xff); |
| 296 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 290 | } | 297 | } |
| 291 | 298 | ||
| 292 | #if defined(CONFIG_DEBUG_FS) | 299 | #if defined(CONFIG_DEBUG_FS) |
diff --git a/drivers/gpu/drm/radeon/rs600.c b/drivers/gpu/drm/radeon/rs600.c index 670b555d2ca2..6acba8017b9a 100644 --- a/drivers/gpu/drm/radeon/rs600.c +++ b/drivers/gpu/drm/radeon/rs600.c | |||
| @@ -847,16 +847,26 @@ void rs600_bandwidth_update(struct radeon_device *rdev) | |||
| 847 | 847 | ||
| 848 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 848 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
| 849 | { | 849 | { |
| 850 | unsigned long flags; | ||
| 851 | u32 r; | ||
| 852 | |||
| 853 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 850 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | | 854 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | |
| 851 | S_000070_MC_IND_CITF_ARB0(1)); | 855 | S_000070_MC_IND_CITF_ARB0(1)); |
| 852 | return RREG32(R_000074_MC_IND_DATA); | 856 | r = RREG32(R_000074_MC_IND_DATA); |
| 857 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 858 | return r; | ||
| 853 | } | 859 | } |
| 854 | 860 | ||
| 855 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 861 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 856 | { | 862 | { |
| 863 | unsigned long flags; | ||
| 864 | |||
| 865 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 857 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | | 866 | WREG32(R_000070_MC_IND_INDEX, S_000070_MC_IND_ADDR(reg) | |
| 858 | S_000070_MC_IND_CITF_ARB0(1) | S_000070_MC_IND_WR_EN(1)); | 867 | S_000070_MC_IND_CITF_ARB0(1) | S_000070_MC_IND_WR_EN(1)); |
| 859 | WREG32(R_000074_MC_IND_DATA, v); | 868 | WREG32(R_000074_MC_IND_DATA, v); |
| 869 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 860 | } | 870 | } |
| 861 | 871 | ||
| 862 | static void rs600_debugfs(struct radeon_device *rdev) | 872 | static void rs600_debugfs(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/rs690.c b/drivers/gpu/drm/radeon/rs690.c index d8ddfb34545d..1447d794c22a 100644 --- a/drivers/gpu/drm/radeon/rs690.c +++ b/drivers/gpu/drm/radeon/rs690.c | |||
| @@ -631,20 +631,27 @@ void rs690_bandwidth_update(struct radeon_device *rdev) | |||
| 631 | 631 | ||
| 632 | uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 632 | uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
| 633 | { | 633 | { |
| 634 | unsigned long flags; | ||
| 634 | uint32_t r; | 635 | uint32_t r; |
| 635 | 636 | ||
| 637 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 636 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg)); | 638 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg)); |
| 637 | r = RREG32(R_00007C_MC_DATA); | 639 | r = RREG32(R_00007C_MC_DATA); |
| 638 | WREG32(R_000078_MC_INDEX, ~C_000078_MC_IND_ADDR); | 640 | WREG32(R_000078_MC_INDEX, ~C_000078_MC_IND_ADDR); |
| 641 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 639 | return r; | 642 | return r; |
| 640 | } | 643 | } |
| 641 | 644 | ||
| 642 | void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 645 | void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 643 | { | 646 | { |
| 647 | unsigned long flags; | ||
| 648 | |||
| 649 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 644 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg) | | 650 | WREG32(R_000078_MC_INDEX, S_000078_MC_IND_ADDR(reg) | |
| 645 | S_000078_MC_IND_WR_EN(1)); | 651 | S_000078_MC_IND_WR_EN(1)); |
| 646 | WREG32(R_00007C_MC_DATA, v); | 652 | WREG32(R_00007C_MC_DATA, v); |
| 647 | WREG32(R_000078_MC_INDEX, 0x7F); | 653 | WREG32(R_000078_MC_INDEX, 0x7F); |
| 654 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 648 | } | 655 | } |
| 649 | 656 | ||
| 650 | static void rs690_mc_program(struct radeon_device *rdev) | 657 | static void rs690_mc_program(struct radeon_device *rdev) |
diff --git a/drivers/gpu/drm/radeon/rs780_dpm.c b/drivers/gpu/drm/radeon/rs780_dpm.c index d1a1ce73bd45..6af8505cf4d2 100644 --- a/drivers/gpu/drm/radeon/rs780_dpm.c +++ b/drivers/gpu/drm/radeon/rs780_dpm.c | |||
| @@ -62,9 +62,7 @@ static void rs780_get_pm_mode_parameters(struct radeon_device *rdev) | |||
| 62 | radeon_crtc = to_radeon_crtc(crtc); | 62 | radeon_crtc = to_radeon_crtc(crtc); |
| 63 | pi->crtc_id = radeon_crtc->crtc_id; | 63 | pi->crtc_id = radeon_crtc->crtc_id; |
| 64 | if (crtc->mode.htotal && crtc->mode.vtotal) | 64 | if (crtc->mode.htotal && crtc->mode.vtotal) |
| 65 | pi->refresh_rate = | 65 | pi->refresh_rate = drm_mode_vrefresh(&crtc->mode); |
| 66 | (crtc->mode.clock * 1000) / | ||
| 67 | (crtc->mode.htotal * crtc->mode.vtotal); | ||
| 68 | break; | 66 | break; |
| 69 | } | 67 | } |
| 70 | } | 68 | } |
| @@ -376,9 +374,8 @@ static void rs780_disable_vbios_powersaving(struct radeon_device *rdev) | |||
| 376 | WREG32_P(CG_INTGFX_MISC, 0, ~0xFFF00000); | 374 | WREG32_P(CG_INTGFX_MISC, 0, ~0xFFF00000); |
| 377 | } | 375 | } |
| 378 | 376 | ||
| 379 | static void rs780_force_voltage_to_high(struct radeon_device *rdev) | 377 | static void rs780_force_voltage(struct radeon_device *rdev, u16 voltage) |
| 380 | { | 378 | { |
| 381 | struct igp_power_info *pi = rs780_get_pi(rdev); | ||
| 382 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); | 379 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); |
| 383 | 380 | ||
| 384 | if ((current_state->max_voltage == RS780_VDDC_LEVEL_HIGH) && | 381 | if ((current_state->max_voltage == RS780_VDDC_LEVEL_HIGH) && |
| @@ -390,7 +387,7 @@ static void rs780_force_voltage_to_high(struct radeon_device *rdev) | |||
| 390 | udelay(1); | 387 | udelay(1); |
| 391 | 388 | ||
| 392 | WREG32_P(FVTHROT_PWM_CTRL_REG0, | 389 | WREG32_P(FVTHROT_PWM_CTRL_REG0, |
| 393 | STARTING_PWM_HIGHTIME(pi->max_voltage), | 390 | STARTING_PWM_HIGHTIME(voltage), |
| 394 | ~STARTING_PWM_HIGHTIME_MASK); | 391 | ~STARTING_PWM_HIGHTIME_MASK); |
| 395 | 392 | ||
| 396 | WREG32_P(FVTHROT_PWM_CTRL_REG0, | 393 | WREG32_P(FVTHROT_PWM_CTRL_REG0, |
| @@ -404,6 +401,26 @@ static void rs780_force_voltage_to_high(struct radeon_device *rdev) | |||
| 404 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | 401 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); |
| 405 | } | 402 | } |
| 406 | 403 | ||
| 404 | static void rs780_force_fbdiv(struct radeon_device *rdev, u32 fb_div) | ||
| 405 | { | ||
| 406 | struct igp_ps *current_state = rs780_get_ps(rdev->pm.dpm.current_ps); | ||
| 407 | |||
| 408 | if (current_state->sclk_low == current_state->sclk_high) | ||
| 409 | return; | ||
| 410 | |||
| 411 | WREG32_P(GFX_MACRO_BYPASS_CNTL, SPLL_BYPASS_CNTL, ~SPLL_BYPASS_CNTL); | ||
| 412 | |||
| 413 | WREG32_P(FVTHROT_FBDIV_REG2, FORCED_FEEDBACK_DIV(fb_div), | ||
| 414 | ~FORCED_FEEDBACK_DIV_MASK); | ||
| 415 | WREG32_P(FVTHROT_FBDIV_REG1, STARTING_FEEDBACK_DIV(fb_div), | ||
| 416 | ~STARTING_FEEDBACK_DIV_MASK); | ||
| 417 | WREG32_P(FVTHROT_FBDIV_REG1, FORCE_FEEDBACK_DIV, ~FORCE_FEEDBACK_DIV); | ||
| 418 | |||
| 419 | udelay(100); | ||
| 420 | |||
| 421 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | ||
| 422 | } | ||
| 423 | |||
| 407 | static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, | 424 | static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, |
| 408 | struct radeon_ps *new_ps, | 425 | struct radeon_ps *new_ps, |
| 409 | struct radeon_ps *old_ps) | 426 | struct radeon_ps *old_ps) |
| @@ -432,17 +449,13 @@ static int rs780_set_engine_clock_scaling(struct radeon_device *rdev, | |||
| 432 | if (ret) | 449 | if (ret) |
| 433 | return ret; | 450 | return ret; |
| 434 | 451 | ||
| 435 | WREG32_P(GFX_MACRO_BYPASS_CNTL, SPLL_BYPASS_CNTL, ~SPLL_BYPASS_CNTL); | 452 | if ((min_dividers.ref_div != max_dividers.ref_div) || |
| 436 | 453 | (min_dividers.post_div != max_dividers.post_div) || | |
| 437 | WREG32_P(FVTHROT_FBDIV_REG2, FORCED_FEEDBACK_DIV(max_dividers.fb_div), | 454 | (max_dividers.ref_div != current_max_dividers.ref_div) || |
| 438 | ~FORCED_FEEDBACK_DIV_MASK); | 455 | (max_dividers.post_div != current_max_dividers.post_div)) |
| 439 | WREG32_P(FVTHROT_FBDIV_REG1, STARTING_FEEDBACK_DIV(max_dividers.fb_div), | 456 | return -EINVAL; |
| 440 | ~STARTING_FEEDBACK_DIV_MASK); | ||
| 441 | WREG32_P(FVTHROT_FBDIV_REG1, FORCE_FEEDBACK_DIV, ~FORCE_FEEDBACK_DIV); | ||
| 442 | |||
| 443 | udelay(100); | ||
| 444 | 457 | ||
| 445 | WREG32_P(GFX_MACRO_BYPASS_CNTL, 0, ~SPLL_BYPASS_CNTL); | 458 | rs780_force_fbdiv(rdev, max_dividers.fb_div); |
| 446 | 459 | ||
| 447 | if (max_dividers.fb_div > min_dividers.fb_div) { | 460 | if (max_dividers.fb_div > min_dividers.fb_div) { |
| 448 | WREG32_P(FVTHROT_FBDIV_REG0, | 461 | WREG32_P(FVTHROT_FBDIV_REG0, |
| @@ -486,6 +499,9 @@ static void rs780_activate_engine_clk_scaling(struct radeon_device *rdev, | |||
| 486 | (new_state->sclk_low == old_state->sclk_low)) | 499 | (new_state->sclk_low == old_state->sclk_low)) |
| 487 | return; | 500 | return; |
| 488 | 501 | ||
| 502 | if (new_state->sclk_high == new_state->sclk_low) | ||
| 503 | return; | ||
| 504 | |||
| 489 | rs780_clk_scaling_enable(rdev, true); | 505 | rs780_clk_scaling_enable(rdev, true); |
| 490 | } | 506 | } |
| 491 | 507 | ||
| @@ -649,7 +665,7 @@ int rs780_dpm_set_power_state(struct radeon_device *rdev) | |||
| 649 | rs780_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); | 665 | rs780_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); |
| 650 | 666 | ||
| 651 | if (pi->voltage_control) { | 667 | if (pi->voltage_control) { |
| 652 | rs780_force_voltage_to_high(rdev); | 668 | rs780_force_voltage(rdev, pi->max_voltage); |
| 653 | mdelay(5); | 669 | mdelay(5); |
| 654 | } | 670 | } |
| 655 | 671 | ||
| @@ -717,14 +733,18 @@ static void rs780_parse_pplib_non_clock_info(struct radeon_device *rdev, | |||
| 717 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { | 733 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { |
| 718 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); | 734 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); |
| 719 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); | 735 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); |
| 720 | } else if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
| 721 | rps->vclk = RS780_DEFAULT_VCLK_FREQ; | ||
| 722 | rps->dclk = RS780_DEFAULT_DCLK_FREQ; | ||
| 723 | } else { | 736 | } else { |
| 724 | rps->vclk = 0; | 737 | rps->vclk = 0; |
| 725 | rps->dclk = 0; | 738 | rps->dclk = 0; |
| 726 | } | 739 | } |
| 727 | 740 | ||
| 741 | if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
| 742 | if ((rps->vclk == 0) || (rps->dclk == 0)) { | ||
| 743 | rps->vclk = RS780_DEFAULT_VCLK_FREQ; | ||
| 744 | rps->dclk = RS780_DEFAULT_DCLK_FREQ; | ||
| 745 | } | ||
| 746 | } | ||
| 747 | |||
| 728 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) | 748 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) |
| 729 | rdev->pm.dpm.boot_ps = rps; | 749 | rdev->pm.dpm.boot_ps = rps; |
| 730 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) | 750 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) |
| @@ -986,3 +1006,55 @@ void rs780_dpm_debugfs_print_current_performance_level(struct radeon_device *rde | |||
| 986 | seq_printf(m, "power level 1 sclk: %u vddc_index: %d\n", | 1006 | seq_printf(m, "power level 1 sclk: %u vddc_index: %d\n", |
| 987 | ps->sclk_high, ps->max_voltage); | 1007 | ps->sclk_high, ps->max_voltage); |
| 988 | } | 1008 | } |
| 1009 | |||
| 1010 | int rs780_dpm_force_performance_level(struct radeon_device *rdev, | ||
| 1011 | enum radeon_dpm_forced_level level) | ||
| 1012 | { | ||
| 1013 | struct igp_power_info *pi = rs780_get_pi(rdev); | ||
| 1014 | struct radeon_ps *rps = rdev->pm.dpm.current_ps; | ||
| 1015 | struct igp_ps *ps = rs780_get_ps(rps); | ||
| 1016 | struct atom_clock_dividers dividers; | ||
| 1017 | int ret; | ||
| 1018 | |||
| 1019 | rs780_clk_scaling_enable(rdev, false); | ||
| 1020 | rs780_voltage_scaling_enable(rdev, false); | ||
| 1021 | |||
| 1022 | if (level == RADEON_DPM_FORCED_LEVEL_HIGH) { | ||
| 1023 | if (pi->voltage_control) | ||
| 1024 | rs780_force_voltage(rdev, pi->max_voltage); | ||
| 1025 | |||
| 1026 | ret = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM, | ||
| 1027 | ps->sclk_high, false, ÷rs); | ||
| 1028 | if (ret) | ||
| 1029 | return ret; | ||
| 1030 | |||
| 1031 | rs780_force_fbdiv(rdev, dividers.fb_div); | ||
| 1032 | } else if (level == RADEON_DPM_FORCED_LEVEL_LOW) { | ||
| 1033 | ret = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM, | ||
| 1034 | ps->sclk_low, false, ÷rs); | ||
| 1035 | if (ret) | ||
| 1036 | return ret; | ||
| 1037 | |||
| 1038 | rs780_force_fbdiv(rdev, dividers.fb_div); | ||
| 1039 | |||
| 1040 | if (pi->voltage_control) | ||
| 1041 | rs780_force_voltage(rdev, pi->min_voltage); | ||
| 1042 | } else { | ||
| 1043 | if (pi->voltage_control) | ||
| 1044 | rs780_force_voltage(rdev, pi->max_voltage); | ||
| 1045 | |||
| 1046 | if (ps->sclk_high != ps->sclk_low) { | ||
| 1047 | WREG32_P(FVTHROT_FBDIV_REG1, 0, ~FORCE_FEEDBACK_DIV); | ||
| 1048 | rs780_clk_scaling_enable(rdev, true); | ||
| 1049 | } | ||
| 1050 | |||
| 1051 | if (pi->voltage_control) { | ||
| 1052 | rs780_voltage_scaling_enable(rdev, true); | ||
| 1053 | rs780_enable_voltage_scaling(rdev, rps); | ||
| 1054 | } | ||
| 1055 | } | ||
| 1056 | |||
| 1057 | rdev->pm.dpm.forced_level = level; | ||
| 1058 | |||
| 1059 | return 0; | ||
| 1060 | } | ||
diff --git a/drivers/gpu/drm/radeon/rv515.c b/drivers/gpu/drm/radeon/rv515.c index 8ea1573ae820..873eb4b193b4 100644 --- a/drivers/gpu/drm/radeon/rv515.c +++ b/drivers/gpu/drm/radeon/rv515.c | |||
| @@ -209,19 +209,27 @@ static void rv515_mc_init(struct radeon_device *rdev) | |||
| 209 | 209 | ||
| 210 | uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg) | 210 | uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg) |
| 211 | { | 211 | { |
| 212 | unsigned long flags; | ||
| 212 | uint32_t r; | 213 | uint32_t r; |
| 213 | 214 | ||
| 215 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 214 | WREG32(MC_IND_INDEX, 0x7f0000 | (reg & 0xffff)); | 216 | WREG32(MC_IND_INDEX, 0x7f0000 | (reg & 0xffff)); |
| 215 | r = RREG32(MC_IND_DATA); | 217 | r = RREG32(MC_IND_DATA); |
| 216 | WREG32(MC_IND_INDEX, 0); | 218 | WREG32(MC_IND_INDEX, 0); |
| 219 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 220 | |||
| 217 | return r; | 221 | return r; |
| 218 | } | 222 | } |
| 219 | 223 | ||
| 220 | void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) | 224 | void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v) |
| 221 | { | 225 | { |
| 226 | unsigned long flags; | ||
| 227 | |||
| 228 | spin_lock_irqsave(&rdev->mc_idx_lock, flags); | ||
| 222 | WREG32(MC_IND_INDEX, 0xff0000 | ((reg) & 0xffff)); | 229 | WREG32(MC_IND_INDEX, 0xff0000 | ((reg) & 0xffff)); |
| 223 | WREG32(MC_IND_DATA, (v)); | 230 | WREG32(MC_IND_DATA, (v)); |
| 224 | WREG32(MC_IND_INDEX, 0); | 231 | WREG32(MC_IND_INDEX, 0); |
| 232 | spin_unlock_irqrestore(&rdev->mc_idx_lock, flags); | ||
| 225 | } | 233 | } |
| 226 | 234 | ||
| 227 | #if defined(CONFIG_DEBUG_FS) | 235 | #if defined(CONFIG_DEBUG_FS) |
diff --git a/drivers/gpu/drm/radeon/rv6xx_dpm.c b/drivers/gpu/drm/radeon/rv6xx_dpm.c index ab1f2016f21e..5811d277a36a 100644 --- a/drivers/gpu/drm/radeon/rv6xx_dpm.c +++ b/drivers/gpu/drm/radeon/rv6xx_dpm.c | |||
| @@ -1758,8 +1758,6 @@ int rv6xx_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1758 | 1758 | ||
| 1759 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1759 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
| 1760 | 1760 | ||
| 1761 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
| 1762 | |||
| 1763 | return 0; | 1761 | return 0; |
| 1764 | } | 1762 | } |
| 1765 | 1763 | ||
diff --git a/drivers/gpu/drm/radeon/rv770_dpm.c b/drivers/gpu/drm/radeon/rv770_dpm.c index 8cbb85dae5aa..913b025ae9b3 100644 --- a/drivers/gpu/drm/radeon/rv770_dpm.c +++ b/drivers/gpu/drm/radeon/rv770_dpm.c | |||
| @@ -2064,12 +2064,6 @@ int rv770_dpm_set_power_state(struct radeon_device *rdev) | |||
| 2064 | rv770_program_dcodt_after_state_switch(rdev, new_ps, old_ps); | 2064 | rv770_program_dcodt_after_state_switch(rdev, new_ps, old_ps); |
| 2065 | rv770_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 2065 | rv770_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
| 2066 | 2066 | ||
| 2067 | ret = rv770_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 2068 | if (ret) { | ||
| 2069 | DRM_ERROR("rv770_dpm_force_performance_level failed\n"); | ||
| 2070 | return ret; | ||
| 2071 | } | ||
| 2072 | |||
| 2073 | return 0; | 2067 | return 0; |
| 2074 | } | 2068 | } |
| 2075 | 2069 | ||
| @@ -2147,14 +2141,18 @@ static void rv7xx_parse_pplib_non_clock_info(struct radeon_device *rdev, | |||
| 2147 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { | 2141 | if (ATOM_PPLIB_NONCLOCKINFO_VER1 < table_rev) { |
| 2148 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); | 2142 | rps->vclk = le32_to_cpu(non_clock_info->ulVCLK); |
| 2149 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); | 2143 | rps->dclk = le32_to_cpu(non_clock_info->ulDCLK); |
| 2150 | } else if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
| 2151 | rps->vclk = RV770_DEFAULT_VCLK_FREQ; | ||
| 2152 | rps->dclk = RV770_DEFAULT_DCLK_FREQ; | ||
| 2153 | } else { | 2144 | } else { |
| 2154 | rps->vclk = 0; | 2145 | rps->vclk = 0; |
| 2155 | rps->dclk = 0; | 2146 | rps->dclk = 0; |
| 2156 | } | 2147 | } |
| 2157 | 2148 | ||
| 2149 | if (r600_is_uvd_state(rps->class, rps->class2)) { | ||
| 2150 | if ((rps->vclk == 0) || (rps->dclk == 0)) { | ||
| 2151 | rps->vclk = RV770_DEFAULT_VCLK_FREQ; | ||
| 2152 | rps->dclk = RV770_DEFAULT_DCLK_FREQ; | ||
| 2153 | } | ||
| 2154 | } | ||
| 2155 | |||
| 2158 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) | 2156 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_BOOT) |
| 2159 | rdev->pm.dpm.boot_ps = rps; | 2157 | rdev->pm.dpm.boot_ps = rps; |
| 2160 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) | 2158 | if (rps->class & ATOM_PPLIB_CLASSIFICATION_UVDSTATE) |
diff --git a/drivers/gpu/drm/radeon/rv770_smc.c b/drivers/gpu/drm/radeon/rv770_smc.c index ab95da570215..b2a224407365 100644 --- a/drivers/gpu/drm/radeon/rv770_smc.c +++ b/drivers/gpu/drm/radeon/rv770_smc.c | |||
| @@ -274,8 +274,8 @@ static const u8 cayman_smc_int_vectors[] = | |||
| 274 | 0x08, 0x72, 0x08, 0x72 | 274 | 0x08, 0x72, 0x08, 0x72 |
| 275 | }; | 275 | }; |
| 276 | 276 | ||
| 277 | int rv770_set_smc_sram_address(struct radeon_device *rdev, | 277 | static int rv770_set_smc_sram_address(struct radeon_device *rdev, |
| 278 | u16 smc_address, u16 limit) | 278 | u16 smc_address, u16 limit) |
| 279 | { | 279 | { |
| 280 | u32 addr; | 280 | u32 addr; |
| 281 | 281 | ||
| @@ -296,9 +296,10 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 296 | u16 smc_start_address, const u8 *src, | 296 | u16 smc_start_address, const u8 *src, |
| 297 | u16 byte_count, u16 limit) | 297 | u16 byte_count, u16 limit) |
| 298 | { | 298 | { |
| 299 | unsigned long flags; | ||
| 299 | u32 data, original_data, extra_shift; | 300 | u32 data, original_data, extra_shift; |
| 300 | u16 addr; | 301 | u16 addr; |
| 301 | int ret; | 302 | int ret = 0; |
| 302 | 303 | ||
| 303 | if (smc_start_address & 3) | 304 | if (smc_start_address & 3) |
| 304 | return -EINVAL; | 305 | return -EINVAL; |
| @@ -307,13 +308,14 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 307 | 308 | ||
| 308 | addr = smc_start_address; | 309 | addr = smc_start_address; |
| 309 | 310 | ||
| 311 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 310 | while (byte_count >= 4) { | 312 | while (byte_count >= 4) { |
| 311 | /* SMC address space is BE */ | 313 | /* SMC address space is BE */ |
| 312 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 314 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
| 313 | 315 | ||
| 314 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 316 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
| 315 | if (ret) | 317 | if (ret) |
| 316 | return ret; | 318 | goto done; |
| 317 | 319 | ||
| 318 | WREG32(SMC_SRAM_DATA, data); | 320 | WREG32(SMC_SRAM_DATA, data); |
| 319 | 321 | ||
| @@ -328,7 +330,7 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 328 | 330 | ||
| 329 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 331 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
| 330 | if (ret) | 332 | if (ret) |
| 331 | return ret; | 333 | goto done; |
| 332 | 334 | ||
| 333 | original_data = RREG32(SMC_SRAM_DATA); | 335 | original_data = RREG32(SMC_SRAM_DATA); |
| 334 | 336 | ||
| @@ -346,12 +348,15 @@ int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 346 | 348 | ||
| 347 | ret = rv770_set_smc_sram_address(rdev, addr, limit); | 349 | ret = rv770_set_smc_sram_address(rdev, addr, limit); |
| 348 | if (ret) | 350 | if (ret) |
| 349 | return ret; | 351 | goto done; |
| 350 | 352 | ||
| 351 | WREG32(SMC_SRAM_DATA, data); | 353 | WREG32(SMC_SRAM_DATA, data); |
| 352 | } | 354 | } |
| 353 | 355 | ||
| 354 | return 0; | 356 | done: |
| 357 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 358 | |||
| 359 | return ret; | ||
| 355 | } | 360 | } |
| 356 | 361 | ||
| 357 | static int rv770_program_interrupt_vectors(struct radeon_device *rdev, | 362 | static int rv770_program_interrupt_vectors(struct radeon_device *rdev, |
| @@ -461,12 +466,15 @@ PPSMC_Result rv770_wait_for_smc_inactive(struct radeon_device *rdev) | |||
| 461 | 466 | ||
| 462 | static void rv770_clear_smc_sram(struct radeon_device *rdev, u16 limit) | 467 | static void rv770_clear_smc_sram(struct radeon_device *rdev, u16 limit) |
| 463 | { | 468 | { |
| 469 | unsigned long flags; | ||
| 464 | u16 i; | 470 | u16 i; |
| 465 | 471 | ||
| 472 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 466 | for (i = 0; i < limit; i += 4) { | 473 | for (i = 0; i < limit; i += 4) { |
| 467 | rv770_set_smc_sram_address(rdev, i, limit); | 474 | rv770_set_smc_sram_address(rdev, i, limit); |
| 468 | WREG32(SMC_SRAM_DATA, 0); | 475 | WREG32(SMC_SRAM_DATA, 0); |
| 469 | } | 476 | } |
| 477 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 470 | } | 478 | } |
| 471 | 479 | ||
| 472 | int rv770_load_smc_ucode(struct radeon_device *rdev, | 480 | int rv770_load_smc_ucode(struct radeon_device *rdev, |
| @@ -595,27 +603,29 @@ int rv770_load_smc_ucode(struct radeon_device *rdev, | |||
| 595 | int rv770_read_smc_sram_dword(struct radeon_device *rdev, | 603 | int rv770_read_smc_sram_dword(struct radeon_device *rdev, |
| 596 | u16 smc_address, u32 *value, u16 limit) | 604 | u16 smc_address, u32 *value, u16 limit) |
| 597 | { | 605 | { |
| 606 | unsigned long flags; | ||
| 598 | int ret; | 607 | int ret; |
| 599 | 608 | ||
| 609 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 600 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); | 610 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); |
| 601 | if (ret) | 611 | if (ret == 0) |
| 602 | return ret; | 612 | *value = RREG32(SMC_SRAM_DATA); |
| 603 | 613 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | |
| 604 | *value = RREG32(SMC_SRAM_DATA); | ||
| 605 | 614 | ||
| 606 | return 0; | 615 | return ret; |
| 607 | } | 616 | } |
| 608 | 617 | ||
| 609 | int rv770_write_smc_sram_dword(struct radeon_device *rdev, | 618 | int rv770_write_smc_sram_dword(struct radeon_device *rdev, |
| 610 | u16 smc_address, u32 value, u16 limit) | 619 | u16 smc_address, u32 value, u16 limit) |
| 611 | { | 620 | { |
| 621 | unsigned long flags; | ||
| 612 | int ret; | 622 | int ret; |
| 613 | 623 | ||
| 624 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 614 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); | 625 | ret = rv770_set_smc_sram_address(rdev, smc_address, limit); |
| 615 | if (ret) | 626 | if (ret == 0) |
| 616 | return ret; | 627 | WREG32(SMC_SRAM_DATA, value); |
| 628 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 617 | 629 | ||
| 618 | WREG32(SMC_SRAM_DATA, value); | 630 | return ret; |
| 619 | |||
| 620 | return 0; | ||
| 621 | } | 631 | } |
diff --git a/drivers/gpu/drm/radeon/rv770_smc.h b/drivers/gpu/drm/radeon/rv770_smc.h index f78d92a4b325..3b2c963c4880 100644 --- a/drivers/gpu/drm/radeon/rv770_smc.h +++ b/drivers/gpu/drm/radeon/rv770_smc.h | |||
| @@ -187,8 +187,6 @@ typedef struct RV770_SMC_STATETABLE RV770_SMC_STATETABLE; | |||
| 187 | #define RV770_SMC_SOFT_REGISTER_uvd_enabled 0x9C | 187 | #define RV770_SMC_SOFT_REGISTER_uvd_enabled 0x9C |
| 188 | #define RV770_SMC_SOFT_REGISTER_is_asic_lombok 0xA0 | 188 | #define RV770_SMC_SOFT_REGISTER_is_asic_lombok 0xA0 |
| 189 | 189 | ||
| 190 | int rv770_set_smc_sram_address(struct radeon_device *rdev, | ||
| 191 | u16 smc_address, u16 limit); | ||
| 192 | int rv770_copy_bytes_to_smc(struct radeon_device *rdev, | 190 | int rv770_copy_bytes_to_smc(struct radeon_device *rdev, |
| 193 | u16 smc_start_address, const u8 *src, | 191 | u16 smc_start_address, const u8 *src, |
| 194 | u16 byte_count, u16 limit); | 192 | u16 byte_count, u16 limit); |
diff --git a/drivers/gpu/drm/radeon/rv770d.h b/drivers/gpu/drm/radeon/rv770d.h index 9fe60e542922..1ae277152cc7 100644 --- a/drivers/gpu/drm/radeon/rv770d.h +++ b/drivers/gpu/drm/radeon/rv770d.h | |||
| @@ -852,7 +852,7 @@ | |||
| 852 | #define AFMT_VBI_PACKET_CONTROL 0x7608 | 852 | #define AFMT_VBI_PACKET_CONTROL 0x7608 |
| 853 | # define AFMT_GENERIC0_UPDATE (1 << 2) | 853 | # define AFMT_GENERIC0_UPDATE (1 << 2) |
| 854 | #define AFMT_INFOFRAME_CONTROL0 0x760c | 854 | #define AFMT_INFOFRAME_CONTROL0 0x760c |
| 855 | # define AFMT_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hmdi regs */ | 855 | # define AFMT_AUDIO_INFO_SOURCE (1 << 6) /* 0 - sound block; 1 - hdmi regs */ |
| 856 | # define AFMT_AUDIO_INFO_UPDATE (1 << 7) | 856 | # define AFMT_AUDIO_INFO_UPDATE (1 << 7) |
| 857 | # define AFMT_MPEG_INFO_UPDATE (1 << 10) | 857 | # define AFMT_MPEG_INFO_UPDATE (1 << 10) |
| 858 | #define AFMT_GENERIC0_7 0x7610 | 858 | #define AFMT_GENERIC0_7 0x7610 |
diff --git a/drivers/gpu/drm/radeon/si.c b/drivers/gpu/drm/radeon/si.c index 3e23b757dcfa..c354c1094967 100644 --- a/drivers/gpu/drm/radeon/si.c +++ b/drivers/gpu/drm/radeon/si.c | |||
| @@ -83,6 +83,8 @@ extern void si_dma_vm_set_page(struct radeon_device *rdev, | |||
| 83 | uint64_t pe, | 83 | uint64_t pe, |
| 84 | uint64_t addr, unsigned count, | 84 | uint64_t addr, unsigned count, |
| 85 | uint32_t incr, uint32_t flags); | 85 | uint32_t incr, uint32_t flags); |
| 86 | static void si_enable_gui_idle_interrupt(struct radeon_device *rdev, | ||
| 87 | bool enable); | ||
| 86 | 88 | ||
| 87 | static const u32 verde_rlc_save_restore_register_list[] = | 89 | static const u32 verde_rlc_save_restore_register_list[] = |
| 88 | { | 90 | { |
| @@ -3386,6 +3388,8 @@ static int si_cp_resume(struct radeon_device *rdev) | |||
| 3386 | u32 rb_bufsz; | 3388 | u32 rb_bufsz; |
| 3387 | int r; | 3389 | int r; |
| 3388 | 3390 | ||
| 3391 | si_enable_gui_idle_interrupt(rdev, false); | ||
| 3392 | |||
| 3389 | WREG32(CP_SEM_WAIT_TIMER, 0x0); | 3393 | WREG32(CP_SEM_WAIT_TIMER, 0x0); |
| 3390 | WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0); | 3394 | WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0); |
| 3391 | 3395 | ||
| @@ -3501,6 +3505,8 @@ static int si_cp_resume(struct radeon_device *rdev) | |||
| 3501 | rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false; | 3505 | rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false; |
| 3502 | } | 3506 | } |
| 3503 | 3507 | ||
| 3508 | si_enable_gui_idle_interrupt(rdev, true); | ||
| 3509 | |||
| 3504 | return 0; | 3510 | return 0; |
| 3505 | } | 3511 | } |
| 3506 | 3512 | ||
| @@ -4888,7 +4894,7 @@ static void si_enable_gfx_cgpg(struct radeon_device *rdev, | |||
| 4888 | { | 4894 | { |
| 4889 | u32 tmp; | 4895 | u32 tmp; |
| 4890 | 4896 | ||
| 4891 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG)) { | 4897 | if (enable && (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG)) { |
| 4892 | tmp = RLC_PUD(0x10) | RLC_PDD(0x10) | RLC_TTPD(0x10) | RLC_MSD(0x10); | 4898 | tmp = RLC_PUD(0x10) | RLC_PDD(0x10) | RLC_TTPD(0x10) | RLC_MSD(0x10); |
| 4893 | WREG32(RLC_TTOP_D, tmp); | 4899 | WREG32(RLC_TTOP_D, tmp); |
| 4894 | 4900 | ||
| @@ -5250,6 +5256,7 @@ void si_update_cg(struct radeon_device *rdev, | |||
| 5250 | u32 block, bool enable) | 5256 | u32 block, bool enable) |
| 5251 | { | 5257 | { |
| 5252 | if (block & RADEON_CG_BLOCK_GFX) { | 5258 | if (block & RADEON_CG_BLOCK_GFX) { |
| 5259 | si_enable_gui_idle_interrupt(rdev, false); | ||
| 5253 | /* order matters! */ | 5260 | /* order matters! */ |
| 5254 | if (enable) { | 5261 | if (enable) { |
| 5255 | si_enable_mgcg(rdev, true); | 5262 | si_enable_mgcg(rdev, true); |
| @@ -5258,6 +5265,7 @@ void si_update_cg(struct radeon_device *rdev, | |||
| 5258 | si_enable_cgcg(rdev, false); | 5265 | si_enable_cgcg(rdev, false); |
| 5259 | si_enable_mgcg(rdev, false); | 5266 | si_enable_mgcg(rdev, false); |
| 5260 | } | 5267 | } |
| 5268 | si_enable_gui_idle_interrupt(rdev, true); | ||
| 5261 | } | 5269 | } |
| 5262 | 5270 | ||
| 5263 | if (block & RADEON_CG_BLOCK_MC) { | 5271 | if (block & RADEON_CG_BLOCK_MC) { |
| @@ -5408,7 +5416,7 @@ static void si_init_pg(struct radeon_device *rdev) | |||
| 5408 | si_init_dma_pg(rdev); | 5416 | si_init_dma_pg(rdev); |
| 5409 | } | 5417 | } |
| 5410 | si_init_ao_cu_mask(rdev); | 5418 | si_init_ao_cu_mask(rdev); |
| 5411 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_CG) { | 5419 | if (rdev->pg_flags & RADEON_PG_SUPPORT_GFX_PG) { |
| 5412 | si_init_gfx_cgpg(rdev); | 5420 | si_init_gfx_cgpg(rdev); |
| 5413 | } | 5421 | } |
| 5414 | si_enable_dma_pg(rdev, true); | 5422 | si_enable_dma_pg(rdev, true); |
| @@ -5560,7 +5568,9 @@ static void si_disable_interrupt_state(struct radeon_device *rdev) | |||
| 5560 | { | 5568 | { |
| 5561 | u32 tmp; | 5569 | u32 tmp; |
| 5562 | 5570 | ||
| 5563 | WREG32(CP_INT_CNTL_RING0, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | 5571 | tmp = RREG32(CP_INT_CNTL_RING0) & |
| 5572 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
| 5573 | WREG32(CP_INT_CNTL_RING0, tmp); | ||
| 5564 | WREG32(CP_INT_CNTL_RING1, 0); | 5574 | WREG32(CP_INT_CNTL_RING1, 0); |
| 5565 | WREG32(CP_INT_CNTL_RING2, 0); | 5575 | WREG32(CP_INT_CNTL_RING2, 0); |
| 5566 | tmp = RREG32(DMA_CNTL + DMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; | 5576 | tmp = RREG32(DMA_CNTL + DMA0_REGISTER_OFFSET) & ~TRAP_ENABLE; |
| @@ -5685,7 +5695,7 @@ static int si_irq_init(struct radeon_device *rdev) | |||
| 5685 | 5695 | ||
| 5686 | int si_irq_set(struct radeon_device *rdev) | 5696 | int si_irq_set(struct radeon_device *rdev) |
| 5687 | { | 5697 | { |
| 5688 | u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE; | 5698 | u32 cp_int_cntl; |
| 5689 | u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0; | 5699 | u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0; |
| 5690 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; | 5700 | u32 crtc1 = 0, crtc2 = 0, crtc3 = 0, crtc4 = 0, crtc5 = 0, crtc6 = 0; |
| 5691 | u32 hpd1 = 0, hpd2 = 0, hpd3 = 0, hpd4 = 0, hpd5 = 0, hpd6 = 0; | 5701 | u32 hpd1 = 0, hpd2 = 0, hpd3 = 0, hpd4 = 0, hpd5 = 0, hpd6 = 0; |
| @@ -5706,6 +5716,9 @@ int si_irq_set(struct radeon_device *rdev) | |||
| 5706 | return 0; | 5716 | return 0; |
| 5707 | } | 5717 | } |
| 5708 | 5718 | ||
| 5719 | cp_int_cntl = RREG32(CP_INT_CNTL_RING0) & | ||
| 5720 | (CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE); | ||
| 5721 | |||
| 5709 | if (!ASIC_IS_NODCE(rdev)) { | 5722 | if (!ASIC_IS_NODCE(rdev)) { |
| 5710 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; | 5723 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; |
| 5711 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; | 5724 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; |
diff --git a/drivers/gpu/drm/radeon/si_dpm.c b/drivers/gpu/drm/radeon/si_dpm.c index 5be9b4e72350..9ace28702c76 100644 --- a/drivers/gpu/drm/radeon/si_dpm.c +++ b/drivers/gpu/drm/radeon/si_dpm.c | |||
| @@ -2910,6 +2910,7 @@ static void si_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 2910 | bool disable_sclk_switching = false; | 2910 | bool disable_sclk_switching = false; |
| 2911 | u32 mclk, sclk; | 2911 | u32 mclk, sclk; |
| 2912 | u16 vddc, vddci; | 2912 | u16 vddc, vddci; |
| 2913 | u32 max_sclk_vddc, max_mclk_vddci, max_mclk_vddc; | ||
| 2913 | int i; | 2914 | int i; |
| 2914 | 2915 | ||
| 2915 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || | 2916 | if ((rdev->pm.dpm.new_active_crtc_count > 1) || |
| @@ -2943,6 +2944,29 @@ static void si_apply_state_adjust_rules(struct radeon_device *rdev, | |||
| 2943 | } | 2944 | } |
| 2944 | } | 2945 | } |
| 2945 | 2946 | ||
| 2947 | /* limit clocks to max supported clocks based on voltage dependency tables */ | ||
| 2948 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk, | ||
| 2949 | &max_sclk_vddc); | ||
| 2950 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk, | ||
| 2951 | &max_mclk_vddci); | ||
| 2952 | btc_get_max_clock_from_voltage_dependency_table(&rdev->pm.dpm.dyn_state.vddc_dependency_on_mclk, | ||
| 2953 | &max_mclk_vddc); | ||
| 2954 | |||
| 2955 | for (i = 0; i < ps->performance_level_count; i++) { | ||
| 2956 | if (max_sclk_vddc) { | ||
| 2957 | if (ps->performance_levels[i].sclk > max_sclk_vddc) | ||
| 2958 | ps->performance_levels[i].sclk = max_sclk_vddc; | ||
| 2959 | } | ||
| 2960 | if (max_mclk_vddci) { | ||
| 2961 | if (ps->performance_levels[i].mclk > max_mclk_vddci) | ||
| 2962 | ps->performance_levels[i].mclk = max_mclk_vddci; | ||
| 2963 | } | ||
| 2964 | if (max_mclk_vddc) { | ||
| 2965 | if (ps->performance_levels[i].mclk > max_mclk_vddc) | ||
| 2966 | ps->performance_levels[i].mclk = max_mclk_vddc; | ||
| 2967 | } | ||
| 2968 | } | ||
| 2969 | |||
| 2946 | /* XXX validate the min clocks required for display */ | 2970 | /* XXX validate the min clocks required for display */ |
| 2947 | 2971 | ||
| 2948 | if (disable_mclk_switching) { | 2972 | if (disable_mclk_switching) { |
| @@ -6075,12 +6099,6 @@ int si_dpm_set_power_state(struct radeon_device *rdev) | |||
| 6075 | return ret; | 6099 | return ret; |
| 6076 | } | 6100 | } |
| 6077 | 6101 | ||
| 6078 | ret = si_dpm_force_performance_level(rdev, RADEON_DPM_FORCED_LEVEL_AUTO); | ||
| 6079 | if (ret) { | ||
| 6080 | DRM_ERROR("si_dpm_force_performance_level failed\n"); | ||
| 6081 | return ret; | ||
| 6082 | } | ||
| 6083 | |||
| 6084 | si_update_cg(rdev, (RADEON_CG_BLOCK_GFX | | 6102 | si_update_cg(rdev, (RADEON_CG_BLOCK_GFX | |
| 6085 | RADEON_CG_BLOCK_MC | | 6103 | RADEON_CG_BLOCK_MC | |
| 6086 | RADEON_CG_BLOCK_SDMA | | 6104 | RADEON_CG_BLOCK_SDMA | |
diff --git a/drivers/gpu/drm/radeon/si_smc.c b/drivers/gpu/drm/radeon/si_smc.c index 5f524c0a541e..d422a1cbf727 100644 --- a/drivers/gpu/drm/radeon/si_smc.c +++ b/drivers/gpu/drm/radeon/si_smc.c | |||
| @@ -29,8 +29,8 @@ | |||
| 29 | #include "ppsmc.h" | 29 | #include "ppsmc.h" |
| 30 | #include "radeon_ucode.h" | 30 | #include "radeon_ucode.h" |
| 31 | 31 | ||
| 32 | int si_set_smc_sram_address(struct radeon_device *rdev, | 32 | static int si_set_smc_sram_address(struct radeon_device *rdev, |
| 33 | u32 smc_address, u32 limit) | 33 | u32 smc_address, u32 limit) |
| 34 | { | 34 | { |
| 35 | if (smc_address & 3) | 35 | if (smc_address & 3) |
| 36 | return -EINVAL; | 36 | return -EINVAL; |
| @@ -47,7 +47,8 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 47 | u32 smc_start_address, | 47 | u32 smc_start_address, |
| 48 | const u8 *src, u32 byte_count, u32 limit) | 48 | const u8 *src, u32 byte_count, u32 limit) |
| 49 | { | 49 | { |
| 50 | int ret; | 50 | unsigned long flags; |
| 51 | int ret = 0; | ||
| 51 | u32 data, original_data, addr, extra_shift; | 52 | u32 data, original_data, addr, extra_shift; |
| 52 | 53 | ||
| 53 | if (smc_start_address & 3) | 54 | if (smc_start_address & 3) |
| @@ -57,13 +58,14 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 57 | 58 | ||
| 58 | addr = smc_start_address; | 59 | addr = smc_start_address; |
| 59 | 60 | ||
| 61 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 60 | while (byte_count >= 4) { | 62 | while (byte_count >= 4) { |
| 61 | /* SMC address space is BE */ | 63 | /* SMC address space is BE */ |
| 62 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; | 64 | data = (src[0] << 24) | (src[1] << 16) | (src[2] << 8) | src[3]; |
| 63 | 65 | ||
| 64 | ret = si_set_smc_sram_address(rdev, addr, limit); | 66 | ret = si_set_smc_sram_address(rdev, addr, limit); |
| 65 | if (ret) | 67 | if (ret) |
| 66 | return ret; | 68 | goto done; |
| 67 | 69 | ||
| 68 | WREG32(SMC_IND_DATA_0, data); | 70 | WREG32(SMC_IND_DATA_0, data); |
| 69 | 71 | ||
| @@ -78,7 +80,7 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 78 | 80 | ||
| 79 | ret = si_set_smc_sram_address(rdev, addr, limit); | 81 | ret = si_set_smc_sram_address(rdev, addr, limit); |
| 80 | if (ret) | 82 | if (ret) |
| 81 | return ret; | 83 | goto done; |
| 82 | 84 | ||
| 83 | original_data = RREG32(SMC_IND_DATA_0); | 85 | original_data = RREG32(SMC_IND_DATA_0); |
| 84 | 86 | ||
| @@ -96,11 +98,15 @@ int si_copy_bytes_to_smc(struct radeon_device *rdev, | |||
| 96 | 98 | ||
| 97 | ret = si_set_smc_sram_address(rdev, addr, limit); | 99 | ret = si_set_smc_sram_address(rdev, addr, limit); |
| 98 | if (ret) | 100 | if (ret) |
| 99 | return ret; | 101 | goto done; |
| 100 | 102 | ||
| 101 | WREG32(SMC_IND_DATA_0, data); | 103 | WREG32(SMC_IND_DATA_0, data); |
| 102 | } | 104 | } |
| 103 | return 0; | 105 | |
| 106 | done: | ||
| 107 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 108 | |||
| 109 | return ret; | ||
| 104 | } | 110 | } |
| 105 | 111 | ||
| 106 | void si_start_smc(struct radeon_device *rdev) | 112 | void si_start_smc(struct radeon_device *rdev) |
| @@ -203,6 +209,7 @@ PPSMC_Result si_wait_for_smc_inactive(struct radeon_device *rdev) | |||
| 203 | 209 | ||
| 204 | int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | 210 | int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) |
| 205 | { | 211 | { |
| 212 | unsigned long flags; | ||
| 206 | u32 ucode_start_address; | 213 | u32 ucode_start_address; |
| 207 | u32 ucode_size; | 214 | u32 ucode_size; |
| 208 | const u8 *src; | 215 | const u8 *src; |
| @@ -241,6 +248,7 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 241 | return -EINVAL; | 248 | return -EINVAL; |
| 242 | 249 | ||
| 243 | src = (const u8 *)rdev->smc_fw->data; | 250 | src = (const u8 *)rdev->smc_fw->data; |
| 251 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 244 | WREG32(SMC_IND_INDEX_0, ucode_start_address); | 252 | WREG32(SMC_IND_INDEX_0, ucode_start_address); |
| 245 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); | 253 | WREG32_P(SMC_IND_ACCESS_CNTL, AUTO_INCREMENT_IND_0, ~AUTO_INCREMENT_IND_0); |
| 246 | while (ucode_size >= 4) { | 254 | while (ucode_size >= 4) { |
| @@ -253,6 +261,7 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 253 | ucode_size -= 4; | 261 | ucode_size -= 4; |
| 254 | } | 262 | } |
| 255 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); | 263 | WREG32_P(SMC_IND_ACCESS_CNTL, 0, ~AUTO_INCREMENT_IND_0); |
| 264 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 256 | 265 | ||
| 257 | return 0; | 266 | return 0; |
| 258 | } | 267 | } |
| @@ -260,25 +269,29 @@ int si_load_smc_ucode(struct radeon_device *rdev, u32 limit) | |||
| 260 | int si_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 269 | int si_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
| 261 | u32 *value, u32 limit) | 270 | u32 *value, u32 limit) |
| 262 | { | 271 | { |
| 272 | unsigned long flags; | ||
| 263 | int ret; | 273 | int ret; |
| 264 | 274 | ||
| 275 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 265 | ret = si_set_smc_sram_address(rdev, smc_address, limit); | 276 | ret = si_set_smc_sram_address(rdev, smc_address, limit); |
| 266 | if (ret) | 277 | if (ret == 0) |
| 267 | return ret; | 278 | *value = RREG32(SMC_IND_DATA_0); |
| 279 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 268 | 280 | ||
| 269 | *value = RREG32(SMC_IND_DATA_0); | 281 | return ret; |
| 270 | return 0; | ||
| 271 | } | 282 | } |
| 272 | 283 | ||
| 273 | int si_write_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, | 284 | int si_write_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, |
| 274 | u32 value, u32 limit) | 285 | u32 value, u32 limit) |
| 275 | { | 286 | { |
| 287 | unsigned long flags; | ||
| 276 | int ret; | 288 | int ret; |
| 277 | 289 | ||
| 290 | spin_lock_irqsave(&rdev->smc_idx_lock, flags); | ||
| 278 | ret = si_set_smc_sram_address(rdev, smc_address, limit); | 291 | ret = si_set_smc_sram_address(rdev, smc_address, limit); |
| 279 | if (ret) | 292 | if (ret == 0) |
| 280 | return ret; | 293 | WREG32(SMC_IND_DATA_0, value); |
| 294 | spin_unlock_irqrestore(&rdev->smc_idx_lock, flags); | ||
| 281 | 295 | ||
| 282 | WREG32(SMC_IND_DATA_0, value); | 296 | return ret; |
| 283 | return 0; | ||
| 284 | } | 297 | } |
diff --git a/drivers/gpu/drm/radeon/sumo_dpm.c b/drivers/gpu/drm/radeon/sumo_dpm.c index 864761c0120e..96ea6db8bf57 100644 --- a/drivers/gpu/drm/radeon/sumo_dpm.c +++ b/drivers/gpu/drm/radeon/sumo_dpm.c | |||
| @@ -1319,8 +1319,6 @@ int sumo_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1319 | if (pi->enable_dpm) | 1319 | if (pi->enable_dpm) |
| 1320 | sumo_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1320 | sumo_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
| 1321 | 1321 | ||
| 1322 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
| 1323 | |||
| 1324 | return 0; | 1322 | return 0; |
| 1325 | } | 1323 | } |
| 1326 | 1324 | ||
diff --git a/drivers/gpu/drm/radeon/trinity_dpm.c b/drivers/gpu/drm/radeon/trinity_dpm.c index b07b7b8f1aff..7f998bf1cc9d 100644 --- a/drivers/gpu/drm/radeon/trinity_dpm.c +++ b/drivers/gpu/drm/radeon/trinity_dpm.c | |||
| @@ -1068,6 +1068,17 @@ static void trinity_update_requested_ps(struct radeon_device *rdev, | |||
| 1068 | pi->requested_rps.ps_priv = &pi->requested_ps; | 1068 | pi->requested_rps.ps_priv = &pi->requested_ps; |
| 1069 | } | 1069 | } |
| 1070 | 1070 | ||
| 1071 | void trinity_dpm_enable_bapm(struct radeon_device *rdev, bool enable) | ||
| 1072 | { | ||
| 1073 | struct trinity_power_info *pi = trinity_get_pi(rdev); | ||
| 1074 | |||
| 1075 | if (pi->enable_bapm) { | ||
| 1076 | trinity_acquire_mutex(rdev); | ||
| 1077 | trinity_dpm_bapm_enable(rdev, enable); | ||
| 1078 | trinity_release_mutex(rdev); | ||
| 1079 | } | ||
| 1080 | } | ||
| 1081 | |||
| 1071 | int trinity_dpm_enable(struct radeon_device *rdev) | 1082 | int trinity_dpm_enable(struct radeon_device *rdev) |
| 1072 | { | 1083 | { |
| 1073 | struct trinity_power_info *pi = trinity_get_pi(rdev); | 1084 | struct trinity_power_info *pi = trinity_get_pi(rdev); |
| @@ -1091,6 +1102,7 @@ int trinity_dpm_enable(struct radeon_device *rdev) | |||
| 1091 | trinity_program_sclk_dpm(rdev); | 1102 | trinity_program_sclk_dpm(rdev); |
| 1092 | trinity_start_dpm(rdev); | 1103 | trinity_start_dpm(rdev); |
| 1093 | trinity_wait_for_dpm_enabled(rdev); | 1104 | trinity_wait_for_dpm_enabled(rdev); |
| 1105 | trinity_dpm_bapm_enable(rdev, false); | ||
| 1094 | trinity_release_mutex(rdev); | 1106 | trinity_release_mutex(rdev); |
| 1095 | 1107 | ||
| 1096 | if (rdev->irq.installed && | 1108 | if (rdev->irq.installed && |
| @@ -1116,6 +1128,7 @@ void trinity_dpm_disable(struct radeon_device *rdev) | |||
| 1116 | trinity_release_mutex(rdev); | 1128 | trinity_release_mutex(rdev); |
| 1117 | return; | 1129 | return; |
| 1118 | } | 1130 | } |
| 1131 | trinity_dpm_bapm_enable(rdev, false); | ||
| 1119 | trinity_disable_clock_power_gating(rdev); | 1132 | trinity_disable_clock_power_gating(rdev); |
| 1120 | sumo_clear_vc(rdev); | 1133 | sumo_clear_vc(rdev); |
| 1121 | trinity_wait_for_level_0(rdev); | 1134 | trinity_wait_for_level_0(rdev); |
| @@ -1212,6 +1225,8 @@ int trinity_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1212 | 1225 | ||
| 1213 | trinity_acquire_mutex(rdev); | 1226 | trinity_acquire_mutex(rdev); |
| 1214 | if (pi->enable_dpm) { | 1227 | if (pi->enable_dpm) { |
| 1228 | if (pi->enable_bapm) | ||
| 1229 | trinity_dpm_bapm_enable(rdev, rdev->pm.dpm.ac_power); | ||
| 1215 | trinity_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); | 1230 | trinity_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); |
| 1216 | trinity_enable_power_level_0(rdev); | 1231 | trinity_enable_power_level_0(rdev); |
| 1217 | trinity_force_level_0(rdev); | 1232 | trinity_force_level_0(rdev); |
| @@ -1221,7 +1236,6 @@ int trinity_dpm_set_power_state(struct radeon_device *rdev) | |||
| 1221 | trinity_force_level_0(rdev); | 1236 | trinity_force_level_0(rdev); |
| 1222 | trinity_unforce_levels(rdev); | 1237 | trinity_unforce_levels(rdev); |
| 1223 | trinity_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1238 | trinity_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
| 1224 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
| 1225 | } | 1239 | } |
| 1226 | trinity_release_mutex(rdev); | 1240 | trinity_release_mutex(rdev); |
| 1227 | 1241 | ||
| @@ -1854,6 +1868,7 @@ int trinity_dpm_init(struct radeon_device *rdev) | |||
| 1854 | for (i = 0; i < SUMO_MAX_HARDWARE_POWERLEVELS; i++) | 1868 | for (i = 0; i < SUMO_MAX_HARDWARE_POWERLEVELS; i++) |
| 1855 | pi->at[i] = TRINITY_AT_DFLT; | 1869 | pi->at[i] = TRINITY_AT_DFLT; |
| 1856 | 1870 | ||
| 1871 | pi->enable_bapm = true; | ||
| 1857 | pi->enable_nbps_policy = true; | 1872 | pi->enable_nbps_policy = true; |
| 1858 | pi->enable_sclk_ds = true; | 1873 | pi->enable_sclk_ds = true; |
| 1859 | pi->enable_gfx_power_gating = true; | 1874 | pi->enable_gfx_power_gating = true; |
diff --git a/drivers/gpu/drm/radeon/trinity_dpm.h b/drivers/gpu/drm/radeon/trinity_dpm.h index e82df071f8b3..c261657750ca 100644 --- a/drivers/gpu/drm/radeon/trinity_dpm.h +++ b/drivers/gpu/drm/radeon/trinity_dpm.h | |||
| @@ -108,6 +108,7 @@ struct trinity_power_info { | |||
| 108 | bool enable_auto_thermal_throttling; | 108 | bool enable_auto_thermal_throttling; |
| 109 | bool enable_dpm; | 109 | bool enable_dpm; |
| 110 | bool enable_sclk_ds; | 110 | bool enable_sclk_ds; |
| 111 | bool enable_bapm; | ||
| 111 | bool uvd_dpm; | 112 | bool uvd_dpm; |
| 112 | struct radeon_ps current_rps; | 113 | struct radeon_ps current_rps; |
| 113 | struct trinity_ps current_ps; | 114 | struct trinity_ps current_ps; |
| @@ -118,6 +119,7 @@ struct trinity_power_info { | |||
| 118 | #define TRINITY_AT_DFLT 30 | 119 | #define TRINITY_AT_DFLT 30 |
| 119 | 120 | ||
| 120 | /* trinity_smc.c */ | 121 | /* trinity_smc.c */ |
| 122 | int trinity_dpm_bapm_enable(struct radeon_device *rdev, bool enable); | ||
| 121 | int trinity_dpm_config(struct radeon_device *rdev, bool enable); | 123 | int trinity_dpm_config(struct radeon_device *rdev, bool enable); |
| 122 | int trinity_uvd_dpm_config(struct radeon_device *rdev); | 124 | int trinity_uvd_dpm_config(struct radeon_device *rdev); |
| 123 | int trinity_dpm_force_state(struct radeon_device *rdev, u32 n); | 125 | int trinity_dpm_force_state(struct radeon_device *rdev, u32 n); |
diff --git a/drivers/gpu/drm/radeon/trinity_smc.c b/drivers/gpu/drm/radeon/trinity_smc.c index a42d89f1830c..9672bcbc7312 100644 --- a/drivers/gpu/drm/radeon/trinity_smc.c +++ b/drivers/gpu/drm/radeon/trinity_smc.c | |||
| @@ -56,6 +56,14 @@ static int trinity_notify_message_to_smu(struct radeon_device *rdev, u32 id) | |||
| 56 | return 0; | 56 | return 0; |
| 57 | } | 57 | } |
| 58 | 58 | ||
| 59 | int trinity_dpm_bapm_enable(struct radeon_device *rdev, bool enable) | ||
| 60 | { | ||
| 61 | if (enable) | ||
| 62 | return trinity_notify_message_to_smu(rdev, PPSMC_MSG_EnableBAPM); | ||
| 63 | else | ||
| 64 | return trinity_notify_message_to_smu(rdev, PPSMC_MSG_DisableBAPM); | ||
| 65 | } | ||
| 66 | |||
| 59 | int trinity_dpm_config(struct radeon_device *rdev, bool enable) | 67 | int trinity_dpm_config(struct radeon_device *rdev, bool enable) |
| 60 | { | 68 | { |
| 61 | if (enable) | 69 | if (enable) |
diff --git a/drivers/gpu/drm/radeon/uvd_v1_0.c b/drivers/gpu/drm/radeon/uvd_v1_0.c index 7266805d9786..3100fa9cb52f 100644 --- a/drivers/gpu/drm/radeon/uvd_v1_0.c +++ b/drivers/gpu/drm/radeon/uvd_v1_0.c | |||
| @@ -212,8 +212,8 @@ int uvd_v1_0_start(struct radeon_device *rdev) | |||
| 212 | /* enable VCPU clock */ | 212 | /* enable VCPU clock */ |
| 213 | WREG32(UVD_VCPU_CNTL, 1 << 9); | 213 | WREG32(UVD_VCPU_CNTL, 1 << 9); |
| 214 | 214 | ||
| 215 | /* enable UMC */ | 215 | /* enable UMC and NC0 */ |
| 216 | WREG32_P(UVD_LMI_CTRL2, 0, ~(1 << 8)); | 216 | WREG32_P(UVD_LMI_CTRL2, 1 << 13, ~((1 << 8) | (1 << 13))); |
| 217 | 217 | ||
| 218 | /* boot up the VCPU */ | 218 | /* boot up the VCPU */ |
| 219 | WREG32(UVD_SOFT_RESET, 0); | 219 | WREG32(UVD_SOFT_RESET, 0); |
diff --git a/drivers/gpu/drm/ttm/ttm_object.c b/drivers/gpu/drm/ttm/ttm_object.c index 58a5f3261c0b..a868176c258a 100644 --- a/drivers/gpu/drm/ttm/ttm_object.c +++ b/drivers/gpu/drm/ttm/ttm_object.c | |||
| @@ -218,7 +218,7 @@ struct ttm_base_object *ttm_base_object_lookup(struct ttm_object_file *tfile, | |||
| 218 | uint32_t key) | 218 | uint32_t key) |
| 219 | { | 219 | { |
| 220 | struct ttm_object_device *tdev = tfile->tdev; | 220 | struct ttm_object_device *tdev = tfile->tdev; |
| 221 | struct ttm_base_object *base; | 221 | struct ttm_base_object *uninitialized_var(base); |
| 222 | struct drm_hash_item *hash; | 222 | struct drm_hash_item *hash; |
| 223 | int ret; | 223 | int ret; |
| 224 | 224 | ||
diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c index 5e93a52d4f2c..210d50365162 100644 --- a/drivers/gpu/drm/ttm/ttm_tt.c +++ b/drivers/gpu/drm/ttm/ttm_tt.c | |||
| @@ -170,7 +170,7 @@ void ttm_tt_destroy(struct ttm_tt *ttm) | |||
| 170 | ttm_tt_unbind(ttm); | 170 | ttm_tt_unbind(ttm); |
| 171 | } | 171 | } |
| 172 | 172 | ||
| 173 | if (likely(ttm->pages != NULL)) { | 173 | if (ttm->state == tt_unbound) { |
| 174 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); | 174 | ttm->bdev->driver->ttm_tt_unpopulate(ttm); |
| 175 | } | 175 | } |
| 176 | 176 | ||
diff --git a/drivers/gpu/drm/udl/udl_gem.c b/drivers/gpu/drm/udl/udl_gem.c index 8dbe9d0ae9a7..8bf646183bac 100644 --- a/drivers/gpu/drm/udl/udl_gem.c +++ b/drivers/gpu/drm/udl/udl_gem.c | |||
| @@ -97,7 +97,6 @@ int udl_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
| 97 | ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page); | 97 | ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page); |
| 98 | switch (ret) { | 98 | switch (ret) { |
| 99 | case -EAGAIN: | 99 | case -EAGAIN: |
| 100 | set_need_resched(); | ||
| 101 | case 0: | 100 | case 0: |
| 102 | case -ERESTARTSYS: | 101 | case -ERESTARTSYS: |
| 103 | return VM_FAULT_NOPAGE; | 102 | return VM_FAULT_NOPAGE; |
