diff options
| author | Ingo Molnar <mingo@elte.hu> | 2010-07-21 15:43:03 -0400 |
|---|---|---|
| committer | Ingo Molnar <mingo@elte.hu> | 2010-07-21 15:43:06 -0400 |
| commit | 9dcdbf7a33d9018ac5d45debcf261be648bdd56a (patch) | |
| tree | bbcc1a018f11ff76cd7ce174ef3ffe2c02da07ee /drivers/gpu | |
| parent | cc5edb0eb9ce892b530e34a5d110382483587942 (diff) | |
| parent | cd5b8f8755a89a57fc8c408d284b8b613f090345 (diff) | |
Merge branch 'linus' into perf/core
Merge reason: Pick up the latest perf fixes.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'drivers/gpu')
| -rw-r--r-- | drivers/gpu/drm/drm_fb_helper.c | 23 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 13 | ||||
| -rw-r--r-- | drivers/gpu/drm/i915/i915_reg.h | 64 | ||||
| -rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_bios.c | 33 | ||||
| -rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_fbcon.c | 3 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/r100.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/r300.c | 3 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/r520.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/r600.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/r600_blit.c | 5 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/radeon.h | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/radeon_atombios.c | 20 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/radeon_connectors.c | 27 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/radeon_device.c | 8 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/radeon_legacy_tv.c | 4 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/rs400.c | 5 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/rs600.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/rs690.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/radeon/rv515.c | 1 | ||||
| -rw-r--r-- | drivers/gpu/drm/ttm/ttm_page_alloc.c | 72 | ||||
| -rw-r--r-- | drivers/gpu/drm/vmwgfx/vmwgfx_kms.c | 1 |
21 files changed, 198 insertions, 90 deletions
diff --git a/drivers/gpu/drm/drm_fb_helper.c b/drivers/gpu/drm/drm_fb_helper.c index 1f2cc6b09623..719662034bbf 100644 --- a/drivers/gpu/drm/drm_fb_helper.c +++ b/drivers/gpu/drm/drm_fb_helper.c | |||
| @@ -315,8 +315,9 @@ static void drm_fb_helper_on(struct fb_info *info) | |||
| 315 | struct drm_device *dev = fb_helper->dev; | 315 | struct drm_device *dev = fb_helper->dev; |
| 316 | struct drm_crtc *crtc; | 316 | struct drm_crtc *crtc; |
| 317 | struct drm_crtc_helper_funcs *crtc_funcs; | 317 | struct drm_crtc_helper_funcs *crtc_funcs; |
| 318 | struct drm_connector *connector; | ||
| 318 | struct drm_encoder *encoder; | 319 | struct drm_encoder *encoder; |
| 319 | int i; | 320 | int i, j; |
| 320 | 321 | ||
| 321 | /* | 322 | /* |
| 322 | * For each CRTC in this fb, turn the crtc on then, | 323 | * For each CRTC in this fb, turn the crtc on then, |
| @@ -332,7 +333,14 @@ static void drm_fb_helper_on(struct fb_info *info) | |||
| 332 | 333 | ||
| 333 | crtc_funcs->dpms(crtc, DRM_MODE_DPMS_ON); | 334 | crtc_funcs->dpms(crtc, DRM_MODE_DPMS_ON); |
| 334 | 335 | ||
| 335 | 336 | /* Walk the connectors & encoders on this fb turning them on */ | |
| 337 | for (j = 0; j < fb_helper->connector_count; j++) { | ||
| 338 | connector = fb_helper->connector_info[j]->connector; | ||
| 339 | connector->dpms = DRM_MODE_DPMS_ON; | ||
| 340 | drm_connector_property_set_value(connector, | ||
| 341 | dev->mode_config.dpms_property, | ||
| 342 | DRM_MODE_DPMS_ON); | ||
| 343 | } | ||
| 336 | /* Found a CRTC on this fb, now find encoders */ | 344 | /* Found a CRTC on this fb, now find encoders */ |
| 337 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { | 345 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { |
| 338 | if (encoder->crtc == crtc) { | 346 | if (encoder->crtc == crtc) { |
| @@ -352,8 +360,9 @@ static void drm_fb_helper_off(struct fb_info *info, int dpms_mode) | |||
| 352 | struct drm_device *dev = fb_helper->dev; | 360 | struct drm_device *dev = fb_helper->dev; |
| 353 | struct drm_crtc *crtc; | 361 | struct drm_crtc *crtc; |
| 354 | struct drm_crtc_helper_funcs *crtc_funcs; | 362 | struct drm_crtc_helper_funcs *crtc_funcs; |
| 363 | struct drm_connector *connector; | ||
| 355 | struct drm_encoder *encoder; | 364 | struct drm_encoder *encoder; |
| 356 | int i; | 365 | int i, j; |
| 357 | 366 | ||
| 358 | /* | 367 | /* |
| 359 | * For each CRTC in this fb, find all associated encoders | 368 | * For each CRTC in this fb, find all associated encoders |
| @@ -367,6 +376,14 @@ static void drm_fb_helper_off(struct fb_info *info, int dpms_mode) | |||
| 367 | if (!crtc->enabled) | 376 | if (!crtc->enabled) |
| 368 | continue; | 377 | continue; |
| 369 | 378 | ||
| 379 | /* Walk the connectors on this fb and mark them off */ | ||
| 380 | for (j = 0; j < fb_helper->connector_count; j++) { | ||
| 381 | connector = fb_helper->connector_info[j]->connector; | ||
| 382 | connector->dpms = dpms_mode; | ||
| 383 | drm_connector_property_set_value(connector, | ||
| 384 | dev->mode_config.dpms_property, | ||
| 385 | dpms_mode); | ||
| 386 | } | ||
| 370 | /* Found a CRTC on this fb, now find encoders */ | 387 | /* Found a CRTC on this fb, now find encoders */ |
| 371 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { | 388 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { |
| 372 | if (encoder->crtc == crtc) { | 389 | if (encoder->crtc == crtc) { |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 074385882ccf..51bd301cf10d 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
| @@ -2241,6 +2241,7 @@ i915_gem_object_get_pages(struct drm_gem_object *obj, | |||
| 2241 | page = read_cache_page_gfp(mapping, i, | 2241 | page = read_cache_page_gfp(mapping, i, |
| 2242 | GFP_HIGHUSER | | 2242 | GFP_HIGHUSER | |
| 2243 | __GFP_COLD | | 2243 | __GFP_COLD | |
| 2244 | __GFP_RECLAIMABLE | | ||
| 2244 | gfpmask); | 2245 | gfpmask); |
| 2245 | if (IS_ERR(page)) | 2246 | if (IS_ERR(page)) |
| 2246 | goto err_pages; | 2247 | goto err_pages; |
| @@ -4741,6 +4742,16 @@ i915_gem_load(struct drm_device *dev) | |||
| 4741 | list_add(&dev_priv->mm.shrink_list, &shrink_list); | 4742 | list_add(&dev_priv->mm.shrink_list, &shrink_list); |
| 4742 | spin_unlock(&shrink_list_lock); | 4743 | spin_unlock(&shrink_list_lock); |
| 4743 | 4744 | ||
| 4745 | /* On GEN3 we really need to make sure the ARB C3 LP bit is set */ | ||
| 4746 | if (IS_GEN3(dev)) { | ||
| 4747 | u32 tmp = I915_READ(MI_ARB_STATE); | ||
| 4748 | if (!(tmp & MI_ARB_C3_LP_WRITE_ENABLE)) { | ||
| 4749 | /* arb state is a masked write, so set bit + bit in mask */ | ||
| 4750 | tmp = MI_ARB_C3_LP_WRITE_ENABLE | (MI_ARB_C3_LP_WRITE_ENABLE << MI_ARB_MASK_SHIFT); | ||
| 4751 | I915_WRITE(MI_ARB_STATE, tmp); | ||
| 4752 | } | ||
| 4753 | } | ||
| 4754 | |||
| 4744 | /* Old X drivers will take 0-2 for front, back, depth buffers */ | 4755 | /* Old X drivers will take 0-2 for front, back, depth buffers */ |
| 4745 | if (!drm_core_check_feature(dev, DRIVER_MODESET)) | 4756 | if (!drm_core_check_feature(dev, DRIVER_MODESET)) |
| 4746 | dev_priv->fence_reg_start = 3; | 4757 | dev_priv->fence_reg_start = 3; |
| @@ -4977,7 +4988,7 @@ i915_gpu_is_active(struct drm_device *dev) | |||
| 4977 | } | 4988 | } |
| 4978 | 4989 | ||
| 4979 | static int | 4990 | static int |
| 4980 | i915_gem_shrink(int nr_to_scan, gfp_t gfp_mask) | 4991 | i915_gem_shrink(struct shrinker *shrink, int nr_to_scan, gfp_t gfp_mask) |
| 4981 | { | 4992 | { |
| 4982 | drm_i915_private_t *dev_priv, *next_dev; | 4993 | drm_i915_private_t *dev_priv, *next_dev; |
| 4983 | struct drm_i915_gem_object *obj_priv, *next_obj; | 4994 | struct drm_i915_gem_object *obj_priv, *next_obj; |
diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h index 150400f40534..6d9b0288272a 100644 --- a/drivers/gpu/drm/i915/i915_reg.h +++ b/drivers/gpu/drm/i915/i915_reg.h | |||
| @@ -359,6 +359,70 @@ | |||
| 359 | #define LM_BURST_LENGTH 0x00000700 | 359 | #define LM_BURST_LENGTH 0x00000700 |
| 360 | #define LM_FIFO_WATERMARK 0x0000001F | 360 | #define LM_FIFO_WATERMARK 0x0000001F |
| 361 | #define MI_ARB_STATE 0x020e4 /* 915+ only */ | 361 | #define MI_ARB_STATE 0x020e4 /* 915+ only */ |
| 362 | #define MI_ARB_MASK_SHIFT 16 /* shift for enable bits */ | ||
| 363 | |||
| 364 | /* Make render/texture TLB fetches lower priorty than associated data | ||
| 365 | * fetches. This is not turned on by default | ||
| 366 | */ | ||
| 367 | #define MI_ARB_RENDER_TLB_LOW_PRIORITY (1 << 15) | ||
| 368 | |||
| 369 | /* Isoch request wait on GTT enable (Display A/B/C streams). | ||
| 370 | * Make isoch requests stall on the TLB update. May cause | ||
| 371 | * display underruns (test mode only) | ||
| 372 | */ | ||
| 373 | #define MI_ARB_ISOCH_WAIT_GTT (1 << 14) | ||
| 374 | |||
| 375 | /* Block grant count for isoch requests when block count is | ||
| 376 | * set to a finite value. | ||
| 377 | */ | ||
| 378 | #define MI_ARB_BLOCK_GRANT_MASK (3 << 12) | ||
| 379 | #define MI_ARB_BLOCK_GRANT_8 (0 << 12) /* for 3 display planes */ | ||
| 380 | #define MI_ARB_BLOCK_GRANT_4 (1 << 12) /* for 2 display planes */ | ||
| 381 | #define MI_ARB_BLOCK_GRANT_2 (2 << 12) /* for 1 display plane */ | ||
| 382 | #define MI_ARB_BLOCK_GRANT_0 (3 << 12) /* don't use */ | ||
| 383 | |||
| 384 | /* Enable render writes to complete in C2/C3/C4 power states. | ||
| 385 | * If this isn't enabled, render writes are prevented in low | ||
| 386 | * power states. That seems bad to me. | ||
| 387 | */ | ||
| 388 | #define MI_ARB_C3_LP_WRITE_ENABLE (1 << 11) | ||
| 389 | |||
| 390 | /* This acknowledges an async flip immediately instead | ||
| 391 | * of waiting for 2TLB fetches. | ||
| 392 | */ | ||
| 393 | #define MI_ARB_ASYNC_FLIP_ACK_IMMEDIATE (1 << 10) | ||
| 394 | |||
| 395 | /* Enables non-sequential data reads through arbiter | ||
| 396 | */ | ||
| 397 | #define MI_ARB_DUAL_DATA_PHASE_DISABLE (1 << 9) | ||
| 398 | |||
| 399 | /* Disable FSB snooping of cacheable write cycles from binner/render | ||
| 400 | * command stream | ||
| 401 | */ | ||
| 402 | #define MI_ARB_CACHE_SNOOP_DISABLE (1 << 8) | ||
| 403 | |||
| 404 | /* Arbiter time slice for non-isoch streams */ | ||
| 405 | #define MI_ARB_TIME_SLICE_MASK (7 << 5) | ||
| 406 | #define MI_ARB_TIME_SLICE_1 (0 << 5) | ||
| 407 | #define MI_ARB_TIME_SLICE_2 (1 << 5) | ||
| 408 | #define MI_ARB_TIME_SLICE_4 (2 << 5) | ||
| 409 | #define MI_ARB_TIME_SLICE_6 (3 << 5) | ||
| 410 | #define MI_ARB_TIME_SLICE_8 (4 << 5) | ||
| 411 | #define MI_ARB_TIME_SLICE_10 (5 << 5) | ||
| 412 | #define MI_ARB_TIME_SLICE_14 (6 << 5) | ||
| 413 | #define MI_ARB_TIME_SLICE_16 (7 << 5) | ||
| 414 | |||
| 415 | /* Low priority grace period page size */ | ||
| 416 | #define MI_ARB_LOW_PRIORITY_GRACE_4KB (0 << 4) /* default */ | ||
| 417 | #define MI_ARB_LOW_PRIORITY_GRACE_8KB (1 << 4) | ||
| 418 | |||
| 419 | /* Disable display A/B trickle feed */ | ||
| 420 | #define MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE (1 << 2) | ||
| 421 | |||
| 422 | /* Set display plane priority */ | ||
| 423 | #define MI_ARB_DISPLAY_PRIORITY_A_B (0 << 0) /* display A > display B */ | ||
| 424 | #define MI_ARB_DISPLAY_PRIORITY_B_A (1 << 0) /* display B > display A */ | ||
| 425 | |||
| 362 | #define CACHE_MODE_0 0x02120 /* 915+ only */ | 426 | #define CACHE_MODE_0 0x02120 /* 915+ only */ |
| 363 | #define CM0_MASK_SHIFT 16 | 427 | #define CM0_MASK_SHIFT 16 |
| 364 | #define CM0_IZ_OPT_DISABLE (1<<6) | 428 | #define CM0_IZ_OPT_DISABLE (1<<6) |
diff --git a/drivers/gpu/drm/nouveau/nouveau_bios.c b/drivers/gpu/drm/nouveau/nouveau_bios.c index fc924b649195..e492919faf44 100644 --- a/drivers/gpu/drm/nouveau/nouveau_bios.c +++ b/drivers/gpu/drm/nouveau/nouveau_bios.c | |||
| @@ -203,36 +203,26 @@ struct methods { | |||
| 203 | const bool rw; | 203 | const bool rw; |
| 204 | }; | 204 | }; |
| 205 | 205 | ||
| 206 | static struct methods nv04_methods[] = { | 206 | static struct methods shadow_methods[] = { |
| 207 | { "PROM", load_vbios_prom, false }, | ||
| 208 | { "PRAMIN", load_vbios_pramin, true }, | ||
| 209 | { "PCIROM", load_vbios_pci, true }, | ||
| 210 | }; | ||
| 211 | |||
| 212 | static struct methods nv50_methods[] = { | ||
| 213 | { "ACPI", load_vbios_acpi, true }, | ||
| 214 | { "PRAMIN", load_vbios_pramin, true }, | 207 | { "PRAMIN", load_vbios_pramin, true }, |
| 215 | { "PROM", load_vbios_prom, false }, | 208 | { "PROM", load_vbios_prom, false }, |
| 216 | { "PCIROM", load_vbios_pci, true }, | 209 | { "PCIROM", load_vbios_pci, true }, |
| 210 | { "ACPI", load_vbios_acpi, true }, | ||
| 217 | }; | 211 | }; |
| 218 | 212 | ||
| 219 | #define METHODCNT 3 | ||
| 220 | |||
| 221 | static bool NVShadowVBIOS(struct drm_device *dev, uint8_t *data) | 213 | static bool NVShadowVBIOS(struct drm_device *dev, uint8_t *data) |
| 222 | { | 214 | { |
| 223 | struct drm_nouveau_private *dev_priv = dev->dev_private; | 215 | const int nr_methods = ARRAY_SIZE(shadow_methods); |
| 224 | struct methods *methods; | 216 | struct methods *methods = shadow_methods; |
| 225 | int i; | ||
| 226 | int testscore = 3; | 217 | int testscore = 3; |
| 227 | int scores[METHODCNT]; | 218 | int scores[nr_methods], i; |
| 228 | 219 | ||
| 229 | if (nouveau_vbios) { | 220 | if (nouveau_vbios) { |
| 230 | methods = nv04_methods; | 221 | for (i = 0; i < nr_methods; i++) |
| 231 | for (i = 0; i < METHODCNT; i++) | ||
| 232 | if (!strcasecmp(nouveau_vbios, methods[i].desc)) | 222 | if (!strcasecmp(nouveau_vbios, methods[i].desc)) |
| 233 | break; | 223 | break; |
| 234 | 224 | ||
| 235 | if (i < METHODCNT) { | 225 | if (i < nr_methods) { |
| 236 | NV_INFO(dev, "Attempting to use BIOS image from %s\n", | 226 | NV_INFO(dev, "Attempting to use BIOS image from %s\n", |
| 237 | methods[i].desc); | 227 | methods[i].desc); |
| 238 | 228 | ||
| @@ -244,12 +234,7 @@ static bool NVShadowVBIOS(struct drm_device *dev, uint8_t *data) | |||
| 244 | NV_ERROR(dev, "VBIOS source \'%s\' invalid\n", nouveau_vbios); | 234 | NV_ERROR(dev, "VBIOS source \'%s\' invalid\n", nouveau_vbios); |
| 245 | } | 235 | } |
| 246 | 236 | ||
| 247 | if (dev_priv->card_type < NV_50) | 237 | for (i = 0; i < nr_methods; i++) { |
| 248 | methods = nv04_methods; | ||
| 249 | else | ||
| 250 | methods = nv50_methods; | ||
| 251 | |||
| 252 | for (i = 0; i < METHODCNT; i++) { | ||
| 253 | NV_TRACE(dev, "Attempting to load BIOS image from %s\n", | 238 | NV_TRACE(dev, "Attempting to load BIOS image from %s\n", |
| 254 | methods[i].desc); | 239 | methods[i].desc); |
| 255 | data[0] = data[1] = 0; /* avoid reuse of previous image */ | 240 | data[0] = data[1] = 0; /* avoid reuse of previous image */ |
| @@ -260,7 +245,7 @@ static bool NVShadowVBIOS(struct drm_device *dev, uint8_t *data) | |||
| 260 | } | 245 | } |
| 261 | 246 | ||
| 262 | while (--testscore > 0) { | 247 | while (--testscore > 0) { |
| 263 | for (i = 0; i < METHODCNT; i++) { | 248 | for (i = 0; i < nr_methods; i++) { |
| 264 | if (scores[i] == testscore) { | 249 | if (scores[i] == testscore) { |
| 265 | NV_TRACE(dev, "Using BIOS image from %s\n", | 250 | NV_TRACE(dev, "Using BIOS image from %s\n", |
| 266 | methods[i].desc); | 251 | methods[i].desc); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fbcon.c b/drivers/gpu/drm/nouveau/nouveau_fbcon.c index c9a4a0d2a115..257ea130ae13 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fbcon.c +++ b/drivers/gpu/drm/nouveau/nouveau_fbcon.c | |||
| @@ -387,7 +387,8 @@ int nouveau_fbcon_init(struct drm_device *dev) | |||
| 387 | dev_priv->nfbdev = nfbdev; | 387 | dev_priv->nfbdev = nfbdev; |
| 388 | nfbdev->helper.funcs = &nouveau_fbcon_helper_funcs; | 388 | nfbdev->helper.funcs = &nouveau_fbcon_helper_funcs; |
| 389 | 389 | ||
| 390 | ret = drm_fb_helper_init(dev, &nfbdev->helper, 2, 4); | 390 | ret = drm_fb_helper_init(dev, &nfbdev->helper, |
| 391 | nv_two_heads(dev) ? 2 : 1, 4); | ||
| 391 | if (ret) { | 392 | if (ret) { |
| 392 | kfree(nfbdev); | 393 | kfree(nfbdev); |
| 393 | return ret; | 394 | return ret; |
diff --git a/drivers/gpu/drm/radeon/r100.c b/drivers/gpu/drm/radeon/r100.c index 3970e62eaab8..aab5ba040bd6 100644 --- a/drivers/gpu/drm/radeon/r100.c +++ b/drivers/gpu/drm/radeon/r100.c | |||
| @@ -2354,6 +2354,7 @@ void r100_mc_init(struct radeon_device *rdev) | |||
| 2354 | if (rdev->flags & RADEON_IS_IGP) | 2354 | if (rdev->flags & RADEON_IS_IGP) |
| 2355 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; | 2355 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; |
| 2356 | radeon_vram_location(rdev, &rdev->mc, base); | 2356 | radeon_vram_location(rdev, &rdev->mc, base); |
| 2357 | rdev->mc.gtt_base_align = 0; | ||
| 2357 | if (!(rdev->flags & RADEON_IS_AGP)) | 2358 | if (!(rdev->flags & RADEON_IS_AGP)) |
| 2358 | radeon_gtt_location(rdev, &rdev->mc); | 2359 | radeon_gtt_location(rdev, &rdev->mc); |
| 2359 | radeon_update_bandwidth_info(rdev); | 2360 | radeon_update_bandwidth_info(rdev); |
diff --git a/drivers/gpu/drm/radeon/r300.c b/drivers/gpu/drm/radeon/r300.c index 7e81db5eb804..19a7ef7ee344 100644 --- a/drivers/gpu/drm/radeon/r300.c +++ b/drivers/gpu/drm/radeon/r300.c | |||
| @@ -481,6 +481,7 @@ void r300_mc_init(struct radeon_device *rdev) | |||
| 481 | if (rdev->flags & RADEON_IS_IGP) | 481 | if (rdev->flags & RADEON_IS_IGP) |
| 482 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; | 482 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; |
| 483 | radeon_vram_location(rdev, &rdev->mc, base); | 483 | radeon_vram_location(rdev, &rdev->mc, base); |
| 484 | rdev->mc.gtt_base_align = 0; | ||
| 484 | if (!(rdev->flags & RADEON_IS_AGP)) | 485 | if (!(rdev->flags & RADEON_IS_AGP)) |
| 485 | radeon_gtt_location(rdev, &rdev->mc); | 486 | radeon_gtt_location(rdev, &rdev->mc); |
| 486 | radeon_update_bandwidth_info(rdev); | 487 | radeon_update_bandwidth_info(rdev); |
| @@ -1176,6 +1177,8 @@ int r300_cs_parse(struct radeon_cs_parser *p) | |||
| 1176 | int r; | 1177 | int r; |
| 1177 | 1178 | ||
| 1178 | track = kzalloc(sizeof(*track), GFP_KERNEL); | 1179 | track = kzalloc(sizeof(*track), GFP_KERNEL); |
| 1180 | if (track == NULL) | ||
| 1181 | return -ENOMEM; | ||
| 1179 | r100_cs_track_clear(p->rdev, track); | 1182 | r100_cs_track_clear(p->rdev, track); |
| 1180 | p->track = track; | 1183 | p->track = track; |
| 1181 | do { | 1184 | do { |
diff --git a/drivers/gpu/drm/radeon/r520.c b/drivers/gpu/drm/radeon/r520.c index 34330df28483..694af7cc23ac 100644 --- a/drivers/gpu/drm/radeon/r520.c +++ b/drivers/gpu/drm/radeon/r520.c | |||
| @@ -125,6 +125,7 @@ void r520_mc_init(struct radeon_device *rdev) | |||
| 125 | r520_vram_get_type(rdev); | 125 | r520_vram_get_type(rdev); |
| 126 | r100_vram_init_sizes(rdev); | 126 | r100_vram_init_sizes(rdev); |
| 127 | radeon_vram_location(rdev, &rdev->mc, 0); | 127 | radeon_vram_location(rdev, &rdev->mc, 0); |
| 128 | rdev->mc.gtt_base_align = 0; | ||
| 128 | if (!(rdev->flags & RADEON_IS_AGP)) | 129 | if (!(rdev->flags & RADEON_IS_AGP)) |
| 129 | radeon_gtt_location(rdev, &rdev->mc); | 130 | radeon_gtt_location(rdev, &rdev->mc); |
| 130 | radeon_update_bandwidth_info(rdev); | 131 | radeon_update_bandwidth_info(rdev); |
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index 3d6645ce2151..e100f69faeec 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c | |||
| @@ -1179,6 +1179,7 @@ void r600_vram_gtt_location(struct radeon_device *rdev, struct radeon_mc *mc) | |||
| 1179 | if (rdev->flags & RADEON_IS_IGP) | 1179 | if (rdev->flags & RADEON_IS_IGP) |
| 1180 | base = (RREG32(MC_VM_FB_LOCATION) & 0xFFFF) << 24; | 1180 | base = (RREG32(MC_VM_FB_LOCATION) & 0xFFFF) << 24; |
| 1181 | radeon_vram_location(rdev, &rdev->mc, base); | 1181 | radeon_vram_location(rdev, &rdev->mc, base); |
| 1182 | rdev->mc.gtt_base_align = 0; | ||
| 1182 | radeon_gtt_location(rdev, mc); | 1183 | radeon_gtt_location(rdev, mc); |
| 1183 | } | 1184 | } |
| 1184 | } | 1185 | } |
diff --git a/drivers/gpu/drm/radeon/r600_blit.c b/drivers/gpu/drm/radeon/r600_blit.c index f4fb88ece2bb..ca5c29f70779 100644 --- a/drivers/gpu/drm/radeon/r600_blit.c +++ b/drivers/gpu/drm/radeon/r600_blit.c | |||
| @@ -538,9 +538,12 @@ int | |||
| 538 | r600_prepare_blit_copy(struct drm_device *dev, struct drm_file *file_priv) | 538 | r600_prepare_blit_copy(struct drm_device *dev, struct drm_file *file_priv) |
| 539 | { | 539 | { |
| 540 | drm_radeon_private_t *dev_priv = dev->dev_private; | 540 | drm_radeon_private_t *dev_priv = dev->dev_private; |
| 541 | int ret; | ||
| 541 | DRM_DEBUG("\n"); | 542 | DRM_DEBUG("\n"); |
| 542 | 543 | ||
| 543 | r600_nomm_get_vb(dev); | 544 | ret = r600_nomm_get_vb(dev); |
| 545 | if (ret) | ||
| 546 | return ret; | ||
| 544 | 547 | ||
| 545 | dev_priv->blit_vb->file_priv = file_priv; | 548 | dev_priv->blit_vb->file_priv = file_priv; |
| 546 | 549 | ||
diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h index ab61aaa887bb..2f94dc66c183 100644 --- a/drivers/gpu/drm/radeon/radeon.h +++ b/drivers/gpu/drm/radeon/radeon.h | |||
| @@ -351,6 +351,7 @@ struct radeon_mc { | |||
| 351 | int vram_mtrr; | 351 | int vram_mtrr; |
| 352 | bool vram_is_ddr; | 352 | bool vram_is_ddr; |
| 353 | bool igp_sideport_enabled; | 353 | bool igp_sideport_enabled; |
| 354 | u64 gtt_base_align; | ||
| 354 | }; | 355 | }; |
| 355 | 356 | ||
| 356 | bool radeon_combios_sideport_present(struct radeon_device *rdev); | 357 | bool radeon_combios_sideport_present(struct radeon_device *rdev); |
diff --git a/drivers/gpu/drm/radeon/radeon_atombios.c b/drivers/gpu/drm/radeon/radeon_atombios.c index 99bd8a9c56b3..10673ae59cfa 100644 --- a/drivers/gpu/drm/radeon/radeon_atombios.c +++ b/drivers/gpu/drm/radeon/radeon_atombios.c | |||
| @@ -280,6 +280,15 @@ static bool radeon_atom_apply_quirks(struct drm_device *dev, | |||
| 280 | } | 280 | } |
| 281 | } | 281 | } |
| 282 | 282 | ||
| 283 | /* ASUS HD 3600 board lists the DVI port as HDMI */ | ||
| 284 | if ((dev->pdev->device == 0x9598) && | ||
| 285 | (dev->pdev->subsystem_vendor == 0x1043) && | ||
| 286 | (dev->pdev->subsystem_device == 0x01e4)) { | ||
| 287 | if (*connector_type == DRM_MODE_CONNECTOR_HDMIA) { | ||
| 288 | *connector_type = DRM_MODE_CONNECTOR_DVII; | ||
| 289 | } | ||
| 290 | } | ||
| 291 | |||
| 283 | /* ASUS HD 3450 board lists the DVI port as HDMI */ | 292 | /* ASUS HD 3450 board lists the DVI port as HDMI */ |
| 284 | if ((dev->pdev->device == 0x95C5) && | 293 | if ((dev->pdev->device == 0x95C5) && |
| 285 | (dev->pdev->subsystem_vendor == 0x1043) && | 294 | (dev->pdev->subsystem_vendor == 0x1043) && |
| @@ -1029,8 +1038,15 @@ bool radeon_atombios_sideport_present(struct radeon_device *rdev) | |||
| 1029 | data_offset); | 1038 | data_offset); |
| 1030 | switch (crev) { | 1039 | switch (crev) { |
| 1031 | case 1: | 1040 | case 1: |
| 1032 | if (igp_info->info.ucMemoryType & 0xf0) | 1041 | /* AMD IGPS */ |
| 1033 | return true; | 1042 | if ((rdev->family == CHIP_RS690) || |
| 1043 | (rdev->family == CHIP_RS740)) { | ||
| 1044 | if (igp_info->info.ulBootUpMemoryClock) | ||
| 1045 | return true; | ||
| 1046 | } else { | ||
| 1047 | if (igp_info->info.ucMemoryType & 0xf0) | ||
| 1048 | return true; | ||
| 1049 | } | ||
| 1034 | break; | 1050 | break; |
| 1035 | case 2: | 1051 | case 2: |
| 1036 | if (igp_info->info_2.ucMemoryType & 0x0f) | 1052 | if (igp_info->info_2.ucMemoryType & 0x0f) |
diff --git a/drivers/gpu/drm/radeon/radeon_connectors.c b/drivers/gpu/drm/radeon/radeon_connectors.c index 0c7ccc6961a3..adccbc2c202c 100644 --- a/drivers/gpu/drm/radeon/radeon_connectors.c +++ b/drivers/gpu/drm/radeon/radeon_connectors.c | |||
| @@ -771,30 +771,27 @@ static enum drm_connector_status radeon_dvi_detect(struct drm_connector *connect | |||
| 771 | } else | 771 | } else |
| 772 | ret = connector_status_connected; | 772 | ret = connector_status_connected; |
| 773 | 773 | ||
| 774 | /* multiple connectors on the same encoder with the same ddc line | 774 | /* This gets complicated. We have boards with VGA + HDMI with a |
| 775 | * This tends to be HDMI and DVI on the same encoder with the | 775 | * shared DDC line and we have boards with DVI-D + HDMI with a shared |
| 776 | * same ddc line. If the edid says HDMI, consider the HDMI port | 776 | * DDC line. The latter is more complex because with DVI<->HDMI adapters |
| 777 | * connected and the DVI port disconnected. If the edid doesn't | 777 | * you don't really know what's connected to which port as both are digital. |
| 778 | * say HDMI, vice versa. | ||
| 779 | */ | 778 | */ |
| 780 | if (radeon_connector->shared_ddc && (ret == connector_status_connected)) { | 779 | if (radeon_connector->shared_ddc && (ret == connector_status_connected)) { |
| 781 | struct drm_device *dev = connector->dev; | 780 | struct drm_device *dev = connector->dev; |
| 781 | struct radeon_device *rdev = dev->dev_private; | ||
| 782 | struct drm_connector *list_connector; | 782 | struct drm_connector *list_connector; |
| 783 | struct radeon_connector *list_radeon_connector; | 783 | struct radeon_connector *list_radeon_connector; |
| 784 | list_for_each_entry(list_connector, &dev->mode_config.connector_list, head) { | 784 | list_for_each_entry(list_connector, &dev->mode_config.connector_list, head) { |
| 785 | if (connector == list_connector) | 785 | if (connector == list_connector) |
| 786 | continue; | 786 | continue; |
| 787 | list_radeon_connector = to_radeon_connector(list_connector); | 787 | list_radeon_connector = to_radeon_connector(list_connector); |
| 788 | if (radeon_connector->devices == list_radeon_connector->devices) { | 788 | if (list_radeon_connector->shared_ddc && |
| 789 | if (drm_detect_hdmi_monitor(radeon_connector->edid)) { | 789 | (list_radeon_connector->ddc_bus->rec.i2c_id == |
| 790 | if (connector->connector_type == DRM_MODE_CONNECTOR_DVID) { | 790 | radeon_connector->ddc_bus->rec.i2c_id)) { |
| 791 | kfree(radeon_connector->edid); | 791 | /* cases where both connectors are digital */ |
| 792 | radeon_connector->edid = NULL; | 792 | if (list_connector->connector_type != DRM_MODE_CONNECTOR_VGA) { |
| 793 | ret = connector_status_disconnected; | 793 | /* hpd is our only option in this case */ |
| 794 | } | 794 | if (!radeon_hpd_sense(rdev, radeon_connector->hpd.hpd)) { |
| 795 | } else { | ||
| 796 | if ((connector->connector_type == DRM_MODE_CONNECTOR_HDMIA) || | ||
| 797 | (connector->connector_type == DRM_MODE_CONNECTOR_HDMIB)) { | ||
| 798 | kfree(radeon_connector->edid); | 795 | kfree(radeon_connector->edid); |
| 799 | radeon_connector->edid = NULL; | 796 | radeon_connector->edid = NULL; |
| 800 | ret = connector_status_disconnected; | 797 | ret = connector_status_disconnected; |
diff --git a/drivers/gpu/drm/radeon/radeon_device.c b/drivers/gpu/drm/radeon/radeon_device.c index 5f317317aba2..dd279da90546 100644 --- a/drivers/gpu/drm/radeon/radeon_device.c +++ b/drivers/gpu/drm/radeon/radeon_device.c | |||
| @@ -226,20 +226,20 @@ void radeon_gtt_location(struct radeon_device *rdev, struct radeon_mc *mc) | |||
| 226 | { | 226 | { |
| 227 | u64 size_af, size_bf; | 227 | u64 size_af, size_bf; |
| 228 | 228 | ||
| 229 | size_af = 0xFFFFFFFF - mc->vram_end; | 229 | size_af = ((0xFFFFFFFF - mc->vram_end) + mc->gtt_base_align) & ~mc->gtt_base_align; |
| 230 | size_bf = mc->vram_start; | 230 | size_bf = mc->vram_start & ~mc->gtt_base_align; |
| 231 | if (size_bf > size_af) { | 231 | if (size_bf > size_af) { |
| 232 | if (mc->gtt_size > size_bf) { | 232 | if (mc->gtt_size > size_bf) { |
| 233 | dev_warn(rdev->dev, "limiting GTT\n"); | 233 | dev_warn(rdev->dev, "limiting GTT\n"); |
| 234 | mc->gtt_size = size_bf; | 234 | mc->gtt_size = size_bf; |
| 235 | } | 235 | } |
| 236 | mc->gtt_start = mc->vram_start - mc->gtt_size; | 236 | mc->gtt_start = (mc->vram_start & ~mc->gtt_base_align) - mc->gtt_size; |
| 237 | } else { | 237 | } else { |
| 238 | if (mc->gtt_size > size_af) { | 238 | if (mc->gtt_size > size_af) { |
| 239 | dev_warn(rdev->dev, "limiting GTT\n"); | 239 | dev_warn(rdev->dev, "limiting GTT\n"); |
| 240 | mc->gtt_size = size_af; | 240 | mc->gtt_size = size_af; |
| 241 | } | 241 | } |
| 242 | mc->gtt_start = mc->vram_end + 1; | 242 | mc->gtt_start = (mc->vram_end + 1 + mc->gtt_base_align) & ~mc->gtt_base_align; |
| 243 | } | 243 | } |
| 244 | mc->gtt_end = mc->gtt_start + mc->gtt_size - 1; | 244 | mc->gtt_end = mc->gtt_start + mc->gtt_size - 1; |
| 245 | dev_info(rdev->dev, "GTT: %lluM 0x%08llX - 0x%08llX\n", | 245 | dev_info(rdev->dev, "GTT: %lluM 0x%08llX - 0x%08llX\n", |
diff --git a/drivers/gpu/drm/radeon/radeon_legacy_tv.c b/drivers/gpu/drm/radeon/radeon_legacy_tv.c index f2ed27c8055b..032040397743 100644 --- a/drivers/gpu/drm/radeon/radeon_legacy_tv.c +++ b/drivers/gpu/drm/radeon/radeon_legacy_tv.c | |||
| @@ -642,8 +642,8 @@ void radeon_legacy_tv_mode_set(struct drm_encoder *encoder, | |||
| 642 | } | 642 | } |
| 643 | flicker_removal = (tmp + 500) / 1000; | 643 | flicker_removal = (tmp + 500) / 1000; |
| 644 | 644 | ||
| 645 | if (flicker_removal < 2) | 645 | if (flicker_removal < 3) |
| 646 | flicker_removal = 2; | 646 | flicker_removal = 3; |
| 647 | for (i = 0; i < ARRAY_SIZE(SLOPE_limit); ++i) { | 647 | for (i = 0; i < ARRAY_SIZE(SLOPE_limit); ++i) { |
| 648 | if (flicker_removal == SLOPE_limit[i]) | 648 | if (flicker_removal == SLOPE_limit[i]) |
| 649 | break; | 649 | break; |
diff --git a/drivers/gpu/drm/radeon/rs400.c b/drivers/gpu/drm/radeon/rs400.c index 9e4240b3bf0b..f454c9a5e7f2 100644 --- a/drivers/gpu/drm/radeon/rs400.c +++ b/drivers/gpu/drm/radeon/rs400.c | |||
| @@ -57,7 +57,9 @@ void rs400_gart_adjust_size(struct radeon_device *rdev) | |||
| 57 | } | 57 | } |
| 58 | if (rdev->family == CHIP_RS400 || rdev->family == CHIP_RS480) { | 58 | if (rdev->family == CHIP_RS400 || rdev->family == CHIP_RS480) { |
| 59 | /* FIXME: RS400 & RS480 seems to have issue with GART size | 59 | /* FIXME: RS400 & RS480 seems to have issue with GART size |
| 60 | * if 4G of system memory (needs more testing) */ | 60 | * if 4G of system memory (needs more testing) |
| 61 | */ | ||
| 62 | /* XXX is this still an issue with proper alignment? */ | ||
| 61 | rdev->mc.gtt_size = 32 * 1024 * 1024; | 63 | rdev->mc.gtt_size = 32 * 1024 * 1024; |
| 62 | DRM_ERROR("Forcing to 32M GART size (because of ASIC bug ?)\n"); | 64 | DRM_ERROR("Forcing to 32M GART size (because of ASIC bug ?)\n"); |
| 63 | } | 65 | } |
| @@ -263,6 +265,7 @@ void rs400_mc_init(struct radeon_device *rdev) | |||
| 263 | r100_vram_init_sizes(rdev); | 265 | r100_vram_init_sizes(rdev); |
| 264 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; | 266 | base = (RREG32(RADEON_NB_TOM) & 0xffff) << 16; |
| 265 | radeon_vram_location(rdev, &rdev->mc, base); | 267 | radeon_vram_location(rdev, &rdev->mc, base); |
| 268 | rdev->mc.gtt_base_align = rdev->mc.gtt_size - 1; | ||
| 266 | radeon_gtt_location(rdev, &rdev->mc); | 269 | radeon_gtt_location(rdev, &rdev->mc); |
| 267 | radeon_update_bandwidth_info(rdev); | 270 | radeon_update_bandwidth_info(rdev); |
| 268 | } | 271 | } |
diff --git a/drivers/gpu/drm/radeon/rs600.c b/drivers/gpu/drm/radeon/rs600.c index 7bb4c3e52f3b..6dc15ea8ba33 100644 --- a/drivers/gpu/drm/radeon/rs600.c +++ b/drivers/gpu/drm/radeon/rs600.c | |||
| @@ -698,6 +698,7 @@ void rs600_mc_init(struct radeon_device *rdev) | |||
| 698 | base = G_000004_MC_FB_START(base) << 16; | 698 | base = G_000004_MC_FB_START(base) << 16; |
| 699 | rdev->mc.igp_sideport_enabled = radeon_atombios_sideport_present(rdev); | 699 | rdev->mc.igp_sideport_enabled = radeon_atombios_sideport_present(rdev); |
| 700 | radeon_vram_location(rdev, &rdev->mc, base); | 700 | radeon_vram_location(rdev, &rdev->mc, base); |
| 701 | rdev->mc.gtt_base_align = 0; | ||
| 701 | radeon_gtt_location(rdev, &rdev->mc); | 702 | radeon_gtt_location(rdev, &rdev->mc); |
| 702 | radeon_update_bandwidth_info(rdev); | 703 | radeon_update_bandwidth_info(rdev); |
| 703 | } | 704 | } |
diff --git a/drivers/gpu/drm/radeon/rs690.c b/drivers/gpu/drm/radeon/rs690.c index f4f0a61bcdce..ce4ecbe10816 100644 --- a/drivers/gpu/drm/radeon/rs690.c +++ b/drivers/gpu/drm/radeon/rs690.c | |||
| @@ -162,6 +162,7 @@ void rs690_mc_init(struct radeon_device *rdev) | |||
| 162 | rs690_pm_info(rdev); | 162 | rs690_pm_info(rdev); |
| 163 | rdev->mc.igp_sideport_enabled = radeon_atombios_sideport_present(rdev); | 163 | rdev->mc.igp_sideport_enabled = radeon_atombios_sideport_present(rdev); |
| 164 | radeon_vram_location(rdev, &rdev->mc, base); | 164 | radeon_vram_location(rdev, &rdev->mc, base); |
| 165 | rdev->mc.gtt_base_align = rdev->mc.gtt_size - 1; | ||
| 165 | radeon_gtt_location(rdev, &rdev->mc); | 166 | radeon_gtt_location(rdev, &rdev->mc); |
| 166 | radeon_update_bandwidth_info(rdev); | 167 | radeon_update_bandwidth_info(rdev); |
| 167 | } | 168 | } |
diff --git a/drivers/gpu/drm/radeon/rv515.c b/drivers/gpu/drm/radeon/rv515.c index 7d9a7b0a180a..0c9c169a6852 100644 --- a/drivers/gpu/drm/radeon/rv515.c +++ b/drivers/gpu/drm/radeon/rv515.c | |||
| @@ -195,6 +195,7 @@ void rv515_mc_init(struct radeon_device *rdev) | |||
| 195 | rv515_vram_get_type(rdev); | 195 | rv515_vram_get_type(rdev); |
| 196 | r100_vram_init_sizes(rdev); | 196 | r100_vram_init_sizes(rdev); |
| 197 | radeon_vram_location(rdev, &rdev->mc, 0); | 197 | radeon_vram_location(rdev, &rdev->mc, 0); |
| 198 | rdev->mc.gtt_base_align = 0; | ||
| 198 | if (!(rdev->flags & RADEON_IS_AGP)) | 199 | if (!(rdev->flags & RADEON_IS_AGP)) |
| 199 | radeon_gtt_location(rdev, &rdev->mc); | 200 | radeon_gtt_location(rdev, &rdev->mc); |
| 200 | radeon_update_bandwidth_info(rdev); | 201 | radeon_update_bandwidth_info(rdev); |
diff --git a/drivers/gpu/drm/ttm/ttm_page_alloc.c b/drivers/gpu/drm/ttm/ttm_page_alloc.c index 2f047577b1e3..d233c65f3f7f 100644 --- a/drivers/gpu/drm/ttm/ttm_page_alloc.c +++ b/drivers/gpu/drm/ttm/ttm_page_alloc.c | |||
| @@ -40,7 +40,9 @@ | |||
| 40 | #include <linux/slab.h> | 40 | #include <linux/slab.h> |
| 41 | 41 | ||
| 42 | #include <asm/atomic.h> | 42 | #include <asm/atomic.h> |
| 43 | #ifdef TTM_HAS_AGP | ||
| 43 | #include <asm/agp.h> | 44 | #include <asm/agp.h> |
| 45 | #endif | ||
| 44 | 46 | ||
| 45 | #include "ttm/ttm_bo_driver.h" | 47 | #include "ttm/ttm_bo_driver.h" |
| 46 | #include "ttm/ttm_page_alloc.h" | 48 | #include "ttm/ttm_page_alloc.h" |
| @@ -104,7 +106,6 @@ struct ttm_pool_opts { | |||
| 104 | struct ttm_pool_manager { | 106 | struct ttm_pool_manager { |
| 105 | struct kobject kobj; | 107 | struct kobject kobj; |
| 106 | struct shrinker mm_shrink; | 108 | struct shrinker mm_shrink; |
| 107 | atomic_t page_alloc_inited; | ||
| 108 | struct ttm_pool_opts options; | 109 | struct ttm_pool_opts options; |
| 109 | 110 | ||
| 110 | union { | 111 | union { |
| @@ -142,7 +143,7 @@ static void ttm_pool_kobj_release(struct kobject *kobj) | |||
| 142 | { | 143 | { |
| 143 | struct ttm_pool_manager *m = | 144 | struct ttm_pool_manager *m = |
| 144 | container_of(kobj, struct ttm_pool_manager, kobj); | 145 | container_of(kobj, struct ttm_pool_manager, kobj); |
| 145 | (void)m; | 146 | kfree(m); |
| 146 | } | 147 | } |
| 147 | 148 | ||
| 148 | static ssize_t ttm_pool_store(struct kobject *kobj, | 149 | static ssize_t ttm_pool_store(struct kobject *kobj, |
| @@ -214,9 +215,7 @@ static struct kobj_type ttm_pool_kobj_type = { | |||
| 214 | .default_attrs = ttm_pool_attrs, | 215 | .default_attrs = ttm_pool_attrs, |
| 215 | }; | 216 | }; |
| 216 | 217 | ||
| 217 | static struct ttm_pool_manager _manager = { | 218 | static struct ttm_pool_manager *_manager; |
| 218 | .page_alloc_inited = ATOMIC_INIT(0) | ||
| 219 | }; | ||
| 220 | 219 | ||
| 221 | #ifndef CONFIG_X86 | 220 | #ifndef CONFIG_X86 |
| 222 | static int set_pages_array_wb(struct page **pages, int addrinarray) | 221 | static int set_pages_array_wb(struct page **pages, int addrinarray) |
| @@ -271,7 +270,7 @@ static struct ttm_page_pool *ttm_get_pool(int flags, | |||
| 271 | if (flags & TTM_PAGE_FLAG_DMA32) | 270 | if (flags & TTM_PAGE_FLAG_DMA32) |
| 272 | pool_index |= 0x2; | 271 | pool_index |= 0x2; |
| 273 | 272 | ||
| 274 | return &_manager.pools[pool_index]; | 273 | return &_manager->pools[pool_index]; |
| 275 | } | 274 | } |
| 276 | 275 | ||
| 277 | /* set memory back to wb and free the pages. */ | 276 | /* set memory back to wb and free the pages. */ |
| @@ -387,7 +386,7 @@ static int ttm_pool_get_num_unused_pages(void) | |||
| 387 | unsigned i; | 386 | unsigned i; |
| 388 | int total = 0; | 387 | int total = 0; |
| 389 | for (i = 0; i < NUM_POOLS; ++i) | 388 | for (i = 0; i < NUM_POOLS; ++i) |
| 390 | total += _manager.pools[i].npages; | 389 | total += _manager->pools[i].npages; |
| 391 | 390 | ||
| 392 | return total; | 391 | return total; |
| 393 | } | 392 | } |
| @@ -395,7 +394,7 @@ static int ttm_pool_get_num_unused_pages(void) | |||
| 395 | /** | 394 | /** |
| 396 | * Callback for mm to request pool to reduce number of page held. | 395 | * Callback for mm to request pool to reduce number of page held. |
| 397 | */ | 396 | */ |
| 398 | static int ttm_pool_mm_shrink(int shrink_pages, gfp_t gfp_mask) | 397 | static int ttm_pool_mm_shrink(struct shrinker *shrink, int shrink_pages, gfp_t gfp_mask) |
| 399 | { | 398 | { |
| 400 | static atomic_t start_pool = ATOMIC_INIT(0); | 399 | static atomic_t start_pool = ATOMIC_INIT(0); |
| 401 | unsigned i; | 400 | unsigned i; |
| @@ -408,7 +407,7 @@ static int ttm_pool_mm_shrink(int shrink_pages, gfp_t gfp_mask) | |||
| 408 | unsigned nr_free = shrink_pages; | 407 | unsigned nr_free = shrink_pages; |
| 409 | if (shrink_pages == 0) | 408 | if (shrink_pages == 0) |
| 410 | break; | 409 | break; |
| 411 | pool = &_manager.pools[(i + pool_offset)%NUM_POOLS]; | 410 | pool = &_manager->pools[(i + pool_offset)%NUM_POOLS]; |
| 412 | shrink_pages = ttm_page_pool_free(pool, nr_free); | 411 | shrink_pages = ttm_page_pool_free(pool, nr_free); |
| 413 | } | 412 | } |
| 414 | /* return estimated number of unused pages in pool */ | 413 | /* return estimated number of unused pages in pool */ |
| @@ -576,10 +575,10 @@ static void ttm_page_pool_fill_locked(struct ttm_page_pool *pool, | |||
| 576 | 575 | ||
| 577 | /* If allocation request is small and there is not enough | 576 | /* If allocation request is small and there is not enough |
| 578 | * pages in pool we fill the pool first */ | 577 | * pages in pool we fill the pool first */ |
| 579 | if (count < _manager.options.small | 578 | if (count < _manager->options.small |
| 580 | && count > pool->npages) { | 579 | && count > pool->npages) { |
| 581 | struct list_head new_pages; | 580 | struct list_head new_pages; |
| 582 | unsigned alloc_size = _manager.options.alloc_size; | 581 | unsigned alloc_size = _manager->options.alloc_size; |
| 583 | 582 | ||
| 584 | /** | 583 | /** |
| 585 | * Can't change page caching if in irqsave context. We have to | 584 | * Can't change page caching if in irqsave context. We have to |
| @@ -759,8 +758,8 @@ void ttm_put_pages(struct list_head *pages, unsigned page_count, int flags, | |||
| 759 | pool->npages += page_count; | 758 | pool->npages += page_count; |
| 760 | /* Check that we don't go over the pool limit */ | 759 | /* Check that we don't go over the pool limit */ |
| 761 | page_count = 0; | 760 | page_count = 0; |
| 762 | if (pool->npages > _manager.options.max_size) { | 761 | if (pool->npages > _manager->options.max_size) { |
| 763 | page_count = pool->npages - _manager.options.max_size; | 762 | page_count = pool->npages - _manager->options.max_size; |
| 764 | /* free at least NUM_PAGES_TO_ALLOC number of pages | 763 | /* free at least NUM_PAGES_TO_ALLOC number of pages |
| 765 | * to reduce calls to set_memory_wb */ | 764 | * to reduce calls to set_memory_wb */ |
| 766 | if (page_count < NUM_PAGES_TO_ALLOC) | 765 | if (page_count < NUM_PAGES_TO_ALLOC) |
| @@ -785,33 +784,36 @@ static void ttm_page_pool_init_locked(struct ttm_page_pool *pool, int flags, | |||
| 785 | int ttm_page_alloc_init(struct ttm_mem_global *glob, unsigned max_pages) | 784 | int ttm_page_alloc_init(struct ttm_mem_global *glob, unsigned max_pages) |
| 786 | { | 785 | { |
| 787 | int ret; | 786 | int ret; |
| 788 | if (atomic_add_return(1, &_manager.page_alloc_inited) > 1) | 787 | |
| 789 | return 0; | 788 | WARN_ON(_manager); |
| 790 | 789 | ||
| 791 | printk(KERN_INFO TTM_PFX "Initializing pool allocator.\n"); | 790 | printk(KERN_INFO TTM_PFX "Initializing pool allocator.\n"); |
| 792 | 791 | ||
| 793 | ttm_page_pool_init_locked(&_manager.wc_pool, GFP_HIGHUSER, "wc"); | 792 | _manager = kzalloc(sizeof(*_manager), GFP_KERNEL); |
| 793 | |||
| 794 | ttm_page_pool_init_locked(&_manager->wc_pool, GFP_HIGHUSER, "wc"); | ||
| 794 | 795 | ||
| 795 | ttm_page_pool_init_locked(&_manager.uc_pool, GFP_HIGHUSER, "uc"); | 796 | ttm_page_pool_init_locked(&_manager->uc_pool, GFP_HIGHUSER, "uc"); |
| 796 | 797 | ||
| 797 | ttm_page_pool_init_locked(&_manager.wc_pool_dma32, GFP_USER | GFP_DMA32, | 798 | ttm_page_pool_init_locked(&_manager->wc_pool_dma32, |
| 798 | "wc dma"); | 799 | GFP_USER | GFP_DMA32, "wc dma"); |
| 799 | 800 | ||
| 800 | ttm_page_pool_init_locked(&_manager.uc_pool_dma32, GFP_USER | GFP_DMA32, | 801 | ttm_page_pool_init_locked(&_manager->uc_pool_dma32, |
| 801 | "uc dma"); | 802 | GFP_USER | GFP_DMA32, "uc dma"); |
| 802 | 803 | ||
| 803 | _manager.options.max_size = max_pages; | 804 | _manager->options.max_size = max_pages; |
| 804 | _manager.options.small = SMALL_ALLOCATION; | 805 | _manager->options.small = SMALL_ALLOCATION; |
| 805 | _manager.options.alloc_size = NUM_PAGES_TO_ALLOC; | 806 | _manager->options.alloc_size = NUM_PAGES_TO_ALLOC; |
| 806 | 807 | ||
| 807 | kobject_init(&_manager.kobj, &ttm_pool_kobj_type); | 808 | ret = kobject_init_and_add(&_manager->kobj, &ttm_pool_kobj_type, |
| 808 | ret = kobject_add(&_manager.kobj, &glob->kobj, "pool"); | 809 | &glob->kobj, "pool"); |
| 809 | if (unlikely(ret != 0)) { | 810 | if (unlikely(ret != 0)) { |
| 810 | kobject_put(&_manager.kobj); | 811 | kobject_put(&_manager->kobj); |
| 812 | _manager = NULL; | ||
| 811 | return ret; | 813 | return ret; |
| 812 | } | 814 | } |
| 813 | 815 | ||
| 814 | ttm_pool_mm_shrink_init(&_manager); | 816 | ttm_pool_mm_shrink_init(_manager); |
| 815 | 817 | ||
| 816 | return 0; | 818 | return 0; |
| 817 | } | 819 | } |
| @@ -820,16 +822,14 @@ void ttm_page_alloc_fini() | |||
| 820 | { | 822 | { |
| 821 | int i; | 823 | int i; |
| 822 | 824 | ||
| 823 | if (atomic_sub_return(1, &_manager.page_alloc_inited) > 0) | ||
| 824 | return; | ||
| 825 | |||
| 826 | printk(KERN_INFO TTM_PFX "Finalizing pool allocator.\n"); | 825 | printk(KERN_INFO TTM_PFX "Finalizing pool allocator.\n"); |
| 827 | ttm_pool_mm_shrink_fini(&_manager); | 826 | ttm_pool_mm_shrink_fini(_manager); |
| 828 | 827 | ||
| 829 | for (i = 0; i < NUM_POOLS; ++i) | 828 | for (i = 0; i < NUM_POOLS; ++i) |
| 830 | ttm_page_pool_free(&_manager.pools[i], FREE_ALL_PAGES); | 829 | ttm_page_pool_free(&_manager->pools[i], FREE_ALL_PAGES); |
| 831 | 830 | ||
| 832 | kobject_put(&_manager.kobj); | 831 | kobject_put(&_manager->kobj); |
| 832 | _manager = NULL; | ||
| 833 | } | 833 | } |
| 834 | 834 | ||
| 835 | int ttm_page_alloc_debugfs(struct seq_file *m, void *data) | 835 | int ttm_page_alloc_debugfs(struct seq_file *m, void *data) |
| @@ -837,14 +837,14 @@ int ttm_page_alloc_debugfs(struct seq_file *m, void *data) | |||
| 837 | struct ttm_page_pool *p; | 837 | struct ttm_page_pool *p; |
| 838 | unsigned i; | 838 | unsigned i; |
| 839 | char *h[] = {"pool", "refills", "pages freed", "size"}; | 839 | char *h[] = {"pool", "refills", "pages freed", "size"}; |
| 840 | if (atomic_read(&_manager.page_alloc_inited) == 0) { | 840 | if (!_manager) { |
| 841 | seq_printf(m, "No pool allocator running.\n"); | 841 | seq_printf(m, "No pool allocator running.\n"); |
| 842 | return 0; | 842 | return 0; |
| 843 | } | 843 | } |
| 844 | seq_printf(m, "%6s %12s %13s %8s\n", | 844 | seq_printf(m, "%6s %12s %13s %8s\n", |
| 845 | h[0], h[1], h[2], h[3]); | 845 | h[0], h[1], h[2], h[3]); |
| 846 | for (i = 0; i < NUM_POOLS; ++i) { | 846 | for (i = 0; i < NUM_POOLS; ++i) { |
| 847 | p = &_manager.pools[i]; | 847 | p = &_manager->pools[i]; |
| 848 | 848 | ||
| 849 | seq_printf(m, "%6s %12ld %13ld %8d\n", | 849 | seq_printf(m, "%6s %12ld %13ld %8d\n", |
| 850 | p->name, p->nrefills, | 850 | p->name, p->nrefills, |
diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c b/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c index f1d626112415..437ac786277a 100644 --- a/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_kms.c | |||
| @@ -972,6 +972,7 @@ int vmw_kms_update_layout_ioctl(struct drm_device *dev, void *data, | |||
| 972 | ret = copy_from_user(rects, user_rects, rects_size); | 972 | ret = copy_from_user(rects, user_rects, rects_size); |
| 973 | if (unlikely(ret != 0)) { | 973 | if (unlikely(ret != 0)) { |
| 974 | DRM_ERROR("Failed to get rects.\n"); | 974 | DRM_ERROR("Failed to get rects.\n"); |
| 975 | ret = -EFAULT; | ||
| 975 | goto out_free; | 976 | goto out_free; |
| 976 | } | 977 | } |
| 977 | 978 | ||
