diff options
122 files changed, 12028 insertions, 2947 deletions
diff --git a/drivers/char/agp/intel-agp.c b/drivers/char/agp/intel-agp.c index 3cb56a049e24..30c36ac2cd00 100644 --- a/drivers/char/agp/intel-agp.c +++ b/drivers/char/agp/intel-agp.c | |||
| @@ -36,10 +36,10 @@ | |||
| 36 | #define PCI_DEVICE_ID_INTEL_82965GME_IG 0x2A12 | 36 | #define PCI_DEVICE_ID_INTEL_82965GME_IG 0x2A12 |
| 37 | #define PCI_DEVICE_ID_INTEL_82945GME_HB 0x27AC | 37 | #define PCI_DEVICE_ID_INTEL_82945GME_HB 0x27AC |
| 38 | #define PCI_DEVICE_ID_INTEL_82945GME_IG 0x27AE | 38 | #define PCI_DEVICE_ID_INTEL_82945GME_IG 0x27AE |
| 39 | #define PCI_DEVICE_ID_INTEL_IGDGM_HB 0xA010 | 39 | #define PCI_DEVICE_ID_INTEL_PINEVIEW_M_HB 0xA010 |
| 40 | #define PCI_DEVICE_ID_INTEL_IGDGM_IG 0xA011 | 40 | #define PCI_DEVICE_ID_INTEL_PINEVIEW_M_IG 0xA011 |
| 41 | #define PCI_DEVICE_ID_INTEL_IGDG_HB 0xA000 | 41 | #define PCI_DEVICE_ID_INTEL_PINEVIEW_HB 0xA000 |
| 42 | #define PCI_DEVICE_ID_INTEL_IGDG_IG 0xA001 | 42 | #define PCI_DEVICE_ID_INTEL_PINEVIEW_IG 0xA001 |
| 43 | #define PCI_DEVICE_ID_INTEL_G33_HB 0x29C0 | 43 | #define PCI_DEVICE_ID_INTEL_G33_HB 0x29C0 |
| 44 | #define PCI_DEVICE_ID_INTEL_G33_IG 0x29C2 | 44 | #define PCI_DEVICE_ID_INTEL_G33_IG 0x29C2 |
| 45 | #define PCI_DEVICE_ID_INTEL_Q35_HB 0x29B0 | 45 | #define PCI_DEVICE_ID_INTEL_Q35_HB 0x29B0 |
| @@ -50,20 +50,20 @@ | |||
| 50 | #define PCI_DEVICE_ID_INTEL_B43_IG 0x2E42 | 50 | #define PCI_DEVICE_ID_INTEL_B43_IG 0x2E42 |
| 51 | #define PCI_DEVICE_ID_INTEL_GM45_HB 0x2A40 | 51 | #define PCI_DEVICE_ID_INTEL_GM45_HB 0x2A40 |
| 52 | #define PCI_DEVICE_ID_INTEL_GM45_IG 0x2A42 | 52 | #define PCI_DEVICE_ID_INTEL_GM45_IG 0x2A42 |
| 53 | #define PCI_DEVICE_ID_INTEL_IGD_E_HB 0x2E00 | 53 | #define PCI_DEVICE_ID_INTEL_EAGLELAKE_HB 0x2E00 |
| 54 | #define PCI_DEVICE_ID_INTEL_IGD_E_IG 0x2E02 | 54 | #define PCI_DEVICE_ID_INTEL_EAGLELAKE_IG 0x2E02 |
| 55 | #define PCI_DEVICE_ID_INTEL_Q45_HB 0x2E10 | 55 | #define PCI_DEVICE_ID_INTEL_Q45_HB 0x2E10 |
| 56 | #define PCI_DEVICE_ID_INTEL_Q45_IG 0x2E12 | 56 | #define PCI_DEVICE_ID_INTEL_Q45_IG 0x2E12 |
| 57 | #define PCI_DEVICE_ID_INTEL_G45_HB 0x2E20 | 57 | #define PCI_DEVICE_ID_INTEL_G45_HB 0x2E20 |
| 58 | #define PCI_DEVICE_ID_INTEL_G45_IG 0x2E22 | 58 | #define PCI_DEVICE_ID_INTEL_G45_IG 0x2E22 |
| 59 | #define PCI_DEVICE_ID_INTEL_G41_HB 0x2E30 | 59 | #define PCI_DEVICE_ID_INTEL_G41_HB 0x2E30 |
| 60 | #define PCI_DEVICE_ID_INTEL_G41_IG 0x2E32 | 60 | #define PCI_DEVICE_ID_INTEL_G41_IG 0x2E32 |
| 61 | #define PCI_DEVICE_ID_INTEL_IGDNG_D_HB 0x0040 | 61 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_D_HB 0x0040 |
| 62 | #define PCI_DEVICE_ID_INTEL_IGDNG_D_IG 0x0042 | 62 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_D_IG 0x0042 |
| 63 | #define PCI_DEVICE_ID_INTEL_IGDNG_M_HB 0x0044 | 63 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_M_HB 0x0044 |
| 64 | #define PCI_DEVICE_ID_INTEL_IGDNG_MA_HB 0x0062 | 64 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_MA_HB 0x0062 |
| 65 | #define PCI_DEVICE_ID_INTEL_IGDNG_MC2_HB 0x006a | 65 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_MC2_HB 0x006a |
| 66 | #define PCI_DEVICE_ID_INTEL_IGDNG_M_IG 0x0046 | 66 | #define PCI_DEVICE_ID_INTEL_IRONLAKE_M_IG 0x0046 |
| 67 | 67 | ||
| 68 | /* cover 915 and 945 variants */ | 68 | /* cover 915 and 945 variants */ |
| 69 | #define IS_I915 (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_E7221_HB || \ | 69 | #define IS_I915 (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_E7221_HB || \ |
| @@ -83,22 +83,22 @@ | |||
| 83 | #define IS_G33 (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G33_HB || \ | 83 | #define IS_G33 (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G33_HB || \ |
| 84 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q35_HB || \ | 84 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q35_HB || \ |
| 85 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q33_HB || \ | 85 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q33_HB || \ |
| 86 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDGM_HB || \ | 86 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_PINEVIEW_M_HB || \ |
| 87 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDG_HB) | 87 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_PINEVIEW_HB) |
| 88 | 88 | ||
| 89 | #define IS_IGD (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDGM_HB || \ | 89 | #define IS_PINEVIEW (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_PINEVIEW_M_HB || \ |
| 90 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDG_HB) | 90 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_PINEVIEW_HB) |
| 91 | 91 | ||
| 92 | #define IS_G4X (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGD_E_HB || \ | 92 | #define IS_G4X (agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_EAGLELAKE_HB || \ |
| 93 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q45_HB || \ | 93 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_Q45_HB || \ |
| 94 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G45_HB || \ | 94 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G45_HB || \ |
| 95 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_GM45_HB || \ | 95 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_GM45_HB || \ |
| 96 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G41_HB || \ | 96 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_G41_HB || \ |
| 97 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_B43_HB || \ | 97 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_B43_HB || \ |
| 98 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDNG_D_HB || \ | 98 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IRONLAKE_D_HB || \ |
| 99 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDNG_M_HB || \ | 99 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IRONLAKE_M_HB || \ |
| 100 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDNG_MA_HB || \ | 100 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IRONLAKE_MA_HB || \ |
| 101 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IGDNG_MC2_HB) | 101 | agp_bridge->dev->device == PCI_DEVICE_ID_INTEL_IRONLAKE_MC2_HB) |
| 102 | 102 | ||
| 103 | extern int agp_memory_reserved; | 103 | extern int agp_memory_reserved; |
| 104 | 104 | ||
| @@ -178,6 +178,7 @@ static struct _intel_private { | |||
| 178 | * popup and for the GTT. | 178 | * popup and for the GTT. |
| 179 | */ | 179 | */ |
| 180 | int gtt_entries; /* i830+ */ | 180 | int gtt_entries; /* i830+ */ |
| 181 | int gtt_total_size; | ||
| 181 | union { | 182 | union { |
| 182 | void __iomem *i9xx_flush_page; | 183 | void __iomem *i9xx_flush_page; |
| 183 | void *i8xx_flush_page; | 184 | void *i8xx_flush_page; |
| @@ -653,7 +654,7 @@ static void intel_i830_init_gtt_entries(void) | |||
| 653 | size = 512; | 654 | size = 512; |
| 654 | } | 655 | } |
| 655 | size += 4; /* add in BIOS popup space */ | 656 | size += 4; /* add in BIOS popup space */ |
| 656 | } else if (IS_G33 && !IS_IGD) { | 657 | } else if (IS_G33 && !IS_PINEVIEW) { |
| 657 | /* G33's GTT size defined in gmch_ctrl */ | 658 | /* G33's GTT size defined in gmch_ctrl */ |
| 658 | switch (gmch_ctrl & G33_PGETBL_SIZE_MASK) { | 659 | switch (gmch_ctrl & G33_PGETBL_SIZE_MASK) { |
| 659 | case G33_PGETBL_SIZE_1M: | 660 | case G33_PGETBL_SIZE_1M: |
| @@ -669,7 +670,7 @@ static void intel_i830_init_gtt_entries(void) | |||
| 669 | size = 512; | 670 | size = 512; |
| 670 | } | 671 | } |
| 671 | size += 4; | 672 | size += 4; |
| 672 | } else if (IS_G4X || IS_IGD) { | 673 | } else if (IS_G4X || IS_PINEVIEW) { |
| 673 | /* On 4 series hardware, GTT stolen is separate from graphics | 674 | /* On 4 series hardware, GTT stolen is separate from graphics |
| 674 | * stolen, ignore it in stolen gtt entries counting. However, | 675 | * stolen, ignore it in stolen gtt entries counting. However, |
| 675 | * 4KB of the stolen memory doesn't get mapped to the GTT. | 676 | * 4KB of the stolen memory doesn't get mapped to the GTT. |
| @@ -1153,7 +1154,7 @@ static int intel_i915_configure(void) | |||
| 1153 | readl(intel_private.registers+I810_PGETBL_CTL); /* PCI Posting. */ | 1154 | readl(intel_private.registers+I810_PGETBL_CTL); /* PCI Posting. */ |
| 1154 | 1155 | ||
| 1155 | if (agp_bridge->driver->needs_scratch_page) { | 1156 | if (agp_bridge->driver->needs_scratch_page) { |
| 1156 | for (i = intel_private.gtt_entries; i < current_size->num_entries; i++) { | 1157 | for (i = intel_private.gtt_entries; i < intel_private.gtt_total_size; i++) { |
| 1157 | writel(agp_bridge->scratch_page, intel_private.gtt+i); | 1158 | writel(agp_bridge->scratch_page, intel_private.gtt+i); |
| 1158 | } | 1159 | } |
| 1159 | readl(intel_private.gtt+i-1); /* PCI Posting. */ | 1160 | readl(intel_private.gtt+i-1); /* PCI Posting. */ |
| @@ -1308,6 +1309,8 @@ static int intel_i915_create_gatt_table(struct agp_bridge_data *bridge) | |||
| 1308 | if (!intel_private.gtt) | 1309 | if (!intel_private.gtt) |
| 1309 | return -ENOMEM; | 1310 | return -ENOMEM; |
| 1310 | 1311 | ||
| 1312 | intel_private.gtt_total_size = gtt_map_size / 4; | ||
| 1313 | |||
| 1311 | temp &= 0xfff80000; | 1314 | temp &= 0xfff80000; |
| 1312 | 1315 | ||
| 1313 | intel_private.registers = ioremap(temp, 128 * 4096); | 1316 | intel_private.registers = ioremap(temp, 128 * 4096); |
| @@ -1352,15 +1355,15 @@ static void intel_i965_get_gtt_range(int *gtt_offset, int *gtt_size) | |||
| 1352 | { | 1355 | { |
| 1353 | switch (agp_bridge->dev->device) { | 1356 | switch (agp_bridge->dev->device) { |
| 1354 | case PCI_DEVICE_ID_INTEL_GM45_HB: | 1357 | case PCI_DEVICE_ID_INTEL_GM45_HB: |
| 1355 | case PCI_DEVICE_ID_INTEL_IGD_E_HB: | 1358 | case PCI_DEVICE_ID_INTEL_EAGLELAKE_HB: |
| 1356 | case PCI_DEVICE_ID_INTEL_Q45_HB: | 1359 | case PCI_DEVICE_ID_INTEL_Q45_HB: |
| 1357 | case PCI_DEVICE_ID_INTEL_G45_HB: | 1360 | case PCI_DEVICE_ID_INTEL_G45_HB: |
| 1358 | case PCI_DEVICE_ID_INTEL_G41_HB: | 1361 | case PCI_DEVICE_ID_INTEL_G41_HB: |
| 1359 | case PCI_DEVICE_ID_INTEL_B43_HB: | 1362 | case PCI_DEVICE_ID_INTEL_B43_HB: |
| 1360 | case PCI_DEVICE_ID_INTEL_IGDNG_D_HB: | 1363 | case PCI_DEVICE_ID_INTEL_IRONLAKE_D_HB: |
| 1361 | case PCI_DEVICE_ID_INTEL_IGDNG_M_HB: | 1364 | case PCI_DEVICE_ID_INTEL_IRONLAKE_M_HB: |
| 1362 | case PCI_DEVICE_ID_INTEL_IGDNG_MA_HB: | 1365 | case PCI_DEVICE_ID_INTEL_IRONLAKE_MA_HB: |
| 1363 | case PCI_DEVICE_ID_INTEL_IGDNG_MC2_HB: | 1366 | case PCI_DEVICE_ID_INTEL_IRONLAKE_MC2_HB: |
| 1364 | *gtt_offset = *gtt_size = MB(2); | 1367 | *gtt_offset = *gtt_size = MB(2); |
| 1365 | break; | 1368 | break; |
| 1366 | default: | 1369 | default: |
| @@ -1395,6 +1398,8 @@ static int intel_i965_create_gatt_table(struct agp_bridge_data *bridge) | |||
| 1395 | if (!intel_private.gtt) | 1398 | if (!intel_private.gtt) |
| 1396 | return -ENOMEM; | 1399 | return -ENOMEM; |
| 1397 | 1400 | ||
| 1401 | intel_private.gtt_total_size = gtt_size / 4; | ||
| 1402 | |||
| 1398 | intel_private.registers = ioremap(temp, 128 * 4096); | 1403 | intel_private.registers = ioremap(temp, 128 * 4096); |
| 1399 | if (!intel_private.registers) { | 1404 | if (!intel_private.registers) { |
| 1400 | iounmap(intel_private.gtt); | 1405 | iounmap(intel_private.gtt); |
| @@ -2340,14 +2345,14 @@ static const struct intel_driver_description { | |||
| 2340 | NULL, &intel_g33_driver }, | 2345 | NULL, &intel_g33_driver }, |
| 2341 | { PCI_DEVICE_ID_INTEL_Q33_HB, PCI_DEVICE_ID_INTEL_Q33_IG, 0, "Q33", | 2346 | { PCI_DEVICE_ID_INTEL_Q33_HB, PCI_DEVICE_ID_INTEL_Q33_IG, 0, "Q33", |
| 2342 | NULL, &intel_g33_driver }, | 2347 | NULL, &intel_g33_driver }, |
| 2343 | { PCI_DEVICE_ID_INTEL_IGDGM_HB, PCI_DEVICE_ID_INTEL_IGDGM_IG, 0, "IGD", | 2348 | { PCI_DEVICE_ID_INTEL_PINEVIEW_M_HB, PCI_DEVICE_ID_INTEL_PINEVIEW_M_IG, 0, "Pineview", |
| 2344 | NULL, &intel_g33_driver }, | 2349 | NULL, &intel_g33_driver }, |
| 2345 | { PCI_DEVICE_ID_INTEL_IGDG_HB, PCI_DEVICE_ID_INTEL_IGDG_IG, 0, "IGD", | 2350 | { PCI_DEVICE_ID_INTEL_PINEVIEW_HB, PCI_DEVICE_ID_INTEL_PINEVIEW_IG, 0, "Pineview", |
| 2346 | NULL, &intel_g33_driver }, | 2351 | NULL, &intel_g33_driver }, |
| 2347 | { PCI_DEVICE_ID_INTEL_GM45_HB, PCI_DEVICE_ID_INTEL_GM45_IG, 0, | 2352 | { PCI_DEVICE_ID_INTEL_GM45_HB, PCI_DEVICE_ID_INTEL_GM45_IG, 0, |
| 2348 | "Mobile Intel® GM45 Express", NULL, &intel_i965_driver }, | 2353 | "GM45", NULL, &intel_i965_driver }, |
| 2349 | { PCI_DEVICE_ID_INTEL_IGD_E_HB, PCI_DEVICE_ID_INTEL_IGD_E_IG, 0, | 2354 | { PCI_DEVICE_ID_INTEL_EAGLELAKE_HB, PCI_DEVICE_ID_INTEL_EAGLELAKE_IG, 0, |
| 2350 | "Intel Integrated Graphics Device", NULL, &intel_i965_driver }, | 2355 | "Eaglelake", NULL, &intel_i965_driver }, |
| 2351 | { PCI_DEVICE_ID_INTEL_Q45_HB, PCI_DEVICE_ID_INTEL_Q45_IG, 0, | 2356 | { PCI_DEVICE_ID_INTEL_Q45_HB, PCI_DEVICE_ID_INTEL_Q45_IG, 0, |
| 2352 | "Q45/Q43", NULL, &intel_i965_driver }, | 2357 | "Q45/Q43", NULL, &intel_i965_driver }, |
| 2353 | { PCI_DEVICE_ID_INTEL_G45_HB, PCI_DEVICE_ID_INTEL_G45_IG, 0, | 2358 | { PCI_DEVICE_ID_INTEL_G45_HB, PCI_DEVICE_ID_INTEL_G45_IG, 0, |
| @@ -2356,14 +2361,14 @@ static const struct intel_driver_description { | |||
| 2356 | "B43", NULL, &intel_i965_driver }, | 2361 | "B43", NULL, &intel_i965_driver }, |
| 2357 | { PCI_DEVICE_ID_INTEL_G41_HB, PCI_DEVICE_ID_INTEL_G41_IG, 0, | 2362 | { PCI_DEVICE_ID_INTEL_G41_HB, PCI_DEVICE_ID_INTEL_G41_IG, 0, |
| 2358 | "G41", NULL, &intel_i965_driver }, | 2363 | "G41", NULL, &intel_i965_driver }, |
| 2359 | { PCI_DEVICE_ID_INTEL_IGDNG_D_HB, PCI_DEVICE_ID_INTEL_IGDNG_D_IG, 0, | 2364 | { PCI_DEVICE_ID_INTEL_IRONLAKE_D_HB, PCI_DEVICE_ID_INTEL_IRONLAKE_D_IG, 0, |
| 2360 | "IGDNG/D", NULL, &intel_i965_driver }, | 2365 | "Ironlake/D", NULL, &intel_i965_driver }, |
| 2361 | { PCI_DEVICE_ID_INTEL_IGDNG_M_HB, PCI_DEVICE_ID_INTEL_IGDNG_M_IG, 0, | 2366 | { PCI_DEVICE_ID_INTEL_IRONLAKE_M_HB, PCI_DEVICE_ID_INTEL_IRONLAKE_M_IG, 0, |
| 2362 | "IGDNG/M", NULL, &intel_i965_driver }, | 2367 | "Ironlake/M", NULL, &intel_i965_driver }, |
| 2363 | { PCI_DEVICE_ID_INTEL_IGDNG_MA_HB, PCI_DEVICE_ID_INTEL_IGDNG_M_IG, 0, | 2368 | { PCI_DEVICE_ID_INTEL_IRONLAKE_MA_HB, PCI_DEVICE_ID_INTEL_IRONLAKE_M_IG, 0, |
| 2364 | "IGDNG/MA", NULL, &intel_i965_driver }, | 2369 | "Ironlake/MA", NULL, &intel_i965_driver }, |
| 2365 | { PCI_DEVICE_ID_INTEL_IGDNG_MC2_HB, PCI_DEVICE_ID_INTEL_IGDNG_M_IG, 0, | 2370 | { PCI_DEVICE_ID_INTEL_IRONLAKE_MC2_HB, PCI_DEVICE_ID_INTEL_IRONLAKE_M_IG, 0, |
| 2366 | "IGDNG/MC2", NULL, &intel_i965_driver }, | 2371 | "Ironlake/MC2", NULL, &intel_i965_driver }, |
| 2367 | { 0, 0, 0, NULL, NULL, NULL } | 2372 | { 0, 0, 0, NULL, NULL, NULL } |
| 2368 | }; | 2373 | }; |
| 2369 | 2374 | ||
| @@ -2545,8 +2550,8 @@ static struct pci_device_id agp_intel_pci_table[] = { | |||
| 2545 | ID(PCI_DEVICE_ID_INTEL_82945G_HB), | 2550 | ID(PCI_DEVICE_ID_INTEL_82945G_HB), |
| 2546 | ID(PCI_DEVICE_ID_INTEL_82945GM_HB), | 2551 | ID(PCI_DEVICE_ID_INTEL_82945GM_HB), |
| 2547 | ID(PCI_DEVICE_ID_INTEL_82945GME_HB), | 2552 | ID(PCI_DEVICE_ID_INTEL_82945GME_HB), |
| 2548 | ID(PCI_DEVICE_ID_INTEL_IGDGM_HB), | 2553 | ID(PCI_DEVICE_ID_INTEL_PINEVIEW_M_HB), |
| 2549 | ID(PCI_DEVICE_ID_INTEL_IGDG_HB), | 2554 | ID(PCI_DEVICE_ID_INTEL_PINEVIEW_HB), |
| 2550 | ID(PCI_DEVICE_ID_INTEL_82946GZ_HB), | 2555 | ID(PCI_DEVICE_ID_INTEL_82946GZ_HB), |
| 2551 | ID(PCI_DEVICE_ID_INTEL_82G35_HB), | 2556 | ID(PCI_DEVICE_ID_INTEL_82G35_HB), |
| 2552 | ID(PCI_DEVICE_ID_INTEL_82965Q_HB), | 2557 | ID(PCI_DEVICE_ID_INTEL_82965Q_HB), |
| @@ -2557,15 +2562,15 @@ static struct pci_device_id agp_intel_pci_table[] = { | |||
| 2557 | ID(PCI_DEVICE_ID_INTEL_Q35_HB), | 2562 | ID(PCI_DEVICE_ID_INTEL_Q35_HB), |
| 2558 | ID(PCI_DEVICE_ID_INTEL_Q33_HB), | 2563 | ID(PCI_DEVICE_ID_INTEL_Q33_HB), |
| 2559 | ID(PCI_DEVICE_ID_INTEL_GM45_HB), | 2564 | ID(PCI_DEVICE_ID_INTEL_GM45_HB), |
| 2560 | ID(PCI_DEVICE_ID_INTEL_IGD_E_HB), | 2565 | ID(PCI_DEVICE_ID_INTEL_EAGLELAKE_HB), |
| 2561 | ID(PCI_DEVICE_ID_INTEL_Q45_HB), | 2566 | ID(PCI_DEVICE_ID_INTEL_Q45_HB), |
| 2562 | ID(PCI_DEVICE_ID_INTEL_G45_HB), | 2567 | ID(PCI_DEVICE_ID_INTEL_G45_HB), |
| 2563 | ID(PCI_DEVICE_ID_INTEL_G41_HB), | 2568 | ID(PCI_DEVICE_ID_INTEL_G41_HB), |
| 2564 | ID(PCI_DEVICE_ID_INTEL_B43_HB), | 2569 | ID(PCI_DEVICE_ID_INTEL_B43_HB), |
| 2565 | ID(PCI_DEVICE_ID_INTEL_IGDNG_D_HB), | 2570 | ID(PCI_DEVICE_ID_INTEL_IRONLAKE_D_HB), |
| 2566 | ID(PCI_DEVICE_ID_INTEL_IGDNG_M_HB), | 2571 | ID(PCI_DEVICE_ID_INTEL_IRONLAKE_M_HB), |
| 2567 | ID(PCI_DEVICE_ID_INTEL_IGDNG_MA_HB), | 2572 | ID(PCI_DEVICE_ID_INTEL_IRONLAKE_MA_HB), |
| 2568 | ID(PCI_DEVICE_ID_INTEL_IGDNG_MC2_HB), | 2573 | ID(PCI_DEVICE_ID_INTEL_IRONLAKE_MC2_HB), |
| 2569 | { } | 2574 | { } |
| 2570 | }; | 2575 | }; |
| 2571 | 2576 | ||
diff --git a/drivers/gpu/drm/Makefile b/drivers/gpu/drm/Makefile index 3c8827a7aabd..91567ac806f1 100644 --- a/drivers/gpu/drm/Makefile +++ b/drivers/gpu/drm/Makefile | |||
| @@ -15,7 +15,7 @@ drm-y := drm_auth.o drm_bufs.o drm_cache.o \ | |||
| 15 | 15 | ||
| 16 | drm-$(CONFIG_COMPAT) += drm_ioc32.o | 16 | drm-$(CONFIG_COMPAT) += drm_ioc32.o |
| 17 | 17 | ||
| 18 | drm_kms_helper-y := drm_fb_helper.o drm_crtc_helper.o | 18 | drm_kms_helper-y := drm_fb_helper.o drm_crtc_helper.o drm_dp_i2c_helper.o |
| 19 | 19 | ||
| 20 | obj-$(CONFIG_DRM_KMS_HELPER) += drm_kms_helper.o | 20 | obj-$(CONFIG_DRM_KMS_HELPER) += drm_kms_helper.o |
| 21 | 21 | ||
diff --git a/drivers/gpu/drm/drm_crtc.c b/drivers/gpu/drm/drm_crtc.c index 3f7c500b2115..5124401f266a 100644 --- a/drivers/gpu/drm/drm_crtc.c +++ b/drivers/gpu/drm/drm_crtc.c | |||
| @@ -125,6 +125,15 @@ static struct drm_prop_enum_list drm_tv_subconnector_enum_list[] = | |||
| 125 | DRM_ENUM_NAME_FN(drm_get_tv_subconnector_name, | 125 | DRM_ENUM_NAME_FN(drm_get_tv_subconnector_name, |
| 126 | drm_tv_subconnector_enum_list) | 126 | drm_tv_subconnector_enum_list) |
| 127 | 127 | ||
| 128 | static struct drm_prop_enum_list drm_dirty_info_enum_list[] = { | ||
| 129 | { DRM_MODE_DIRTY_OFF, "Off" }, | ||
| 130 | { DRM_MODE_DIRTY_ON, "On" }, | ||
| 131 | { DRM_MODE_DIRTY_ANNOTATE, "Annotate" }, | ||
| 132 | }; | ||
| 133 | |||
| 134 | DRM_ENUM_NAME_FN(drm_get_dirty_info_name, | ||
| 135 | drm_dirty_info_enum_list) | ||
| 136 | |||
| 128 | struct drm_conn_prop_enum_list { | 137 | struct drm_conn_prop_enum_list { |
| 129 | int type; | 138 | int type; |
| 130 | char *name; | 139 | char *name; |
| @@ -247,7 +256,8 @@ static void drm_mode_object_put(struct drm_device *dev, | |||
| 247 | mutex_unlock(&dev->mode_config.idr_mutex); | 256 | mutex_unlock(&dev->mode_config.idr_mutex); |
| 248 | } | 257 | } |
| 249 | 258 | ||
| 250 | void *drm_mode_object_find(struct drm_device *dev, uint32_t id, uint32_t type) | 259 | struct drm_mode_object *drm_mode_object_find(struct drm_device *dev, |
| 260 | uint32_t id, uint32_t type) | ||
| 251 | { | 261 | { |
| 252 | struct drm_mode_object *obj = NULL; | 262 | struct drm_mode_object *obj = NULL; |
| 253 | 263 | ||
| @@ -802,6 +812,36 @@ int drm_mode_create_dithering_property(struct drm_device *dev) | |||
| 802 | EXPORT_SYMBOL(drm_mode_create_dithering_property); | 812 | EXPORT_SYMBOL(drm_mode_create_dithering_property); |
| 803 | 813 | ||
| 804 | /** | 814 | /** |
| 815 | * drm_mode_create_dirty_property - create dirty property | ||
| 816 | * @dev: DRM device | ||
| 817 | * | ||
| 818 | * Called by a driver the first time it's needed, must be attached to desired | ||
| 819 | * connectors. | ||
| 820 | */ | ||
| 821 | int drm_mode_create_dirty_info_property(struct drm_device *dev) | ||
| 822 | { | ||
| 823 | struct drm_property *dirty_info; | ||
| 824 | int i; | ||
| 825 | |||
| 826 | if (dev->mode_config.dirty_info_property) | ||
| 827 | return 0; | ||
| 828 | |||
| 829 | dirty_info = | ||
| 830 | drm_property_create(dev, DRM_MODE_PROP_ENUM | | ||
| 831 | DRM_MODE_PROP_IMMUTABLE, | ||
| 832 | "dirty", | ||
| 833 | ARRAY_SIZE(drm_dirty_info_enum_list)); | ||
| 834 | for (i = 0; i < ARRAY_SIZE(drm_dirty_info_enum_list); i++) | ||
| 835 | drm_property_add_enum(dirty_info, i, | ||
| 836 | drm_dirty_info_enum_list[i].type, | ||
| 837 | drm_dirty_info_enum_list[i].name); | ||
| 838 | dev->mode_config.dirty_info_property = dirty_info; | ||
| 839 | |||
| 840 | return 0; | ||
| 841 | } | ||
| 842 | EXPORT_SYMBOL(drm_mode_create_dirty_info_property); | ||
| 843 | |||
| 844 | /** | ||
| 805 | * drm_mode_config_init - initialize DRM mode_configuration structure | 845 | * drm_mode_config_init - initialize DRM mode_configuration structure |
| 806 | * @dev: DRM device | 846 | * @dev: DRM device |
| 807 | * | 847 | * |
| @@ -1753,6 +1793,71 @@ out: | |||
| 1753 | return ret; | 1793 | return ret; |
| 1754 | } | 1794 | } |
| 1755 | 1795 | ||
| 1796 | int drm_mode_dirtyfb_ioctl(struct drm_device *dev, | ||
| 1797 | void *data, struct drm_file *file_priv) | ||
| 1798 | { | ||
| 1799 | struct drm_clip_rect __user *clips_ptr; | ||
| 1800 | struct drm_clip_rect *clips = NULL; | ||
| 1801 | struct drm_mode_fb_dirty_cmd *r = data; | ||
| 1802 | struct drm_mode_object *obj; | ||
| 1803 | struct drm_framebuffer *fb; | ||
| 1804 | unsigned flags; | ||
| 1805 | int num_clips; | ||
| 1806 | int ret = 0; | ||
| 1807 | |||
| 1808 | mutex_lock(&dev->mode_config.mutex); | ||
| 1809 | obj = drm_mode_object_find(dev, r->fb_id, DRM_MODE_OBJECT_FB); | ||
| 1810 | if (!obj) { | ||
| 1811 | DRM_ERROR("invalid framebuffer id\n"); | ||
| 1812 | ret = -EINVAL; | ||
| 1813 | goto out_err1; | ||
| 1814 | } | ||
| 1815 | fb = obj_to_fb(obj); | ||
| 1816 | |||
| 1817 | num_clips = r->num_clips; | ||
| 1818 | clips_ptr = (struct drm_clip_rect *)(unsigned long)r->clips_ptr; | ||
| 1819 | |||
| 1820 | if (!num_clips != !clips_ptr) { | ||
| 1821 | ret = -EINVAL; | ||
| 1822 | goto out_err1; | ||
| 1823 | } | ||
| 1824 | |||
| 1825 | flags = DRM_MODE_FB_DIRTY_FLAGS & r->flags; | ||
| 1826 | |||
| 1827 | /* If userspace annotates copy, clips must come in pairs */ | ||
| 1828 | if (flags & DRM_MODE_FB_DIRTY_ANNOTATE_COPY && (num_clips % 2)) { | ||
| 1829 | ret = -EINVAL; | ||
| 1830 | goto out_err1; | ||
| 1831 | } | ||
| 1832 | |||
| 1833 | if (num_clips && clips_ptr) { | ||
| 1834 | clips = kzalloc(num_clips * sizeof(*clips), GFP_KERNEL); | ||
| 1835 | if (!clips) { | ||
| 1836 | ret = -ENOMEM; | ||
| 1837 | goto out_err1; | ||
| 1838 | } | ||
| 1839 | |||
| 1840 | ret = copy_from_user(clips, clips_ptr, | ||
| 1841 | num_clips * sizeof(*clips)); | ||
| 1842 | if (ret) | ||
| 1843 | goto out_err2; | ||
| 1844 | } | ||
| 1845 | |||
| 1846 | if (fb->funcs->dirty) { | ||
| 1847 | ret = fb->funcs->dirty(fb, flags, r->color, clips, num_clips); | ||
| 1848 | } else { | ||
| 1849 | ret = -ENOSYS; | ||
| 1850 | goto out_err2; | ||
| 1851 | } | ||
| 1852 | |||
| 1853 | out_err2: | ||
| 1854 | kfree(clips); | ||
| 1855 | out_err1: | ||
| 1856 | mutex_unlock(&dev->mode_config.mutex); | ||
| 1857 | return ret; | ||
| 1858 | } | ||
| 1859 | |||
| 1860 | |||
| 1756 | /** | 1861 | /** |
| 1757 | * drm_fb_release - remove and free the FBs on this file | 1862 | * drm_fb_release - remove and free the FBs on this file |
| 1758 | * @filp: file * from the ioctl | 1863 | * @filp: file * from the ioctl |
| @@ -2478,3 +2583,72 @@ out: | |||
| 2478 | mutex_unlock(&dev->mode_config.mutex); | 2583 | mutex_unlock(&dev->mode_config.mutex); |
| 2479 | return ret; | 2584 | return ret; |
| 2480 | } | 2585 | } |
| 2586 | |||
| 2587 | int drm_mode_page_flip_ioctl(struct drm_device *dev, | ||
| 2588 | void *data, struct drm_file *file_priv) | ||
| 2589 | { | ||
| 2590 | struct drm_mode_crtc_page_flip *page_flip = data; | ||
| 2591 | struct drm_mode_object *obj; | ||
| 2592 | struct drm_crtc *crtc; | ||
| 2593 | struct drm_framebuffer *fb; | ||
| 2594 | struct drm_pending_vblank_event *e = NULL; | ||
| 2595 | unsigned long flags; | ||
| 2596 | int ret = -EINVAL; | ||
| 2597 | |||
| 2598 | if (page_flip->flags & ~DRM_MODE_PAGE_FLIP_FLAGS || | ||
| 2599 | page_flip->reserved != 0) | ||
| 2600 | return -EINVAL; | ||
| 2601 | |||
| 2602 | mutex_lock(&dev->mode_config.mutex); | ||
| 2603 | obj = drm_mode_object_find(dev, page_flip->crtc_id, DRM_MODE_OBJECT_CRTC); | ||
| 2604 | if (!obj) | ||
| 2605 | goto out; | ||
| 2606 | crtc = obj_to_crtc(obj); | ||
| 2607 | |||
| 2608 | if (crtc->funcs->page_flip == NULL) | ||
| 2609 | goto out; | ||
| 2610 | |||
| 2611 | obj = drm_mode_object_find(dev, page_flip->fb_id, DRM_MODE_OBJECT_FB); | ||
| 2612 | if (!obj) | ||
| 2613 | goto out; | ||
| 2614 | fb = obj_to_fb(obj); | ||
| 2615 | |||
| 2616 | if (page_flip->flags & DRM_MODE_PAGE_FLIP_EVENT) { | ||
| 2617 | ret = -ENOMEM; | ||
| 2618 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 2619 | if (file_priv->event_space < sizeof e->event) { | ||
| 2620 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 2621 | goto out; | ||
| 2622 | } | ||
| 2623 | file_priv->event_space -= sizeof e->event; | ||
| 2624 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 2625 | |||
| 2626 | e = kzalloc(sizeof *e, GFP_KERNEL); | ||
| 2627 | if (e == NULL) { | ||
| 2628 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 2629 | file_priv->event_space += sizeof e->event; | ||
| 2630 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 2631 | goto out; | ||
| 2632 | } | ||
| 2633 | |||
| 2634 | e->event.base.type = DRM_EVENT_FLIP_COMPLETE; | ||
| 2635 | e->event.base.length = sizeof e->event; | ||
| 2636 | e->event.user_data = page_flip->user_data; | ||
| 2637 | e->base.event = &e->event.base; | ||
| 2638 | e->base.file_priv = file_priv; | ||
| 2639 | e->base.destroy = | ||
| 2640 | (void (*) (struct drm_pending_event *)) kfree; | ||
| 2641 | } | ||
| 2642 | |||
| 2643 | ret = crtc->funcs->page_flip(crtc, fb, e); | ||
| 2644 | if (ret) { | ||
| 2645 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 2646 | file_priv->event_space += sizeof e->event; | ||
| 2647 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 2648 | kfree(e); | ||
| 2649 | } | ||
| 2650 | |||
| 2651 | out: | ||
| 2652 | mutex_unlock(&dev->mode_config.mutex); | ||
| 2653 | return ret; | ||
| 2654 | } | ||
diff --git a/drivers/gpu/drm/drm_crtc_helper.c b/drivers/gpu/drm/drm_crtc_helper.c index bbfd110a7168..4231d6db72ec 100644 --- a/drivers/gpu/drm/drm_crtc_helper.c +++ b/drivers/gpu/drm/drm_crtc_helper.c | |||
| @@ -109,7 +109,7 @@ int drm_helper_probe_single_connector_modes(struct drm_connector *connector, | |||
| 109 | 109 | ||
| 110 | count = (*connector_funcs->get_modes)(connector); | 110 | count = (*connector_funcs->get_modes)(connector); |
| 111 | if (!count) { | 111 | if (!count) { |
| 112 | count = drm_add_modes_noedid(connector, 800, 600); | 112 | count = drm_add_modes_noedid(connector, 1024, 768); |
| 113 | if (!count) | 113 | if (!count) |
| 114 | return 0; | 114 | return 0; |
| 115 | } | 115 | } |
| @@ -1020,6 +1020,9 @@ bool drm_helper_initial_config(struct drm_device *dev) | |||
| 1020 | { | 1020 | { |
| 1021 | int count = 0; | 1021 | int count = 0; |
| 1022 | 1022 | ||
| 1023 | /* disable all the possible outputs/crtcs before entering KMS mode */ | ||
| 1024 | drm_helper_disable_unused_functions(dev); | ||
| 1025 | |||
| 1023 | drm_fb_helper_parse_command_line(dev); | 1026 | drm_fb_helper_parse_command_line(dev); |
| 1024 | 1027 | ||
| 1025 | count = drm_helper_probe_connector_modes(dev, | 1028 | count = drm_helper_probe_connector_modes(dev, |
diff --git a/drivers/gpu/drm/i915/intel_dp_i2c.c b/drivers/gpu/drm/drm_dp_i2c_helper.c index a63b6f57d2d4..548887c8506f 100644 --- a/drivers/gpu/drm/i915/intel_dp_i2c.c +++ b/drivers/gpu/drm/drm_dp_i2c_helper.c | |||
| @@ -28,84 +28,20 @@ | |||
| 28 | #include <linux/errno.h> | 28 | #include <linux/errno.h> |
| 29 | #include <linux/sched.h> | 29 | #include <linux/sched.h> |
| 30 | #include <linux/i2c.h> | 30 | #include <linux/i2c.h> |
| 31 | #include "intel_dp.h" | 31 | #include "drm_dp_helper.h" |
| 32 | #include "drmP.h" | 32 | #include "drmP.h" |
| 33 | 33 | ||
| 34 | /* Run a single AUX_CH I2C transaction, writing/reading data as necessary */ | 34 | /* Run a single AUX_CH I2C transaction, writing/reading data as necessary */ |
| 35 | |||
| 36 | #define MODE_I2C_START 1 | ||
| 37 | #define MODE_I2C_WRITE 2 | ||
| 38 | #define MODE_I2C_READ 4 | ||
| 39 | #define MODE_I2C_STOP 8 | ||
| 40 | |||
| 41 | static int | 35 | static int |
| 42 | i2c_algo_dp_aux_transaction(struct i2c_adapter *adapter, int mode, | 36 | i2c_algo_dp_aux_transaction(struct i2c_adapter *adapter, int mode, |
| 43 | uint8_t write_byte, uint8_t *read_byte) | 37 | uint8_t write_byte, uint8_t *read_byte) |
| 44 | { | 38 | { |
| 45 | struct i2c_algo_dp_aux_data *algo_data = adapter->algo_data; | 39 | struct i2c_algo_dp_aux_data *algo_data = adapter->algo_data; |
| 46 | uint16_t address = algo_data->address; | ||
| 47 | uint8_t msg[5]; | ||
| 48 | uint8_t reply[2]; | ||
| 49 | int msg_bytes; | ||
| 50 | int reply_bytes; | ||
| 51 | int ret; | 40 | int ret; |
| 52 | 41 | ||
| 53 | /* Set up the command byte */ | 42 | ret = (*algo_data->aux_ch)(adapter, mode, |
| 54 | if (mode & MODE_I2C_READ) | 43 | write_byte, read_byte); |
| 55 | msg[0] = AUX_I2C_READ << 4; | 44 | return ret; |
| 56 | else | ||
| 57 | msg[0] = AUX_I2C_WRITE << 4; | ||
| 58 | |||
| 59 | if (!(mode & MODE_I2C_STOP)) | ||
| 60 | msg[0] |= AUX_I2C_MOT << 4; | ||
| 61 | |||
| 62 | msg[1] = address >> 8; | ||
| 63 | msg[2] = address; | ||
| 64 | |||
| 65 | switch (mode) { | ||
| 66 | case MODE_I2C_WRITE: | ||
| 67 | msg[3] = 0; | ||
| 68 | msg[4] = write_byte; | ||
| 69 | msg_bytes = 5; | ||
| 70 | reply_bytes = 1; | ||
| 71 | break; | ||
| 72 | case MODE_I2C_READ: | ||
| 73 | msg[3] = 0; | ||
| 74 | msg_bytes = 4; | ||
| 75 | reply_bytes = 2; | ||
| 76 | break; | ||
| 77 | default: | ||
| 78 | msg_bytes = 3; | ||
| 79 | reply_bytes = 1; | ||
| 80 | break; | ||
| 81 | } | ||
| 82 | |||
| 83 | for (;;) { | ||
| 84 | ret = (*algo_data->aux_ch)(adapter, | ||
| 85 | msg, msg_bytes, | ||
| 86 | reply, reply_bytes); | ||
| 87 | if (ret < 0) { | ||
| 88 | DRM_DEBUG("aux_ch failed %d\n", ret); | ||
| 89 | return ret; | ||
| 90 | } | ||
| 91 | switch (reply[0] & AUX_I2C_REPLY_MASK) { | ||
| 92 | case AUX_I2C_REPLY_ACK: | ||
| 93 | if (mode == MODE_I2C_READ) { | ||
| 94 | *read_byte = reply[1]; | ||
| 95 | } | ||
| 96 | return reply_bytes - 1; | ||
| 97 | case AUX_I2C_REPLY_NACK: | ||
| 98 | DRM_DEBUG("aux_ch nack\n"); | ||
| 99 | return -EREMOTEIO; | ||
| 100 | case AUX_I2C_REPLY_DEFER: | ||
| 101 | DRM_DEBUG("aux_ch defer\n"); | ||
| 102 | udelay(100); | ||
| 103 | break; | ||
| 104 | default: | ||
| 105 | DRM_ERROR("aux_ch invalid reply 0x%02x\n", reply[0]); | ||
| 106 | return -EREMOTEIO; | ||
| 107 | } | ||
| 108 | } | ||
| 109 | } | 45 | } |
| 110 | 46 | ||
| 111 | /* | 47 | /* |
| @@ -224,7 +160,7 @@ i2c_algo_dp_aux_xfer(struct i2c_adapter *adapter, | |||
| 224 | if (ret >= 0) | 160 | if (ret >= 0) |
| 225 | ret = num; | 161 | ret = num; |
| 226 | i2c_algo_dp_aux_stop(adapter, reading); | 162 | i2c_algo_dp_aux_stop(adapter, reading); |
| 227 | DRM_DEBUG("dp_aux_xfer return %d\n", ret); | 163 | DRM_DEBUG_KMS("dp_aux_xfer return %d\n", ret); |
| 228 | return ret; | 164 | return ret; |
| 229 | } | 165 | } |
| 230 | 166 | ||
diff --git a/drivers/gpu/drm/drm_drv.c b/drivers/gpu/drm/drm_drv.c index a75ca63deea6..ff2f1042cb44 100644 --- a/drivers/gpu/drm/drm_drv.c +++ b/drivers/gpu/drm/drm_drv.c | |||
| @@ -145,6 +145,8 @@ static struct drm_ioctl_desc drm_ioctls[] = { | |||
| 145 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_GETFB, drm_mode_getfb, DRM_MASTER|DRM_CONTROL_ALLOW), | 145 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_GETFB, drm_mode_getfb, DRM_MASTER|DRM_CONTROL_ALLOW), |
| 146 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_ADDFB, drm_mode_addfb, DRM_MASTER|DRM_CONTROL_ALLOW), | 146 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_ADDFB, drm_mode_addfb, DRM_MASTER|DRM_CONTROL_ALLOW), |
| 147 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_RMFB, drm_mode_rmfb, DRM_MASTER|DRM_CONTROL_ALLOW), | 147 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_RMFB, drm_mode_rmfb, DRM_MASTER|DRM_CONTROL_ALLOW), |
| 148 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_PAGE_FLIP, drm_mode_page_flip_ioctl, DRM_MASTER|DRM_CONTROL_ALLOW), | ||
| 149 | DRM_IOCTL_DEF(DRM_IOCTL_MODE_DIRTYFB, drm_mode_dirtyfb_ioctl, DRM_MASTER|DRM_CONTROL_ALLOW) | ||
| 148 | }; | 150 | }; |
| 149 | 151 | ||
| 150 | #define DRM_CORE_IOCTL_COUNT ARRAY_SIZE( drm_ioctls ) | 152 | #define DRM_CORE_IOCTL_COUNT ARRAY_SIZE( drm_ioctls ) |
| @@ -366,6 +368,29 @@ module_init(drm_core_init); | |||
| 366 | module_exit(drm_core_exit); | 368 | module_exit(drm_core_exit); |
| 367 | 369 | ||
| 368 | /** | 370 | /** |
| 371 | * Copy and IOCTL return string to user space | ||
| 372 | */ | ||
| 373 | static int drm_copy_field(char *buf, size_t *buf_len, const char *value) | ||
| 374 | { | ||
| 375 | int len; | ||
| 376 | |||
| 377 | /* don't overflow userbuf */ | ||
| 378 | len = strlen(value); | ||
| 379 | if (len > *buf_len) | ||
| 380 | len = *buf_len; | ||
| 381 | |||
| 382 | /* let userspace know exact length of driver value (which could be | ||
| 383 | * larger than the userspace-supplied buffer) */ | ||
| 384 | *buf_len = strlen(value); | ||
| 385 | |||
| 386 | /* finally, try filling in the userbuf */ | ||
| 387 | if (len && buf) | ||
| 388 | if (copy_to_user(buf, value, len)) | ||
| 389 | return -EFAULT; | ||
| 390 | return 0; | ||
| 391 | } | ||
| 392 | |||
| 393 | /** | ||
| 369 | * Get version information | 394 | * Get version information |
| 370 | * | 395 | * |
| 371 | * \param inode device inode. | 396 | * \param inode device inode. |
| @@ -380,16 +405,21 @@ static int drm_version(struct drm_device *dev, void *data, | |||
| 380 | struct drm_file *file_priv) | 405 | struct drm_file *file_priv) |
| 381 | { | 406 | { |
| 382 | struct drm_version *version = data; | 407 | struct drm_version *version = data; |
| 383 | int len; | 408 | int err; |
| 384 | 409 | ||
| 385 | version->version_major = dev->driver->major; | 410 | version->version_major = dev->driver->major; |
| 386 | version->version_minor = dev->driver->minor; | 411 | version->version_minor = dev->driver->minor; |
| 387 | version->version_patchlevel = dev->driver->patchlevel; | 412 | version->version_patchlevel = dev->driver->patchlevel; |
| 388 | DRM_COPY(version->name, dev->driver->name); | 413 | err = drm_copy_field(version->name, &version->name_len, |
| 389 | DRM_COPY(version->date, dev->driver->date); | 414 | dev->driver->name); |
| 390 | DRM_COPY(version->desc, dev->driver->desc); | 415 | if (!err) |
| 391 | 416 | err = drm_copy_field(version->date, &version->date_len, | |
| 392 | return 0; | 417 | dev->driver->date); |
| 418 | if (!err) | ||
| 419 | err = drm_copy_field(version->desc, &version->desc_len, | ||
| 420 | dev->driver->desc); | ||
| 421 | |||
| 422 | return err; | ||
| 393 | } | 423 | } |
| 394 | 424 | ||
| 395 | /** | 425 | /** |
diff --git a/drivers/gpu/drm/drm_edid.c b/drivers/gpu/drm/drm_edid.c index b54ba63d506e..c39b26f1abed 100644 --- a/drivers/gpu/drm/drm_edid.c +++ b/drivers/gpu/drm/drm_edid.c | |||
| @@ -123,18 +123,20 @@ static const u8 edid_header[] = { | |||
| 123 | */ | 123 | */ |
| 124 | static bool edid_is_valid(struct edid *edid) | 124 | static bool edid_is_valid(struct edid *edid) |
| 125 | { | 125 | { |
| 126 | int i; | 126 | int i, score = 0; |
| 127 | u8 csum = 0; | 127 | u8 csum = 0; |
| 128 | u8 *raw_edid = (u8 *)edid; | 128 | u8 *raw_edid = (u8 *)edid; |
| 129 | 129 | ||
| 130 | if (memcmp(edid->header, edid_header, sizeof(edid_header))) | 130 | for (i = 0; i < sizeof(edid_header); i++) |
| 131 | goto bad; | 131 | if (raw_edid[i] == edid_header[i]) |
| 132 | if (edid->version != 1) { | 132 | score++; |
| 133 | DRM_ERROR("EDID has major version %d, instead of 1\n", edid->version); | 133 | |
| 134 | if (score == 8) ; | ||
| 135 | else if (score >= 6) { | ||
| 136 | DRM_DEBUG("Fixing EDID header, your hardware may be failing\n"); | ||
| 137 | memcpy(raw_edid, edid_header, sizeof(edid_header)); | ||
| 138 | } else | ||
| 134 | goto bad; | 139 | goto bad; |
| 135 | } | ||
| 136 | if (edid->revision > 4) | ||
| 137 | DRM_DEBUG("EDID minor > 4, assuming backward compatibility\n"); | ||
| 138 | 140 | ||
| 139 | for (i = 0; i < EDID_LENGTH; i++) | 141 | for (i = 0; i < EDID_LENGTH; i++) |
| 140 | csum += raw_edid[i]; | 142 | csum += raw_edid[i]; |
| @@ -143,6 +145,14 @@ static bool edid_is_valid(struct edid *edid) | |||
| 143 | goto bad; | 145 | goto bad; |
| 144 | } | 146 | } |
| 145 | 147 | ||
| 148 | if (edid->version != 1) { | ||
| 149 | DRM_ERROR("EDID has major version %d, instead of 1\n", edid->version); | ||
| 150 | goto bad; | ||
| 151 | } | ||
| 152 | |||
| 153 | if (edid->revision > 4) | ||
| 154 | DRM_DEBUG("EDID minor > 4, assuming backward compatibility\n"); | ||
| 155 | |||
| 146 | return 1; | 156 | return 1; |
| 147 | 157 | ||
| 148 | bad: | 158 | bad: |
| @@ -481,16 +491,17 @@ static struct drm_display_mode drm_dmt_modes[] = { | |||
| 481 | 3048, 3536, 0, 1600, 1603, 1609, 1682, 0, | 491 | 3048, 3536, 0, 1600, 1603, 1609, 1682, 0, |
| 482 | DRM_MODE_FLAG_NHSYNC | DRM_MODE_FLAG_PVSYNC) }, | 492 | DRM_MODE_FLAG_NHSYNC | DRM_MODE_FLAG_PVSYNC) }, |
| 483 | }; | 493 | }; |
| 494 | static const int drm_num_dmt_modes = | ||
| 495 | sizeof(drm_dmt_modes) / sizeof(struct drm_display_mode); | ||
| 484 | 496 | ||
| 485 | static struct drm_display_mode *drm_find_dmt(struct drm_device *dev, | 497 | static struct drm_display_mode *drm_find_dmt(struct drm_device *dev, |
| 486 | int hsize, int vsize, int fresh) | 498 | int hsize, int vsize, int fresh) |
| 487 | { | 499 | { |
| 488 | int i, count; | 500 | int i; |
| 489 | struct drm_display_mode *ptr, *mode; | 501 | struct drm_display_mode *ptr, *mode; |
| 490 | 502 | ||
| 491 | count = sizeof(drm_dmt_modes) / sizeof(struct drm_display_mode); | ||
| 492 | mode = NULL; | 503 | mode = NULL; |
| 493 | for (i = 0; i < count; i++) { | 504 | for (i = 0; i < drm_num_dmt_modes; i++) { |
| 494 | ptr = &drm_dmt_modes[i]; | 505 | ptr = &drm_dmt_modes[i]; |
| 495 | if (hsize == ptr->hdisplay && | 506 | if (hsize == ptr->hdisplay && |
| 496 | vsize == ptr->vdisplay && | 507 | vsize == ptr->vdisplay && |
| @@ -834,8 +845,165 @@ static int add_standard_modes(struct drm_connector *connector, struct edid *edid | |||
| 834 | return modes; | 845 | return modes; |
| 835 | } | 846 | } |
| 836 | 847 | ||
| 848 | /* | ||
| 849 | * XXX fix this for: | ||
| 850 | * - GTF secondary curve formula | ||
| 851 | * - EDID 1.4 range offsets | ||
| 852 | * - CVT extended bits | ||
| 853 | */ | ||
| 854 | static bool | ||
| 855 | mode_in_range(struct drm_display_mode *mode, struct detailed_timing *timing) | ||
| 856 | { | ||
| 857 | struct detailed_data_monitor_range *range; | ||
| 858 | int hsync, vrefresh; | ||
| 859 | |||
| 860 | range = &timing->data.other_data.data.range; | ||
| 861 | |||
| 862 | hsync = drm_mode_hsync(mode); | ||
| 863 | vrefresh = drm_mode_vrefresh(mode); | ||
| 864 | |||
| 865 | if (hsync < range->min_hfreq_khz || hsync > range->max_hfreq_khz) | ||
| 866 | return false; | ||
| 867 | |||
| 868 | if (vrefresh < range->min_vfreq || vrefresh > range->max_vfreq) | ||
| 869 | return false; | ||
| 870 | |||
| 871 | if (range->pixel_clock_mhz && range->pixel_clock_mhz != 0xff) { | ||
| 872 | /* be forgiving since it's in units of 10MHz */ | ||
| 873 | int max_clock = range->pixel_clock_mhz * 10 + 9; | ||
| 874 | max_clock *= 1000; | ||
| 875 | if (mode->clock > max_clock) | ||
| 876 | return false; | ||
| 877 | } | ||
| 878 | |||
| 879 | return true; | ||
| 880 | } | ||
| 881 | |||
| 882 | /* | ||
| 883 | * XXX If drm_dmt_modes ever regrows the CVT-R modes (and it will) this will | ||
| 884 | * need to account for them. | ||
| 885 | */ | ||
| 886 | static int drm_gtf_modes_for_range(struct drm_connector *connector, | ||
| 887 | struct detailed_timing *timing) | ||
| 888 | { | ||
| 889 | int i, modes = 0; | ||
| 890 | struct drm_display_mode *newmode; | ||
| 891 | struct drm_device *dev = connector->dev; | ||
| 892 | |||
| 893 | for (i = 0; i < drm_num_dmt_modes; i++) { | ||
| 894 | if (mode_in_range(drm_dmt_modes + i, timing)) { | ||
| 895 | newmode = drm_mode_duplicate(dev, &drm_dmt_modes[i]); | ||
| 896 | if (newmode) { | ||
| 897 | drm_mode_probed_add(connector, newmode); | ||
| 898 | modes++; | ||
| 899 | } | ||
| 900 | } | ||
| 901 | } | ||
| 902 | |||
| 903 | return modes; | ||
| 904 | } | ||
| 905 | |||
| 906 | static int drm_cvt_modes(struct drm_connector *connector, | ||
| 907 | struct detailed_timing *timing) | ||
| 908 | { | ||
| 909 | int i, j, modes = 0; | ||
| 910 | struct drm_display_mode *newmode; | ||
| 911 | struct drm_device *dev = connector->dev; | ||
| 912 | struct cvt_timing *cvt; | ||
| 913 | const int rates[] = { 60, 85, 75, 60, 50 }; | ||
| 914 | |||
| 915 | for (i = 0; i < 4; i++) { | ||
| 916 | int width, height; | ||
| 917 | cvt = &(timing->data.other_data.data.cvt[i]); | ||
| 918 | |||
| 919 | height = (cvt->code[0] + ((cvt->code[1] & 0xf0) << 8) + 1) * 2; | ||
| 920 | switch (cvt->code[1] & 0xc0) { | ||
| 921 | case 0x00: | ||
| 922 | width = height * 4 / 3; | ||
| 923 | break; | ||
| 924 | case 0x40: | ||
| 925 | width = height * 16 / 9; | ||
| 926 | break; | ||
| 927 | case 0x80: | ||
| 928 | width = height * 16 / 10; | ||
| 929 | break; | ||
| 930 | case 0xc0: | ||
| 931 | width = height * 15 / 9; | ||
| 932 | break; | ||
| 933 | } | ||
| 934 | |||
| 935 | for (j = 1; j < 5; j++) { | ||
| 936 | if (cvt->code[2] & (1 << j)) { | ||
| 937 | newmode = drm_cvt_mode(dev, width, height, | ||
| 938 | rates[j], j == 0, | ||
| 939 | false, false); | ||
| 940 | if (newmode) { | ||
| 941 | drm_mode_probed_add(connector, newmode); | ||
| 942 | modes++; | ||
| 943 | } | ||
| 944 | } | ||
| 945 | } | ||
| 946 | } | ||
| 947 | |||
| 948 | return modes; | ||
| 949 | } | ||
| 950 | |||
| 951 | static int add_detailed_modes(struct drm_connector *connector, | ||
| 952 | struct detailed_timing *timing, | ||
| 953 | struct edid *edid, u32 quirks, int preferred) | ||
| 954 | { | ||
| 955 | int i, modes = 0; | ||
| 956 | struct detailed_non_pixel *data = &timing->data.other_data; | ||
| 957 | int timing_level = standard_timing_level(edid); | ||
| 958 | int gtf = (edid->features & DRM_EDID_FEATURE_DEFAULT_GTF); | ||
| 959 | struct drm_display_mode *newmode; | ||
| 960 | struct drm_device *dev = connector->dev; | ||
| 961 | |||
| 962 | if (timing->pixel_clock) { | ||
| 963 | newmode = drm_mode_detailed(dev, edid, timing, quirks); | ||
| 964 | if (!newmode) | ||
| 965 | return 0; | ||
| 966 | |||
| 967 | if (preferred) | ||
| 968 | newmode->type |= DRM_MODE_TYPE_PREFERRED; | ||
| 969 | |||
| 970 | drm_mode_probed_add(connector, newmode); | ||
| 971 | return 1; | ||
| 972 | } | ||
| 973 | |||
| 974 | /* other timing types */ | ||
| 975 | switch (data->type) { | ||
| 976 | case EDID_DETAIL_MONITOR_RANGE: | ||
| 977 | if (gtf) | ||
| 978 | modes += drm_gtf_modes_for_range(connector, timing); | ||
| 979 | break; | ||
| 980 | case EDID_DETAIL_STD_MODES: | ||
| 981 | /* Six modes per detailed section */ | ||
| 982 | for (i = 0; i < 6; i++) { | ||
| 983 | struct std_timing *std; | ||
| 984 | struct drm_display_mode *newmode; | ||
| 985 | |||
| 986 | std = &data->data.timings[i]; | ||
| 987 | newmode = drm_mode_std(dev, std, edid->revision, | ||
| 988 | timing_level); | ||
| 989 | if (newmode) { | ||
| 990 | drm_mode_probed_add(connector, newmode); | ||
| 991 | modes++; | ||
| 992 | } | ||
| 993 | } | ||
| 994 | break; | ||
| 995 | case EDID_DETAIL_CVT_3BYTE: | ||
| 996 | modes += drm_cvt_modes(connector, timing); | ||
| 997 | break; | ||
| 998 | default: | ||
| 999 | break; | ||
| 1000 | } | ||
| 1001 | |||
| 1002 | return modes; | ||
| 1003 | } | ||
| 1004 | |||
| 837 | /** | 1005 | /** |
| 838 | * add_detailed_modes - get detailed mode info from EDID data | 1006 | * add_detailed_info - get detailed mode info from EDID data |
| 839 | * @connector: attached connector | 1007 | * @connector: attached connector |
| 840 | * @edid: EDID block to scan | 1008 | * @edid: EDID block to scan |
| 841 | * @quirks: quirks to apply | 1009 | * @quirks: quirks to apply |
| @@ -846,67 +1014,24 @@ static int add_standard_modes(struct drm_connector *connector, struct edid *edid | |||
| 846 | static int add_detailed_info(struct drm_connector *connector, | 1014 | static int add_detailed_info(struct drm_connector *connector, |
| 847 | struct edid *edid, u32 quirks) | 1015 | struct edid *edid, u32 quirks) |
| 848 | { | 1016 | { |
| 849 | struct drm_device *dev = connector->dev; | 1017 | int i, modes = 0; |
| 850 | int i, j, modes = 0; | ||
| 851 | int timing_level; | ||
| 852 | |||
| 853 | timing_level = standard_timing_level(edid); | ||
| 854 | 1018 | ||
| 855 | for (i = 0; i < EDID_DETAILED_TIMINGS; i++) { | 1019 | for (i = 0; i < EDID_DETAILED_TIMINGS; i++) { |
| 856 | struct detailed_timing *timing = &edid->detailed_timings[i]; | 1020 | struct detailed_timing *timing = &edid->detailed_timings[i]; |
| 857 | struct detailed_non_pixel *data = &timing->data.other_data; | 1021 | int preferred = (i == 0) && (edid->features & DRM_EDID_FEATURE_PREFERRED_TIMING); |
| 858 | struct drm_display_mode *newmode; | ||
| 859 | |||
| 860 | /* X server check is version 1.1 or higher */ | ||
| 861 | if (edid->version == 1 && edid->revision >= 1 && | ||
| 862 | !timing->pixel_clock) { | ||
| 863 | /* Other timing or info */ | ||
| 864 | switch (data->type) { | ||
| 865 | case EDID_DETAIL_MONITOR_SERIAL: | ||
| 866 | break; | ||
| 867 | case EDID_DETAIL_MONITOR_STRING: | ||
| 868 | break; | ||
| 869 | case EDID_DETAIL_MONITOR_RANGE: | ||
| 870 | /* Get monitor range data */ | ||
| 871 | break; | ||
| 872 | case EDID_DETAIL_MONITOR_NAME: | ||
| 873 | break; | ||
| 874 | case EDID_DETAIL_MONITOR_CPDATA: | ||
| 875 | break; | ||
| 876 | case EDID_DETAIL_STD_MODES: | ||
| 877 | for (j = 0; j < 6; i++) { | ||
| 878 | struct std_timing *std; | ||
| 879 | struct drm_display_mode *newmode; | ||
| 880 | |||
| 881 | std = &data->data.timings[j]; | ||
| 882 | newmode = drm_mode_std(dev, std, | ||
| 883 | edid->revision, | ||
| 884 | timing_level); | ||
| 885 | if (newmode) { | ||
| 886 | drm_mode_probed_add(connector, newmode); | ||
| 887 | modes++; | ||
| 888 | } | ||
| 889 | } | ||
| 890 | break; | ||
| 891 | default: | ||
| 892 | break; | ||
| 893 | } | ||
| 894 | } else { | ||
| 895 | newmode = drm_mode_detailed(dev, edid, timing, quirks); | ||
| 896 | if (!newmode) | ||
| 897 | continue; | ||
| 898 | 1022 | ||
| 899 | /* First detailed mode is preferred */ | 1023 | /* In 1.0, only timings are allowed */ |
| 900 | if (i == 0 && (edid->features & DRM_EDID_FEATURE_PREFERRED_TIMING)) | 1024 | if (!timing->pixel_clock && edid->version == 1 && |
| 901 | newmode->type |= DRM_MODE_TYPE_PREFERRED; | 1025 | edid->revision == 0) |
| 902 | drm_mode_probed_add(connector, newmode); | 1026 | continue; |
| 903 | 1027 | ||
| 904 | modes++; | 1028 | modes += add_detailed_modes(connector, timing, edid, quirks, |
| 905 | } | 1029 | preferred); |
| 906 | } | 1030 | } |
| 907 | 1031 | ||
| 908 | return modes; | 1032 | return modes; |
| 909 | } | 1033 | } |
| 1034 | |||
| 910 | /** | 1035 | /** |
| 911 | * add_detailed_mode_eedid - get detailed mode info from addtional timing | 1036 | * add_detailed_mode_eedid - get detailed mode info from addtional timing |
| 912 | * EDID block | 1037 | * EDID block |
| @@ -920,12 +1045,9 @@ static int add_detailed_info(struct drm_connector *connector, | |||
| 920 | static int add_detailed_info_eedid(struct drm_connector *connector, | 1045 | static int add_detailed_info_eedid(struct drm_connector *connector, |
| 921 | struct edid *edid, u32 quirks) | 1046 | struct edid *edid, u32 quirks) |
| 922 | { | 1047 | { |
| 923 | struct drm_device *dev = connector->dev; | 1048 | int i, modes = 0; |
| 924 | int i, j, modes = 0; | ||
| 925 | char *edid_ext = NULL; | 1049 | char *edid_ext = NULL; |
| 926 | struct detailed_timing *timing; | 1050 | struct detailed_timing *timing; |
| 927 | struct detailed_non_pixel *data; | ||
| 928 | struct drm_display_mode *newmode; | ||
| 929 | int edid_ext_num; | 1051 | int edid_ext_num; |
| 930 | int start_offset, end_offset; | 1052 | int start_offset, end_offset; |
| 931 | int timing_level; | 1053 | int timing_level; |
| @@ -976,51 +1098,7 @@ static int add_detailed_info_eedid(struct drm_connector *connector, | |||
| 976 | for (i = start_offset; i < end_offset; | 1098 | for (i = start_offset; i < end_offset; |
| 977 | i += sizeof(struct detailed_timing)) { | 1099 | i += sizeof(struct detailed_timing)) { |
| 978 | timing = (struct detailed_timing *)(edid_ext + i); | 1100 | timing = (struct detailed_timing *)(edid_ext + i); |
| 979 | data = &timing->data.other_data; | 1101 | modes += add_detailed_modes(connector, timing, edid, quirks, 0); |
| 980 | /* Detailed mode timing */ | ||
| 981 | if (timing->pixel_clock) { | ||
| 982 | newmode = drm_mode_detailed(dev, edid, timing, quirks); | ||
| 983 | if (!newmode) | ||
| 984 | continue; | ||
| 985 | |||
| 986 | drm_mode_probed_add(connector, newmode); | ||
| 987 | |||
| 988 | modes++; | ||
| 989 | continue; | ||
| 990 | } | ||
| 991 | |||
| 992 | /* Other timing or info */ | ||
| 993 | switch (data->type) { | ||
| 994 | case EDID_DETAIL_MONITOR_SERIAL: | ||
| 995 | break; | ||
| 996 | case EDID_DETAIL_MONITOR_STRING: | ||
| 997 | break; | ||
| 998 | case EDID_DETAIL_MONITOR_RANGE: | ||
| 999 | /* Get monitor range data */ | ||
| 1000 | break; | ||
| 1001 | case EDID_DETAIL_MONITOR_NAME: | ||
| 1002 | break; | ||
| 1003 | case EDID_DETAIL_MONITOR_CPDATA: | ||
| 1004 | break; | ||
| 1005 | case EDID_DETAIL_STD_MODES: | ||
| 1006 | /* Five modes per detailed section */ | ||
| 1007 | for (j = 0; j < 5; i++) { | ||
| 1008 | struct std_timing *std; | ||
| 1009 | struct drm_display_mode *newmode; | ||
| 1010 | |||
| 1011 | std = &data->data.timings[j]; | ||
| 1012 | newmode = drm_mode_std(dev, std, | ||
| 1013 | edid->revision, | ||
| 1014 | timing_level); | ||
| 1015 | if (newmode) { | ||
| 1016 | drm_mode_probed_add(connector, newmode); | ||
| 1017 | modes++; | ||
| 1018 | } | ||
| 1019 | } | ||
| 1020 | break; | ||
| 1021 | default: | ||
| 1022 | break; | ||
| 1023 | } | ||
| 1024 | } | 1102 | } |
| 1025 | 1103 | ||
| 1026 | return modes; | 1104 | return modes; |
| @@ -1066,19 +1144,19 @@ static int drm_ddc_read_edid(struct drm_connector *connector, | |||
| 1066 | struct i2c_adapter *adapter, | 1144 | struct i2c_adapter *adapter, |
| 1067 | char *buf, int len) | 1145 | char *buf, int len) |
| 1068 | { | 1146 | { |
| 1069 | int ret; | 1147 | int i; |
| 1070 | 1148 | ||
| 1071 | ret = drm_do_probe_ddc_edid(adapter, buf, len); | 1149 | for (i = 0; i < 4; i++) { |
| 1072 | if (ret != 0) { | 1150 | if (drm_do_probe_ddc_edid(adapter, buf, len)) |
| 1073 | goto end; | 1151 | return -1; |
| 1074 | } | 1152 | if (edid_is_valid((struct edid *)buf)) |
| 1075 | if (!edid_is_valid((struct edid *)buf)) { | 1153 | return 0; |
| 1076 | dev_warn(&connector->dev->pdev->dev, "%s: EDID invalid.\n", | ||
| 1077 | drm_get_connector_name(connector)); | ||
| 1078 | ret = -1; | ||
| 1079 | } | 1154 | } |
| 1080 | end: | 1155 | |
| 1081 | return ret; | 1156 | /* repeated checksum failures; warn, but carry on */ |
| 1157 | dev_warn(&connector->dev->pdev->dev, "%s: EDID invalid.\n", | ||
| 1158 | drm_get_connector_name(connector)); | ||
| 1159 | return -1; | ||
| 1082 | } | 1160 | } |
| 1083 | 1161 | ||
| 1084 | /** | 1162 | /** |
| @@ -1296,6 +1374,8 @@ int drm_add_modes_noedid(struct drm_connector *connector, | |||
| 1296 | ptr->vdisplay > vdisplay) | 1374 | ptr->vdisplay > vdisplay) |
| 1297 | continue; | 1375 | continue; |
| 1298 | } | 1376 | } |
| 1377 | if (drm_mode_vrefresh(ptr) > 61) | ||
| 1378 | continue; | ||
| 1299 | mode = drm_mode_duplicate(dev, ptr); | 1379 | mode = drm_mode_duplicate(dev, ptr); |
| 1300 | if (mode) { | 1380 | if (mode) { |
| 1301 | drm_mode_probed_add(connector, mode); | 1381 | drm_mode_probed_add(connector, mode); |
diff --git a/drivers/gpu/drm/drm_fb_helper.c b/drivers/gpu/drm/drm_fb_helper.c index 65ef011fa8ba..1b49fa055f4f 100644 --- a/drivers/gpu/drm/drm_fb_helper.c +++ b/drivers/gpu/drm/drm_fb_helper.c | |||
| @@ -373,11 +373,9 @@ static void drm_fb_helper_off(struct fb_info *info, int dpms_mode) | |||
| 373 | mutex_unlock(&dev->mode_config.mutex); | 373 | mutex_unlock(&dev->mode_config.mutex); |
| 374 | } | 374 | } |
| 375 | } | 375 | } |
| 376 | if (dpms_mode == DRM_MODE_DPMS_OFF) { | 376 | mutex_lock(&dev->mode_config.mutex); |
| 377 | mutex_lock(&dev->mode_config.mutex); | 377 | crtc_funcs->dpms(crtc, DRM_MODE_DPMS_OFF); |
| 378 | crtc_funcs->dpms(crtc, dpms_mode); | 378 | mutex_unlock(&dev->mode_config.mutex); |
| 379 | mutex_unlock(&dev->mode_config.mutex); | ||
| 380 | } | ||
| 381 | } | 379 | } |
| 382 | } | 380 | } |
| 383 | } | 381 | } |
| @@ -385,18 +383,23 @@ static void drm_fb_helper_off(struct fb_info *info, int dpms_mode) | |||
| 385 | int drm_fb_helper_blank(int blank, struct fb_info *info) | 383 | int drm_fb_helper_blank(int blank, struct fb_info *info) |
| 386 | { | 384 | { |
| 387 | switch (blank) { | 385 | switch (blank) { |
| 386 | /* Display: On; HSync: On, VSync: On */ | ||
| 388 | case FB_BLANK_UNBLANK: | 387 | case FB_BLANK_UNBLANK: |
| 389 | drm_fb_helper_on(info); | 388 | drm_fb_helper_on(info); |
| 390 | break; | 389 | break; |
| 390 | /* Display: Off; HSync: On, VSync: On */ | ||
| 391 | case FB_BLANK_NORMAL: | 391 | case FB_BLANK_NORMAL: |
| 392 | drm_fb_helper_off(info, DRM_MODE_DPMS_STANDBY); | 392 | drm_fb_helper_off(info, DRM_MODE_DPMS_ON); |
| 393 | break; | 393 | break; |
| 394 | /* Display: Off; HSync: Off, VSync: On */ | ||
| 394 | case FB_BLANK_HSYNC_SUSPEND: | 395 | case FB_BLANK_HSYNC_SUSPEND: |
| 395 | drm_fb_helper_off(info, DRM_MODE_DPMS_STANDBY); | 396 | drm_fb_helper_off(info, DRM_MODE_DPMS_STANDBY); |
| 396 | break; | 397 | break; |
| 398 | /* Display: Off; HSync: On, VSync: Off */ | ||
| 397 | case FB_BLANK_VSYNC_SUSPEND: | 399 | case FB_BLANK_VSYNC_SUSPEND: |
| 398 | drm_fb_helper_off(info, DRM_MODE_DPMS_SUSPEND); | 400 | drm_fb_helper_off(info, DRM_MODE_DPMS_SUSPEND); |
| 399 | break; | 401 | break; |
| 402 | /* Display: Off; HSync: Off, VSync: Off */ | ||
| 400 | case FB_BLANK_POWERDOWN: | 403 | case FB_BLANK_POWERDOWN: |
| 401 | drm_fb_helper_off(info, DRM_MODE_DPMS_OFF); | 404 | drm_fb_helper_off(info, DRM_MODE_DPMS_OFF); |
| 402 | break; | 405 | break; |
| @@ -905,8 +908,13 @@ int drm_fb_helper_single_fb_probe(struct drm_device *dev, | |||
| 905 | 908 | ||
| 906 | if (new_fb) { | 909 | if (new_fb) { |
| 907 | info->var.pixclock = 0; | 910 | info->var.pixclock = 0; |
| 908 | if (register_framebuffer(info) < 0) | 911 | ret = fb_alloc_cmap(&info->cmap, modeset->crtc->gamma_size, 0); |
| 912 | if (ret) | ||
| 913 | return ret; | ||
| 914 | if (register_framebuffer(info) < 0) { | ||
| 915 | fb_dealloc_cmap(&info->cmap); | ||
| 909 | return -EINVAL; | 916 | return -EINVAL; |
| 917 | } | ||
| 910 | } else { | 918 | } else { |
| 911 | drm_fb_helper_set_par(info); | 919 | drm_fb_helper_set_par(info); |
| 912 | } | 920 | } |
| @@ -936,6 +944,7 @@ void drm_fb_helper_free(struct drm_fb_helper *helper) | |||
| 936 | unregister_sysrq_key('v', &sysrq_drm_fb_helper_restore_op); | 944 | unregister_sysrq_key('v', &sysrq_drm_fb_helper_restore_op); |
| 937 | } | 945 | } |
| 938 | drm_fb_helper_crtc_free(helper); | 946 | drm_fb_helper_crtc_free(helper); |
| 947 | fb_dealloc_cmap(&helper->fb->fbdev->cmap); | ||
| 939 | } | 948 | } |
| 940 | EXPORT_SYMBOL(drm_fb_helper_free); | 949 | EXPORT_SYMBOL(drm_fb_helper_free); |
| 941 | 950 | ||
diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c index 251bc0e3b5ec..08d14df3bb42 100644 --- a/drivers/gpu/drm/drm_fops.c +++ b/drivers/gpu/drm/drm_fops.c | |||
| @@ -257,6 +257,9 @@ static int drm_open_helper(struct inode *inode, struct file *filp, | |||
| 257 | 257 | ||
| 258 | INIT_LIST_HEAD(&priv->lhead); | 258 | INIT_LIST_HEAD(&priv->lhead); |
| 259 | INIT_LIST_HEAD(&priv->fbs); | 259 | INIT_LIST_HEAD(&priv->fbs); |
| 260 | INIT_LIST_HEAD(&priv->event_list); | ||
| 261 | init_waitqueue_head(&priv->event_wait); | ||
| 262 | priv->event_space = 4096; /* set aside 4k for event buffer */ | ||
| 260 | 263 | ||
| 261 | if (dev->driver->driver_features & DRIVER_GEM) | 264 | if (dev->driver->driver_features & DRIVER_GEM) |
| 262 | drm_gem_open(dev, priv); | 265 | drm_gem_open(dev, priv); |
| @@ -297,6 +300,18 @@ static int drm_open_helper(struct inode *inode, struct file *filp, | |||
| 297 | goto out_free; | 300 | goto out_free; |
| 298 | } | 301 | } |
| 299 | } | 302 | } |
| 303 | mutex_lock(&dev->struct_mutex); | ||
| 304 | if (dev->driver->master_set) { | ||
| 305 | ret = dev->driver->master_set(dev, priv, true); | ||
| 306 | if (ret) { | ||
| 307 | /* drop both references if this fails */ | ||
| 308 | drm_master_put(&priv->minor->master); | ||
| 309 | drm_master_put(&priv->master); | ||
| 310 | mutex_unlock(&dev->struct_mutex); | ||
| 311 | goto out_free; | ||
| 312 | } | ||
| 313 | } | ||
| 314 | mutex_unlock(&dev->struct_mutex); | ||
| 300 | } else { | 315 | } else { |
| 301 | /* get a reference to the master */ | 316 | /* get a reference to the master */ |
| 302 | priv->master = drm_master_get(priv->minor->master); | 317 | priv->master = drm_master_get(priv->minor->master); |
| @@ -413,6 +428,30 @@ static void drm_master_release(struct drm_device *dev, struct file *filp) | |||
| 413 | } | 428 | } |
| 414 | } | 429 | } |
| 415 | 430 | ||
| 431 | static void drm_events_release(struct drm_file *file_priv) | ||
| 432 | { | ||
| 433 | struct drm_device *dev = file_priv->minor->dev; | ||
| 434 | struct drm_pending_event *e, *et; | ||
| 435 | struct drm_pending_vblank_event *v, *vt; | ||
| 436 | unsigned long flags; | ||
| 437 | |||
| 438 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 439 | |||
| 440 | /* Remove pending flips */ | ||
| 441 | list_for_each_entry_safe(v, vt, &dev->vblank_event_list, base.link) | ||
| 442 | if (v->base.file_priv == file_priv) { | ||
| 443 | list_del(&v->base.link); | ||
| 444 | drm_vblank_put(dev, v->pipe); | ||
| 445 | v->base.destroy(&v->base); | ||
| 446 | } | ||
| 447 | |||
| 448 | /* Remove unconsumed events */ | ||
| 449 | list_for_each_entry_safe(e, et, &file_priv->event_list, link) | ||
| 450 | e->destroy(e); | ||
| 451 | |||
| 452 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 453 | } | ||
| 454 | |||
| 416 | /** | 455 | /** |
| 417 | * Release file. | 456 | * Release file. |
| 418 | * | 457 | * |
| @@ -451,6 +490,8 @@ int drm_release(struct inode *inode, struct file *filp) | |||
| 451 | if (file_priv->minor->master) | 490 | if (file_priv->minor->master) |
| 452 | drm_master_release(dev, filp); | 491 | drm_master_release(dev, filp); |
| 453 | 492 | ||
| 493 | drm_events_release(file_priv); | ||
| 494 | |||
| 454 | if (dev->driver->driver_features & DRIVER_GEM) | 495 | if (dev->driver->driver_features & DRIVER_GEM) |
| 455 | drm_gem_release(dev, file_priv); | 496 | drm_gem_release(dev, file_priv); |
| 456 | 497 | ||
| @@ -504,6 +545,8 @@ int drm_release(struct inode *inode, struct file *filp) | |||
| 504 | 545 | ||
| 505 | if (file_priv->minor->master == file_priv->master) { | 546 | if (file_priv->minor->master == file_priv->master) { |
| 506 | /* drop the reference held my the minor */ | 547 | /* drop the reference held my the minor */ |
| 548 | if (dev->driver->master_drop) | ||
| 549 | dev->driver->master_drop(dev, file_priv, true); | ||
| 507 | drm_master_put(&file_priv->minor->master); | 550 | drm_master_put(&file_priv->minor->master); |
| 508 | } | 551 | } |
| 509 | } | 552 | } |
| @@ -544,9 +587,74 @@ int drm_release(struct inode *inode, struct file *filp) | |||
| 544 | } | 587 | } |
| 545 | EXPORT_SYMBOL(drm_release); | 588 | EXPORT_SYMBOL(drm_release); |
| 546 | 589 | ||
| 547 | /** No-op. */ | 590 | static bool |
| 591 | drm_dequeue_event(struct drm_file *file_priv, | ||
| 592 | size_t total, size_t max, struct drm_pending_event **out) | ||
| 593 | { | ||
| 594 | struct drm_device *dev = file_priv->minor->dev; | ||
| 595 | struct drm_pending_event *e; | ||
| 596 | unsigned long flags; | ||
| 597 | bool ret = false; | ||
| 598 | |||
| 599 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 600 | |||
| 601 | *out = NULL; | ||
| 602 | if (list_empty(&file_priv->event_list)) | ||
| 603 | goto out; | ||
| 604 | e = list_first_entry(&file_priv->event_list, | ||
| 605 | struct drm_pending_event, link); | ||
| 606 | if (e->event->length + total > max) | ||
| 607 | goto out; | ||
| 608 | |||
| 609 | file_priv->event_space += e->event->length; | ||
| 610 | list_del(&e->link); | ||
| 611 | *out = e; | ||
| 612 | ret = true; | ||
| 613 | |||
| 614 | out: | ||
| 615 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 616 | return ret; | ||
| 617 | } | ||
| 618 | |||
| 619 | ssize_t drm_read(struct file *filp, char __user *buffer, | ||
| 620 | size_t count, loff_t *offset) | ||
| 621 | { | ||
| 622 | struct drm_file *file_priv = filp->private_data; | ||
| 623 | struct drm_pending_event *e; | ||
| 624 | size_t total; | ||
| 625 | ssize_t ret; | ||
| 626 | |||
| 627 | ret = wait_event_interruptible(file_priv->event_wait, | ||
| 628 | !list_empty(&file_priv->event_list)); | ||
| 629 | if (ret < 0) | ||
| 630 | return ret; | ||
| 631 | |||
| 632 | total = 0; | ||
| 633 | while (drm_dequeue_event(file_priv, total, count, &e)) { | ||
| 634 | if (copy_to_user(buffer + total, | ||
| 635 | e->event, e->event->length)) { | ||
| 636 | total = -EFAULT; | ||
| 637 | break; | ||
| 638 | } | ||
| 639 | |||
| 640 | total += e->event->length; | ||
| 641 | e->destroy(e); | ||
| 642 | } | ||
| 643 | |||
| 644 | return total; | ||
| 645 | } | ||
| 646 | EXPORT_SYMBOL(drm_read); | ||
| 647 | |||
| 548 | unsigned int drm_poll(struct file *filp, struct poll_table_struct *wait) | 648 | unsigned int drm_poll(struct file *filp, struct poll_table_struct *wait) |
| 549 | { | 649 | { |
| 550 | return 0; | 650 | struct drm_file *file_priv = filp->private_data; |
| 651 | unsigned int mask = 0; | ||
| 652 | |||
| 653 | poll_wait(filp, &file_priv->event_wait, wait); | ||
| 654 | |||
| 655 | if (!list_empty(&file_priv->event_list)) | ||
| 656 | mask |= POLLIN | POLLRDNORM; | ||
| 657 | |||
| 658 | return mask; | ||
| 551 | } | 659 | } |
| 552 | EXPORT_SYMBOL(drm_poll); | 660 | EXPORT_SYMBOL(drm_poll); |
diff --git a/drivers/gpu/drm/drm_irq.c b/drivers/gpu/drm/drm_irq.c index 0a6f0b3bdc78..7998ee66b317 100644 --- a/drivers/gpu/drm/drm_irq.c +++ b/drivers/gpu/drm/drm_irq.c | |||
| @@ -429,15 +429,21 @@ int drm_vblank_get(struct drm_device *dev, int crtc) | |||
| 429 | 429 | ||
| 430 | spin_lock_irqsave(&dev->vbl_lock, irqflags); | 430 | spin_lock_irqsave(&dev->vbl_lock, irqflags); |
| 431 | /* Going from 0->1 means we have to enable interrupts again */ | 431 | /* Going from 0->1 means we have to enable interrupts again */ |
| 432 | if (atomic_add_return(1, &dev->vblank_refcount[crtc]) == 1 && | 432 | if (atomic_add_return(1, &dev->vblank_refcount[crtc]) == 1) { |
| 433 | !dev->vblank_enabled[crtc]) { | 433 | if (!dev->vblank_enabled[crtc]) { |
| 434 | ret = dev->driver->enable_vblank(dev, crtc); | 434 | ret = dev->driver->enable_vblank(dev, crtc); |
| 435 | DRM_DEBUG("enabling vblank on crtc %d, ret: %d\n", crtc, ret); | 435 | DRM_DEBUG("enabling vblank on crtc %d, ret: %d\n", crtc, ret); |
| 436 | if (ret) | 436 | if (ret) |
| 437 | atomic_dec(&dev->vblank_refcount[crtc]); | ||
| 438 | else { | ||
| 439 | dev->vblank_enabled[crtc] = 1; | ||
| 440 | drm_update_vblank_count(dev, crtc); | ||
| 441 | } | ||
| 442 | } | ||
| 443 | } else { | ||
| 444 | if (!dev->vblank_enabled[crtc]) { | ||
| 437 | atomic_dec(&dev->vblank_refcount[crtc]); | 445 | atomic_dec(&dev->vblank_refcount[crtc]); |
| 438 | else { | 446 | ret = -EINVAL; |
| 439 | dev->vblank_enabled[crtc] = 1; | ||
| 440 | drm_update_vblank_count(dev, crtc); | ||
| 441 | } | 447 | } |
| 442 | } | 448 | } |
| 443 | spin_unlock_irqrestore(&dev->vbl_lock, irqflags); | 449 | spin_unlock_irqrestore(&dev->vbl_lock, irqflags); |
| @@ -464,6 +470,18 @@ void drm_vblank_put(struct drm_device *dev, int crtc) | |||
| 464 | } | 470 | } |
| 465 | EXPORT_SYMBOL(drm_vblank_put); | 471 | EXPORT_SYMBOL(drm_vblank_put); |
| 466 | 472 | ||
| 473 | void drm_vblank_off(struct drm_device *dev, int crtc) | ||
| 474 | { | ||
| 475 | unsigned long irqflags; | ||
| 476 | |||
| 477 | spin_lock_irqsave(&dev->vbl_lock, irqflags); | ||
| 478 | DRM_WAKEUP(&dev->vbl_queue[crtc]); | ||
| 479 | dev->vblank_enabled[crtc] = 0; | ||
| 480 | dev->last_vblank[crtc] = dev->driver->get_vblank_counter(dev, crtc); | ||
| 481 | spin_unlock_irqrestore(&dev->vbl_lock, irqflags); | ||
| 482 | } | ||
| 483 | EXPORT_SYMBOL(drm_vblank_off); | ||
| 484 | |||
| 467 | /** | 485 | /** |
| 468 | * drm_vblank_pre_modeset - account for vblanks across mode sets | 486 | * drm_vblank_pre_modeset - account for vblanks across mode sets |
| 469 | * @dev: DRM device | 487 | * @dev: DRM device |
| @@ -550,6 +568,63 @@ out: | |||
| 550 | return ret; | 568 | return ret; |
| 551 | } | 569 | } |
| 552 | 570 | ||
| 571 | static int drm_queue_vblank_event(struct drm_device *dev, int pipe, | ||
| 572 | union drm_wait_vblank *vblwait, | ||
| 573 | struct drm_file *file_priv) | ||
| 574 | { | ||
| 575 | struct drm_pending_vblank_event *e; | ||
| 576 | struct timeval now; | ||
| 577 | unsigned long flags; | ||
| 578 | unsigned int seq; | ||
| 579 | |||
| 580 | e = kzalloc(sizeof *e, GFP_KERNEL); | ||
| 581 | if (e == NULL) | ||
| 582 | return -ENOMEM; | ||
| 583 | |||
| 584 | e->pipe = pipe; | ||
| 585 | e->event.base.type = DRM_EVENT_VBLANK; | ||
| 586 | e->event.base.length = sizeof e->event; | ||
| 587 | e->event.user_data = vblwait->request.signal; | ||
| 588 | e->base.event = &e->event.base; | ||
| 589 | e->base.file_priv = file_priv; | ||
| 590 | e->base.destroy = (void (*) (struct drm_pending_event *)) kfree; | ||
| 591 | |||
| 592 | do_gettimeofday(&now); | ||
| 593 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 594 | |||
| 595 | if (file_priv->event_space < sizeof e->event) { | ||
| 596 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 597 | kfree(e); | ||
| 598 | return -ENOMEM; | ||
| 599 | } | ||
| 600 | |||
| 601 | file_priv->event_space -= sizeof e->event; | ||
| 602 | seq = drm_vblank_count(dev, pipe); | ||
| 603 | if ((vblwait->request.type & _DRM_VBLANK_NEXTONMISS) && | ||
| 604 | (seq - vblwait->request.sequence) <= (1 << 23)) { | ||
| 605 | vblwait->request.sequence = seq + 1; | ||
| 606 | vblwait->reply.sequence = vblwait->request.sequence; | ||
| 607 | } | ||
| 608 | |||
| 609 | DRM_DEBUG("event on vblank count %d, current %d, crtc %d\n", | ||
| 610 | vblwait->request.sequence, seq, pipe); | ||
| 611 | |||
| 612 | e->event.sequence = vblwait->request.sequence; | ||
| 613 | if ((seq - vblwait->request.sequence) <= (1 << 23)) { | ||
| 614 | e->event.tv_sec = now.tv_sec; | ||
| 615 | e->event.tv_usec = now.tv_usec; | ||
| 616 | drm_vblank_put(dev, e->pipe); | ||
| 617 | list_add_tail(&e->base.link, &e->base.file_priv->event_list); | ||
| 618 | wake_up_interruptible(&e->base.file_priv->event_wait); | ||
| 619 | } else { | ||
| 620 | list_add_tail(&e->base.link, &dev->vblank_event_list); | ||
| 621 | } | ||
| 622 | |||
| 623 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 624 | |||
| 625 | return 0; | ||
| 626 | } | ||
| 627 | |||
| 553 | /** | 628 | /** |
| 554 | * Wait for VBLANK. | 629 | * Wait for VBLANK. |
| 555 | * | 630 | * |
| @@ -609,6 +684,9 @@ int drm_wait_vblank(struct drm_device *dev, void *data, | |||
| 609 | goto done; | 684 | goto done; |
| 610 | } | 685 | } |
| 611 | 686 | ||
| 687 | if (flags & _DRM_VBLANK_EVENT) | ||
| 688 | return drm_queue_vblank_event(dev, crtc, vblwait, file_priv); | ||
| 689 | |||
| 612 | if ((flags & _DRM_VBLANK_NEXTONMISS) && | 690 | if ((flags & _DRM_VBLANK_NEXTONMISS) && |
| 613 | (seq - vblwait->request.sequence) <= (1<<23)) { | 691 | (seq - vblwait->request.sequence) <= (1<<23)) { |
| 614 | vblwait->request.sequence = seq + 1; | 692 | vblwait->request.sequence = seq + 1; |
| @@ -641,6 +719,38 @@ done: | |||
| 641 | return ret; | 719 | return ret; |
| 642 | } | 720 | } |
| 643 | 721 | ||
| 722 | void drm_handle_vblank_events(struct drm_device *dev, int crtc) | ||
| 723 | { | ||
| 724 | struct drm_pending_vblank_event *e, *t; | ||
| 725 | struct timeval now; | ||
| 726 | unsigned long flags; | ||
| 727 | unsigned int seq; | ||
| 728 | |||
| 729 | do_gettimeofday(&now); | ||
| 730 | seq = drm_vblank_count(dev, crtc); | ||
| 731 | |||
| 732 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 733 | |||
| 734 | list_for_each_entry_safe(e, t, &dev->vblank_event_list, base.link) { | ||
| 735 | if (e->pipe != crtc) | ||
| 736 | continue; | ||
| 737 | if ((seq - e->event.sequence) > (1<<23)) | ||
| 738 | continue; | ||
| 739 | |||
| 740 | DRM_DEBUG("vblank event on %d, current %d\n", | ||
| 741 | e->event.sequence, seq); | ||
| 742 | |||
| 743 | e->event.sequence = seq; | ||
| 744 | e->event.tv_sec = now.tv_sec; | ||
| 745 | e->event.tv_usec = now.tv_usec; | ||
| 746 | drm_vblank_put(dev, e->pipe); | ||
| 747 | list_move_tail(&e->base.link, &e->base.file_priv->event_list); | ||
| 748 | wake_up_interruptible(&e->base.file_priv->event_wait); | ||
| 749 | } | ||
| 750 | |||
| 751 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 752 | } | ||
| 753 | |||
| 644 | /** | 754 | /** |
| 645 | * drm_handle_vblank - handle a vblank event | 755 | * drm_handle_vblank - handle a vblank event |
| 646 | * @dev: DRM device | 756 | * @dev: DRM device |
| @@ -651,7 +761,11 @@ done: | |||
| 651 | */ | 761 | */ |
| 652 | void drm_handle_vblank(struct drm_device *dev, int crtc) | 762 | void drm_handle_vblank(struct drm_device *dev, int crtc) |
| 653 | { | 763 | { |
| 764 | if (!dev->num_crtcs) | ||
| 765 | return; | ||
| 766 | |||
| 654 | atomic_inc(&dev->_vblank_count[crtc]); | 767 | atomic_inc(&dev->_vblank_count[crtc]); |
| 655 | DRM_WAKEUP(&dev->vbl_queue[crtc]); | 768 | DRM_WAKEUP(&dev->vbl_queue[crtc]); |
| 769 | drm_handle_vblank_events(dev, crtc); | ||
| 656 | } | 770 | } |
| 657 | EXPORT_SYMBOL(drm_handle_vblank); | 771 | EXPORT_SYMBOL(drm_handle_vblank); |
diff --git a/drivers/gpu/drm/drm_mm.c b/drivers/gpu/drm/drm_mm.c index 97dc5a4f0de4..d7d7eac3ddd2 100644 --- a/drivers/gpu/drm/drm_mm.c +++ b/drivers/gpu/drm/drm_mm.c | |||
| @@ -226,6 +226,44 @@ struct drm_mm_node *drm_mm_get_block_generic(struct drm_mm_node *node, | |||
| 226 | } | 226 | } |
| 227 | EXPORT_SYMBOL(drm_mm_get_block_generic); | 227 | EXPORT_SYMBOL(drm_mm_get_block_generic); |
| 228 | 228 | ||
| 229 | struct drm_mm_node *drm_mm_get_block_range_generic(struct drm_mm_node *node, | ||
| 230 | unsigned long size, | ||
| 231 | unsigned alignment, | ||
| 232 | unsigned long start, | ||
| 233 | unsigned long end, | ||
| 234 | int atomic) | ||
| 235 | { | ||
| 236 | struct drm_mm_node *align_splitoff = NULL; | ||
| 237 | unsigned tmp = 0; | ||
| 238 | unsigned wasted = 0; | ||
| 239 | |||
| 240 | if (node->start < start) | ||
| 241 | wasted += start - node->start; | ||
| 242 | if (alignment) | ||
| 243 | tmp = ((node->start + wasted) % alignment); | ||
| 244 | |||
| 245 | if (tmp) | ||
| 246 | wasted += alignment - tmp; | ||
| 247 | if (wasted) { | ||
| 248 | align_splitoff = drm_mm_split_at_start(node, wasted, atomic); | ||
| 249 | if (unlikely(align_splitoff == NULL)) | ||
| 250 | return NULL; | ||
| 251 | } | ||
| 252 | |||
| 253 | if (node->size == size) { | ||
| 254 | list_del_init(&node->fl_entry); | ||
| 255 | node->free = 0; | ||
| 256 | } else { | ||
| 257 | node = drm_mm_split_at_start(node, size, atomic); | ||
| 258 | } | ||
| 259 | |||
| 260 | if (align_splitoff) | ||
| 261 | drm_mm_put_block(align_splitoff); | ||
| 262 | |||
| 263 | return node; | ||
| 264 | } | ||
| 265 | EXPORT_SYMBOL(drm_mm_get_block_range_generic); | ||
| 266 | |||
| 229 | /* | 267 | /* |
| 230 | * Put a block. Merge with the previous and / or next block if they are free. | 268 | * Put a block. Merge with the previous and / or next block if they are free. |
| 231 | * Otherwise add to the free stack. | 269 | * Otherwise add to the free stack. |
| @@ -331,6 +369,56 @@ struct drm_mm_node *drm_mm_search_free(const struct drm_mm *mm, | |||
| 331 | } | 369 | } |
| 332 | EXPORT_SYMBOL(drm_mm_search_free); | 370 | EXPORT_SYMBOL(drm_mm_search_free); |
| 333 | 371 | ||
| 372 | struct drm_mm_node *drm_mm_search_free_in_range(const struct drm_mm *mm, | ||
| 373 | unsigned long size, | ||
| 374 | unsigned alignment, | ||
| 375 | unsigned long start, | ||
| 376 | unsigned long end, | ||
| 377 | int best_match) | ||
| 378 | { | ||
| 379 | struct list_head *list; | ||
| 380 | const struct list_head *free_stack = &mm->fl_entry; | ||
| 381 | struct drm_mm_node *entry; | ||
| 382 | struct drm_mm_node *best; | ||
| 383 | unsigned long best_size; | ||
| 384 | unsigned wasted; | ||
| 385 | |||
| 386 | best = NULL; | ||
| 387 | best_size = ~0UL; | ||
| 388 | |||
| 389 | list_for_each(list, free_stack) { | ||
| 390 | entry = list_entry(list, struct drm_mm_node, fl_entry); | ||
| 391 | wasted = 0; | ||
| 392 | |||
| 393 | if (entry->size < size) | ||
| 394 | continue; | ||
| 395 | |||
| 396 | if (entry->start > end || (entry->start+entry->size) < start) | ||
| 397 | continue; | ||
| 398 | |||
| 399 | if (entry->start < start) | ||
| 400 | wasted += start - entry->start; | ||
| 401 | |||
| 402 | if (alignment) { | ||
| 403 | register unsigned tmp = (entry->start + wasted) % alignment; | ||
| 404 | if (tmp) | ||
| 405 | wasted += alignment - tmp; | ||
| 406 | } | ||
| 407 | |||
| 408 | if (entry->size >= size + wasted) { | ||
| 409 | if (!best_match) | ||
| 410 | return entry; | ||
| 411 | if (size < best_size) { | ||
| 412 | best = entry; | ||
| 413 | best_size = entry->size; | ||
| 414 | } | ||
| 415 | } | ||
| 416 | } | ||
| 417 | |||
| 418 | return best; | ||
| 419 | } | ||
| 420 | EXPORT_SYMBOL(drm_mm_search_free_in_range); | ||
| 421 | |||
| 334 | int drm_mm_clean(struct drm_mm * mm) | 422 | int drm_mm_clean(struct drm_mm * mm) |
| 335 | { | 423 | { |
| 336 | struct list_head *head = &mm->ml_entry; | 424 | struct list_head *head = &mm->ml_entry; |
| @@ -381,6 +469,26 @@ void drm_mm_takedown(struct drm_mm * mm) | |||
| 381 | } | 469 | } |
| 382 | EXPORT_SYMBOL(drm_mm_takedown); | 470 | EXPORT_SYMBOL(drm_mm_takedown); |
| 383 | 471 | ||
| 472 | void drm_mm_debug_table(struct drm_mm *mm, const char *prefix) | ||
| 473 | { | ||
| 474 | struct drm_mm_node *entry; | ||
| 475 | int total_used = 0, total_free = 0, total = 0; | ||
| 476 | |||
| 477 | list_for_each_entry(entry, &mm->ml_entry, ml_entry) { | ||
| 478 | printk(KERN_DEBUG "%s 0x%08lx-0x%08lx: %8ld: %s\n", | ||
| 479 | prefix, entry->start, entry->start + entry->size, | ||
| 480 | entry->size, entry->free ? "free" : "used"); | ||
| 481 | total += entry->size; | ||
| 482 | if (entry->free) | ||
| 483 | total_free += entry->size; | ||
| 484 | else | ||
| 485 | total_used += entry->size; | ||
| 486 | } | ||
| 487 | printk(KERN_DEBUG "%s total: %d, used %d free %d\n", prefix, total, | ||
| 488 | total_used, total_free); | ||
| 489 | } | ||
| 490 | EXPORT_SYMBOL(drm_mm_debug_table); | ||
| 491 | |||
| 384 | #if defined(CONFIG_DEBUG_FS) | 492 | #if defined(CONFIG_DEBUG_FS) |
| 385 | int drm_mm_dump_table(struct seq_file *m, struct drm_mm *mm) | 493 | int drm_mm_dump_table(struct seq_file *m, struct drm_mm *mm) |
| 386 | { | 494 | { |
| @@ -395,7 +503,7 @@ int drm_mm_dump_table(struct seq_file *m, struct drm_mm *mm) | |||
| 395 | else | 503 | else |
| 396 | total_used += entry->size; | 504 | total_used += entry->size; |
| 397 | } | 505 | } |
| 398 | seq_printf(m, "total: %d, used %d free %d\n", total, total_free, total_used); | 506 | seq_printf(m, "total: %d, used %d free %d\n", total, total_used, total_free); |
| 399 | return 0; | 507 | return 0; |
| 400 | } | 508 | } |
| 401 | EXPORT_SYMBOL(drm_mm_dump_table); | 509 | EXPORT_SYMBOL(drm_mm_dump_table); |
diff --git a/drivers/gpu/drm/drm_modes.c b/drivers/gpu/drm/drm_modes.c index 51f677215f1d..6d81a02463a3 100644 --- a/drivers/gpu/drm/drm_modes.c +++ b/drivers/gpu/drm/drm_modes.c | |||
| @@ -553,6 +553,32 @@ int drm_mode_height(struct drm_display_mode *mode) | |||
| 553 | } | 553 | } |
| 554 | EXPORT_SYMBOL(drm_mode_height); | 554 | EXPORT_SYMBOL(drm_mode_height); |
| 555 | 555 | ||
| 556 | /** drm_mode_hsync - get the hsync of a mode | ||
| 557 | * @mode: mode | ||
| 558 | * | ||
| 559 | * LOCKING: | ||
| 560 | * None. | ||
| 561 | * | ||
| 562 | * Return @modes's hsync rate in kHz, rounded to the nearest int. | ||
| 563 | */ | ||
| 564 | int drm_mode_hsync(struct drm_display_mode *mode) | ||
| 565 | { | ||
| 566 | unsigned int calc_val; | ||
| 567 | |||
| 568 | if (mode->hsync) | ||
| 569 | return mode->hsync; | ||
| 570 | |||
| 571 | if (mode->htotal < 0) | ||
| 572 | return 0; | ||
| 573 | |||
| 574 | calc_val = (mode->clock * 1000) / mode->htotal; /* hsync in Hz */ | ||
| 575 | calc_val += 500; /* round to 1000Hz */ | ||
| 576 | calc_val /= 1000; /* truncate to kHz */ | ||
| 577 | |||
| 578 | return calc_val; | ||
| 579 | } | ||
| 580 | EXPORT_SYMBOL(drm_mode_hsync); | ||
| 581 | |||
| 556 | /** | 582 | /** |
| 557 | * drm_mode_vrefresh - get the vrefresh of a mode | 583 | * drm_mode_vrefresh - get the vrefresh of a mode |
| 558 | * @mode: mode | 584 | * @mode: mode |
| @@ -560,7 +586,7 @@ EXPORT_SYMBOL(drm_mode_height); | |||
| 560 | * LOCKING: | 586 | * LOCKING: |
| 561 | * None. | 587 | * None. |
| 562 | * | 588 | * |
| 563 | * Return @mode's vrefresh rate or calculate it if necessary. | 589 | * Return @mode's vrefresh rate in Hz or calculate it if necessary. |
| 564 | * | 590 | * |
| 565 | * FIXME: why is this needed? shouldn't vrefresh be set already? | 591 | * FIXME: why is this needed? shouldn't vrefresh be set already? |
| 566 | * | 592 | * |
diff --git a/drivers/gpu/drm/drm_stub.c b/drivers/gpu/drm/drm_stub.c index 55bb8a82d612..ad73e141afdb 100644 --- a/drivers/gpu/drm/drm_stub.c +++ b/drivers/gpu/drm/drm_stub.c | |||
| @@ -128,6 +128,7 @@ struct drm_master *drm_master_get(struct drm_master *master) | |||
| 128 | kref_get(&master->refcount); | 128 | kref_get(&master->refcount); |
| 129 | return master; | 129 | return master; |
| 130 | } | 130 | } |
| 131 | EXPORT_SYMBOL(drm_master_get); | ||
| 131 | 132 | ||
| 132 | static void drm_master_destroy(struct kref *kref) | 133 | static void drm_master_destroy(struct kref *kref) |
| 133 | { | 134 | { |
| @@ -170,10 +171,13 @@ void drm_master_put(struct drm_master **master) | |||
| 170 | kref_put(&(*master)->refcount, drm_master_destroy); | 171 | kref_put(&(*master)->refcount, drm_master_destroy); |
| 171 | *master = NULL; | 172 | *master = NULL; |
| 172 | } | 173 | } |
| 174 | EXPORT_SYMBOL(drm_master_put); | ||
| 173 | 175 | ||
| 174 | int drm_setmaster_ioctl(struct drm_device *dev, void *data, | 176 | int drm_setmaster_ioctl(struct drm_device *dev, void *data, |
| 175 | struct drm_file *file_priv) | 177 | struct drm_file *file_priv) |
| 176 | { | 178 | { |
| 179 | int ret = 0; | ||
| 180 | |||
| 177 | if (file_priv->is_master) | 181 | if (file_priv->is_master) |
| 178 | return 0; | 182 | return 0; |
| 179 | 183 | ||
| @@ -188,6 +192,13 @@ int drm_setmaster_ioctl(struct drm_device *dev, void *data, | |||
| 188 | mutex_lock(&dev->struct_mutex); | 192 | mutex_lock(&dev->struct_mutex); |
| 189 | file_priv->minor->master = drm_master_get(file_priv->master); | 193 | file_priv->minor->master = drm_master_get(file_priv->master); |
| 190 | file_priv->is_master = 1; | 194 | file_priv->is_master = 1; |
| 195 | if (dev->driver->master_set) { | ||
| 196 | ret = dev->driver->master_set(dev, file_priv, false); | ||
| 197 | if (unlikely(ret != 0)) { | ||
| 198 | file_priv->is_master = 0; | ||
| 199 | drm_master_put(&file_priv->minor->master); | ||
| 200 | } | ||
| 201 | } | ||
| 191 | mutex_unlock(&dev->struct_mutex); | 202 | mutex_unlock(&dev->struct_mutex); |
| 192 | } | 203 | } |
| 193 | 204 | ||
| @@ -204,6 +215,8 @@ int drm_dropmaster_ioctl(struct drm_device *dev, void *data, | |||
| 204 | return -EINVAL; | 215 | return -EINVAL; |
| 205 | 216 | ||
| 206 | mutex_lock(&dev->struct_mutex); | 217 | mutex_lock(&dev->struct_mutex); |
| 218 | if (dev->driver->master_drop) | ||
| 219 | dev->driver->master_drop(dev, file_priv, false); | ||
| 207 | drm_master_put(&file_priv->minor->master); | 220 | drm_master_put(&file_priv->minor->master); |
| 208 | file_priv->is_master = 0; | 221 | file_priv->is_master = 0; |
| 209 | mutex_unlock(&dev->struct_mutex); | 222 | mutex_unlock(&dev->struct_mutex); |
| @@ -220,9 +233,11 @@ static int drm_fill_in_dev(struct drm_device * dev, struct pci_dev *pdev, | |||
| 220 | INIT_LIST_HEAD(&dev->ctxlist); | 233 | INIT_LIST_HEAD(&dev->ctxlist); |
| 221 | INIT_LIST_HEAD(&dev->vmalist); | 234 | INIT_LIST_HEAD(&dev->vmalist); |
| 222 | INIT_LIST_HEAD(&dev->maplist); | 235 | INIT_LIST_HEAD(&dev->maplist); |
| 236 | INIT_LIST_HEAD(&dev->vblank_event_list); | ||
| 223 | 237 | ||
| 224 | spin_lock_init(&dev->count_lock); | 238 | spin_lock_init(&dev->count_lock); |
| 225 | spin_lock_init(&dev->drw_lock); | 239 | spin_lock_init(&dev->drw_lock); |
| 240 | spin_lock_init(&dev->event_lock); | ||
| 226 | init_timer(&dev->timer); | 241 | init_timer(&dev->timer); |
| 227 | mutex_init(&dev->struct_mutex); | 242 | mutex_init(&dev->struct_mutex); |
| 228 | mutex_init(&dev->ctxlist_mutex); | 243 | mutex_init(&dev->ctxlist_mutex); |
diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile index fa7b9be096bc..9929f84ec3e1 100644 --- a/drivers/gpu/drm/i915/Makefile +++ b/drivers/gpu/drm/i915/Makefile | |||
| @@ -15,7 +15,6 @@ i915-y := i915_drv.o i915_dma.o i915_irq.o i915_mem.o \ | |||
| 15 | intel_lvds.o \ | 15 | intel_lvds.o \ |
| 16 | intel_bios.o \ | 16 | intel_bios.o \ |
| 17 | intel_dp.o \ | 17 | intel_dp.o \ |
| 18 | intel_dp_i2c.o \ | ||
| 19 | intel_hdmi.o \ | 18 | intel_hdmi.o \ |
| 20 | intel_sdvo.o \ | 19 | intel_sdvo.o \ |
| 21 | intel_modes.o \ | 20 | intel_modes.o \ |
| @@ -23,6 +22,7 @@ i915-y := i915_drv.o i915_dma.o i915_irq.o i915_mem.o \ | |||
| 23 | intel_fb.o \ | 22 | intel_fb.o \ |
| 24 | intel_tv.o \ | 23 | intel_tv.o \ |
| 25 | intel_dvo.o \ | 24 | intel_dvo.o \ |
| 25 | intel_overlay.o \ | ||
| 26 | dvo_ch7xxx.o \ | 26 | dvo_ch7xxx.o \ |
| 27 | dvo_ch7017.o \ | 27 | dvo_ch7017.o \ |
| 28 | dvo_ivch.o \ | 28 | dvo_ivch.o \ |
diff --git a/drivers/gpu/drm/i915/dvo_ch7017.c b/drivers/gpu/drm/i915/dvo_ch7017.c index 621815b531db..1184c14ba87d 100644 --- a/drivers/gpu/drm/i915/dvo_ch7017.c +++ b/drivers/gpu/drm/i915/dvo_ch7017.c | |||
| @@ -249,7 +249,8 @@ static bool ch7017_init(struct intel_dvo_device *dvo, | |||
| 249 | if (val != CH7017_DEVICE_ID_VALUE && | 249 | if (val != CH7017_DEVICE_ID_VALUE && |
| 250 | val != CH7018_DEVICE_ID_VALUE && | 250 | val != CH7018_DEVICE_ID_VALUE && |
| 251 | val != CH7019_DEVICE_ID_VALUE) { | 251 | val != CH7019_DEVICE_ID_VALUE) { |
| 252 | DRM_DEBUG("ch701x not detected, got %d: from %s Slave %d.\n", | 252 | DRM_DEBUG_KMS("ch701x not detected, got %d: from %s " |
| 253 | "Slave %d.\n", | ||
| 253 | val, i2cbus->adapter.name,dvo->slave_addr); | 254 | val, i2cbus->adapter.name,dvo->slave_addr); |
| 254 | goto fail; | 255 | goto fail; |
| 255 | } | 256 | } |
| @@ -284,7 +285,7 @@ static void ch7017_mode_set(struct intel_dvo_device *dvo, | |||
| 284 | uint8_t horizontal_active_pixel_output, vertical_active_line_output; | 285 | uint8_t horizontal_active_pixel_output, vertical_active_line_output; |
| 285 | uint8_t active_input_line_output; | 286 | uint8_t active_input_line_output; |
| 286 | 287 | ||
| 287 | DRM_DEBUG("Registers before mode setting\n"); | 288 | DRM_DEBUG_KMS("Registers before mode setting\n"); |
| 288 | ch7017_dump_regs(dvo); | 289 | ch7017_dump_regs(dvo); |
| 289 | 290 | ||
| 290 | /* LVDS PLL settings from page 75 of 7017-7017ds.pdf*/ | 291 | /* LVDS PLL settings from page 75 of 7017-7017ds.pdf*/ |
| @@ -346,7 +347,7 @@ static void ch7017_mode_set(struct intel_dvo_device *dvo, | |||
| 346 | /* Turn the LVDS back on with new settings. */ | 347 | /* Turn the LVDS back on with new settings. */ |
| 347 | ch7017_write(dvo, CH7017_LVDS_POWER_DOWN, lvds_power_down); | 348 | ch7017_write(dvo, CH7017_LVDS_POWER_DOWN, lvds_power_down); |
| 348 | 349 | ||
| 349 | DRM_DEBUG("Registers after mode setting\n"); | 350 | DRM_DEBUG_KMS("Registers after mode setting\n"); |
| 350 | ch7017_dump_regs(dvo); | 351 | ch7017_dump_regs(dvo); |
| 351 | } | 352 | } |
| 352 | 353 | ||
| @@ -386,7 +387,7 @@ static void ch7017_dump_regs(struct intel_dvo_device *dvo) | |||
| 386 | #define DUMP(reg) \ | 387 | #define DUMP(reg) \ |
| 387 | do { \ | 388 | do { \ |
| 388 | ch7017_read(dvo, reg, &val); \ | 389 | ch7017_read(dvo, reg, &val); \ |
| 389 | DRM_DEBUG(#reg ": %02x\n", val); \ | 390 | DRM_DEBUG_KMS(#reg ": %02x\n", val); \ |
| 390 | } while (0) | 391 | } while (0) |
| 391 | 392 | ||
| 392 | DUMP(CH7017_HORIZONTAL_ACTIVE_PIXEL_INPUT); | 393 | DUMP(CH7017_HORIZONTAL_ACTIVE_PIXEL_INPUT); |
diff --git a/drivers/gpu/drm/i915/dvo_ch7xxx.c b/drivers/gpu/drm/i915/dvo_ch7xxx.c index a9b896289680..d56ff5cc22b2 100644 --- a/drivers/gpu/drm/i915/dvo_ch7xxx.c +++ b/drivers/gpu/drm/i915/dvo_ch7xxx.c | |||
| @@ -152,7 +152,7 @@ static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, uint8_t *ch) | |||
| 152 | }; | 152 | }; |
| 153 | 153 | ||
| 154 | if (!ch7xxx->quiet) { | 154 | if (!ch7xxx->quiet) { |
| 155 | DRM_DEBUG("Unable to read register 0x%02x from %s:%02x.\n", | 155 | DRM_DEBUG_KMS("Unable to read register 0x%02x from %s:%02x.\n", |
| 156 | addr, i2cbus->adapter.name, dvo->slave_addr); | 156 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 157 | } | 157 | } |
| 158 | return false; | 158 | return false; |
| @@ -179,7 +179,7 @@ static bool ch7xxx_writeb(struct intel_dvo_device *dvo, int addr, uint8_t ch) | |||
| 179 | return true; | 179 | return true; |
| 180 | 180 | ||
| 181 | if (!ch7xxx->quiet) { | 181 | if (!ch7xxx->quiet) { |
| 182 | DRM_DEBUG("Unable to write register 0x%02x to %s:%d.\n", | 182 | DRM_DEBUG_KMS("Unable to write register 0x%02x to %s:%d.\n", |
| 183 | addr, i2cbus->adapter.name, dvo->slave_addr); | 183 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 184 | } | 184 | } |
| 185 | 185 | ||
| @@ -207,7 +207,8 @@ static bool ch7xxx_init(struct intel_dvo_device *dvo, | |||
| 207 | 207 | ||
| 208 | name = ch7xxx_get_id(vendor); | 208 | name = ch7xxx_get_id(vendor); |
| 209 | if (!name) { | 209 | if (!name) { |
| 210 | DRM_DEBUG("ch7xxx not detected; got 0x%02x from %s slave %d.\n", | 210 | DRM_DEBUG_KMS("ch7xxx not detected; got 0x%02x from %s " |
| 211 | "slave %d.\n", | ||
| 211 | vendor, adapter->name, dvo->slave_addr); | 212 | vendor, adapter->name, dvo->slave_addr); |
| 212 | goto out; | 213 | goto out; |
| 213 | } | 214 | } |
| @@ -217,13 +218,14 @@ static bool ch7xxx_init(struct intel_dvo_device *dvo, | |||
| 217 | goto out; | 218 | goto out; |
| 218 | 219 | ||
| 219 | if (device != CH7xxx_DID) { | 220 | if (device != CH7xxx_DID) { |
| 220 | DRM_DEBUG("ch7xxx not detected; got 0x%02x from %s slave %d.\n", | 221 | DRM_DEBUG_KMS("ch7xxx not detected; got 0x%02x from %s " |
| 222 | "slave %d.\n", | ||
| 221 | vendor, adapter->name, dvo->slave_addr); | 223 | vendor, adapter->name, dvo->slave_addr); |
| 222 | goto out; | 224 | goto out; |
| 223 | } | 225 | } |
| 224 | 226 | ||
| 225 | ch7xxx->quiet = false; | 227 | ch7xxx->quiet = false; |
| 226 | DRM_DEBUG("Detected %s chipset, vendor/device ID 0x%02x/0x%02x\n", | 228 | DRM_DEBUG_KMS("Detected %s chipset, vendor/device ID 0x%02x/0x%02x\n", |
| 227 | name, vendor, device); | 229 | name, vendor, device); |
| 228 | return true; | 230 | return true; |
| 229 | out: | 231 | out: |
| @@ -315,8 +317,8 @@ static void ch7xxx_dump_regs(struct intel_dvo_device *dvo) | |||
| 315 | 317 | ||
| 316 | for (i = 0; i < CH7xxx_NUM_REGS; i++) { | 318 | for (i = 0; i < CH7xxx_NUM_REGS; i++) { |
| 317 | if ((i % 8) == 0 ) | 319 | if ((i % 8) == 0 ) |
| 318 | DRM_DEBUG("\n %02X: ", i); | 320 | DRM_LOG_KMS("\n %02X: ", i); |
| 319 | DRM_DEBUG("%02X ", ch7xxx->mode_reg.regs[i]); | 321 | DRM_LOG_KMS("%02X ", ch7xxx->mode_reg.regs[i]); |
| 320 | } | 322 | } |
| 321 | } | 323 | } |
| 322 | 324 | ||
diff --git a/drivers/gpu/drm/i915/dvo_ivch.c b/drivers/gpu/drm/i915/dvo_ivch.c index aa176f9921fe..24169e528f0f 100644 --- a/drivers/gpu/drm/i915/dvo_ivch.c +++ b/drivers/gpu/drm/i915/dvo_ivch.c | |||
| @@ -202,7 +202,8 @@ static bool ivch_read(struct intel_dvo_device *dvo, int addr, uint16_t *data) | |||
| 202 | }; | 202 | }; |
| 203 | 203 | ||
| 204 | if (!priv->quiet) { | 204 | if (!priv->quiet) { |
| 205 | DRM_DEBUG("Unable to read register 0x%02x from %s:%02x.\n", | 205 | DRM_DEBUG_KMS("Unable to read register 0x%02x from " |
| 206 | "%s:%02x.\n", | ||
| 206 | addr, i2cbus->adapter.name, dvo->slave_addr); | 207 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 207 | } | 208 | } |
| 208 | return false; | 209 | return false; |
| @@ -230,7 +231,7 @@ static bool ivch_write(struct intel_dvo_device *dvo, int addr, uint16_t data) | |||
| 230 | return true; | 231 | return true; |
| 231 | 232 | ||
| 232 | if (!priv->quiet) { | 233 | if (!priv->quiet) { |
| 233 | DRM_DEBUG("Unable to write register 0x%02x to %s:%d.\n", | 234 | DRM_DEBUG_KMS("Unable to write register 0x%02x to %s:%d.\n", |
| 234 | addr, i2cbus->adapter.name, dvo->slave_addr); | 235 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 235 | } | 236 | } |
| 236 | 237 | ||
| @@ -261,7 +262,7 @@ static bool ivch_init(struct intel_dvo_device *dvo, | |||
| 261 | * the address it's responding on. | 262 | * the address it's responding on. |
| 262 | */ | 263 | */ |
| 263 | if ((temp & VR00_BASE_ADDRESS_MASK) != dvo->slave_addr) { | 264 | if ((temp & VR00_BASE_ADDRESS_MASK) != dvo->slave_addr) { |
| 264 | DRM_DEBUG("ivch detect failed due to address mismatch " | 265 | DRM_DEBUG_KMS("ivch detect failed due to address mismatch " |
| 265 | "(%d vs %d)\n", | 266 | "(%d vs %d)\n", |
| 266 | (temp & VR00_BASE_ADDRESS_MASK), dvo->slave_addr); | 267 | (temp & VR00_BASE_ADDRESS_MASK), dvo->slave_addr); |
| 267 | goto out; | 268 | goto out; |
| @@ -367,41 +368,41 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo) | |||
| 367 | uint16_t val; | 368 | uint16_t val; |
| 368 | 369 | ||
| 369 | ivch_read(dvo, VR00, &val); | 370 | ivch_read(dvo, VR00, &val); |
| 370 | DRM_DEBUG("VR00: 0x%04x\n", val); | 371 | DRM_LOG_KMS("VR00: 0x%04x\n", val); |
| 371 | ivch_read(dvo, VR01, &val); | 372 | ivch_read(dvo, VR01, &val); |
| 372 | DRM_DEBUG("VR01: 0x%04x\n", val); | 373 | DRM_LOG_KMS("VR01: 0x%04x\n", val); |
| 373 | ivch_read(dvo, VR30, &val); | 374 | ivch_read(dvo, VR30, &val); |
| 374 | DRM_DEBUG("VR30: 0x%04x\n", val); | 375 | DRM_LOG_KMS("VR30: 0x%04x\n", val); |
| 375 | ivch_read(dvo, VR40, &val); | 376 | ivch_read(dvo, VR40, &val); |
| 376 | DRM_DEBUG("VR40: 0x%04x\n", val); | 377 | DRM_LOG_KMS("VR40: 0x%04x\n", val); |
| 377 | 378 | ||
| 378 | /* GPIO registers */ | 379 | /* GPIO registers */ |
| 379 | ivch_read(dvo, VR80, &val); | 380 | ivch_read(dvo, VR80, &val); |
| 380 | DRM_DEBUG("VR80: 0x%04x\n", val); | 381 | DRM_LOG_KMS("VR80: 0x%04x\n", val); |
| 381 | ivch_read(dvo, VR81, &val); | 382 | ivch_read(dvo, VR81, &val); |
| 382 | DRM_DEBUG("VR81: 0x%04x\n", val); | 383 | DRM_LOG_KMS("VR81: 0x%04x\n", val); |
| 383 | ivch_read(dvo, VR82, &val); | 384 | ivch_read(dvo, VR82, &val); |
| 384 | DRM_DEBUG("VR82: 0x%04x\n", val); | 385 | DRM_LOG_KMS("VR82: 0x%04x\n", val); |
| 385 | ivch_read(dvo, VR83, &val); | 386 | ivch_read(dvo, VR83, &val); |
| 386 | DRM_DEBUG("VR83: 0x%04x\n", val); | 387 | DRM_LOG_KMS("VR83: 0x%04x\n", val); |
| 387 | ivch_read(dvo, VR84, &val); | 388 | ivch_read(dvo, VR84, &val); |
| 388 | DRM_DEBUG("VR84: 0x%04x\n", val); | 389 | DRM_LOG_KMS("VR84: 0x%04x\n", val); |
| 389 | ivch_read(dvo, VR85, &val); | 390 | ivch_read(dvo, VR85, &val); |
| 390 | DRM_DEBUG("VR85: 0x%04x\n", val); | 391 | DRM_LOG_KMS("VR85: 0x%04x\n", val); |
| 391 | ivch_read(dvo, VR86, &val); | 392 | ivch_read(dvo, VR86, &val); |
| 392 | DRM_DEBUG("VR86: 0x%04x\n", val); | 393 | DRM_LOG_KMS("VR86: 0x%04x\n", val); |
| 393 | ivch_read(dvo, VR87, &val); | 394 | ivch_read(dvo, VR87, &val); |
| 394 | DRM_DEBUG("VR87: 0x%04x\n", val); | 395 | DRM_LOG_KMS("VR87: 0x%04x\n", val); |
| 395 | ivch_read(dvo, VR88, &val); | 396 | ivch_read(dvo, VR88, &val); |
| 396 | DRM_DEBUG("VR88: 0x%04x\n", val); | 397 | DRM_LOG_KMS("VR88: 0x%04x\n", val); |
| 397 | 398 | ||
| 398 | /* Scratch register 0 - AIM Panel type */ | 399 | /* Scratch register 0 - AIM Panel type */ |
| 399 | ivch_read(dvo, VR8E, &val); | 400 | ivch_read(dvo, VR8E, &val); |
| 400 | DRM_DEBUG("VR8E: 0x%04x\n", val); | 401 | DRM_LOG_KMS("VR8E: 0x%04x\n", val); |
| 401 | 402 | ||
| 402 | /* Scratch register 1 - Status register */ | 403 | /* Scratch register 1 - Status register */ |
| 403 | ivch_read(dvo, VR8F, &val); | 404 | ivch_read(dvo, VR8F, &val); |
| 404 | DRM_DEBUG("VR8F: 0x%04x\n", val); | 405 | DRM_LOG_KMS("VR8F: 0x%04x\n", val); |
| 405 | } | 406 | } |
| 406 | 407 | ||
| 407 | static void ivch_save(struct intel_dvo_device *dvo) | 408 | static void ivch_save(struct intel_dvo_device *dvo) |
diff --git a/drivers/gpu/drm/i915/dvo_sil164.c b/drivers/gpu/drm/i915/dvo_sil164.c index e1c1f7341e5c..0001c13f0a80 100644 --- a/drivers/gpu/drm/i915/dvo_sil164.c +++ b/drivers/gpu/drm/i915/dvo_sil164.c | |||
| @@ -105,7 +105,7 @@ static bool sil164_readb(struct intel_dvo_device *dvo, int addr, uint8_t *ch) | |||
| 105 | }; | 105 | }; |
| 106 | 106 | ||
| 107 | if (!sil->quiet) { | 107 | if (!sil->quiet) { |
| 108 | DRM_DEBUG("Unable to read register 0x%02x from %s:%02x.\n", | 108 | DRM_DEBUG_KMS("Unable to read register 0x%02x from %s:%02x.\n", |
| 109 | addr, i2cbus->adapter.name, dvo->slave_addr); | 109 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 110 | } | 110 | } |
| 111 | return false; | 111 | return false; |
| @@ -131,7 +131,7 @@ static bool sil164_writeb(struct intel_dvo_device *dvo, int addr, uint8_t ch) | |||
| 131 | return true; | 131 | return true; |
| 132 | 132 | ||
| 133 | if (!sil->quiet) { | 133 | if (!sil->quiet) { |
| 134 | DRM_DEBUG("Unable to write register 0x%02x to %s:%d.\n", | 134 | DRM_DEBUG_KMS("Unable to write register 0x%02x to %s:%d.\n", |
| 135 | addr, i2cbus->adapter.name, dvo->slave_addr); | 135 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 136 | } | 136 | } |
| 137 | 137 | ||
| @@ -158,7 +158,7 @@ static bool sil164_init(struct intel_dvo_device *dvo, | |||
| 158 | goto out; | 158 | goto out; |
| 159 | 159 | ||
| 160 | if (ch != (SIL164_VID & 0xff)) { | 160 | if (ch != (SIL164_VID & 0xff)) { |
| 161 | DRM_DEBUG("sil164 not detected got %d: from %s Slave %d.\n", | 161 | DRM_DEBUG_KMS("sil164 not detected got %d: from %s Slave %d.\n", |
| 162 | ch, adapter->name, dvo->slave_addr); | 162 | ch, adapter->name, dvo->slave_addr); |
| 163 | goto out; | 163 | goto out; |
| 164 | } | 164 | } |
| @@ -167,13 +167,13 @@ static bool sil164_init(struct intel_dvo_device *dvo, | |||
| 167 | goto out; | 167 | goto out; |
| 168 | 168 | ||
| 169 | if (ch != (SIL164_DID & 0xff)) { | 169 | if (ch != (SIL164_DID & 0xff)) { |
| 170 | DRM_DEBUG("sil164 not detected got %d: from %s Slave %d.\n", | 170 | DRM_DEBUG_KMS("sil164 not detected got %d: from %s Slave %d.\n", |
| 171 | ch, adapter->name, dvo->slave_addr); | 171 | ch, adapter->name, dvo->slave_addr); |
| 172 | goto out; | 172 | goto out; |
| 173 | } | 173 | } |
| 174 | sil->quiet = false; | 174 | sil->quiet = false; |
| 175 | 175 | ||
| 176 | DRM_DEBUG("init sil164 dvo controller successfully!\n"); | 176 | DRM_DEBUG_KMS("init sil164 dvo controller successfully!\n"); |
| 177 | return true; | 177 | return true; |
| 178 | 178 | ||
| 179 | out: | 179 | out: |
| @@ -241,15 +241,15 @@ static void sil164_dump_regs(struct intel_dvo_device *dvo) | |||
| 241 | uint8_t val; | 241 | uint8_t val; |
| 242 | 242 | ||
| 243 | sil164_readb(dvo, SIL164_FREQ_LO, &val); | 243 | sil164_readb(dvo, SIL164_FREQ_LO, &val); |
| 244 | DRM_DEBUG("SIL164_FREQ_LO: 0x%02x\n", val); | 244 | DRM_LOG_KMS("SIL164_FREQ_LO: 0x%02x\n", val); |
| 245 | sil164_readb(dvo, SIL164_FREQ_HI, &val); | 245 | sil164_readb(dvo, SIL164_FREQ_HI, &val); |
| 246 | DRM_DEBUG("SIL164_FREQ_HI: 0x%02x\n", val); | 246 | DRM_LOG_KMS("SIL164_FREQ_HI: 0x%02x\n", val); |
| 247 | sil164_readb(dvo, SIL164_REG8, &val); | 247 | sil164_readb(dvo, SIL164_REG8, &val); |
| 248 | DRM_DEBUG("SIL164_REG8: 0x%02x\n", val); | 248 | DRM_LOG_KMS("SIL164_REG8: 0x%02x\n", val); |
| 249 | sil164_readb(dvo, SIL164_REG9, &val); | 249 | sil164_readb(dvo, SIL164_REG9, &val); |
| 250 | DRM_DEBUG("SIL164_REG9: 0x%02x\n", val); | 250 | DRM_LOG_KMS("SIL164_REG9: 0x%02x\n", val); |
| 251 | sil164_readb(dvo, SIL164_REGC, &val); | 251 | sil164_readb(dvo, SIL164_REGC, &val); |
| 252 | DRM_DEBUG("SIL164_REGC: 0x%02x\n", val); | 252 | DRM_LOG_KMS("SIL164_REGC: 0x%02x\n", val); |
| 253 | } | 253 | } |
| 254 | 254 | ||
| 255 | static void sil164_save(struct intel_dvo_device *dvo) | 255 | static void sil164_save(struct intel_dvo_device *dvo) |
diff --git a/drivers/gpu/drm/i915/dvo_tfp410.c b/drivers/gpu/drm/i915/dvo_tfp410.c index 9ecc907384ec..c7c391bc116a 100644 --- a/drivers/gpu/drm/i915/dvo_tfp410.c +++ b/drivers/gpu/drm/i915/dvo_tfp410.c | |||
| @@ -130,7 +130,7 @@ static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, uint8_t *ch) | |||
| 130 | }; | 130 | }; |
| 131 | 131 | ||
| 132 | if (!tfp->quiet) { | 132 | if (!tfp->quiet) { |
| 133 | DRM_DEBUG("Unable to read register 0x%02x from %s:%02x.\n", | 133 | DRM_DEBUG_KMS("Unable to read register 0x%02x from %s:%02x.\n", |
| 134 | addr, i2cbus->adapter.name, dvo->slave_addr); | 134 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 135 | } | 135 | } |
| 136 | return false; | 136 | return false; |
| @@ -156,7 +156,7 @@ static bool tfp410_writeb(struct intel_dvo_device *dvo, int addr, uint8_t ch) | |||
| 156 | return true; | 156 | return true; |
| 157 | 157 | ||
| 158 | if (!tfp->quiet) { | 158 | if (!tfp->quiet) { |
| 159 | DRM_DEBUG("Unable to write register 0x%02x to %s:%d.\n", | 159 | DRM_DEBUG_KMS("Unable to write register 0x%02x to %s:%d.\n", |
| 160 | addr, i2cbus->adapter.name, dvo->slave_addr); | 160 | addr, i2cbus->adapter.name, dvo->slave_addr); |
| 161 | } | 161 | } |
| 162 | 162 | ||
| @@ -191,13 +191,15 @@ static bool tfp410_init(struct intel_dvo_device *dvo, | |||
| 191 | tfp->quiet = true; | 191 | tfp->quiet = true; |
| 192 | 192 | ||
| 193 | if ((id = tfp410_getid(dvo, TFP410_VID_LO)) != TFP410_VID) { | 193 | if ((id = tfp410_getid(dvo, TFP410_VID_LO)) != TFP410_VID) { |
| 194 | DRM_DEBUG("tfp410 not detected got VID %X: from %s Slave %d.\n", | 194 | DRM_DEBUG_KMS("tfp410 not detected got VID %X: from %s " |
| 195 | "Slave %d.\n", | ||
| 195 | id, adapter->name, dvo->slave_addr); | 196 | id, adapter->name, dvo->slave_addr); |
| 196 | goto out; | 197 | goto out; |
| 197 | } | 198 | } |
| 198 | 199 | ||
| 199 | if ((id = tfp410_getid(dvo, TFP410_DID_LO)) != TFP410_DID) { | 200 | if ((id = tfp410_getid(dvo, TFP410_DID_LO)) != TFP410_DID) { |
| 200 | DRM_DEBUG("tfp410 not detected got DID %X: from %s Slave %d.\n", | 201 | DRM_DEBUG_KMS("tfp410 not detected got DID %X: from %s " |
| 202 | "Slave %d.\n", | ||
| 201 | id, adapter->name, dvo->slave_addr); | 203 | id, adapter->name, dvo->slave_addr); |
| 202 | goto out; | 204 | goto out; |
| 203 | } | 205 | } |
| @@ -262,33 +264,33 @@ static void tfp410_dump_regs(struct intel_dvo_device *dvo) | |||
| 262 | uint8_t val, val2; | 264 | uint8_t val, val2; |
| 263 | 265 | ||
| 264 | tfp410_readb(dvo, TFP410_REV, &val); | 266 | tfp410_readb(dvo, TFP410_REV, &val); |
| 265 | DRM_DEBUG("TFP410_REV: 0x%02X\n", val); | 267 | DRM_LOG_KMS("TFP410_REV: 0x%02X\n", val); |
| 266 | tfp410_readb(dvo, TFP410_CTL_1, &val); | 268 | tfp410_readb(dvo, TFP410_CTL_1, &val); |
| 267 | DRM_DEBUG("TFP410_CTL1: 0x%02X\n", val); | 269 | DRM_LOG_KMS("TFP410_CTL1: 0x%02X\n", val); |
| 268 | tfp410_readb(dvo, TFP410_CTL_2, &val); | 270 | tfp410_readb(dvo, TFP410_CTL_2, &val); |
| 269 | DRM_DEBUG("TFP410_CTL2: 0x%02X\n", val); | 271 | DRM_LOG_KMS("TFP410_CTL2: 0x%02X\n", val); |
| 270 | tfp410_readb(dvo, TFP410_CTL_3, &val); | 272 | tfp410_readb(dvo, TFP410_CTL_3, &val); |
| 271 | DRM_DEBUG("TFP410_CTL3: 0x%02X\n", val); | 273 | DRM_LOG_KMS("TFP410_CTL3: 0x%02X\n", val); |
| 272 | tfp410_readb(dvo, TFP410_USERCFG, &val); | 274 | tfp410_readb(dvo, TFP410_USERCFG, &val); |
| 273 | DRM_DEBUG("TFP410_USERCFG: 0x%02X\n", val); | 275 | DRM_LOG_KMS("TFP410_USERCFG: 0x%02X\n", val); |
| 274 | tfp410_readb(dvo, TFP410_DE_DLY, &val); | 276 | tfp410_readb(dvo, TFP410_DE_DLY, &val); |
| 275 | DRM_DEBUG("TFP410_DE_DLY: 0x%02X\n", val); | 277 | DRM_LOG_KMS("TFP410_DE_DLY: 0x%02X\n", val); |
| 276 | tfp410_readb(dvo, TFP410_DE_CTL, &val); | 278 | tfp410_readb(dvo, TFP410_DE_CTL, &val); |
| 277 | DRM_DEBUG("TFP410_DE_CTL: 0x%02X\n", val); | 279 | DRM_LOG_KMS("TFP410_DE_CTL: 0x%02X\n", val); |
| 278 | tfp410_readb(dvo, TFP410_DE_TOP, &val); | 280 | tfp410_readb(dvo, TFP410_DE_TOP, &val); |
| 279 | DRM_DEBUG("TFP410_DE_TOP: 0x%02X\n", val); | 281 | DRM_LOG_KMS("TFP410_DE_TOP: 0x%02X\n", val); |
| 280 | tfp410_readb(dvo, TFP410_DE_CNT_LO, &val); | 282 | tfp410_readb(dvo, TFP410_DE_CNT_LO, &val); |
| 281 | tfp410_readb(dvo, TFP410_DE_CNT_HI, &val2); | 283 | tfp410_readb(dvo, TFP410_DE_CNT_HI, &val2); |
| 282 | DRM_DEBUG("TFP410_DE_CNT: 0x%02X%02X\n", val2, val); | 284 | DRM_LOG_KMS("TFP410_DE_CNT: 0x%02X%02X\n", val2, val); |
| 283 | tfp410_readb(dvo, TFP410_DE_LIN_LO, &val); | 285 | tfp410_readb(dvo, TFP410_DE_LIN_LO, &val); |
| 284 | tfp410_readb(dvo, TFP410_DE_LIN_HI, &val2); | 286 | tfp410_readb(dvo, TFP410_DE_LIN_HI, &val2); |
| 285 | DRM_DEBUG("TFP410_DE_LIN: 0x%02X%02X\n", val2, val); | 287 | DRM_LOG_KMS("TFP410_DE_LIN: 0x%02X%02X\n", val2, val); |
| 286 | tfp410_readb(dvo, TFP410_H_RES_LO, &val); | 288 | tfp410_readb(dvo, TFP410_H_RES_LO, &val); |
| 287 | tfp410_readb(dvo, TFP410_H_RES_HI, &val2); | 289 | tfp410_readb(dvo, TFP410_H_RES_HI, &val2); |
| 288 | DRM_DEBUG("TFP410_H_RES: 0x%02X%02X\n", val2, val); | 290 | DRM_LOG_KMS("TFP410_H_RES: 0x%02X%02X\n", val2, val); |
| 289 | tfp410_readb(dvo, TFP410_V_RES_LO, &val); | 291 | tfp410_readb(dvo, TFP410_V_RES_LO, &val); |
| 290 | tfp410_readb(dvo, TFP410_V_RES_HI, &val2); | 292 | tfp410_readb(dvo, TFP410_V_RES_HI, &val2); |
| 291 | DRM_DEBUG("TFP410_V_RES: 0x%02X%02X\n", val2, val); | 293 | DRM_LOG_KMS("TFP410_V_RES: 0x%02X%02X\n", val2, val); |
| 292 | } | 294 | } |
| 293 | 295 | ||
| 294 | static void tfp410_save(struct intel_dvo_device *dvo) | 296 | static void tfp410_save(struct intel_dvo_device *dvo) |
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index 26bf0552b3cb..18476bf0b580 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c | |||
| @@ -27,6 +27,7 @@ | |||
| 27 | */ | 27 | */ |
| 28 | 28 | ||
| 29 | #include <linux/seq_file.h> | 29 | #include <linux/seq_file.h> |
| 30 | #include <linux/debugfs.h> | ||
| 30 | #include "drmP.h" | 31 | #include "drmP.h" |
| 31 | #include "drm.h" | 32 | #include "drm.h" |
| 32 | #include "i915_drm.h" | 33 | #include "i915_drm.h" |
| @@ -96,13 +97,14 @@ static int i915_gem_object_list_info(struct seq_file *m, void *data) | |||
| 96 | { | 97 | { |
| 97 | struct drm_gem_object *obj = obj_priv->obj; | 98 | struct drm_gem_object *obj = obj_priv->obj; |
| 98 | 99 | ||
| 99 | seq_printf(m, " %p: %s %8zd %08x %08x %d %s", | 100 | seq_printf(m, " %p: %s %8zd %08x %08x %d%s%s", |
| 100 | obj, | 101 | obj, |
| 101 | get_pin_flag(obj_priv), | 102 | get_pin_flag(obj_priv), |
| 102 | obj->size, | 103 | obj->size, |
| 103 | obj->read_domains, obj->write_domain, | 104 | obj->read_domains, obj->write_domain, |
| 104 | obj_priv->last_rendering_seqno, | 105 | obj_priv->last_rendering_seqno, |
| 105 | obj_priv->dirty ? "dirty" : ""); | 106 | obj_priv->dirty ? " dirty" : "", |
| 107 | obj_priv->madv == I915_MADV_DONTNEED ? " purgeable" : ""); | ||
| 106 | 108 | ||
| 107 | if (obj->name) | 109 | if (obj->name) |
| 108 | seq_printf(m, " (name: %d)", obj->name); | 110 | seq_printf(m, " (name: %d)", obj->name); |
| @@ -160,7 +162,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) | |||
| 160 | struct drm_device *dev = node->minor->dev; | 162 | struct drm_device *dev = node->minor->dev; |
| 161 | drm_i915_private_t *dev_priv = dev->dev_private; | 163 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 162 | 164 | ||
| 163 | if (!IS_IGDNG(dev)) { | 165 | if (!IS_IRONLAKE(dev)) { |
| 164 | seq_printf(m, "Interrupt enable: %08x\n", | 166 | seq_printf(m, "Interrupt enable: %08x\n", |
| 165 | I915_READ(IER)); | 167 | I915_READ(IER)); |
| 166 | seq_printf(m, "Interrupt identity: %08x\n", | 168 | seq_printf(m, "Interrupt identity: %08x\n", |
| @@ -412,6 +414,109 @@ static int i915_registers_info(struct seq_file *m, void *data) { | |||
| 412 | return 0; | 414 | return 0; |
| 413 | } | 415 | } |
| 414 | 416 | ||
| 417 | static int | ||
| 418 | i915_wedged_open(struct inode *inode, | ||
| 419 | struct file *filp) | ||
| 420 | { | ||
| 421 | filp->private_data = inode->i_private; | ||
| 422 | return 0; | ||
| 423 | } | ||
| 424 | |||
| 425 | static ssize_t | ||
| 426 | i915_wedged_read(struct file *filp, | ||
| 427 | char __user *ubuf, | ||
| 428 | size_t max, | ||
| 429 | loff_t *ppos) | ||
| 430 | { | ||
| 431 | struct drm_device *dev = filp->private_data; | ||
| 432 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 433 | char buf[80]; | ||
| 434 | int len; | ||
| 435 | |||
| 436 | len = snprintf(buf, sizeof (buf), | ||
| 437 | "wedged : %d\n", | ||
| 438 | atomic_read(&dev_priv->mm.wedged)); | ||
| 439 | |||
| 440 | return simple_read_from_buffer(ubuf, max, ppos, buf, len); | ||
| 441 | } | ||
| 442 | |||
| 443 | static ssize_t | ||
| 444 | i915_wedged_write(struct file *filp, | ||
| 445 | const char __user *ubuf, | ||
| 446 | size_t cnt, | ||
| 447 | loff_t *ppos) | ||
| 448 | { | ||
| 449 | struct drm_device *dev = filp->private_data; | ||
| 450 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 451 | char buf[20]; | ||
| 452 | int val = 1; | ||
| 453 | |||
| 454 | if (cnt > 0) { | ||
| 455 | if (cnt > sizeof (buf) - 1) | ||
| 456 | return -EINVAL; | ||
| 457 | |||
| 458 | if (copy_from_user(buf, ubuf, cnt)) | ||
| 459 | return -EFAULT; | ||
| 460 | buf[cnt] = 0; | ||
| 461 | |||
| 462 | val = simple_strtoul(buf, NULL, 0); | ||
| 463 | } | ||
| 464 | |||
| 465 | DRM_INFO("Manually setting wedged to %d\n", val); | ||
| 466 | |||
| 467 | atomic_set(&dev_priv->mm.wedged, val); | ||
| 468 | if (val) { | ||
| 469 | DRM_WAKEUP(&dev_priv->irq_queue); | ||
| 470 | queue_work(dev_priv->wq, &dev_priv->error_work); | ||
| 471 | } | ||
| 472 | |||
| 473 | return cnt; | ||
| 474 | } | ||
| 475 | |||
| 476 | static const struct file_operations i915_wedged_fops = { | ||
| 477 | .owner = THIS_MODULE, | ||
| 478 | .open = i915_wedged_open, | ||
| 479 | .read = i915_wedged_read, | ||
| 480 | .write = i915_wedged_write, | ||
| 481 | }; | ||
| 482 | |||
| 483 | /* As the drm_debugfs_init() routines are called before dev->dev_private is | ||
| 484 | * allocated we need to hook into the minor for release. */ | ||
| 485 | static int | ||
| 486 | drm_add_fake_info_node(struct drm_minor *minor, | ||
| 487 | struct dentry *ent, | ||
| 488 | const void *key) | ||
| 489 | { | ||
| 490 | struct drm_info_node *node; | ||
| 491 | |||
| 492 | node = kmalloc(sizeof(struct drm_info_node), GFP_KERNEL); | ||
| 493 | if (node == NULL) { | ||
| 494 | debugfs_remove(ent); | ||
| 495 | return -ENOMEM; | ||
| 496 | } | ||
| 497 | |||
| 498 | node->minor = minor; | ||
| 499 | node->dent = ent; | ||
| 500 | node->info_ent = (void *) key; | ||
| 501 | list_add(&node->list, &minor->debugfs_nodes.list); | ||
| 502 | |||
| 503 | return 0; | ||
| 504 | } | ||
| 505 | |||
| 506 | static int i915_wedged_create(struct dentry *root, struct drm_minor *minor) | ||
| 507 | { | ||
| 508 | struct drm_device *dev = minor->dev; | ||
| 509 | struct dentry *ent; | ||
| 510 | |||
| 511 | ent = debugfs_create_file("i915_wedged", | ||
| 512 | S_IRUGO | S_IWUSR, | ||
| 513 | root, dev, | ||
| 514 | &i915_wedged_fops); | ||
| 515 | if (IS_ERR(ent)) | ||
| 516 | return PTR_ERR(ent); | ||
| 517 | |||
| 518 | return drm_add_fake_info_node(minor, ent, &i915_wedged_fops); | ||
| 519 | } | ||
| 415 | 520 | ||
| 416 | static struct drm_info_list i915_debugfs_list[] = { | 521 | static struct drm_info_list i915_debugfs_list[] = { |
| 417 | {"i915_regs", i915_registers_info, 0}, | 522 | {"i915_regs", i915_registers_info, 0}, |
| @@ -432,6 +537,12 @@ static struct drm_info_list i915_debugfs_list[] = { | |||
| 432 | 537 | ||
| 433 | int i915_debugfs_init(struct drm_minor *minor) | 538 | int i915_debugfs_init(struct drm_minor *minor) |
| 434 | { | 539 | { |
| 540 | int ret; | ||
| 541 | |||
| 542 | ret = i915_wedged_create(minor->debugfs_root, minor); | ||
| 543 | if (ret) | ||
| 544 | return ret; | ||
| 545 | |||
| 435 | return drm_debugfs_create_files(i915_debugfs_list, | 546 | return drm_debugfs_create_files(i915_debugfs_list, |
| 436 | I915_DEBUGFS_ENTRIES, | 547 | I915_DEBUGFS_ENTRIES, |
| 437 | minor->debugfs_root, minor); | 548 | minor->debugfs_root, minor); |
| @@ -441,7 +552,8 @@ void i915_debugfs_cleanup(struct drm_minor *minor) | |||
| 441 | { | 552 | { |
| 442 | drm_debugfs_remove_files(i915_debugfs_list, | 553 | drm_debugfs_remove_files(i915_debugfs_list, |
| 443 | I915_DEBUGFS_ENTRIES, minor); | 554 | I915_DEBUGFS_ENTRIES, minor); |
| 555 | drm_debugfs_remove_files((struct drm_info_list *) &i915_wedged_fops, | ||
| 556 | 1, minor); | ||
| 444 | } | 557 | } |
| 445 | 558 | ||
| 446 | #endif /* CONFIG_DEBUG_FS */ | 559 | #endif /* CONFIG_DEBUG_FS */ |
| 447 | |||
diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c index e5b138be45fa..701bfeac7f57 100644 --- a/drivers/gpu/drm/i915/i915_dma.c +++ b/drivers/gpu/drm/i915/i915_dma.c | |||
| @@ -807,6 +807,12 @@ static int i915_getparam(struct drm_device *dev, void *data, | |||
| 807 | case I915_PARAM_NUM_FENCES_AVAIL: | 807 | case I915_PARAM_NUM_FENCES_AVAIL: |
| 808 | value = dev_priv->num_fence_regs - dev_priv->fence_reg_start; | 808 | value = dev_priv->num_fence_regs - dev_priv->fence_reg_start; |
| 809 | break; | 809 | break; |
| 810 | case I915_PARAM_HAS_OVERLAY: | ||
| 811 | value = dev_priv->overlay ? 1 : 0; | ||
| 812 | break; | ||
| 813 | case I915_PARAM_HAS_PAGEFLIPPING: | ||
| 814 | value = 1; | ||
| 815 | break; | ||
| 810 | default: | 816 | default: |
| 811 | DRM_DEBUG_DRIVER("Unknown parameter %d\n", | 817 | DRM_DEBUG_DRIVER("Unknown parameter %d\n", |
| 812 | param->param); | 818 | param->param); |
| @@ -962,7 +968,7 @@ static int i915_probe_agp(struct drm_device *dev, uint32_t *aperture_size, | |||
| 962 | * Some of the preallocated space is taken by the GTT | 968 | * Some of the preallocated space is taken by the GTT |
| 963 | * and popup. GTT is 1K per MB of aperture size, and popup is 4K. | 969 | * and popup. GTT is 1K per MB of aperture size, and popup is 4K. |
| 964 | */ | 970 | */ |
| 965 | if (IS_G4X(dev) || IS_IGD(dev) || IS_IGDNG(dev)) | 971 | if (IS_G4X(dev) || IS_PINEVIEW(dev) || IS_IRONLAKE(dev)) |
| 966 | overhead = 4096; | 972 | overhead = 4096; |
| 967 | else | 973 | else |
| 968 | overhead = (*aperture_size / 1024) + 4096; | 974 | overhead = (*aperture_size / 1024) + 4096; |
| @@ -1048,7 +1054,7 @@ static unsigned long i915_gtt_to_phys(struct drm_device *dev, | |||
| 1048 | int gtt_offset, gtt_size; | 1054 | int gtt_offset, gtt_size; |
| 1049 | 1055 | ||
| 1050 | if (IS_I965G(dev)) { | 1056 | if (IS_I965G(dev)) { |
| 1051 | if (IS_G4X(dev) || IS_IGDNG(dev)) { | 1057 | if (IS_G4X(dev) || IS_IRONLAKE(dev)) { |
| 1052 | gtt_offset = 2*1024*1024; | 1058 | gtt_offset = 2*1024*1024; |
| 1053 | gtt_size = 2*1024*1024; | 1059 | gtt_size = 2*1024*1024; |
| 1054 | } else { | 1060 | } else { |
| @@ -1070,7 +1076,7 @@ static unsigned long i915_gtt_to_phys(struct drm_device *dev, | |||
| 1070 | 1076 | ||
| 1071 | entry = *(volatile u32 *)(gtt + (gtt_addr / 1024)); | 1077 | entry = *(volatile u32 *)(gtt + (gtt_addr / 1024)); |
| 1072 | 1078 | ||
| 1073 | DRM_DEBUG("GTT addr: 0x%08lx, PTE: 0x%08lx\n", gtt_addr, entry); | 1079 | DRM_DEBUG_DRIVER("GTT addr: 0x%08lx, PTE: 0x%08lx\n", gtt_addr, entry); |
| 1074 | 1080 | ||
| 1075 | /* Mask out these reserved bits on this hardware. */ | 1081 | /* Mask out these reserved bits on this hardware. */ |
| 1076 | if (!IS_I9XX(dev) || IS_I915G(dev) || IS_I915GM(dev) || | 1082 | if (!IS_I9XX(dev) || IS_I915G(dev) || IS_I915GM(dev) || |
| @@ -1096,7 +1102,7 @@ static unsigned long i915_gtt_to_phys(struct drm_device *dev, | |||
| 1096 | phys =(entry & PTE_ADDRESS_MASK) | | 1102 | phys =(entry & PTE_ADDRESS_MASK) | |
| 1097 | ((uint64_t)(entry & PTE_ADDRESS_MASK_HIGH) << (32 - 4)); | 1103 | ((uint64_t)(entry & PTE_ADDRESS_MASK_HIGH) << (32 - 4)); |
| 1098 | 1104 | ||
| 1099 | DRM_DEBUG("GTT addr: 0x%08lx, phys addr: 0x%08lx\n", gtt_addr, phys); | 1105 | DRM_DEBUG_DRIVER("GTT addr: 0x%08lx, phys addr: 0x%08lx\n", gtt_addr, phys); |
| 1100 | 1106 | ||
| 1101 | return phys; | 1107 | return phys; |
| 1102 | } | 1108 | } |
| @@ -1306,7 +1312,7 @@ static void i915_get_mem_freq(struct drm_device *dev) | |||
| 1306 | drm_i915_private_t *dev_priv = dev->dev_private; | 1312 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1307 | u32 tmp; | 1313 | u32 tmp; |
| 1308 | 1314 | ||
| 1309 | if (!IS_IGD(dev)) | 1315 | if (!IS_PINEVIEW(dev)) |
| 1310 | return; | 1316 | return; |
| 1311 | 1317 | ||
| 1312 | tmp = I915_READ(CLKCFG); | 1318 | tmp = I915_READ(CLKCFG); |
| @@ -1413,7 +1419,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
| 1413 | if (ret) | 1419 | if (ret) |
| 1414 | goto out_iomapfree; | 1420 | goto out_iomapfree; |
| 1415 | 1421 | ||
| 1416 | dev_priv->wq = create_workqueue("i915"); | 1422 | dev_priv->wq = create_singlethread_workqueue("i915"); |
| 1417 | if (dev_priv->wq == NULL) { | 1423 | if (dev_priv->wq == NULL) { |
| 1418 | DRM_ERROR("Failed to create our workqueue.\n"); | 1424 | DRM_ERROR("Failed to create our workqueue.\n"); |
| 1419 | ret = -ENOMEM; | 1425 | ret = -ENOMEM; |
| @@ -1434,7 +1440,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
| 1434 | 1440 | ||
| 1435 | dev->driver->get_vblank_counter = i915_get_vblank_counter; | 1441 | dev->driver->get_vblank_counter = i915_get_vblank_counter; |
| 1436 | dev->max_vblank_count = 0xffffff; /* only 24 bits of frame count */ | 1442 | dev->max_vblank_count = 0xffffff; /* only 24 bits of frame count */ |
| 1437 | if (IS_G4X(dev) || IS_IGDNG(dev)) { | 1443 | if (IS_G4X(dev) || IS_IRONLAKE(dev)) { |
| 1438 | dev->max_vblank_count = 0xffffffff; /* full 32 bit counter */ | 1444 | dev->max_vblank_count = 0xffffffff; /* full 32 bit counter */ |
| 1439 | dev->driver->get_vblank_counter = gm45_get_vblank_counter; | 1445 | dev->driver->get_vblank_counter = gm45_get_vblank_counter; |
| 1440 | } | 1446 | } |
| @@ -1489,9 +1495,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
| 1489 | } | 1495 | } |
| 1490 | 1496 | ||
| 1491 | /* Must be done after probing outputs */ | 1497 | /* Must be done after probing outputs */ |
| 1492 | /* FIXME: verify on IGDNG */ | 1498 | intel_opregion_init(dev, 0); |
| 1493 | if (!IS_IGDNG(dev)) | ||
| 1494 | intel_opregion_init(dev, 0); | ||
| 1495 | 1499 | ||
| 1496 | setup_timer(&dev_priv->hangcheck_timer, i915_hangcheck_elapsed, | 1500 | setup_timer(&dev_priv->hangcheck_timer, i915_hangcheck_elapsed, |
| 1497 | (unsigned long) dev); | 1501 | (unsigned long) dev); |
| @@ -1525,6 +1529,15 @@ int i915_driver_unload(struct drm_device *dev) | |||
| 1525 | } | 1529 | } |
| 1526 | 1530 | ||
| 1527 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { | 1531 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { |
| 1532 | /* | ||
| 1533 | * free the memory space allocated for the child device | ||
| 1534 | * config parsed from VBT | ||
| 1535 | */ | ||
| 1536 | if (dev_priv->child_dev && dev_priv->child_dev_num) { | ||
| 1537 | kfree(dev_priv->child_dev); | ||
| 1538 | dev_priv->child_dev = NULL; | ||
| 1539 | dev_priv->child_dev_num = 0; | ||
| 1540 | } | ||
| 1528 | drm_irq_uninstall(dev); | 1541 | drm_irq_uninstall(dev); |
| 1529 | vga_client_register(dev->pdev, NULL, NULL, NULL); | 1542 | vga_client_register(dev->pdev, NULL, NULL, NULL); |
| 1530 | } | 1543 | } |
| @@ -1535,8 +1548,7 @@ int i915_driver_unload(struct drm_device *dev) | |||
| 1535 | if (dev_priv->regs != NULL) | 1548 | if (dev_priv->regs != NULL) |
| 1536 | iounmap(dev_priv->regs); | 1549 | iounmap(dev_priv->regs); |
| 1537 | 1550 | ||
| 1538 | if (!IS_IGDNG(dev)) | 1551 | intel_opregion_free(dev, 0); |
| 1539 | intel_opregion_free(dev, 0); | ||
| 1540 | 1552 | ||
| 1541 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { | 1553 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { |
| 1542 | intel_modeset_cleanup(dev); | 1554 | intel_modeset_cleanup(dev); |
| @@ -1548,6 +1560,8 @@ int i915_driver_unload(struct drm_device *dev) | |||
| 1548 | mutex_unlock(&dev->struct_mutex); | 1560 | mutex_unlock(&dev->struct_mutex); |
| 1549 | drm_mm_takedown(&dev_priv->vram); | 1561 | drm_mm_takedown(&dev_priv->vram); |
| 1550 | i915_gem_lastclose(dev); | 1562 | i915_gem_lastclose(dev); |
| 1563 | |||
| 1564 | intel_cleanup_overlay(dev); | ||
| 1551 | } | 1565 | } |
| 1552 | 1566 | ||
| 1553 | pci_dev_put(dev_priv->bridge_dev); | 1567 | pci_dev_put(dev_priv->bridge_dev); |
| @@ -1656,6 +1670,8 @@ struct drm_ioctl_desc i915_ioctls[] = { | |||
| 1656 | DRM_IOCTL_DEF(DRM_I915_GEM_GET_APERTURE, i915_gem_get_aperture_ioctl, 0), | 1670 | DRM_IOCTL_DEF(DRM_I915_GEM_GET_APERTURE, i915_gem_get_aperture_ioctl, 0), |
| 1657 | DRM_IOCTL_DEF(DRM_I915_GET_PIPE_FROM_CRTC_ID, intel_get_pipe_from_crtc_id, 0), | 1671 | DRM_IOCTL_DEF(DRM_I915_GET_PIPE_FROM_CRTC_ID, intel_get_pipe_from_crtc_id, 0), |
| 1658 | DRM_IOCTL_DEF(DRM_I915_GEM_MADVISE, i915_gem_madvise_ioctl, 0), | 1672 | DRM_IOCTL_DEF(DRM_I915_GEM_MADVISE, i915_gem_madvise_ioctl, 0), |
| 1673 | DRM_IOCTL_DEF(DRM_I915_OVERLAY_PUT_IMAGE, intel_overlay_put_image, DRM_MASTER|DRM_CONTROL_ALLOW), | ||
| 1674 | DRM_IOCTL_DEF(DRM_I915_OVERLAY_ATTRS, intel_overlay_attrs, DRM_MASTER|DRM_CONTROL_ALLOW), | ||
| 1659 | }; | 1675 | }; |
| 1660 | 1676 | ||
| 1661 | int i915_max_ioctl = DRM_ARRAY_SIZE(i915_ioctls); | 1677 | int i915_max_ioctl = DRM_ARRAY_SIZE(i915_ioctls); |
diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c index 7f436ec075f6..2fa217862058 100644 --- a/drivers/gpu/drm/i915/i915_drv.c +++ b/drivers/gpu/drm/i915/i915_drv.c | |||
| @@ -333,6 +333,7 @@ static struct drm_driver driver = { | |||
| 333 | .mmap = drm_gem_mmap, | 333 | .mmap = drm_gem_mmap, |
| 334 | .poll = drm_poll, | 334 | .poll = drm_poll, |
| 335 | .fasync = drm_fasync, | 335 | .fasync = drm_fasync, |
| 336 | .read = drm_read, | ||
| 336 | #ifdef CONFIG_COMPAT | 337 | #ifdef CONFIG_COMPAT |
| 337 | .compat_ioctl = i915_compat_ioctl, | 338 | .compat_ioctl = i915_compat_ioctl, |
| 338 | #endif | 339 | #endif |
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index a725f6591192..fbecac72f5bb 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
| @@ -170,6 +170,8 @@ struct drm_i915_display_funcs { | |||
| 170 | /* clock gating init */ | 170 | /* clock gating init */ |
| 171 | }; | 171 | }; |
| 172 | 172 | ||
| 173 | struct intel_overlay; | ||
| 174 | |||
| 173 | typedef struct drm_i915_private { | 175 | typedef struct drm_i915_private { |
| 174 | struct drm_device *dev; | 176 | struct drm_device *dev; |
| 175 | 177 | ||
| @@ -187,6 +189,7 @@ typedef struct drm_i915_private { | |||
| 187 | unsigned int status_gfx_addr; | 189 | unsigned int status_gfx_addr; |
| 188 | drm_local_map_t hws_map; | 190 | drm_local_map_t hws_map; |
| 189 | struct drm_gem_object *hws_obj; | 191 | struct drm_gem_object *hws_obj; |
| 192 | struct drm_gem_object *pwrctx; | ||
| 190 | 193 | ||
| 191 | struct resource mch_res; | 194 | struct resource mch_res; |
| 192 | 195 | ||
| @@ -206,11 +209,13 @@ typedef struct drm_i915_private { | |||
| 206 | /** Cached value of IMR to avoid reads in updating the bitfield */ | 209 | /** Cached value of IMR to avoid reads in updating the bitfield */ |
| 207 | u32 irq_mask_reg; | 210 | u32 irq_mask_reg; |
| 208 | u32 pipestat[2]; | 211 | u32 pipestat[2]; |
| 209 | /** splitted irq regs for graphics and display engine on IGDNG, | 212 | /** splitted irq regs for graphics and display engine on Ironlake, |
| 210 | irq_mask_reg is still used for display irq. */ | 213 | irq_mask_reg is still used for display irq. */ |
| 211 | u32 gt_irq_mask_reg; | 214 | u32 gt_irq_mask_reg; |
| 212 | u32 gt_irq_enable_reg; | 215 | u32 gt_irq_enable_reg; |
| 213 | u32 de_irq_enable_reg; | 216 | u32 de_irq_enable_reg; |
| 217 | u32 pch_irq_mask_reg; | ||
| 218 | u32 pch_irq_enable_reg; | ||
| 214 | 219 | ||
| 215 | u32 hotplug_supported_mask; | 220 | u32 hotplug_supported_mask; |
| 216 | struct work_struct hotplug_work; | 221 | struct work_struct hotplug_work; |
| @@ -240,6 +245,9 @@ typedef struct drm_i915_private { | |||
| 240 | 245 | ||
| 241 | struct intel_opregion opregion; | 246 | struct intel_opregion opregion; |
| 242 | 247 | ||
| 248 | /* overlay */ | ||
| 249 | struct intel_overlay *overlay; | ||
| 250 | |||
| 243 | /* LVDS info */ | 251 | /* LVDS info */ |
| 244 | int backlight_duty_cycle; /* restore backlight to this value */ | 252 | int backlight_duty_cycle; /* restore backlight to this value */ |
| 245 | bool panel_wants_dither; | 253 | bool panel_wants_dither; |
| @@ -258,7 +266,7 @@ typedef struct drm_i915_private { | |||
| 258 | 266 | ||
| 259 | struct notifier_block lid_notifier; | 267 | struct notifier_block lid_notifier; |
| 260 | 268 | ||
| 261 | int crt_ddc_bus; /* -1 = unknown, else GPIO to use for CRT DDC */ | 269 | int crt_ddc_bus; /* 0 = unknown, else GPIO to use for CRT DDC */ |
| 262 | struct drm_i915_fence_reg fence_regs[16]; /* assume 965 */ | 270 | struct drm_i915_fence_reg fence_regs[16]; /* assume 965 */ |
| 263 | int fence_reg_start; /* 4 if userland hasn't ioctl'd us yet */ | 271 | int fence_reg_start; /* 4 if userland hasn't ioctl'd us yet */ |
| 264 | int num_fence_regs; /* 8 on pre-965, 16 otherwise */ | 272 | int num_fence_regs; /* 8 on pre-965, 16 otherwise */ |
| @@ -280,6 +288,7 @@ typedef struct drm_i915_private { | |||
| 280 | u32 saveDSPBCNTR; | 288 | u32 saveDSPBCNTR; |
| 281 | u32 saveDSPARB; | 289 | u32 saveDSPARB; |
| 282 | u32 saveRENDERSTANDBY; | 290 | u32 saveRENDERSTANDBY; |
| 291 | u32 savePWRCTXA; | ||
| 283 | u32 saveHWS; | 292 | u32 saveHWS; |
| 284 | u32 savePIPEACONF; | 293 | u32 savePIPEACONF; |
| 285 | u32 savePIPEBCONF; | 294 | u32 savePIPEBCONF; |
| @@ -374,8 +383,6 @@ typedef struct drm_i915_private { | |||
| 374 | u32 saveFDI_RXA_IMR; | 383 | u32 saveFDI_RXA_IMR; |
| 375 | u32 saveFDI_RXB_IMR; | 384 | u32 saveFDI_RXB_IMR; |
| 376 | u32 saveCACHE_MODE_0; | 385 | u32 saveCACHE_MODE_0; |
| 377 | u32 saveD_STATE; | ||
| 378 | u32 saveDSPCLK_GATE_D; | ||
| 379 | u32 saveMI_ARB_STATE; | 386 | u32 saveMI_ARB_STATE; |
| 380 | u32 saveSWF0[16]; | 387 | u32 saveSWF0[16]; |
| 381 | u32 saveSWF1[16]; | 388 | u32 saveSWF1[16]; |
| @@ -539,13 +546,21 @@ typedef struct drm_i915_private { | |||
| 539 | /* indicate whether the LVDS_BORDER should be enabled or not */ | 546 | /* indicate whether the LVDS_BORDER should be enabled or not */ |
| 540 | unsigned int lvds_border_bits; | 547 | unsigned int lvds_border_bits; |
| 541 | 548 | ||
| 549 | struct drm_crtc *plane_to_crtc_mapping[2]; | ||
| 550 | struct drm_crtc *pipe_to_crtc_mapping[2]; | ||
| 551 | wait_queue_head_t pending_flip_queue; | ||
| 552 | |||
| 542 | /* Reclocking support */ | 553 | /* Reclocking support */ |
| 543 | bool render_reclock_avail; | 554 | bool render_reclock_avail; |
| 544 | bool lvds_downclock_avail; | 555 | bool lvds_downclock_avail; |
| 556 | /* indicates the reduced downclock for LVDS*/ | ||
| 557 | int lvds_downclock; | ||
| 545 | struct work_struct idle_work; | 558 | struct work_struct idle_work; |
| 546 | struct timer_list idle_timer; | 559 | struct timer_list idle_timer; |
| 547 | bool busy; | 560 | bool busy; |
| 548 | u16 orig_clock; | 561 | u16 orig_clock; |
| 562 | int child_dev_num; | ||
| 563 | struct child_device_config *child_dev; | ||
| 549 | } drm_i915_private_t; | 564 | } drm_i915_private_t; |
| 550 | 565 | ||
| 551 | /** driver private structure attached to each drm_gem_object */ | 566 | /** driver private structure attached to each drm_gem_object */ |
| @@ -638,6 +653,13 @@ struct drm_i915_gem_object { | |||
| 638 | * Advice: are the backing pages purgeable? | 653 | * Advice: are the backing pages purgeable? |
| 639 | */ | 654 | */ |
| 640 | int madv; | 655 | int madv; |
| 656 | |||
| 657 | /** | ||
| 658 | * Number of crtcs where this object is currently the fb, but | ||
| 659 | * will be page flipped away on the next vblank. When it | ||
| 660 | * reaches 0, dev_priv->pending_flip_queue will be woken up. | ||
| 661 | */ | ||
| 662 | atomic_t pending_flip; | ||
| 641 | }; | 663 | }; |
| 642 | 664 | ||
| 643 | /** | 665 | /** |
| @@ -738,6 +760,8 @@ i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); | |||
| 738 | void | 760 | void |
| 739 | i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); | 761 | i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
| 740 | 762 | ||
| 763 | void intel_enable_asle (struct drm_device *dev); | ||
| 764 | |||
| 741 | 765 | ||
| 742 | /* i915_mem.c */ | 766 | /* i915_mem.c */ |
| 743 | extern int i915_mem_alloc(struct drm_device *dev, void *data, | 767 | extern int i915_mem_alloc(struct drm_device *dev, void *data, |
| @@ -813,6 +837,9 @@ void i915_gem_cleanup_ringbuffer(struct drm_device *dev); | |||
| 813 | int i915_gem_do_init(struct drm_device *dev, unsigned long start, | 837 | int i915_gem_do_init(struct drm_device *dev, unsigned long start, |
| 814 | unsigned long end); | 838 | unsigned long end); |
| 815 | int i915_gem_idle(struct drm_device *dev); | 839 | int i915_gem_idle(struct drm_device *dev); |
| 840 | uint32_t i915_add_request(struct drm_device *dev, struct drm_file *file_priv, | ||
| 841 | uint32_t flush_domains); | ||
| 842 | int i915_do_wait_request(struct drm_device *dev, uint32_t seqno, int interruptible); | ||
| 816 | int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); | 843 | int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); |
| 817 | int i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, | 844 | int i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj, |
| 818 | int write); | 845 | int write); |
| @@ -824,6 +851,7 @@ void i915_gem_free_all_phys_object(struct drm_device *dev); | |||
| 824 | int i915_gem_object_get_pages(struct drm_gem_object *obj); | 851 | int i915_gem_object_get_pages(struct drm_gem_object *obj); |
| 825 | void i915_gem_object_put_pages(struct drm_gem_object *obj); | 852 | void i915_gem_object_put_pages(struct drm_gem_object *obj); |
| 826 | void i915_gem_release(struct drm_device * dev, struct drm_file *file_priv); | 853 | void i915_gem_release(struct drm_device * dev, struct drm_file *file_priv); |
| 854 | void i915_gem_object_flush_write_domain(struct drm_gem_object *obj); | ||
| 827 | 855 | ||
| 828 | void i915_gem_shrinker_init(void); | 856 | void i915_gem_shrinker_init(void); |
| 829 | void i915_gem_shrinker_exit(void); | 857 | void i915_gem_shrinker_exit(void); |
| @@ -863,11 +891,13 @@ extern int i915_restore_state(struct drm_device *dev); | |||
| 863 | extern int intel_opregion_init(struct drm_device *dev, int resume); | 891 | extern int intel_opregion_init(struct drm_device *dev, int resume); |
| 864 | extern void intel_opregion_free(struct drm_device *dev, int suspend); | 892 | extern void intel_opregion_free(struct drm_device *dev, int suspend); |
| 865 | extern void opregion_asle_intr(struct drm_device *dev); | 893 | extern void opregion_asle_intr(struct drm_device *dev); |
| 894 | extern void ironlake_opregion_gse_intr(struct drm_device *dev); | ||
| 866 | extern void opregion_enable_asle(struct drm_device *dev); | 895 | extern void opregion_enable_asle(struct drm_device *dev); |
| 867 | #else | 896 | #else |
| 868 | static inline int intel_opregion_init(struct drm_device *dev, int resume) { return 0; } | 897 | static inline int intel_opregion_init(struct drm_device *dev, int resume) { return 0; } |
| 869 | static inline void intel_opregion_free(struct drm_device *dev, int suspend) { return; } | 898 | static inline void intel_opregion_free(struct drm_device *dev, int suspend) { return; } |
| 870 | static inline void opregion_asle_intr(struct drm_device *dev) { return; } | 899 | static inline void opregion_asle_intr(struct drm_device *dev) { return; } |
| 900 | static inline void ironlake_opregion_gse_intr(struct drm_device *dev) { return; } | ||
| 871 | static inline void opregion_enable_asle(struct drm_device *dev) { return; } | 901 | static inline void opregion_enable_asle(struct drm_device *dev) { return; } |
| 872 | #endif | 902 | #endif |
| 873 | 903 | ||
| @@ -955,8 +985,8 @@ extern int i915_wait_ring(struct drm_device * dev, int n, const char *caller); | |||
| 955 | #define IS_I830(dev) ((dev)->pci_device == 0x3577) | 985 | #define IS_I830(dev) ((dev)->pci_device == 0x3577) |
| 956 | #define IS_845G(dev) ((dev)->pci_device == 0x2562) | 986 | #define IS_845G(dev) ((dev)->pci_device == 0x2562) |
| 957 | #define IS_I85X(dev) ((dev)->pci_device == 0x3582) | 987 | #define IS_I85X(dev) ((dev)->pci_device == 0x3582) |
| 958 | #define IS_I855(dev) ((dev)->pci_device == 0x3582) | ||
| 959 | #define IS_I865G(dev) ((dev)->pci_device == 0x2572) | 988 | #define IS_I865G(dev) ((dev)->pci_device == 0x2572) |
| 989 | #define IS_I8XX(dev) (IS_I830(dev) || IS_845G(dev) || IS_I85X(dev) || IS_I865G(dev)) | ||
| 960 | 990 | ||
| 961 | #define IS_I915G(dev) ((dev)->pci_device == 0x2582 || (dev)->pci_device == 0x258a) | 991 | #define IS_I915G(dev) ((dev)->pci_device == 0x2582 || (dev)->pci_device == 0x258a) |
| 962 | #define IS_I915GM(dev) ((dev)->pci_device == 0x2592) | 992 | #define IS_I915GM(dev) ((dev)->pci_device == 0x2592) |
| @@ -990,47 +1020,51 @@ extern int i915_wait_ring(struct drm_device * dev, int n, const char *caller); | |||
| 990 | (dev)->pci_device == 0x2E42 || \ | 1020 | (dev)->pci_device == 0x2E42 || \ |
| 991 | IS_GM45(dev)) | 1021 | IS_GM45(dev)) |
| 992 | 1022 | ||
| 993 | #define IS_IGDG(dev) ((dev)->pci_device == 0xa001) | 1023 | #define IS_PINEVIEW_G(dev) ((dev)->pci_device == 0xa001) |
| 994 | #define IS_IGDGM(dev) ((dev)->pci_device == 0xa011) | 1024 | #define IS_PINEVIEW_M(dev) ((dev)->pci_device == 0xa011) |
| 995 | #define IS_IGD(dev) (IS_IGDG(dev) || IS_IGDGM(dev)) | 1025 | #define IS_PINEVIEW(dev) (IS_PINEVIEW_G(dev) || IS_PINEVIEW_M(dev)) |
| 996 | 1026 | ||
| 997 | #define IS_G33(dev) ((dev)->pci_device == 0x29C2 || \ | 1027 | #define IS_G33(dev) ((dev)->pci_device == 0x29C2 || \ |
| 998 | (dev)->pci_device == 0x29B2 || \ | 1028 | (dev)->pci_device == 0x29B2 || \ |
| 999 | (dev)->pci_device == 0x29D2 || \ | 1029 | (dev)->pci_device == 0x29D2 || \ |
| 1000 | (IS_IGD(dev))) | 1030 | (IS_PINEVIEW(dev))) |
| 1001 | 1031 | ||
| 1002 | #define IS_IGDNG_D(dev) ((dev)->pci_device == 0x0042) | 1032 | #define IS_IRONLAKE_D(dev) ((dev)->pci_device == 0x0042) |
| 1003 | #define IS_IGDNG_M(dev) ((dev)->pci_device == 0x0046) | 1033 | #define IS_IRONLAKE_M(dev) ((dev)->pci_device == 0x0046) |
| 1004 | #define IS_IGDNG(dev) (IS_IGDNG_D(dev) || IS_IGDNG_M(dev)) | 1034 | #define IS_IRONLAKE(dev) (IS_IRONLAKE_D(dev) || IS_IRONLAKE_M(dev)) |
| 1005 | 1035 | ||
| 1006 | #define IS_I9XX(dev) (IS_I915G(dev) || IS_I915GM(dev) || IS_I945G(dev) || \ | 1036 | #define IS_I9XX(dev) (IS_I915G(dev) || IS_I915GM(dev) || IS_I945G(dev) || \ |
| 1007 | IS_I945GM(dev) || IS_I965G(dev) || IS_G33(dev) || \ | 1037 | IS_I945GM(dev) || IS_I965G(dev) || IS_G33(dev) || \ |
| 1008 | IS_IGDNG(dev)) | 1038 | IS_IRONLAKE(dev)) |
| 1009 | 1039 | ||
| 1010 | #define IS_MOBILE(dev) (IS_I830(dev) || IS_I85X(dev) || IS_I915GM(dev) || \ | 1040 | #define IS_MOBILE(dev) (IS_I830(dev) || IS_I85X(dev) || IS_I915GM(dev) || \ |
| 1011 | IS_I945GM(dev) || IS_I965GM(dev) || IS_GM45(dev) || \ | 1041 | IS_I945GM(dev) || IS_I965GM(dev) || IS_GM45(dev) || \ |
| 1012 | IS_IGD(dev) || IS_IGDNG_M(dev)) | 1042 | IS_PINEVIEW(dev) || IS_IRONLAKE_M(dev)) |
| 1013 | 1043 | ||
| 1014 | #define I915_NEED_GFX_HWS(dev) (IS_G33(dev) || IS_GM45(dev) || IS_G4X(dev) || \ | 1044 | #define I915_NEED_GFX_HWS(dev) (IS_G33(dev) || IS_GM45(dev) || IS_G4X(dev) || \ |
| 1015 | IS_IGDNG(dev)) | 1045 | IS_IRONLAKE(dev)) |
| 1016 | /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte | 1046 | /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte |
| 1017 | * rows, which changed the alignment requirements and fence programming. | 1047 | * rows, which changed the alignment requirements and fence programming. |
| 1018 | */ | 1048 | */ |
| 1019 | #define HAS_128_BYTE_Y_TILING(dev) (IS_I9XX(dev) && !(IS_I915G(dev) || \ | 1049 | #define HAS_128_BYTE_Y_TILING(dev) (IS_I9XX(dev) && !(IS_I915G(dev) || \ |
| 1020 | IS_I915GM(dev))) | 1050 | IS_I915GM(dev))) |
| 1021 | #define SUPPORTS_INTEGRATED_HDMI(dev) (IS_G4X(dev) || IS_IGDNG(dev)) | 1051 | #define SUPPORTS_DIGITAL_OUTPUTS(dev) (IS_I9XX(dev) && !IS_PINEVIEW(dev)) |
| 1022 | #define SUPPORTS_INTEGRATED_DP(dev) (IS_G4X(dev) || IS_IGDNG(dev)) | 1052 | #define SUPPORTS_INTEGRATED_HDMI(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
| 1023 | #define SUPPORTS_EDP(dev) (IS_IGDNG_M(dev)) | 1053 | #define SUPPORTS_INTEGRATED_DP(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
| 1054 | #define SUPPORTS_EDP(dev) (IS_IRONLAKE_M(dev)) | ||
| 1055 | #define SUPPORTS_TV(dev) (IS_I9XX(dev) && IS_MOBILE(dev) && \ | ||
| 1056 | !IS_IRONLAKE(dev) && !IS_PINEVIEW(dev)) | ||
| 1024 | #define I915_HAS_HOTPLUG(dev) (IS_I945G(dev) || IS_I945GM(dev) || IS_G33(dev) || IS_I965G(dev)) | 1057 | #define I915_HAS_HOTPLUG(dev) (IS_I945G(dev) || IS_I945GM(dev) || IS_G33(dev) || IS_I965G(dev)) |
| 1025 | /* dsparb controlled by hw only */ | 1058 | /* dsparb controlled by hw only */ |
| 1026 | #define DSPARB_HWCONTROL(dev) (IS_G4X(dev) || IS_IGDNG(dev)) | 1059 | #define DSPARB_HWCONTROL(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
| 1027 | 1060 | ||
| 1028 | #define HAS_FW_BLC(dev) (IS_I9XX(dev) || IS_G4X(dev) || IS_IGDNG(dev)) | 1061 | #define HAS_FW_BLC(dev) (IS_I9XX(dev) || IS_G4X(dev) || IS_IRONLAKE(dev)) |
| 1029 | #define HAS_PIPE_CXSR(dev) (IS_G4X(dev) || IS_IGDNG(dev)) | 1062 | #define HAS_PIPE_CXSR(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
| 1030 | #define I915_HAS_FBC(dev) (IS_MOBILE(dev) && \ | 1063 | #define I915_HAS_FBC(dev) (IS_MOBILE(dev) && \ |
| 1031 | (IS_I9XX(dev) || IS_GM45(dev)) && \ | 1064 | (IS_I9XX(dev) || IS_GM45(dev)) && \ |
| 1032 | !IS_IGD(dev) && \ | 1065 | !IS_PINEVIEW(dev) && \ |
| 1033 | !IS_IGDNG(dev)) | 1066 | !IS_IRONLAKE(dev)) |
| 1067 | #define I915_HAS_RC6(dev) (IS_I965GM(dev) || IS_GM45(dev) || IS_IRONLAKE_M(dev)) | ||
| 1034 | 1068 | ||
| 1035 | #define PRIMARY_RINGBUFFER_SIZE (128*1024) | 1069 | #define PRIMARY_RINGBUFFER_SIZE (128*1024) |
| 1036 | 1070 | ||
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index a2a3fa599923..8c463cf2050a 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
| @@ -1288,6 +1288,7 @@ i915_gem_create_mmap_offset(struct drm_gem_object *obj) | |||
| 1288 | list->hash.key = list->file_offset_node->start; | 1288 | list->hash.key = list->file_offset_node->start; |
| 1289 | if (drm_ht_insert_item(&mm->offset_hash, &list->hash)) { | 1289 | if (drm_ht_insert_item(&mm->offset_hash, &list->hash)) { |
| 1290 | DRM_ERROR("failed to add to map hash\n"); | 1290 | DRM_ERROR("failed to add to map hash\n"); |
| 1291 | ret = -ENOMEM; | ||
| 1291 | goto out_free_mm; | 1292 | goto out_free_mm; |
| 1292 | } | 1293 | } |
| 1293 | 1294 | ||
| @@ -1583,7 +1584,7 @@ i915_gem_object_move_to_inactive(struct drm_gem_object *obj) | |||
| 1583 | * | 1584 | * |
| 1584 | * Returned sequence numbers are nonzero on success. | 1585 | * Returned sequence numbers are nonzero on success. |
| 1585 | */ | 1586 | */ |
| 1586 | static uint32_t | 1587 | uint32_t |
| 1587 | i915_add_request(struct drm_device *dev, struct drm_file *file_priv, | 1588 | i915_add_request(struct drm_device *dev, struct drm_file *file_priv, |
| 1588 | uint32_t flush_domains) | 1589 | uint32_t flush_domains) |
| 1589 | { | 1590 | { |
| @@ -1617,7 +1618,7 @@ i915_add_request(struct drm_device *dev, struct drm_file *file_priv, | |||
| 1617 | OUT_RING(MI_USER_INTERRUPT); | 1618 | OUT_RING(MI_USER_INTERRUPT); |
| 1618 | ADVANCE_LP_RING(); | 1619 | ADVANCE_LP_RING(); |
| 1619 | 1620 | ||
| 1620 | DRM_DEBUG("%d\n", seqno); | 1621 | DRM_DEBUG_DRIVER("%d\n", seqno); |
| 1621 | 1622 | ||
| 1622 | request->seqno = seqno; | 1623 | request->seqno = seqno; |
| 1623 | request->emitted_jiffies = jiffies; | 1624 | request->emitted_jiffies = jiffies; |
| @@ -1820,12 +1821,8 @@ i915_gem_retire_work_handler(struct work_struct *work) | |||
| 1820 | mutex_unlock(&dev->struct_mutex); | 1821 | mutex_unlock(&dev->struct_mutex); |
| 1821 | } | 1822 | } |
| 1822 | 1823 | ||
| 1823 | /** | 1824 | int |
| 1824 | * Waits for a sequence number to be signaled, and cleans up the | 1825 | i915_do_wait_request(struct drm_device *dev, uint32_t seqno, int interruptible) |
| 1825 | * request and object lists appropriately for that event. | ||
| 1826 | */ | ||
| 1827 | static int | ||
| 1828 | i915_wait_request(struct drm_device *dev, uint32_t seqno) | ||
| 1829 | { | 1826 | { |
| 1830 | drm_i915_private_t *dev_priv = dev->dev_private; | 1827 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 1831 | u32 ier; | 1828 | u32 ier; |
| @@ -1837,7 +1834,7 @@ i915_wait_request(struct drm_device *dev, uint32_t seqno) | |||
| 1837 | return -EIO; | 1834 | return -EIO; |
| 1838 | 1835 | ||
| 1839 | if (!i915_seqno_passed(i915_get_gem_seqno(dev), seqno)) { | 1836 | if (!i915_seqno_passed(i915_get_gem_seqno(dev), seqno)) { |
| 1840 | if (IS_IGDNG(dev)) | 1837 | if (IS_IRONLAKE(dev)) |
| 1841 | ier = I915_READ(DEIER) | I915_READ(GTIER); | 1838 | ier = I915_READ(DEIER) | I915_READ(GTIER); |
| 1842 | else | 1839 | else |
| 1843 | ier = I915_READ(IER); | 1840 | ier = I915_READ(IER); |
| @@ -1852,10 +1849,15 @@ i915_wait_request(struct drm_device *dev, uint32_t seqno) | |||
| 1852 | 1849 | ||
| 1853 | dev_priv->mm.waiting_gem_seqno = seqno; | 1850 | dev_priv->mm.waiting_gem_seqno = seqno; |
| 1854 | i915_user_irq_get(dev); | 1851 | i915_user_irq_get(dev); |
| 1855 | ret = wait_event_interruptible(dev_priv->irq_queue, | 1852 | if (interruptible) |
| 1856 | i915_seqno_passed(i915_get_gem_seqno(dev), | 1853 | ret = wait_event_interruptible(dev_priv->irq_queue, |
| 1857 | seqno) || | 1854 | i915_seqno_passed(i915_get_gem_seqno(dev), seqno) || |
| 1858 | atomic_read(&dev_priv->mm.wedged)); | 1855 | atomic_read(&dev_priv->mm.wedged)); |
| 1856 | else | ||
| 1857 | wait_event(dev_priv->irq_queue, | ||
| 1858 | i915_seqno_passed(i915_get_gem_seqno(dev), seqno) || | ||
| 1859 | atomic_read(&dev_priv->mm.wedged)); | ||
| 1860 | |||
| 1859 | i915_user_irq_put(dev); | 1861 | i915_user_irq_put(dev); |
| 1860 | dev_priv->mm.waiting_gem_seqno = 0; | 1862 | dev_priv->mm.waiting_gem_seqno = 0; |
| 1861 | 1863 | ||
| @@ -1879,6 +1881,16 @@ i915_wait_request(struct drm_device *dev, uint32_t seqno) | |||
| 1879 | return ret; | 1881 | return ret; |
| 1880 | } | 1882 | } |
| 1881 | 1883 | ||
| 1884 | /** | ||
| 1885 | * Waits for a sequence number to be signaled, and cleans up the | ||
| 1886 | * request and object lists appropriately for that event. | ||
| 1887 | */ | ||
| 1888 | static int | ||
| 1889 | i915_wait_request(struct drm_device *dev, uint32_t seqno) | ||
| 1890 | { | ||
| 1891 | return i915_do_wait_request(dev, seqno, 1); | ||
| 1892 | } | ||
| 1893 | |||
| 1882 | static void | 1894 | static void |
| 1883 | i915_gem_flush(struct drm_device *dev, | 1895 | i915_gem_flush(struct drm_device *dev, |
| 1884 | uint32_t invalidate_domains, | 1896 | uint32_t invalidate_domains, |
| @@ -1947,7 +1959,7 @@ i915_gem_flush(struct drm_device *dev, | |||
| 1947 | #endif | 1959 | #endif |
| 1948 | BEGIN_LP_RING(2); | 1960 | BEGIN_LP_RING(2); |
| 1949 | OUT_RING(cmd); | 1961 | OUT_RING(cmd); |
| 1950 | OUT_RING(0); /* noop */ | 1962 | OUT_RING(MI_NOOP); |
| 1951 | ADVANCE_LP_RING(); | 1963 | ADVANCE_LP_RING(); |
| 1952 | } | 1964 | } |
| 1953 | } | 1965 | } |
| @@ -2760,6 +2772,22 @@ i915_gem_object_flush_cpu_write_domain(struct drm_gem_object *obj) | |||
| 2760 | old_write_domain); | 2772 | old_write_domain); |
| 2761 | } | 2773 | } |
| 2762 | 2774 | ||
| 2775 | void | ||
| 2776 | i915_gem_object_flush_write_domain(struct drm_gem_object *obj) | ||
| 2777 | { | ||
| 2778 | switch (obj->write_domain) { | ||
| 2779 | case I915_GEM_DOMAIN_GTT: | ||
| 2780 | i915_gem_object_flush_gtt_write_domain(obj); | ||
| 2781 | break; | ||
| 2782 | case I915_GEM_DOMAIN_CPU: | ||
| 2783 | i915_gem_object_flush_cpu_write_domain(obj); | ||
| 2784 | break; | ||
| 2785 | default: | ||
| 2786 | i915_gem_object_flush_gpu_write_domain(obj); | ||
| 2787 | break; | ||
| 2788 | } | ||
| 2789 | } | ||
| 2790 | |||
| 2763 | /** | 2791 | /** |
| 2764 | * Moves a single object to the GTT read, and possibly write domain. | 2792 | * Moves a single object to the GTT read, and possibly write domain. |
| 2765 | * | 2793 | * |
| @@ -3525,6 +3553,41 @@ i915_gem_check_execbuffer (struct drm_i915_gem_execbuffer *exec, | |||
| 3525 | return 0; | 3553 | return 0; |
| 3526 | } | 3554 | } |
| 3527 | 3555 | ||
| 3556 | static int | ||
| 3557 | i915_gem_wait_for_pending_flip(struct drm_device *dev, | ||
| 3558 | struct drm_gem_object **object_list, | ||
| 3559 | int count) | ||
| 3560 | { | ||
| 3561 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 3562 | struct drm_i915_gem_object *obj_priv; | ||
| 3563 | DEFINE_WAIT(wait); | ||
| 3564 | int i, ret = 0; | ||
| 3565 | |||
| 3566 | for (;;) { | ||
| 3567 | prepare_to_wait(&dev_priv->pending_flip_queue, | ||
| 3568 | &wait, TASK_INTERRUPTIBLE); | ||
| 3569 | for (i = 0; i < count; i++) { | ||
| 3570 | obj_priv = object_list[i]->driver_private; | ||
| 3571 | if (atomic_read(&obj_priv->pending_flip) > 0) | ||
| 3572 | break; | ||
| 3573 | } | ||
| 3574 | if (i == count) | ||
| 3575 | break; | ||
| 3576 | |||
| 3577 | if (!signal_pending(current)) { | ||
| 3578 | mutex_unlock(&dev->struct_mutex); | ||
| 3579 | schedule(); | ||
| 3580 | mutex_lock(&dev->struct_mutex); | ||
| 3581 | continue; | ||
| 3582 | } | ||
| 3583 | ret = -ERESTARTSYS; | ||
| 3584 | break; | ||
| 3585 | } | ||
| 3586 | finish_wait(&dev_priv->pending_flip_queue, &wait); | ||
| 3587 | |||
| 3588 | return ret; | ||
| 3589 | } | ||
| 3590 | |||
| 3528 | int | 3591 | int |
| 3529 | i915_gem_execbuffer(struct drm_device *dev, void *data, | 3592 | i915_gem_execbuffer(struct drm_device *dev, void *data, |
| 3530 | struct drm_file *file_priv) | 3593 | struct drm_file *file_priv) |
| @@ -3540,7 +3603,7 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, | |||
| 3540 | int ret, ret2, i, pinned = 0; | 3603 | int ret, ret2, i, pinned = 0; |
| 3541 | uint64_t exec_offset; | 3604 | uint64_t exec_offset; |
| 3542 | uint32_t seqno, flush_domains, reloc_index; | 3605 | uint32_t seqno, flush_domains, reloc_index; |
| 3543 | int pin_tries; | 3606 | int pin_tries, flips; |
| 3544 | 3607 | ||
| 3545 | #if WATCH_EXEC | 3608 | #if WATCH_EXEC |
| 3546 | DRM_INFO("buffers_ptr %d buffer_count %d len %08x\n", | 3609 | DRM_INFO("buffers_ptr %d buffer_count %d len %08x\n", |
| @@ -3552,8 +3615,8 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, | |||
| 3552 | return -EINVAL; | 3615 | return -EINVAL; |
| 3553 | } | 3616 | } |
| 3554 | /* Copy in the exec list from userland */ | 3617 | /* Copy in the exec list from userland */ |
| 3555 | exec_list = drm_calloc_large(sizeof(*exec_list), args->buffer_count); | 3618 | exec_list = drm_malloc_ab(sizeof(*exec_list), args->buffer_count); |
| 3556 | object_list = drm_calloc_large(sizeof(*object_list), args->buffer_count); | 3619 | object_list = drm_malloc_ab(sizeof(*object_list), args->buffer_count); |
| 3557 | if (exec_list == NULL || object_list == NULL) { | 3620 | if (exec_list == NULL || object_list == NULL) { |
| 3558 | DRM_ERROR("Failed to allocate exec or object list " | 3621 | DRM_ERROR("Failed to allocate exec or object list " |
| 3559 | "for %d buffers\n", | 3622 | "for %d buffers\n", |
| @@ -3598,20 +3661,19 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, | |||
| 3598 | i915_verify_inactive(dev, __FILE__, __LINE__); | 3661 | i915_verify_inactive(dev, __FILE__, __LINE__); |
| 3599 | 3662 | ||
| 3600 | if (atomic_read(&dev_priv->mm.wedged)) { | 3663 | if (atomic_read(&dev_priv->mm.wedged)) { |
| 3601 | DRM_ERROR("Execbuf while wedged\n"); | ||
| 3602 | mutex_unlock(&dev->struct_mutex); | 3664 | mutex_unlock(&dev->struct_mutex); |
| 3603 | ret = -EIO; | 3665 | ret = -EIO; |
| 3604 | goto pre_mutex_err; | 3666 | goto pre_mutex_err; |
| 3605 | } | 3667 | } |
| 3606 | 3668 | ||
| 3607 | if (dev_priv->mm.suspended) { | 3669 | if (dev_priv->mm.suspended) { |
| 3608 | DRM_ERROR("Execbuf while VT-switched.\n"); | ||
| 3609 | mutex_unlock(&dev->struct_mutex); | 3670 | mutex_unlock(&dev->struct_mutex); |
| 3610 | ret = -EBUSY; | 3671 | ret = -EBUSY; |
| 3611 | goto pre_mutex_err; | 3672 | goto pre_mutex_err; |
| 3612 | } | 3673 | } |
| 3613 | 3674 | ||
| 3614 | /* Look up object handles */ | 3675 | /* Look up object handles */ |
| 3676 | flips = 0; | ||
| 3615 | for (i = 0; i < args->buffer_count; i++) { | 3677 | for (i = 0; i < args->buffer_count; i++) { |
| 3616 | object_list[i] = drm_gem_object_lookup(dev, file_priv, | 3678 | object_list[i] = drm_gem_object_lookup(dev, file_priv, |
| 3617 | exec_list[i].handle); | 3679 | exec_list[i].handle); |
| @@ -3630,6 +3692,14 @@ i915_gem_execbuffer(struct drm_device *dev, void *data, | |||
| 3630 | goto err; | 3692 | goto err; |
| 3631 | } | 3693 | } |
| 3632 | obj_priv->in_execbuffer = true; | 3694 | obj_priv->in_execbuffer = true; |
| 3695 | flips += atomic_read(&obj_priv->pending_flip); | ||
| 3696 | } | ||
| 3697 | |||
| 3698 | if (flips > 0) { | ||
| 3699 | ret = i915_gem_wait_for_pending_flip(dev, object_list, | ||
| 3700 | args->buffer_count); | ||
| 3701 | if (ret) | ||
| 3702 | goto err; | ||
| 3633 | } | 3703 | } |
| 3634 | 3704 | ||
| 3635 | /* Pin and relocate */ | 3705 | /* Pin and relocate */ |
| @@ -4356,7 +4426,7 @@ i915_gem_init_hws(struct drm_device *dev) | |||
| 4356 | memset(dev_priv->hw_status_page, 0, PAGE_SIZE); | 4426 | memset(dev_priv->hw_status_page, 0, PAGE_SIZE); |
| 4357 | I915_WRITE(HWS_PGA, dev_priv->status_gfx_addr); | 4427 | I915_WRITE(HWS_PGA, dev_priv->status_gfx_addr); |
| 4358 | I915_READ(HWS_PGA); /* posting read */ | 4428 | I915_READ(HWS_PGA); /* posting read */ |
| 4359 | DRM_DEBUG("hws offset: 0x%08x\n", dev_priv->status_gfx_addr); | 4429 | DRM_DEBUG_DRIVER("hws offset: 0x%08x\n", dev_priv->status_gfx_addr); |
| 4360 | 4430 | ||
| 4361 | return 0; | 4431 | return 0; |
| 4362 | } | 4432 | } |
| @@ -4614,8 +4684,8 @@ i915_gem_load(struct drm_device *dev) | |||
| 4614 | for (i = 0; i < 8; i++) | 4684 | for (i = 0; i < 8; i++) |
| 4615 | I915_WRITE(FENCE_REG_945_8 + (i * 4), 0); | 4685 | I915_WRITE(FENCE_REG_945_8 + (i * 4), 0); |
| 4616 | } | 4686 | } |
| 4617 | |||
| 4618 | i915_gem_detect_bit_6_swizzle(dev); | 4687 | i915_gem_detect_bit_6_swizzle(dev); |
| 4688 | init_waitqueue_head(&dev_priv->pending_flip_queue); | ||
| 4619 | } | 4689 | } |
| 4620 | 4690 | ||
| 4621 | /* | 4691 | /* |
| @@ -4790,7 +4860,7 @@ i915_gem_phys_pwrite(struct drm_device *dev, struct drm_gem_object *obj, | |||
| 4790 | user_data = (char __user *) (uintptr_t) args->data_ptr; | 4860 | user_data = (char __user *) (uintptr_t) args->data_ptr; |
| 4791 | obj_addr = obj_priv->phys_obj->handle->vaddr + args->offset; | 4861 | obj_addr = obj_priv->phys_obj->handle->vaddr + args->offset; |
| 4792 | 4862 | ||
| 4793 | DRM_DEBUG("obj_addr %p, %lld\n", obj_addr, args->size); | 4863 | DRM_DEBUG_DRIVER("obj_addr %p, %lld\n", obj_addr, args->size); |
| 4794 | ret = copy_from_user(obj_addr, user_data, args->size); | 4864 | ret = copy_from_user(obj_addr, user_data, args->size); |
| 4795 | if (ret) | 4865 | if (ret) |
| 4796 | return -EFAULT; | 4866 | return -EFAULT; |
diff --git a/drivers/gpu/drm/i915/i915_gem_tiling.c b/drivers/gpu/drm/i915/i915_gem_tiling.c index 200e398453ca..30d6af6c09bb 100644 --- a/drivers/gpu/drm/i915/i915_gem_tiling.c +++ b/drivers/gpu/drm/i915/i915_gem_tiling.c | |||
| @@ -121,7 +121,7 @@ intel_alloc_mchbar_resource(struct drm_device *dev) | |||
| 121 | 0, pcibios_align_resource, | 121 | 0, pcibios_align_resource, |
| 122 | dev_priv->bridge_dev); | 122 | dev_priv->bridge_dev); |
| 123 | if (ret) { | 123 | if (ret) { |
| 124 | DRM_DEBUG("failed bus alloc: %d\n", ret); | 124 | DRM_DEBUG_DRIVER("failed bus alloc: %d\n", ret); |
| 125 | dev_priv->mch_res.start = 0; | 125 | dev_priv->mch_res.start = 0; |
| 126 | goto out; | 126 | goto out; |
| 127 | } | 127 | } |
| @@ -209,8 +209,8 @@ i915_gem_detect_bit_6_swizzle(struct drm_device *dev) | |||
| 209 | uint32_t swizzle_y = I915_BIT_6_SWIZZLE_UNKNOWN; | 209 | uint32_t swizzle_y = I915_BIT_6_SWIZZLE_UNKNOWN; |
| 210 | bool need_disable; | 210 | bool need_disable; |
| 211 | 211 | ||
| 212 | if (IS_IGDNG(dev)) { | 212 | if (IS_IRONLAKE(dev)) { |
| 213 | /* On IGDNG whatever DRAM config, GPU always do | 213 | /* On Ironlake whatever DRAM config, GPU always do |
| 214 | * same swizzling setup. | 214 | * same swizzling setup. |
| 215 | */ | 215 | */ |
| 216 | swizzle_x = I915_BIT_6_SWIZZLE_9_10; | 216 | swizzle_x = I915_BIT_6_SWIZZLE_9_10; |
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c index aa7fd82aa6eb..85f4c5de97e2 100644 --- a/drivers/gpu/drm/i915/i915_irq.c +++ b/drivers/gpu/drm/i915/i915_irq.c | |||
| @@ -43,10 +43,13 @@ | |||
| 43 | * we leave them always unmasked in IMR and then control enabling them through | 43 | * we leave them always unmasked in IMR and then control enabling them through |
| 44 | * PIPESTAT alone. | 44 | * PIPESTAT alone. |
| 45 | */ | 45 | */ |
| 46 | #define I915_INTERRUPT_ENABLE_FIX (I915_ASLE_INTERRUPT | \ | 46 | #define I915_INTERRUPT_ENABLE_FIX \ |
| 47 | I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | \ | 47 | (I915_ASLE_INTERRUPT | \ |
| 48 | I915_DISPLAY_PIPE_B_EVENT_INTERRUPT | \ | 48 | I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | \ |
| 49 | I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT) | 49 | I915_DISPLAY_PIPE_B_EVENT_INTERRUPT | \ |
| 50 | I915_DISPLAY_PLANE_A_FLIP_PENDING_INTERRUPT | \ | ||
| 51 | I915_DISPLAY_PLANE_B_FLIP_PENDING_INTERRUPT | \ | ||
| 52 | I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT) | ||
| 50 | 53 | ||
| 51 | /** Interrupts that we mask and unmask at runtime. */ | 54 | /** Interrupts that we mask and unmask at runtime. */ |
| 52 | #define I915_INTERRUPT_ENABLE_VAR (I915_USER_INTERRUPT) | 55 | #define I915_INTERRUPT_ENABLE_VAR (I915_USER_INTERRUPT) |
| @@ -61,7 +64,7 @@ | |||
| 61 | DRM_I915_VBLANK_PIPE_B) | 64 | DRM_I915_VBLANK_PIPE_B) |
| 62 | 65 | ||
| 63 | void | 66 | void |
| 64 | igdng_enable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | 67 | ironlake_enable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) |
| 65 | { | 68 | { |
| 66 | if ((dev_priv->gt_irq_mask_reg & mask) != 0) { | 69 | if ((dev_priv->gt_irq_mask_reg & mask) != 0) { |
| 67 | dev_priv->gt_irq_mask_reg &= ~mask; | 70 | dev_priv->gt_irq_mask_reg &= ~mask; |
| @@ -71,7 +74,7 @@ igdng_enable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | |||
| 71 | } | 74 | } |
| 72 | 75 | ||
| 73 | static inline void | 76 | static inline void |
| 74 | igdng_disable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | 77 | ironlake_disable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) |
| 75 | { | 78 | { |
| 76 | if ((dev_priv->gt_irq_mask_reg & mask) != mask) { | 79 | if ((dev_priv->gt_irq_mask_reg & mask) != mask) { |
| 77 | dev_priv->gt_irq_mask_reg |= mask; | 80 | dev_priv->gt_irq_mask_reg |= mask; |
| @@ -82,7 +85,7 @@ igdng_disable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask) | |||
| 82 | 85 | ||
| 83 | /* For display hotplug interrupt */ | 86 | /* For display hotplug interrupt */ |
| 84 | void | 87 | void |
| 85 | igdng_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask) | 88 | ironlake_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask) |
| 86 | { | 89 | { |
| 87 | if ((dev_priv->irq_mask_reg & mask) != 0) { | 90 | if ((dev_priv->irq_mask_reg & mask) != 0) { |
| 88 | dev_priv->irq_mask_reg &= ~mask; | 91 | dev_priv->irq_mask_reg &= ~mask; |
| @@ -92,7 +95,7 @@ igdng_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask) | |||
| 92 | } | 95 | } |
| 93 | 96 | ||
| 94 | static inline void | 97 | static inline void |
| 95 | igdng_disable_display_irq(drm_i915_private_t *dev_priv, u32 mask) | 98 | ironlake_disable_display_irq(drm_i915_private_t *dev_priv, u32 mask) |
| 96 | { | 99 | { |
| 97 | if ((dev_priv->irq_mask_reg & mask) != mask) { | 100 | if ((dev_priv->irq_mask_reg & mask) != mask) { |
| 98 | dev_priv->irq_mask_reg |= mask; | 101 | dev_priv->irq_mask_reg |= mask; |
| @@ -157,6 +160,20 @@ i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask) | |||
| 157 | } | 160 | } |
| 158 | 161 | ||
| 159 | /** | 162 | /** |
| 163 | * intel_enable_asle - enable ASLE interrupt for OpRegion | ||
| 164 | */ | ||
| 165 | void intel_enable_asle (struct drm_device *dev) | ||
| 166 | { | ||
| 167 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | ||
| 168 | |||
| 169 | if (IS_IRONLAKE(dev)) | ||
| 170 | ironlake_enable_display_irq(dev_priv, DE_GSE); | ||
| 171 | else | ||
| 172 | i915_enable_pipestat(dev_priv, 1, | ||
| 173 | I915_LEGACY_BLC_EVENT_ENABLE); | ||
| 174 | } | ||
| 175 | |||
| 176 | /** | ||
| 160 | * i915_pipe_enabled - check if a pipe is enabled | 177 | * i915_pipe_enabled - check if a pipe is enabled |
| 161 | * @dev: DRM device | 178 | * @dev: DRM device |
| 162 | * @pipe: pipe to check | 179 | * @pipe: pipe to check |
| @@ -191,7 +208,8 @@ u32 i915_get_vblank_counter(struct drm_device *dev, int pipe) | |||
| 191 | low_frame = pipe ? PIPEBFRAMEPIXEL : PIPEAFRAMEPIXEL; | 208 | low_frame = pipe ? PIPEBFRAMEPIXEL : PIPEAFRAMEPIXEL; |
| 192 | 209 | ||
| 193 | if (!i915_pipe_enabled(dev, pipe)) { | 210 | if (!i915_pipe_enabled(dev, pipe)) { |
| 194 | DRM_DEBUG("trying to get vblank count for disabled pipe %d\n", pipe); | 211 | DRM_DEBUG_DRIVER("trying to get vblank count for disabled " |
| 212 | "pipe %d\n", pipe); | ||
| 195 | return 0; | 213 | return 0; |
| 196 | } | 214 | } |
| 197 | 215 | ||
| @@ -220,7 +238,8 @@ u32 gm45_get_vblank_counter(struct drm_device *dev, int pipe) | |||
| 220 | int reg = pipe ? PIPEB_FRMCOUNT_GM45 : PIPEA_FRMCOUNT_GM45; | 238 | int reg = pipe ? PIPEB_FRMCOUNT_GM45 : PIPEA_FRMCOUNT_GM45; |
| 221 | 239 | ||
| 222 | if (!i915_pipe_enabled(dev, pipe)) { | 240 | if (!i915_pipe_enabled(dev, pipe)) { |
| 223 | DRM_DEBUG("trying to get vblank count for disabled pipe %d\n", pipe); | 241 | DRM_DEBUG_DRIVER("trying to get vblank count for disabled " |
| 242 | "pipe %d\n", pipe); | ||
| 224 | return 0; | 243 | return 0; |
| 225 | } | 244 | } |
| 226 | 245 | ||
| @@ -250,12 +269,12 @@ static void i915_hotplug_work_func(struct work_struct *work) | |||
| 250 | drm_sysfs_hotplug_event(dev); | 269 | drm_sysfs_hotplug_event(dev); |
| 251 | } | 270 | } |
| 252 | 271 | ||
| 253 | irqreturn_t igdng_irq_handler(struct drm_device *dev) | 272 | irqreturn_t ironlake_irq_handler(struct drm_device *dev) |
| 254 | { | 273 | { |
| 255 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 274 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| 256 | int ret = IRQ_NONE; | 275 | int ret = IRQ_NONE; |
| 257 | u32 de_iir, gt_iir, de_ier; | 276 | u32 de_iir, gt_iir, de_ier, pch_iir; |
| 258 | u32 new_de_iir, new_gt_iir; | 277 | u32 new_de_iir, new_gt_iir, new_pch_iir; |
| 259 | struct drm_i915_master_private *master_priv; | 278 | struct drm_i915_master_private *master_priv; |
| 260 | 279 | ||
| 261 | /* disable master interrupt before clearing iir */ | 280 | /* disable master interrupt before clearing iir */ |
| @@ -265,13 +284,18 @@ irqreturn_t igdng_irq_handler(struct drm_device *dev) | |||
| 265 | 284 | ||
| 266 | de_iir = I915_READ(DEIIR); | 285 | de_iir = I915_READ(DEIIR); |
| 267 | gt_iir = I915_READ(GTIIR); | 286 | gt_iir = I915_READ(GTIIR); |
| 287 | pch_iir = I915_READ(SDEIIR); | ||
| 268 | 288 | ||
| 269 | for (;;) { | 289 | for (;;) { |
| 270 | if (de_iir == 0 && gt_iir == 0) | 290 | if (de_iir == 0 && gt_iir == 0 && pch_iir == 0) |
| 271 | break; | 291 | break; |
| 272 | 292 | ||
| 273 | ret = IRQ_HANDLED; | 293 | ret = IRQ_HANDLED; |
| 274 | 294 | ||
| 295 | /* should clear PCH hotplug event before clear CPU irq */ | ||
| 296 | I915_WRITE(SDEIIR, pch_iir); | ||
| 297 | new_pch_iir = I915_READ(SDEIIR); | ||
| 298 | |||
| 275 | I915_WRITE(DEIIR, de_iir); | 299 | I915_WRITE(DEIIR, de_iir); |
| 276 | new_de_iir = I915_READ(DEIIR); | 300 | new_de_iir = I915_READ(DEIIR); |
| 277 | I915_WRITE(GTIIR, gt_iir); | 301 | I915_WRITE(GTIIR, gt_iir); |
| @@ -291,8 +315,18 @@ irqreturn_t igdng_irq_handler(struct drm_device *dev) | |||
| 291 | DRM_WAKEUP(&dev_priv->irq_queue); | 315 | DRM_WAKEUP(&dev_priv->irq_queue); |
| 292 | } | 316 | } |
| 293 | 317 | ||
| 318 | if (de_iir & DE_GSE) | ||
| 319 | ironlake_opregion_gse_intr(dev); | ||
| 320 | |||
| 321 | /* check event from PCH */ | ||
| 322 | if ((de_iir & DE_PCH_EVENT) && | ||
| 323 | (pch_iir & SDE_HOTPLUG_MASK)) { | ||
| 324 | queue_work(dev_priv->wq, &dev_priv->hotplug_work); | ||
| 325 | } | ||
| 326 | |||
| 294 | de_iir = new_de_iir; | 327 | de_iir = new_de_iir; |
| 295 | gt_iir = new_gt_iir; | 328 | gt_iir = new_gt_iir; |
| 329 | pch_iir = new_pch_iir; | ||
| 296 | } | 330 | } |
| 297 | 331 | ||
| 298 | I915_WRITE(DEIER, de_ier); | 332 | I915_WRITE(DEIER, de_ier); |
| @@ -317,19 +351,19 @@ static void i915_error_work_func(struct work_struct *work) | |||
| 317 | char *reset_event[] = { "RESET=1", NULL }; | 351 | char *reset_event[] = { "RESET=1", NULL }; |
| 318 | char *reset_done_event[] = { "ERROR=0", NULL }; | 352 | char *reset_done_event[] = { "ERROR=0", NULL }; |
| 319 | 353 | ||
| 320 | DRM_DEBUG("generating error event\n"); | 354 | DRM_DEBUG_DRIVER("generating error event\n"); |
| 321 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); | 355 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event); |
| 322 | 356 | ||
| 323 | if (atomic_read(&dev_priv->mm.wedged)) { | 357 | if (atomic_read(&dev_priv->mm.wedged)) { |
| 324 | if (IS_I965G(dev)) { | 358 | if (IS_I965G(dev)) { |
| 325 | DRM_DEBUG("resetting chip\n"); | 359 | DRM_DEBUG_DRIVER("resetting chip\n"); |
| 326 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_event); | 360 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_event); |
| 327 | if (!i965_reset(dev, GDRST_RENDER)) { | 361 | if (!i965_reset(dev, GDRST_RENDER)) { |
| 328 | atomic_set(&dev_priv->mm.wedged, 0); | 362 | atomic_set(&dev_priv->mm.wedged, 0); |
| 329 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_done_event); | 363 | kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_done_event); |
| 330 | } | 364 | } |
| 331 | } else { | 365 | } else { |
| 332 | printk("reboot required\n"); | 366 | DRM_DEBUG_DRIVER("reboot required\n"); |
| 333 | } | 367 | } |
| 334 | } | 368 | } |
| 335 | } | 369 | } |
| @@ -355,7 +389,7 @@ static void i915_capture_error_state(struct drm_device *dev) | |||
| 355 | 389 | ||
| 356 | error = kmalloc(sizeof(*error), GFP_ATOMIC); | 390 | error = kmalloc(sizeof(*error), GFP_ATOMIC); |
| 357 | if (!error) { | 391 | if (!error) { |
| 358 | DRM_DEBUG("out ot memory, not capturing error state\n"); | 392 | DRM_DEBUG_DRIVER("out ot memory, not capturing error state\n"); |
| 359 | goto out; | 393 | goto out; |
| 360 | } | 394 | } |
| 361 | 395 | ||
| @@ -512,7 +546,6 @@ static void i915_handle_error(struct drm_device *dev, bool wedged) | |||
| 512 | /* | 546 | /* |
| 513 | * Wakeup waiting processes so they don't hang | 547 | * Wakeup waiting processes so they don't hang |
| 514 | */ | 548 | */ |
| 515 | printk("i915: Waking up sleeping processes\n"); | ||
| 516 | DRM_WAKEUP(&dev_priv->irq_queue); | 549 | DRM_WAKEUP(&dev_priv->irq_queue); |
| 517 | } | 550 | } |
| 518 | 551 | ||
| @@ -535,8 +568,8 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) | |||
| 535 | 568 | ||
| 536 | atomic_inc(&dev_priv->irq_received); | 569 | atomic_inc(&dev_priv->irq_received); |
| 537 | 570 | ||
| 538 | if (IS_IGDNG(dev)) | 571 | if (IS_IRONLAKE(dev)) |
| 539 | return igdng_irq_handler(dev); | 572 | return ironlake_irq_handler(dev); |
| 540 | 573 | ||
| 541 | iir = I915_READ(IIR); | 574 | iir = I915_READ(IIR); |
| 542 | 575 | ||
| @@ -568,14 +601,14 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) | |||
| 568 | */ | 601 | */ |
| 569 | if (pipea_stats & 0x8000ffff) { | 602 | if (pipea_stats & 0x8000ffff) { |
| 570 | if (pipea_stats & PIPE_FIFO_UNDERRUN_STATUS) | 603 | if (pipea_stats & PIPE_FIFO_UNDERRUN_STATUS) |
| 571 | DRM_DEBUG("pipe a underrun\n"); | 604 | DRM_DEBUG_DRIVER("pipe a underrun\n"); |
| 572 | I915_WRITE(PIPEASTAT, pipea_stats); | 605 | I915_WRITE(PIPEASTAT, pipea_stats); |
| 573 | irq_received = 1; | 606 | irq_received = 1; |
| 574 | } | 607 | } |
| 575 | 608 | ||
| 576 | if (pipeb_stats & 0x8000ffff) { | 609 | if (pipeb_stats & 0x8000ffff) { |
| 577 | if (pipeb_stats & PIPE_FIFO_UNDERRUN_STATUS) | 610 | if (pipeb_stats & PIPE_FIFO_UNDERRUN_STATUS) |
| 578 | DRM_DEBUG("pipe b underrun\n"); | 611 | DRM_DEBUG_DRIVER("pipe b underrun\n"); |
| 579 | I915_WRITE(PIPEBSTAT, pipeb_stats); | 612 | I915_WRITE(PIPEBSTAT, pipeb_stats); |
| 580 | irq_received = 1; | 613 | irq_received = 1; |
| 581 | } | 614 | } |
| @@ -591,7 +624,7 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) | |||
| 591 | (iir & I915_DISPLAY_PORT_INTERRUPT)) { | 624 | (iir & I915_DISPLAY_PORT_INTERRUPT)) { |
| 592 | u32 hotplug_status = I915_READ(PORT_HOTPLUG_STAT); | 625 | u32 hotplug_status = I915_READ(PORT_HOTPLUG_STAT); |
| 593 | 626 | ||
| 594 | DRM_DEBUG("hotplug event received, stat 0x%08x\n", | 627 | DRM_DEBUG_DRIVER("hotplug event received, stat 0x%08x\n", |
| 595 | hotplug_status); | 628 | hotplug_status); |
| 596 | if (hotplug_status & dev_priv->hotplug_supported_mask) | 629 | if (hotplug_status & dev_priv->hotplug_supported_mask) |
| 597 | queue_work(dev_priv->wq, | 630 | queue_work(dev_priv->wq, |
| @@ -599,27 +632,6 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) | |||
| 599 | 632 | ||
| 600 | I915_WRITE(PORT_HOTPLUG_STAT, hotplug_status); | 633 | I915_WRITE(PORT_HOTPLUG_STAT, hotplug_status); |
| 601 | I915_READ(PORT_HOTPLUG_STAT); | 634 | I915_READ(PORT_HOTPLUG_STAT); |
| 602 | |||
| 603 | /* EOS interrupts occurs */ | ||
| 604 | if (IS_IGD(dev) && | ||
| 605 | (hotplug_status & CRT_EOS_INT_STATUS)) { | ||
| 606 | u32 temp; | ||
| 607 | |||
| 608 | DRM_DEBUG("EOS interrupt occurs\n"); | ||
| 609 | /* status is already cleared */ | ||
| 610 | temp = I915_READ(ADPA); | ||
| 611 | temp &= ~ADPA_DAC_ENABLE; | ||
| 612 | I915_WRITE(ADPA, temp); | ||
| 613 | |||
| 614 | temp = I915_READ(PORT_HOTPLUG_EN); | ||
| 615 | temp &= ~CRT_EOS_INT_EN; | ||
| 616 | I915_WRITE(PORT_HOTPLUG_EN, temp); | ||
| 617 | |||
| 618 | temp = I915_READ(PORT_HOTPLUG_STAT); | ||
| 619 | if (temp & CRT_EOS_INT_STATUS) | ||
| 620 | I915_WRITE(PORT_HOTPLUG_STAT, | ||
| 621 | CRT_EOS_INT_STATUS); | ||
| 622 | } | ||
| 623 | } | 635 | } |
| 624 | 636 | ||
| 625 | I915_WRITE(IIR, iir); | 637 | I915_WRITE(IIR, iir); |
| @@ -641,14 +653,22 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS) | |||
| 641 | mod_timer(&dev_priv->hangcheck_timer, jiffies + DRM_I915_HANGCHECK_PERIOD); | 653 | mod_timer(&dev_priv->hangcheck_timer, jiffies + DRM_I915_HANGCHECK_PERIOD); |
| 642 | } | 654 | } |
| 643 | 655 | ||
| 656 | if (iir & I915_DISPLAY_PLANE_A_FLIP_PENDING_INTERRUPT) | ||
| 657 | intel_prepare_page_flip(dev, 0); | ||
| 658 | |||
| 659 | if (iir & I915_DISPLAY_PLANE_B_FLIP_PENDING_INTERRUPT) | ||
| 660 | intel_prepare_page_flip(dev, 1); | ||
| 661 | |||
| 644 | if (pipea_stats & vblank_status) { | 662 | if (pipea_stats & vblank_status) { |
| 645 | vblank++; | 663 | vblank++; |
| 646 | drm_handle_vblank(dev, 0); | 664 | drm_handle_vblank(dev, 0); |
| 665 | intel_finish_page_flip(dev, 0); | ||
| 647 | } | 666 | } |
| 648 | 667 | ||
| 649 | if (pipeb_stats & vblank_status) { | 668 | if (pipeb_stats & vblank_status) { |
| 650 | vblank++; | 669 | vblank++; |
| 651 | drm_handle_vblank(dev, 1); | 670 | drm_handle_vblank(dev, 1); |
| 671 | intel_finish_page_flip(dev, 1); | ||
| 652 | } | 672 | } |
| 653 | 673 | ||
| 654 | if ((pipeb_stats & I915_LEGACY_BLC_EVENT_STATUS) || | 674 | if ((pipeb_stats & I915_LEGACY_BLC_EVENT_STATUS) || |
| @@ -684,7 +704,7 @@ static int i915_emit_irq(struct drm_device * dev) | |||
| 684 | 704 | ||
| 685 | i915_kernel_lost_context(dev); | 705 | i915_kernel_lost_context(dev); |
| 686 | 706 | ||
| 687 | DRM_DEBUG("\n"); | 707 | DRM_DEBUG_DRIVER("\n"); |
| 688 | 708 | ||
| 689 | dev_priv->counter++; | 709 | dev_priv->counter++; |
| 690 | if (dev_priv->counter > 0x7FFFFFFFUL) | 710 | if (dev_priv->counter > 0x7FFFFFFFUL) |
| @@ -709,8 +729,8 @@ void i915_user_irq_get(struct drm_device *dev) | |||
| 709 | 729 | ||
| 710 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); | 730 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); |
| 711 | if (dev->irq_enabled && (++dev_priv->user_irq_refcount == 1)) { | 731 | if (dev->irq_enabled && (++dev_priv->user_irq_refcount == 1)) { |
| 712 | if (IS_IGDNG(dev)) | 732 | if (IS_IRONLAKE(dev)) |
| 713 | igdng_enable_graphics_irq(dev_priv, GT_USER_INTERRUPT); | 733 | ironlake_enable_graphics_irq(dev_priv, GT_USER_INTERRUPT); |
| 714 | else | 734 | else |
| 715 | i915_enable_irq(dev_priv, I915_USER_INTERRUPT); | 735 | i915_enable_irq(dev_priv, I915_USER_INTERRUPT); |
| 716 | } | 736 | } |
| @@ -725,8 +745,8 @@ void i915_user_irq_put(struct drm_device *dev) | |||
| 725 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); | 745 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); |
| 726 | BUG_ON(dev->irq_enabled && dev_priv->user_irq_refcount <= 0); | 746 | BUG_ON(dev->irq_enabled && dev_priv->user_irq_refcount <= 0); |
| 727 | if (dev->irq_enabled && (--dev_priv->user_irq_refcount == 0)) { | 747 | if (dev->irq_enabled && (--dev_priv->user_irq_refcount == 0)) { |
| 728 | if (IS_IGDNG(dev)) | 748 | if (IS_IRONLAKE(dev)) |
| 729 | igdng_disable_graphics_irq(dev_priv, GT_USER_INTERRUPT); | 749 | ironlake_disable_graphics_irq(dev_priv, GT_USER_INTERRUPT); |
| 730 | else | 750 | else |
| 731 | i915_disable_irq(dev_priv, I915_USER_INTERRUPT); | 751 | i915_disable_irq(dev_priv, I915_USER_INTERRUPT); |
| 732 | } | 752 | } |
| @@ -749,7 +769,7 @@ static int i915_wait_irq(struct drm_device * dev, int irq_nr) | |||
| 749 | struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv; | 769 | struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv; |
| 750 | int ret = 0; | 770 | int ret = 0; |
| 751 | 771 | ||
| 752 | DRM_DEBUG("irq_nr=%d breadcrumb=%d\n", irq_nr, | 772 | DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr, |
| 753 | READ_BREADCRUMB(dev_priv)); | 773 | READ_BREADCRUMB(dev_priv)); |
| 754 | 774 | ||
| 755 | if (READ_BREADCRUMB(dev_priv) >= irq_nr) { | 775 | if (READ_BREADCRUMB(dev_priv) >= irq_nr) { |
| @@ -832,7 +852,7 @@ int i915_enable_vblank(struct drm_device *dev, int pipe) | |||
| 832 | if (!(pipeconf & PIPEACONF_ENABLE)) | 852 | if (!(pipeconf & PIPEACONF_ENABLE)) |
| 833 | return -EINVAL; | 853 | return -EINVAL; |
| 834 | 854 | ||
| 835 | if (IS_IGDNG(dev)) | 855 | if (IS_IRONLAKE(dev)) |
| 836 | return 0; | 856 | return 0; |
| 837 | 857 | ||
| 838 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); | 858 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); |
| @@ -854,7 +874,7 @@ void i915_disable_vblank(struct drm_device *dev, int pipe) | |||
| 854 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 874 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| 855 | unsigned long irqflags; | 875 | unsigned long irqflags; |
| 856 | 876 | ||
| 857 | if (IS_IGDNG(dev)) | 877 | if (IS_IRONLAKE(dev)) |
| 858 | return; | 878 | return; |
| 859 | 879 | ||
| 860 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); | 880 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); |
| @@ -868,7 +888,7 @@ void i915_enable_interrupt (struct drm_device *dev) | |||
| 868 | { | 888 | { |
| 869 | struct drm_i915_private *dev_priv = dev->dev_private; | 889 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 870 | 890 | ||
| 871 | if (!IS_IGDNG(dev)) | 891 | if (!IS_IRONLAKE(dev)) |
| 872 | opregion_enable_asle(dev); | 892 | opregion_enable_asle(dev); |
| 873 | dev_priv->irq_enabled = 1; | 893 | dev_priv->irq_enabled = 1; |
| 874 | } | 894 | } |
| @@ -976,7 +996,7 @@ void i915_hangcheck_elapsed(unsigned long data) | |||
| 976 | 996 | ||
| 977 | /* drm_dma.h hooks | 997 | /* drm_dma.h hooks |
| 978 | */ | 998 | */ |
| 979 | static void igdng_irq_preinstall(struct drm_device *dev) | 999 | static void ironlake_irq_preinstall(struct drm_device *dev) |
| 980 | { | 1000 | { |
| 981 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 1001 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| 982 | 1002 | ||
| @@ -992,14 +1012,21 @@ static void igdng_irq_preinstall(struct drm_device *dev) | |||
| 992 | I915_WRITE(GTIMR, 0xffffffff); | 1012 | I915_WRITE(GTIMR, 0xffffffff); |
| 993 | I915_WRITE(GTIER, 0x0); | 1013 | I915_WRITE(GTIER, 0x0); |
| 994 | (void) I915_READ(GTIER); | 1014 | (void) I915_READ(GTIER); |
| 1015 | |||
| 1016 | /* south display irq */ | ||
| 1017 | I915_WRITE(SDEIMR, 0xffffffff); | ||
| 1018 | I915_WRITE(SDEIER, 0x0); | ||
| 1019 | (void) I915_READ(SDEIER); | ||
| 995 | } | 1020 | } |
| 996 | 1021 | ||
| 997 | static int igdng_irq_postinstall(struct drm_device *dev) | 1022 | static int ironlake_irq_postinstall(struct drm_device *dev) |
| 998 | { | 1023 | { |
| 999 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 1024 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| 1000 | /* enable kind of interrupts always enabled */ | 1025 | /* enable kind of interrupts always enabled */ |
| 1001 | u32 display_mask = DE_MASTER_IRQ_CONTROL /*| DE_PCH_EVENT */; | 1026 | u32 display_mask = DE_MASTER_IRQ_CONTROL | DE_GSE | DE_PCH_EVENT; |
| 1002 | u32 render_mask = GT_USER_INTERRUPT; | 1027 | u32 render_mask = GT_USER_INTERRUPT; |
| 1028 | u32 hotplug_mask = SDE_CRT_HOTPLUG | SDE_PORTB_HOTPLUG | | ||
| 1029 | SDE_PORTC_HOTPLUG | SDE_PORTD_HOTPLUG; | ||
| 1003 | 1030 | ||
| 1004 | dev_priv->irq_mask_reg = ~display_mask; | 1031 | dev_priv->irq_mask_reg = ~display_mask; |
| 1005 | dev_priv->de_irq_enable_reg = display_mask; | 1032 | dev_priv->de_irq_enable_reg = display_mask; |
| @@ -1019,6 +1046,14 @@ static int igdng_irq_postinstall(struct drm_device *dev) | |||
| 1019 | I915_WRITE(GTIER, dev_priv->gt_irq_enable_reg); | 1046 | I915_WRITE(GTIER, dev_priv->gt_irq_enable_reg); |
| 1020 | (void) I915_READ(GTIER); | 1047 | (void) I915_READ(GTIER); |
| 1021 | 1048 | ||
| 1049 | dev_priv->pch_irq_mask_reg = ~hotplug_mask; | ||
| 1050 | dev_priv->pch_irq_enable_reg = hotplug_mask; | ||
| 1051 | |||
| 1052 | I915_WRITE(SDEIIR, I915_READ(SDEIIR)); | ||
| 1053 | I915_WRITE(SDEIMR, dev_priv->pch_irq_mask_reg); | ||
| 1054 | I915_WRITE(SDEIER, dev_priv->pch_irq_enable_reg); | ||
| 1055 | (void) I915_READ(SDEIER); | ||
| 1056 | |||
| 1022 | return 0; | 1057 | return 0; |
| 1023 | } | 1058 | } |
| 1024 | 1059 | ||
| @@ -1031,8 +1066,8 @@ void i915_driver_irq_preinstall(struct drm_device * dev) | |||
| 1031 | INIT_WORK(&dev_priv->hotplug_work, i915_hotplug_work_func); | 1066 | INIT_WORK(&dev_priv->hotplug_work, i915_hotplug_work_func); |
| 1032 | INIT_WORK(&dev_priv->error_work, i915_error_work_func); | 1067 | INIT_WORK(&dev_priv->error_work, i915_error_work_func); |
| 1033 | 1068 | ||
| 1034 | if (IS_IGDNG(dev)) { | 1069 | if (IS_IRONLAKE(dev)) { |
| 1035 | igdng_irq_preinstall(dev); | 1070 | ironlake_irq_preinstall(dev); |
| 1036 | return; | 1071 | return; |
| 1037 | } | 1072 | } |
| 1038 | 1073 | ||
| @@ -1059,8 +1094,8 @@ int i915_driver_irq_postinstall(struct drm_device *dev) | |||
| 1059 | 1094 | ||
| 1060 | dev_priv->vblank_pipe = DRM_I915_VBLANK_PIPE_A | DRM_I915_VBLANK_PIPE_B; | 1095 | dev_priv->vblank_pipe = DRM_I915_VBLANK_PIPE_A | DRM_I915_VBLANK_PIPE_B; |
| 1061 | 1096 | ||
| 1062 | if (IS_IGDNG(dev)) | 1097 | if (IS_IRONLAKE(dev)) |
| 1063 | return igdng_irq_postinstall(dev); | 1098 | return ironlake_irq_postinstall(dev); |
| 1064 | 1099 | ||
| 1065 | /* Unmask the interrupts that we always want on. */ | 1100 | /* Unmask the interrupts that we always want on. */ |
| 1066 | dev_priv->irq_mask_reg = ~I915_INTERRUPT_ENABLE_FIX; | 1101 | dev_priv->irq_mask_reg = ~I915_INTERRUPT_ENABLE_FIX; |
| @@ -1120,7 +1155,7 @@ int i915_driver_irq_postinstall(struct drm_device *dev) | |||
| 1120 | return 0; | 1155 | return 0; |
| 1121 | } | 1156 | } |
| 1122 | 1157 | ||
| 1123 | static void igdng_irq_uninstall(struct drm_device *dev) | 1158 | static void ironlake_irq_uninstall(struct drm_device *dev) |
| 1124 | { | 1159 | { |
| 1125 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; | 1160 | drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private; |
| 1126 | I915_WRITE(HWSTAM, 0xffffffff); | 1161 | I915_WRITE(HWSTAM, 0xffffffff); |
| @@ -1143,8 +1178,8 @@ void i915_driver_irq_uninstall(struct drm_device * dev) | |||
| 1143 | 1178 | ||
| 1144 | dev_priv->vblank_pipe = 0; | 1179 | dev_priv->vblank_pipe = 0; |
| 1145 | 1180 | ||
| 1146 | if (IS_IGDNG(dev)) { | 1181 | if (IS_IRONLAKE(dev)) { |
| 1147 | igdng_irq_uninstall(dev); | 1182 | ironlake_irq_uninstall(dev); |
| 1148 | return; | 1183 | return; |
| 1149 | } | 1184 | } |
| 1150 | 1185 | ||
diff --git a/drivers/gpu/drm/i915/i915_opregion.c b/drivers/gpu/drm/i915/i915_opregion.c index 2d5193556d3f..7cc8410239cb 100644 --- a/drivers/gpu/drm/i915/i915_opregion.c +++ b/drivers/gpu/drm/i915/i915_opregion.c | |||
| @@ -118,6 +118,10 @@ struct opregion_asle { | |||
| 118 | #define ASLE_BACKLIGHT_FAIL (2<<12) | 118 | #define ASLE_BACKLIGHT_FAIL (2<<12) |
| 119 | #define ASLE_PFIT_FAIL (2<<14) | 119 | #define ASLE_PFIT_FAIL (2<<14) |
| 120 | #define ASLE_PWM_FREQ_FAIL (2<<16) | 120 | #define ASLE_PWM_FREQ_FAIL (2<<16) |
| 121 | #define ASLE_ALS_ILLUM_FAILED (1<<10) | ||
| 122 | #define ASLE_BACKLIGHT_FAILED (1<<12) | ||
| 123 | #define ASLE_PFIT_FAILED (1<<14) | ||
| 124 | #define ASLE_PWM_FREQ_FAILED (1<<16) | ||
| 121 | 125 | ||
| 122 | /* ASLE backlight brightness to set */ | 126 | /* ASLE backlight brightness to set */ |
| 123 | #define ASLE_BCLP_VALID (1<<31) | 127 | #define ASLE_BCLP_VALID (1<<31) |
| @@ -163,7 +167,7 @@ static u32 asle_set_backlight(struct drm_device *dev, u32 bclp) | |||
| 163 | if (IS_I965G(dev) && (blc_pwm_ctl2 & BLM_COMBINATION_MODE)) | 167 | if (IS_I965G(dev) && (blc_pwm_ctl2 & BLM_COMBINATION_MODE)) |
| 164 | pci_write_config_dword(dev->pdev, PCI_LBPC, bclp); | 168 | pci_write_config_dword(dev->pdev, PCI_LBPC, bclp); |
| 165 | else { | 169 | else { |
| 166 | if (IS_IGD(dev)) { | 170 | if (IS_PINEVIEW(dev)) { |
| 167 | blc_pwm_ctl &= ~(BACKLIGHT_DUTY_CYCLE_MASK - 1); | 171 | blc_pwm_ctl &= ~(BACKLIGHT_DUTY_CYCLE_MASK - 1); |
| 168 | max_backlight = (blc_pwm_ctl & BACKLIGHT_MODULATION_FREQ_MASK) >> | 172 | max_backlight = (blc_pwm_ctl & BACKLIGHT_MODULATION_FREQ_MASK) >> |
| 169 | BACKLIGHT_MODULATION_FREQ_SHIFT; | 173 | BACKLIGHT_MODULATION_FREQ_SHIFT; |
| @@ -224,7 +228,7 @@ void opregion_asle_intr(struct drm_device *dev) | |||
| 224 | asle_req = asle->aslc & ASLE_REQ_MSK; | 228 | asle_req = asle->aslc & ASLE_REQ_MSK; |
| 225 | 229 | ||
| 226 | if (!asle_req) { | 230 | if (!asle_req) { |
| 227 | DRM_DEBUG("non asle set request??\n"); | 231 | DRM_DEBUG_DRIVER("non asle set request??\n"); |
| 228 | return; | 232 | return; |
| 229 | } | 233 | } |
| 230 | 234 | ||
| @@ -243,6 +247,73 @@ void opregion_asle_intr(struct drm_device *dev) | |||
| 243 | asle->aslc = asle_stat; | 247 | asle->aslc = asle_stat; |
| 244 | } | 248 | } |
| 245 | 249 | ||
| 250 | static u32 asle_set_backlight_ironlake(struct drm_device *dev, u32 bclp) | ||
| 251 | { | ||
| 252 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 253 | struct opregion_asle *asle = dev_priv->opregion.asle; | ||
| 254 | u32 cpu_pwm_ctl, pch_pwm_ctl2; | ||
| 255 | u32 max_backlight, level; | ||
| 256 | |||
| 257 | if (!(bclp & ASLE_BCLP_VALID)) | ||
| 258 | return ASLE_BACKLIGHT_FAILED; | ||
| 259 | |||
| 260 | bclp &= ASLE_BCLP_MSK; | ||
| 261 | if (bclp < 0 || bclp > 255) | ||
| 262 | return ASLE_BACKLIGHT_FAILED; | ||
| 263 | |||
| 264 | cpu_pwm_ctl = I915_READ(BLC_PWM_CPU_CTL); | ||
| 265 | pch_pwm_ctl2 = I915_READ(BLC_PWM_PCH_CTL2); | ||
| 266 | /* get the max PWM frequency */ | ||
| 267 | max_backlight = (pch_pwm_ctl2 >> 16) & BACKLIGHT_DUTY_CYCLE_MASK; | ||
| 268 | /* calculate the expected PMW frequency */ | ||
| 269 | level = (bclp * max_backlight) / 255; | ||
| 270 | /* reserve the high 16 bits */ | ||
| 271 | cpu_pwm_ctl &= ~(BACKLIGHT_DUTY_CYCLE_MASK); | ||
| 272 | /* write the updated PWM frequency */ | ||
| 273 | I915_WRITE(BLC_PWM_CPU_CTL, cpu_pwm_ctl | level); | ||
| 274 | |||
| 275 | asle->cblv = (bclp*0x64)/0xff | ASLE_CBLV_VALID; | ||
| 276 | |||
| 277 | return 0; | ||
| 278 | } | ||
| 279 | |||
| 280 | void ironlake_opregion_gse_intr(struct drm_device *dev) | ||
| 281 | { | ||
| 282 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 283 | struct opregion_asle *asle = dev_priv->opregion.asle; | ||
| 284 | u32 asle_stat = 0; | ||
| 285 | u32 asle_req; | ||
| 286 | |||
| 287 | if (!asle) | ||
| 288 | return; | ||
| 289 | |||
| 290 | asle_req = asle->aslc & ASLE_REQ_MSK; | ||
| 291 | |||
| 292 | if (!asle_req) { | ||
| 293 | DRM_DEBUG_DRIVER("non asle set request??\n"); | ||
| 294 | return; | ||
| 295 | } | ||
| 296 | |||
| 297 | if (asle_req & ASLE_SET_ALS_ILLUM) { | ||
| 298 | DRM_DEBUG_DRIVER("Illum is not supported\n"); | ||
| 299 | asle_stat |= ASLE_ALS_ILLUM_FAILED; | ||
| 300 | } | ||
| 301 | |||
| 302 | if (asle_req & ASLE_SET_BACKLIGHT) | ||
| 303 | asle_stat |= asle_set_backlight_ironlake(dev, asle->bclp); | ||
| 304 | |||
| 305 | if (asle_req & ASLE_SET_PFIT) { | ||
| 306 | DRM_DEBUG_DRIVER("Pfit is not supported\n"); | ||
| 307 | asle_stat |= ASLE_PFIT_FAILED; | ||
| 308 | } | ||
| 309 | |||
| 310 | if (asle_req & ASLE_SET_PWM_FREQ) { | ||
| 311 | DRM_DEBUG_DRIVER("PWM freq is not supported\n"); | ||
| 312 | asle_stat |= ASLE_PWM_FREQ_FAILED; | ||
| 313 | } | ||
| 314 | |||
| 315 | asle->aslc = asle_stat; | ||
| 316 | } | ||
| 246 | #define ASLE_ALS_EN (1<<0) | 317 | #define ASLE_ALS_EN (1<<0) |
| 247 | #define ASLE_BLC_EN (1<<1) | 318 | #define ASLE_BLC_EN (1<<1) |
| 248 | #define ASLE_PFIT_EN (1<<2) | 319 | #define ASLE_PFIT_EN (1<<2) |
| @@ -258,8 +329,7 @@ void opregion_enable_asle(struct drm_device *dev) | |||
| 258 | unsigned long irqflags; | 329 | unsigned long irqflags; |
| 259 | 330 | ||
| 260 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); | 331 | spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags); |
| 261 | i915_enable_pipestat(dev_priv, 1, | 332 | intel_enable_asle(dev); |
| 262 | I915_LEGACY_BLC_EVENT_ENABLE); | ||
| 263 | spin_unlock_irqrestore(&dev_priv->user_irq_lock, | 333 | spin_unlock_irqrestore(&dev_priv->user_irq_lock, |
| 264 | irqflags); | 334 | irqflags); |
| 265 | } | 335 | } |
| @@ -361,9 +431,9 @@ int intel_opregion_init(struct drm_device *dev, int resume) | |||
| 361 | int err = 0; | 431 | int err = 0; |
| 362 | 432 | ||
| 363 | pci_read_config_dword(dev->pdev, PCI_ASLS, &asls); | 433 | pci_read_config_dword(dev->pdev, PCI_ASLS, &asls); |
| 364 | DRM_DEBUG("graphic opregion physical addr: 0x%x\n", asls); | 434 | DRM_DEBUG_DRIVER("graphic opregion physical addr: 0x%x\n", asls); |
| 365 | if (asls == 0) { | 435 | if (asls == 0) { |
| 366 | DRM_DEBUG("ACPI OpRegion not supported!\n"); | 436 | DRM_DEBUG_DRIVER("ACPI OpRegion not supported!\n"); |
| 367 | return -ENOTSUPP; | 437 | return -ENOTSUPP; |
| 368 | } | 438 | } |
| 369 | 439 | ||
| @@ -373,30 +443,30 @@ int intel_opregion_init(struct drm_device *dev, int resume) | |||
| 373 | 443 | ||
| 374 | opregion->header = base; | 444 | opregion->header = base; |
| 375 | if (memcmp(opregion->header->signature, OPREGION_SIGNATURE, 16)) { | 445 | if (memcmp(opregion->header->signature, OPREGION_SIGNATURE, 16)) { |
| 376 | DRM_DEBUG("opregion signature mismatch\n"); | 446 | DRM_DEBUG_DRIVER("opregion signature mismatch\n"); |
| 377 | err = -EINVAL; | 447 | err = -EINVAL; |
| 378 | goto err_out; | 448 | goto err_out; |
| 379 | } | 449 | } |
| 380 | 450 | ||
| 381 | mboxes = opregion->header->mboxes; | 451 | mboxes = opregion->header->mboxes; |
| 382 | if (mboxes & MBOX_ACPI) { | 452 | if (mboxes & MBOX_ACPI) { |
| 383 | DRM_DEBUG("Public ACPI methods supported\n"); | 453 | DRM_DEBUG_DRIVER("Public ACPI methods supported\n"); |
| 384 | opregion->acpi = base + OPREGION_ACPI_OFFSET; | 454 | opregion->acpi = base + OPREGION_ACPI_OFFSET; |
| 385 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | 455 | if (drm_core_check_feature(dev, DRIVER_MODESET)) |
| 386 | intel_didl_outputs(dev); | 456 | intel_didl_outputs(dev); |
| 387 | } else { | 457 | } else { |
| 388 | DRM_DEBUG("Public ACPI methods not supported\n"); | 458 | DRM_DEBUG_DRIVER("Public ACPI methods not supported\n"); |
| 389 | err = -ENOTSUPP; | 459 | err = -ENOTSUPP; |
| 390 | goto err_out; | 460 | goto err_out; |
| 391 | } | 461 | } |
| 392 | opregion->enabled = 1; | 462 | opregion->enabled = 1; |
| 393 | 463 | ||
| 394 | if (mboxes & MBOX_SWSCI) { | 464 | if (mboxes & MBOX_SWSCI) { |
| 395 | DRM_DEBUG("SWSCI supported\n"); | 465 | DRM_DEBUG_DRIVER("SWSCI supported\n"); |
| 396 | opregion->swsci = base + OPREGION_SWSCI_OFFSET; | 466 | opregion->swsci = base + OPREGION_SWSCI_OFFSET; |
| 397 | } | 467 | } |
| 398 | if (mboxes & MBOX_ASLE) { | 468 | if (mboxes & MBOX_ASLE) { |
| 399 | DRM_DEBUG("ASLE supported\n"); | 469 | DRM_DEBUG_DRIVER("ASLE supported\n"); |
| 400 | opregion->asle = base + OPREGION_ASLE_OFFSET; | 470 | opregion->asle = base + OPREGION_ASLE_OFFSET; |
| 401 | opregion_enable_asle(dev); | 471 | opregion_enable_asle(dev); |
| 402 | } | 472 | } |
diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h index 1687edf68795..974b3cf70618 100644 --- a/drivers/gpu/drm/i915/i915_reg.h +++ b/drivers/gpu/drm/i915/i915_reg.h | |||
| @@ -140,6 +140,7 @@ | |||
| 140 | #define MI_NOOP MI_INSTR(0, 0) | 140 | #define MI_NOOP MI_INSTR(0, 0) |
| 141 | #define MI_USER_INTERRUPT MI_INSTR(0x02, 0) | 141 | #define MI_USER_INTERRUPT MI_INSTR(0x02, 0) |
| 142 | #define MI_WAIT_FOR_EVENT MI_INSTR(0x03, 0) | 142 | #define MI_WAIT_FOR_EVENT MI_INSTR(0x03, 0) |
| 143 | #define MI_WAIT_FOR_OVERLAY_FLIP (1<<16) | ||
| 143 | #define MI_WAIT_FOR_PLANE_B_FLIP (1<<6) | 144 | #define MI_WAIT_FOR_PLANE_B_FLIP (1<<6) |
| 144 | #define MI_WAIT_FOR_PLANE_A_FLIP (1<<2) | 145 | #define MI_WAIT_FOR_PLANE_A_FLIP (1<<2) |
| 145 | #define MI_WAIT_FOR_PLANE_A_SCANLINES (1<<1) | 146 | #define MI_WAIT_FOR_PLANE_A_SCANLINES (1<<1) |
| @@ -151,7 +152,13 @@ | |||
| 151 | #define MI_END_SCENE (1 << 4) /* flush binner and incr scene count */ | 152 | #define MI_END_SCENE (1 << 4) /* flush binner and incr scene count */ |
| 152 | #define MI_BATCH_BUFFER_END MI_INSTR(0x0a, 0) | 153 | #define MI_BATCH_BUFFER_END MI_INSTR(0x0a, 0) |
| 153 | #define MI_REPORT_HEAD MI_INSTR(0x07, 0) | 154 | #define MI_REPORT_HEAD MI_INSTR(0x07, 0) |
| 155 | #define MI_OVERLAY_FLIP MI_INSTR(0x11,0) | ||
| 156 | #define MI_OVERLAY_CONTINUE (0x0<<21) | ||
| 157 | #define MI_OVERLAY_ON (0x1<<21) | ||
| 158 | #define MI_OVERLAY_OFF (0x2<<21) | ||
| 154 | #define MI_LOAD_SCAN_LINES_INCL MI_INSTR(0x12, 0) | 159 | #define MI_LOAD_SCAN_LINES_INCL MI_INSTR(0x12, 0) |
| 160 | #define MI_DISPLAY_FLIP MI_INSTR(0x14, 2) | ||
| 161 | #define MI_DISPLAY_FLIP_PLANE(n) ((n) << 20) | ||
| 155 | #define MI_STORE_DWORD_IMM MI_INSTR(0x20, 1) | 162 | #define MI_STORE_DWORD_IMM MI_INSTR(0x20, 1) |
| 156 | #define MI_MEM_VIRTUAL (1 << 22) /* 965+ only */ | 163 | #define MI_MEM_VIRTUAL (1 << 22) /* 965+ only */ |
| 157 | #define MI_STORE_DWORD_INDEX MI_INSTR(0x21, 1) | 164 | #define MI_STORE_DWORD_INDEX MI_INSTR(0x21, 1) |
| @@ -260,6 +267,8 @@ | |||
| 260 | #define HWS_PGA 0x02080 | 267 | #define HWS_PGA 0x02080 |
| 261 | #define HWS_ADDRESS_MASK 0xfffff000 | 268 | #define HWS_ADDRESS_MASK 0xfffff000 |
| 262 | #define HWS_START_ADDRESS_SHIFT 4 | 269 | #define HWS_START_ADDRESS_SHIFT 4 |
| 270 | #define PWRCTXA 0x2088 /* 965GM+ only */ | ||
| 271 | #define PWRCTX_EN (1<<0) | ||
| 263 | #define IPEIR 0x02088 | 272 | #define IPEIR 0x02088 |
| 264 | #define IPEHR 0x0208c | 273 | #define IPEHR 0x0208c |
| 265 | #define INSTDONE 0x02090 | 274 | #define INSTDONE 0x02090 |
| @@ -405,6 +414,13 @@ | |||
| 405 | # define GPIO_DATA_VAL_IN (1 << 12) | 414 | # define GPIO_DATA_VAL_IN (1 << 12) |
| 406 | # define GPIO_DATA_PULLUP_DISABLE (1 << 13) | 415 | # define GPIO_DATA_PULLUP_DISABLE (1 << 13) |
| 407 | 416 | ||
| 417 | #define GMBUS0 0x5100 | ||
| 418 | #define GMBUS1 0x5104 | ||
| 419 | #define GMBUS2 0x5108 | ||
| 420 | #define GMBUS3 0x510c | ||
| 421 | #define GMBUS4 0x5110 | ||
| 422 | #define GMBUS5 0x5120 | ||
| 423 | |||
| 408 | /* | 424 | /* |
| 409 | * Clock control & power management | 425 | * Clock control & power management |
| 410 | */ | 426 | */ |
| @@ -435,7 +451,7 @@ | |||
| 435 | #define DPLLB_LVDS_P2_CLOCK_DIV_7 (1 << 24) /* i915 */ | 451 | #define DPLLB_LVDS_P2_CLOCK_DIV_7 (1 << 24) /* i915 */ |
| 436 | #define DPLL_P2_CLOCK_DIV_MASK 0x03000000 /* i915 */ | 452 | #define DPLL_P2_CLOCK_DIV_MASK 0x03000000 /* i915 */ |
| 437 | #define DPLL_FPA01_P1_POST_DIV_MASK 0x00ff0000 /* i915 */ | 453 | #define DPLL_FPA01_P1_POST_DIV_MASK 0x00ff0000 /* i915 */ |
| 438 | #define DPLL_FPA01_P1_POST_DIV_MASK_IGD 0x00ff8000 /* IGD */ | 454 | #define DPLL_FPA01_P1_POST_DIV_MASK_PINEVIEW 0x00ff8000 /* Pineview */ |
| 439 | 455 | ||
| 440 | #define I915_FIFO_UNDERRUN_STATUS (1UL<<31) | 456 | #define I915_FIFO_UNDERRUN_STATUS (1UL<<31) |
| 441 | #define I915_CRC_ERROR_ENABLE (1UL<<29) | 457 | #define I915_CRC_ERROR_ENABLE (1UL<<29) |
| @@ -512,7 +528,7 @@ | |||
| 512 | */ | 528 | */ |
| 513 | #define DPLL_FPA01_P1_POST_DIV_MASK_I830_LVDS 0x003f0000 | 529 | #define DPLL_FPA01_P1_POST_DIV_MASK_I830_LVDS 0x003f0000 |
| 514 | #define DPLL_FPA01_P1_POST_DIV_SHIFT 16 | 530 | #define DPLL_FPA01_P1_POST_DIV_SHIFT 16 |
| 515 | #define DPLL_FPA01_P1_POST_DIV_SHIFT_IGD 15 | 531 | #define DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW 15 |
| 516 | /* i830, required in DVO non-gang */ | 532 | /* i830, required in DVO non-gang */ |
| 517 | #define PLL_P2_DIVIDE_BY_4 (1 << 23) | 533 | #define PLL_P2_DIVIDE_BY_4 (1 << 23) |
| 518 | #define PLL_P1_DIVIDE_BY_TWO (1 << 21) /* i830 */ | 534 | #define PLL_P1_DIVIDE_BY_TWO (1 << 21) /* i830 */ |
| @@ -522,7 +538,7 @@ | |||
| 522 | #define PLLB_REF_INPUT_SPREADSPECTRUMIN (3 << 13) | 538 | #define PLLB_REF_INPUT_SPREADSPECTRUMIN (3 << 13) |
| 523 | #define PLL_REF_INPUT_MASK (3 << 13) | 539 | #define PLL_REF_INPUT_MASK (3 << 13) |
| 524 | #define PLL_LOAD_PULSE_PHASE_SHIFT 9 | 540 | #define PLL_LOAD_PULSE_PHASE_SHIFT 9 |
| 525 | /* IGDNG */ | 541 | /* Ironlake */ |
| 526 | # define PLL_REF_SDVO_HDMI_MULTIPLIER_SHIFT 9 | 542 | # define PLL_REF_SDVO_HDMI_MULTIPLIER_SHIFT 9 |
| 527 | # define PLL_REF_SDVO_HDMI_MULTIPLIER_MASK (7 << 9) | 543 | # define PLL_REF_SDVO_HDMI_MULTIPLIER_MASK (7 << 9) |
| 528 | # define PLL_REF_SDVO_HDMI_MULTIPLIER(x) (((x)-1) << 9) | 544 | # define PLL_REF_SDVO_HDMI_MULTIPLIER(x) (((x)-1) << 9) |
| @@ -586,12 +602,12 @@ | |||
| 586 | #define FPB0 0x06048 | 602 | #define FPB0 0x06048 |
| 587 | #define FPB1 0x0604c | 603 | #define FPB1 0x0604c |
| 588 | #define FP_N_DIV_MASK 0x003f0000 | 604 | #define FP_N_DIV_MASK 0x003f0000 |
| 589 | #define FP_N_IGD_DIV_MASK 0x00ff0000 | 605 | #define FP_N_PINEVIEW_DIV_MASK 0x00ff0000 |
| 590 | #define FP_N_DIV_SHIFT 16 | 606 | #define FP_N_DIV_SHIFT 16 |
| 591 | #define FP_M1_DIV_MASK 0x00003f00 | 607 | #define FP_M1_DIV_MASK 0x00003f00 |
| 592 | #define FP_M1_DIV_SHIFT 8 | 608 | #define FP_M1_DIV_SHIFT 8 |
| 593 | #define FP_M2_DIV_MASK 0x0000003f | 609 | #define FP_M2_DIV_MASK 0x0000003f |
| 594 | #define FP_M2_IGD_DIV_MASK 0x000000ff | 610 | #define FP_M2_PINEVIEW_DIV_MASK 0x000000ff |
| 595 | #define FP_M2_DIV_SHIFT 0 | 611 | #define FP_M2_DIV_SHIFT 0 |
| 596 | #define DPLL_TEST 0x606c | 612 | #define DPLL_TEST 0x606c |
| 597 | #define DPLLB_TEST_SDVO_DIV_1 (0 << 22) | 613 | #define DPLLB_TEST_SDVO_DIV_1 (0 << 22) |
| @@ -769,7 +785,8 @@ | |||
| 769 | 785 | ||
| 770 | /** GM965 GM45 render standby register */ | 786 | /** GM965 GM45 render standby register */ |
| 771 | #define MCHBAR_RENDER_STANDBY 0x111B8 | 787 | #define MCHBAR_RENDER_STANDBY 0x111B8 |
| 772 | 788 | #define RCX_SW_EXIT (1<<23) | |
| 789 | #define RSX_STATUS_MASK 0x00700000 | ||
| 773 | #define PEG_BAND_GAP_DATA 0x14d68 | 790 | #define PEG_BAND_GAP_DATA 0x14d68 |
| 774 | 791 | ||
| 775 | /* | 792 | /* |
| @@ -844,7 +861,6 @@ | |||
| 844 | #define SDVOB_HOTPLUG_INT_EN (1 << 26) | 861 | #define SDVOB_HOTPLUG_INT_EN (1 << 26) |
| 845 | #define SDVOC_HOTPLUG_INT_EN (1 << 25) | 862 | #define SDVOC_HOTPLUG_INT_EN (1 << 25) |
| 846 | #define TV_HOTPLUG_INT_EN (1 << 18) | 863 | #define TV_HOTPLUG_INT_EN (1 << 18) |
| 847 | #define CRT_EOS_INT_EN (1 << 10) | ||
| 848 | #define CRT_HOTPLUG_INT_EN (1 << 9) | 864 | #define CRT_HOTPLUG_INT_EN (1 << 9) |
| 849 | #define CRT_HOTPLUG_FORCE_DETECT (1 << 3) | 865 | #define CRT_HOTPLUG_FORCE_DETECT (1 << 3) |
| 850 | #define CRT_HOTPLUG_ACTIVATION_PERIOD_32 (0 << 8) | 866 | #define CRT_HOTPLUG_ACTIVATION_PERIOD_32 (0 << 8) |
| @@ -868,7 +884,6 @@ | |||
| 868 | HDMID_HOTPLUG_INT_EN | \ | 884 | HDMID_HOTPLUG_INT_EN | \ |
| 869 | SDVOB_HOTPLUG_INT_EN | \ | 885 | SDVOB_HOTPLUG_INT_EN | \ |
| 870 | SDVOC_HOTPLUG_INT_EN | \ | 886 | SDVOC_HOTPLUG_INT_EN | \ |
| 871 | TV_HOTPLUG_INT_EN | \ | ||
| 872 | CRT_HOTPLUG_INT_EN) | 887 | CRT_HOTPLUG_INT_EN) |
| 873 | 888 | ||
| 874 | 889 | ||
| @@ -879,7 +894,6 @@ | |||
| 879 | #define DPC_HOTPLUG_INT_STATUS (1 << 28) | 894 | #define DPC_HOTPLUG_INT_STATUS (1 << 28) |
| 880 | #define HDMID_HOTPLUG_INT_STATUS (1 << 27) | 895 | #define HDMID_HOTPLUG_INT_STATUS (1 << 27) |
| 881 | #define DPD_HOTPLUG_INT_STATUS (1 << 27) | 896 | #define DPD_HOTPLUG_INT_STATUS (1 << 27) |
| 882 | #define CRT_EOS_INT_STATUS (1 << 12) | ||
| 883 | #define CRT_HOTPLUG_INT_STATUS (1 << 11) | 897 | #define CRT_HOTPLUG_INT_STATUS (1 << 11) |
| 884 | #define TV_HOTPLUG_INT_STATUS (1 << 10) | 898 | #define TV_HOTPLUG_INT_STATUS (1 << 10) |
| 885 | #define CRT_HOTPLUG_MONITOR_MASK (3 << 8) | 899 | #define CRT_HOTPLUG_MONITOR_MASK (3 << 8) |
| @@ -1620,7 +1634,7 @@ | |||
| 1620 | #define DP_CLOCK_OUTPUT_ENABLE (1 << 13) | 1634 | #define DP_CLOCK_OUTPUT_ENABLE (1 << 13) |
| 1621 | 1635 | ||
| 1622 | #define DP_SCRAMBLING_DISABLE (1 << 12) | 1636 | #define DP_SCRAMBLING_DISABLE (1 << 12) |
| 1623 | #define DP_SCRAMBLING_DISABLE_IGDNG (1 << 7) | 1637 | #define DP_SCRAMBLING_DISABLE_IRONLAKE (1 << 7) |
| 1624 | 1638 | ||
| 1625 | /** limit RGB values to avoid confusing TVs */ | 1639 | /** limit RGB values to avoid confusing TVs */ |
| 1626 | #define DP_COLOR_RANGE_16_235 (1 << 8) | 1640 | #define DP_COLOR_RANGE_16_235 (1 << 8) |
| @@ -1808,7 +1822,7 @@ | |||
| 1808 | #define DSPFW3 0x7003c | 1822 | #define DSPFW3 0x7003c |
| 1809 | #define DSPFW_HPLL_SR_EN (1<<31) | 1823 | #define DSPFW_HPLL_SR_EN (1<<31) |
| 1810 | #define DSPFW_CURSOR_SR_SHIFT 24 | 1824 | #define DSPFW_CURSOR_SR_SHIFT 24 |
| 1811 | #define IGD_SELF_REFRESH_EN (1<<30) | 1825 | #define PINEVIEW_SELF_REFRESH_EN (1<<30) |
| 1812 | 1826 | ||
| 1813 | /* FIFO watermark sizes etc */ | 1827 | /* FIFO watermark sizes etc */ |
| 1814 | #define G4X_FIFO_LINE_SIZE 64 | 1828 | #define G4X_FIFO_LINE_SIZE 64 |
| @@ -1824,16 +1838,16 @@ | |||
| 1824 | #define G4X_MAX_WM 0x3f | 1838 | #define G4X_MAX_WM 0x3f |
| 1825 | #define I915_MAX_WM 0x3f | 1839 | #define I915_MAX_WM 0x3f |
| 1826 | 1840 | ||
| 1827 | #define IGD_DISPLAY_FIFO 512 /* in 64byte unit */ | 1841 | #define PINEVIEW_DISPLAY_FIFO 512 /* in 64byte unit */ |
| 1828 | #define IGD_FIFO_LINE_SIZE 64 | 1842 | #define PINEVIEW_FIFO_LINE_SIZE 64 |
| 1829 | #define IGD_MAX_WM 0x1ff | 1843 | #define PINEVIEW_MAX_WM 0x1ff |
| 1830 | #define IGD_DFT_WM 0x3f | 1844 | #define PINEVIEW_DFT_WM 0x3f |
| 1831 | #define IGD_DFT_HPLLOFF_WM 0 | 1845 | #define PINEVIEW_DFT_HPLLOFF_WM 0 |
| 1832 | #define IGD_GUARD_WM 10 | 1846 | #define PINEVIEW_GUARD_WM 10 |
| 1833 | #define IGD_CURSOR_FIFO 64 | 1847 | #define PINEVIEW_CURSOR_FIFO 64 |
| 1834 | #define IGD_CURSOR_MAX_WM 0x3f | 1848 | #define PINEVIEW_CURSOR_MAX_WM 0x3f |
| 1835 | #define IGD_CURSOR_DFT_WM 0 | 1849 | #define PINEVIEW_CURSOR_DFT_WM 0 |
| 1836 | #define IGD_CURSOR_GUARD_WM 5 | 1850 | #define PINEVIEW_CURSOR_GUARD_WM 5 |
| 1837 | 1851 | ||
| 1838 | /* | 1852 | /* |
| 1839 | * The two pipe frame counter registers are not synchronized, so | 1853 | * The two pipe frame counter registers are not synchronized, so |
| @@ -1907,6 +1921,7 @@ | |||
| 1907 | #define DISPPLANE_16BPP (0x5<<26) | 1921 | #define DISPPLANE_16BPP (0x5<<26) |
| 1908 | #define DISPPLANE_32BPP_NO_ALPHA (0x6<<26) | 1922 | #define DISPPLANE_32BPP_NO_ALPHA (0x6<<26) |
| 1909 | #define DISPPLANE_32BPP (0x7<<26) | 1923 | #define DISPPLANE_32BPP (0x7<<26) |
| 1924 | #define DISPPLANE_32BPP_30BIT_NO_ALPHA (0xa<<26) | ||
| 1910 | #define DISPPLANE_STEREO_ENABLE (1<<25) | 1925 | #define DISPPLANE_STEREO_ENABLE (1<<25) |
| 1911 | #define DISPPLANE_STEREO_DISABLE 0 | 1926 | #define DISPPLANE_STEREO_DISABLE 0 |
| 1912 | #define DISPPLANE_SEL_PIPE_MASK (1<<24) | 1927 | #define DISPPLANE_SEL_PIPE_MASK (1<<24) |
| @@ -1918,7 +1933,7 @@ | |||
| 1918 | #define DISPPLANE_NO_LINE_DOUBLE 0 | 1933 | #define DISPPLANE_NO_LINE_DOUBLE 0 |
| 1919 | #define DISPPLANE_STEREO_POLARITY_FIRST 0 | 1934 | #define DISPPLANE_STEREO_POLARITY_FIRST 0 |
| 1920 | #define DISPPLANE_STEREO_POLARITY_SECOND (1<<18) | 1935 | #define DISPPLANE_STEREO_POLARITY_SECOND (1<<18) |
| 1921 | #define DISPPLANE_TRICKLE_FEED_DISABLE (1<<14) /* IGDNG */ | 1936 | #define DISPPLANE_TRICKLE_FEED_DISABLE (1<<14) /* Ironlake */ |
| 1922 | #define DISPPLANE_TILED (1<<10) | 1937 | #define DISPPLANE_TILED (1<<10) |
| 1923 | #define DSPAADDR 0x70184 | 1938 | #define DSPAADDR 0x70184 |
| 1924 | #define DSPASTRIDE 0x70188 | 1939 | #define DSPASTRIDE 0x70188 |
| @@ -1971,7 +1986,7 @@ | |||
| 1971 | # define VGA_2X_MODE (1 << 30) | 1986 | # define VGA_2X_MODE (1 << 30) |
| 1972 | # define VGA_PIPE_B_SELECT (1 << 29) | 1987 | # define VGA_PIPE_B_SELECT (1 << 29) |
| 1973 | 1988 | ||
| 1974 | /* IGDNG */ | 1989 | /* Ironlake */ |
| 1975 | 1990 | ||
| 1976 | #define CPU_VGACNTRL 0x41000 | 1991 | #define CPU_VGACNTRL 0x41000 |
| 1977 | 1992 | ||
| @@ -2117,6 +2132,7 @@ | |||
| 2117 | #define SDE_PORTC_HOTPLUG (1 << 9) | 2132 | #define SDE_PORTC_HOTPLUG (1 << 9) |
| 2118 | #define SDE_PORTB_HOTPLUG (1 << 8) | 2133 | #define SDE_PORTB_HOTPLUG (1 << 8) |
| 2119 | #define SDE_SDVOB_HOTPLUG (1 << 6) | 2134 | #define SDE_SDVOB_HOTPLUG (1 << 6) |
| 2135 | #define SDE_HOTPLUG_MASK (0xf << 8) | ||
| 2120 | 2136 | ||
| 2121 | #define SDEISR 0xc4000 | 2137 | #define SDEISR 0xc4000 |
| 2122 | #define SDEIMR 0xc4004 | 2138 | #define SDEIMR 0xc4004 |
| @@ -2157,6 +2173,13 @@ | |||
| 2157 | #define PCH_GPIOE 0xc5020 | 2173 | #define PCH_GPIOE 0xc5020 |
| 2158 | #define PCH_GPIOF 0xc5024 | 2174 | #define PCH_GPIOF 0xc5024 |
| 2159 | 2175 | ||
| 2176 | #define PCH_GMBUS0 0xc5100 | ||
| 2177 | #define PCH_GMBUS1 0xc5104 | ||
| 2178 | #define PCH_GMBUS2 0xc5108 | ||
| 2179 | #define PCH_GMBUS3 0xc510c | ||
| 2180 | #define PCH_GMBUS4 0xc5110 | ||
| 2181 | #define PCH_GMBUS5 0xc5120 | ||
| 2182 | |||
| 2160 | #define PCH_DPLL_A 0xc6014 | 2183 | #define PCH_DPLL_A 0xc6014 |
| 2161 | #define PCH_DPLL_B 0xc6018 | 2184 | #define PCH_DPLL_B 0xc6018 |
| 2162 | 2185 | ||
| @@ -2292,7 +2315,7 @@ | |||
| 2292 | #define FDI_DP_PORT_WIDTH_X3 (2<<19) | 2315 | #define FDI_DP_PORT_WIDTH_X3 (2<<19) |
| 2293 | #define FDI_DP_PORT_WIDTH_X4 (3<<19) | 2316 | #define FDI_DP_PORT_WIDTH_X4 (3<<19) |
| 2294 | #define FDI_TX_ENHANCE_FRAME_ENABLE (1<<18) | 2317 | #define FDI_TX_ENHANCE_FRAME_ENABLE (1<<18) |
| 2295 | /* IGDNG: hardwired to 1 */ | 2318 | /* Ironlake: hardwired to 1 */ |
| 2296 | #define FDI_TX_PLL_ENABLE (1<<14) | 2319 | #define FDI_TX_PLL_ENABLE (1<<14) |
| 2297 | /* both Tx and Rx */ | 2320 | /* both Tx and Rx */ |
| 2298 | #define FDI_SCRAMBLING_ENABLE (0<<7) | 2321 | #define FDI_SCRAMBLING_ENABLE (0<<7) |
diff --git a/drivers/gpu/drm/i915/i915_suspend.c b/drivers/gpu/drm/i915/i915_suspend.c index 6eec8171a44e..d5ebb00a9d49 100644 --- a/drivers/gpu/drm/i915/i915_suspend.c +++ b/drivers/gpu/drm/i915/i915_suspend.c | |||
| @@ -27,14 +27,14 @@ | |||
| 27 | #include "drmP.h" | 27 | #include "drmP.h" |
| 28 | #include "drm.h" | 28 | #include "drm.h" |
| 29 | #include "i915_drm.h" | 29 | #include "i915_drm.h" |
| 30 | #include "i915_drv.h" | 30 | #include "intel_drv.h" |
| 31 | 31 | ||
| 32 | static bool i915_pipe_enabled(struct drm_device *dev, enum pipe pipe) | 32 | static bool i915_pipe_enabled(struct drm_device *dev, enum pipe pipe) |
| 33 | { | 33 | { |
| 34 | struct drm_i915_private *dev_priv = dev->dev_private; | 34 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 35 | u32 dpll_reg; | 35 | u32 dpll_reg; |
| 36 | 36 | ||
| 37 | if (IS_IGDNG(dev)) { | 37 | if (IS_IRONLAKE(dev)) { |
| 38 | dpll_reg = (pipe == PIPE_A) ? PCH_DPLL_A: PCH_DPLL_B; | 38 | dpll_reg = (pipe == PIPE_A) ? PCH_DPLL_A: PCH_DPLL_B; |
| 39 | } else { | 39 | } else { |
| 40 | dpll_reg = (pipe == PIPE_A) ? DPLL_A: DPLL_B; | 40 | dpll_reg = (pipe == PIPE_A) ? DPLL_A: DPLL_B; |
| @@ -53,7 +53,7 @@ static void i915_save_palette(struct drm_device *dev, enum pipe pipe) | |||
| 53 | if (!i915_pipe_enabled(dev, pipe)) | 53 | if (!i915_pipe_enabled(dev, pipe)) |
| 54 | return; | 54 | return; |
| 55 | 55 | ||
| 56 | if (IS_IGDNG(dev)) | 56 | if (IS_IRONLAKE(dev)) |
| 57 | reg = (pipe == PIPE_A) ? LGC_PALETTE_A : LGC_PALETTE_B; | 57 | reg = (pipe == PIPE_A) ? LGC_PALETTE_A : LGC_PALETTE_B; |
| 58 | 58 | ||
| 59 | if (pipe == PIPE_A) | 59 | if (pipe == PIPE_A) |
| @@ -75,7 +75,7 @@ static void i915_restore_palette(struct drm_device *dev, enum pipe pipe) | |||
| 75 | if (!i915_pipe_enabled(dev, pipe)) | 75 | if (!i915_pipe_enabled(dev, pipe)) |
| 76 | return; | 76 | return; |
| 77 | 77 | ||
| 78 | if (IS_IGDNG(dev)) | 78 | if (IS_IRONLAKE(dev)) |
| 79 | reg = (pipe == PIPE_A) ? LGC_PALETTE_A : LGC_PALETTE_B; | 79 | reg = (pipe == PIPE_A) ? LGC_PALETTE_A : LGC_PALETTE_B; |
| 80 | 80 | ||
| 81 | if (pipe == PIPE_A) | 81 | if (pipe == PIPE_A) |
| @@ -239,7 +239,7 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 239 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | 239 | if (drm_core_check_feature(dev, DRIVER_MODESET)) |
| 240 | return; | 240 | return; |
| 241 | 241 | ||
| 242 | if (IS_IGDNG(dev)) { | 242 | if (IS_IRONLAKE(dev)) { |
| 243 | dev_priv->savePCH_DREF_CONTROL = I915_READ(PCH_DREF_CONTROL); | 243 | dev_priv->savePCH_DREF_CONTROL = I915_READ(PCH_DREF_CONTROL); |
| 244 | dev_priv->saveDISP_ARB_CTL = I915_READ(DISP_ARB_CTL); | 244 | dev_priv->saveDISP_ARB_CTL = I915_READ(DISP_ARB_CTL); |
| 245 | } | 245 | } |
| @@ -247,7 +247,7 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 247 | /* Pipe & plane A info */ | 247 | /* Pipe & plane A info */ |
| 248 | dev_priv->savePIPEACONF = I915_READ(PIPEACONF); | 248 | dev_priv->savePIPEACONF = I915_READ(PIPEACONF); |
| 249 | dev_priv->savePIPEASRC = I915_READ(PIPEASRC); | 249 | dev_priv->savePIPEASRC = I915_READ(PIPEASRC); |
| 250 | if (IS_IGDNG(dev)) { | 250 | if (IS_IRONLAKE(dev)) { |
| 251 | dev_priv->saveFPA0 = I915_READ(PCH_FPA0); | 251 | dev_priv->saveFPA0 = I915_READ(PCH_FPA0); |
| 252 | dev_priv->saveFPA1 = I915_READ(PCH_FPA1); | 252 | dev_priv->saveFPA1 = I915_READ(PCH_FPA1); |
| 253 | dev_priv->saveDPLL_A = I915_READ(PCH_DPLL_A); | 253 | dev_priv->saveDPLL_A = I915_READ(PCH_DPLL_A); |
| @@ -256,7 +256,7 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 256 | dev_priv->saveFPA1 = I915_READ(FPA1); | 256 | dev_priv->saveFPA1 = I915_READ(FPA1); |
| 257 | dev_priv->saveDPLL_A = I915_READ(DPLL_A); | 257 | dev_priv->saveDPLL_A = I915_READ(DPLL_A); |
| 258 | } | 258 | } |
| 259 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 259 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 260 | dev_priv->saveDPLL_A_MD = I915_READ(DPLL_A_MD); | 260 | dev_priv->saveDPLL_A_MD = I915_READ(DPLL_A_MD); |
| 261 | dev_priv->saveHTOTAL_A = I915_READ(HTOTAL_A); | 261 | dev_priv->saveHTOTAL_A = I915_READ(HTOTAL_A); |
| 262 | dev_priv->saveHBLANK_A = I915_READ(HBLANK_A); | 262 | dev_priv->saveHBLANK_A = I915_READ(HBLANK_A); |
| @@ -264,10 +264,10 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 264 | dev_priv->saveVTOTAL_A = I915_READ(VTOTAL_A); | 264 | dev_priv->saveVTOTAL_A = I915_READ(VTOTAL_A); |
| 265 | dev_priv->saveVBLANK_A = I915_READ(VBLANK_A); | 265 | dev_priv->saveVBLANK_A = I915_READ(VBLANK_A); |
| 266 | dev_priv->saveVSYNC_A = I915_READ(VSYNC_A); | 266 | dev_priv->saveVSYNC_A = I915_READ(VSYNC_A); |
| 267 | if (!IS_IGDNG(dev)) | 267 | if (!IS_IRONLAKE(dev)) |
| 268 | dev_priv->saveBCLRPAT_A = I915_READ(BCLRPAT_A); | 268 | dev_priv->saveBCLRPAT_A = I915_READ(BCLRPAT_A); |
| 269 | 269 | ||
| 270 | if (IS_IGDNG(dev)) { | 270 | if (IS_IRONLAKE(dev)) { |
| 271 | dev_priv->savePIPEA_DATA_M1 = I915_READ(PIPEA_DATA_M1); | 271 | dev_priv->savePIPEA_DATA_M1 = I915_READ(PIPEA_DATA_M1); |
| 272 | dev_priv->savePIPEA_DATA_N1 = I915_READ(PIPEA_DATA_N1); | 272 | dev_priv->savePIPEA_DATA_N1 = I915_READ(PIPEA_DATA_N1); |
| 273 | dev_priv->savePIPEA_LINK_M1 = I915_READ(PIPEA_LINK_M1); | 273 | dev_priv->savePIPEA_LINK_M1 = I915_READ(PIPEA_LINK_M1); |
| @@ -304,7 +304,7 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 304 | /* Pipe & plane B info */ | 304 | /* Pipe & plane B info */ |
| 305 | dev_priv->savePIPEBCONF = I915_READ(PIPEBCONF); | 305 | dev_priv->savePIPEBCONF = I915_READ(PIPEBCONF); |
| 306 | dev_priv->savePIPEBSRC = I915_READ(PIPEBSRC); | 306 | dev_priv->savePIPEBSRC = I915_READ(PIPEBSRC); |
| 307 | if (IS_IGDNG(dev)) { | 307 | if (IS_IRONLAKE(dev)) { |
| 308 | dev_priv->saveFPB0 = I915_READ(PCH_FPB0); | 308 | dev_priv->saveFPB0 = I915_READ(PCH_FPB0); |
| 309 | dev_priv->saveFPB1 = I915_READ(PCH_FPB1); | 309 | dev_priv->saveFPB1 = I915_READ(PCH_FPB1); |
| 310 | dev_priv->saveDPLL_B = I915_READ(PCH_DPLL_B); | 310 | dev_priv->saveDPLL_B = I915_READ(PCH_DPLL_B); |
| @@ -313,7 +313,7 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 313 | dev_priv->saveFPB1 = I915_READ(FPB1); | 313 | dev_priv->saveFPB1 = I915_READ(FPB1); |
| 314 | dev_priv->saveDPLL_B = I915_READ(DPLL_B); | 314 | dev_priv->saveDPLL_B = I915_READ(DPLL_B); |
| 315 | } | 315 | } |
| 316 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 316 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 317 | dev_priv->saveDPLL_B_MD = I915_READ(DPLL_B_MD); | 317 | dev_priv->saveDPLL_B_MD = I915_READ(DPLL_B_MD); |
| 318 | dev_priv->saveHTOTAL_B = I915_READ(HTOTAL_B); | 318 | dev_priv->saveHTOTAL_B = I915_READ(HTOTAL_B); |
| 319 | dev_priv->saveHBLANK_B = I915_READ(HBLANK_B); | 319 | dev_priv->saveHBLANK_B = I915_READ(HBLANK_B); |
| @@ -321,10 +321,10 @@ static void i915_save_modeset_reg(struct drm_device *dev) | |||
| 321 | dev_priv->saveVTOTAL_B = I915_READ(VTOTAL_B); | 321 | dev_priv->saveVTOTAL_B = I915_READ(VTOTAL_B); |
| 322 | dev_priv->saveVBLANK_B = I915_READ(VBLANK_B); | 322 | dev_priv->saveVBLANK_B = I915_READ(VBLANK_B); |
| 323 | dev_priv->saveVSYNC_B = I915_READ(VSYNC_B); | 323 | dev_priv->saveVSYNC_B = I915_READ(VSYNC_B); |
| 324 | if (!IS_IGDNG(dev)) | 324 | if (!IS_IRONLAKE(dev)) |
| 325 | dev_priv->saveBCLRPAT_B = I915_READ(BCLRPAT_B); | 325 | dev_priv->saveBCLRPAT_B = I915_READ(BCLRPAT_B); |
| 326 | 326 | ||
| 327 | if (IS_IGDNG(dev)) { | 327 | if (IS_IRONLAKE(dev)) { |
| 328 | dev_priv->savePIPEB_DATA_M1 = I915_READ(PIPEB_DATA_M1); | 328 | dev_priv->savePIPEB_DATA_M1 = I915_READ(PIPEB_DATA_M1); |
| 329 | dev_priv->savePIPEB_DATA_N1 = I915_READ(PIPEB_DATA_N1); | 329 | dev_priv->savePIPEB_DATA_N1 = I915_READ(PIPEB_DATA_N1); |
| 330 | dev_priv->savePIPEB_LINK_M1 = I915_READ(PIPEB_LINK_M1); | 330 | dev_priv->savePIPEB_LINK_M1 = I915_READ(PIPEB_LINK_M1); |
| @@ -369,7 +369,7 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 369 | if (drm_core_check_feature(dev, DRIVER_MODESET)) | 369 | if (drm_core_check_feature(dev, DRIVER_MODESET)) |
| 370 | return; | 370 | return; |
| 371 | 371 | ||
| 372 | if (IS_IGDNG(dev)) { | 372 | if (IS_IRONLAKE(dev)) { |
| 373 | dpll_a_reg = PCH_DPLL_A; | 373 | dpll_a_reg = PCH_DPLL_A; |
| 374 | dpll_b_reg = PCH_DPLL_B; | 374 | dpll_b_reg = PCH_DPLL_B; |
| 375 | fpa0_reg = PCH_FPA0; | 375 | fpa0_reg = PCH_FPA0; |
| @@ -385,7 +385,7 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 385 | fpb1_reg = FPB1; | 385 | fpb1_reg = FPB1; |
| 386 | } | 386 | } |
| 387 | 387 | ||
| 388 | if (IS_IGDNG(dev)) { | 388 | if (IS_IRONLAKE(dev)) { |
| 389 | I915_WRITE(PCH_DREF_CONTROL, dev_priv->savePCH_DREF_CONTROL); | 389 | I915_WRITE(PCH_DREF_CONTROL, dev_priv->savePCH_DREF_CONTROL); |
| 390 | I915_WRITE(DISP_ARB_CTL, dev_priv->saveDISP_ARB_CTL); | 390 | I915_WRITE(DISP_ARB_CTL, dev_priv->saveDISP_ARB_CTL); |
| 391 | } | 391 | } |
| @@ -402,7 +402,7 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 402 | /* Actually enable it */ | 402 | /* Actually enable it */ |
| 403 | I915_WRITE(dpll_a_reg, dev_priv->saveDPLL_A); | 403 | I915_WRITE(dpll_a_reg, dev_priv->saveDPLL_A); |
| 404 | DRM_UDELAY(150); | 404 | DRM_UDELAY(150); |
| 405 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 405 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 406 | I915_WRITE(DPLL_A_MD, dev_priv->saveDPLL_A_MD); | 406 | I915_WRITE(DPLL_A_MD, dev_priv->saveDPLL_A_MD); |
| 407 | DRM_UDELAY(150); | 407 | DRM_UDELAY(150); |
| 408 | 408 | ||
| @@ -413,10 +413,10 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 413 | I915_WRITE(VTOTAL_A, dev_priv->saveVTOTAL_A); | 413 | I915_WRITE(VTOTAL_A, dev_priv->saveVTOTAL_A); |
| 414 | I915_WRITE(VBLANK_A, dev_priv->saveVBLANK_A); | 414 | I915_WRITE(VBLANK_A, dev_priv->saveVBLANK_A); |
| 415 | I915_WRITE(VSYNC_A, dev_priv->saveVSYNC_A); | 415 | I915_WRITE(VSYNC_A, dev_priv->saveVSYNC_A); |
| 416 | if (!IS_IGDNG(dev)) | 416 | if (!IS_IRONLAKE(dev)) |
| 417 | I915_WRITE(BCLRPAT_A, dev_priv->saveBCLRPAT_A); | 417 | I915_WRITE(BCLRPAT_A, dev_priv->saveBCLRPAT_A); |
| 418 | 418 | ||
| 419 | if (IS_IGDNG(dev)) { | 419 | if (IS_IRONLAKE(dev)) { |
| 420 | I915_WRITE(PIPEA_DATA_M1, dev_priv->savePIPEA_DATA_M1); | 420 | I915_WRITE(PIPEA_DATA_M1, dev_priv->savePIPEA_DATA_M1); |
| 421 | I915_WRITE(PIPEA_DATA_N1, dev_priv->savePIPEA_DATA_N1); | 421 | I915_WRITE(PIPEA_DATA_N1, dev_priv->savePIPEA_DATA_N1); |
| 422 | I915_WRITE(PIPEA_LINK_M1, dev_priv->savePIPEA_LINK_M1); | 422 | I915_WRITE(PIPEA_LINK_M1, dev_priv->savePIPEA_LINK_M1); |
| @@ -467,7 +467,7 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 467 | /* Actually enable it */ | 467 | /* Actually enable it */ |
| 468 | I915_WRITE(dpll_b_reg, dev_priv->saveDPLL_B); | 468 | I915_WRITE(dpll_b_reg, dev_priv->saveDPLL_B); |
| 469 | DRM_UDELAY(150); | 469 | DRM_UDELAY(150); |
| 470 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 470 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 471 | I915_WRITE(DPLL_B_MD, dev_priv->saveDPLL_B_MD); | 471 | I915_WRITE(DPLL_B_MD, dev_priv->saveDPLL_B_MD); |
| 472 | DRM_UDELAY(150); | 472 | DRM_UDELAY(150); |
| 473 | 473 | ||
| @@ -478,10 +478,10 @@ static void i915_restore_modeset_reg(struct drm_device *dev) | |||
| 478 | I915_WRITE(VTOTAL_B, dev_priv->saveVTOTAL_B); | 478 | I915_WRITE(VTOTAL_B, dev_priv->saveVTOTAL_B); |
| 479 | I915_WRITE(VBLANK_B, dev_priv->saveVBLANK_B); | 479 | I915_WRITE(VBLANK_B, dev_priv->saveVBLANK_B); |
| 480 | I915_WRITE(VSYNC_B, dev_priv->saveVSYNC_B); | 480 | I915_WRITE(VSYNC_B, dev_priv->saveVSYNC_B); |
| 481 | if (!IS_IGDNG(dev)) | 481 | if (!IS_IRONLAKE(dev)) |
| 482 | I915_WRITE(BCLRPAT_B, dev_priv->saveBCLRPAT_B); | 482 | I915_WRITE(BCLRPAT_B, dev_priv->saveBCLRPAT_B); |
| 483 | 483 | ||
| 484 | if (IS_IGDNG(dev)) { | 484 | if (IS_IRONLAKE(dev)) { |
| 485 | I915_WRITE(PIPEB_DATA_M1, dev_priv->savePIPEB_DATA_M1); | 485 | I915_WRITE(PIPEB_DATA_M1, dev_priv->savePIPEB_DATA_M1); |
| 486 | I915_WRITE(PIPEB_DATA_N1, dev_priv->savePIPEB_DATA_N1); | 486 | I915_WRITE(PIPEB_DATA_N1, dev_priv->savePIPEB_DATA_N1); |
| 487 | I915_WRITE(PIPEB_LINK_M1, dev_priv->savePIPEB_LINK_M1); | 487 | I915_WRITE(PIPEB_LINK_M1, dev_priv->savePIPEB_LINK_M1); |
| @@ -546,14 +546,14 @@ void i915_save_display(struct drm_device *dev) | |||
| 546 | dev_priv->saveCURSIZE = I915_READ(CURSIZE); | 546 | dev_priv->saveCURSIZE = I915_READ(CURSIZE); |
| 547 | 547 | ||
| 548 | /* CRT state */ | 548 | /* CRT state */ |
| 549 | if (IS_IGDNG(dev)) { | 549 | if (IS_IRONLAKE(dev)) { |
| 550 | dev_priv->saveADPA = I915_READ(PCH_ADPA); | 550 | dev_priv->saveADPA = I915_READ(PCH_ADPA); |
| 551 | } else { | 551 | } else { |
| 552 | dev_priv->saveADPA = I915_READ(ADPA); | 552 | dev_priv->saveADPA = I915_READ(ADPA); |
| 553 | } | 553 | } |
| 554 | 554 | ||
| 555 | /* LVDS state */ | 555 | /* LVDS state */ |
| 556 | if (IS_IGDNG(dev)) { | 556 | if (IS_IRONLAKE(dev)) { |
| 557 | dev_priv->savePP_CONTROL = I915_READ(PCH_PP_CONTROL); | 557 | dev_priv->savePP_CONTROL = I915_READ(PCH_PP_CONTROL); |
| 558 | dev_priv->saveBLC_PWM_CTL = I915_READ(BLC_PWM_PCH_CTL1); | 558 | dev_priv->saveBLC_PWM_CTL = I915_READ(BLC_PWM_PCH_CTL1); |
| 559 | dev_priv->saveBLC_PWM_CTL2 = I915_READ(BLC_PWM_PCH_CTL2); | 559 | dev_priv->saveBLC_PWM_CTL2 = I915_READ(BLC_PWM_PCH_CTL2); |
| @@ -571,10 +571,10 @@ void i915_save_display(struct drm_device *dev) | |||
| 571 | dev_priv->saveLVDS = I915_READ(LVDS); | 571 | dev_priv->saveLVDS = I915_READ(LVDS); |
| 572 | } | 572 | } |
| 573 | 573 | ||
| 574 | if (!IS_I830(dev) && !IS_845G(dev) && !IS_IGDNG(dev)) | 574 | if (!IS_I830(dev) && !IS_845G(dev) && !IS_IRONLAKE(dev)) |
| 575 | dev_priv->savePFIT_CONTROL = I915_READ(PFIT_CONTROL); | 575 | dev_priv->savePFIT_CONTROL = I915_READ(PFIT_CONTROL); |
| 576 | 576 | ||
| 577 | if (IS_IGDNG(dev)) { | 577 | if (IS_IRONLAKE(dev)) { |
| 578 | dev_priv->savePP_ON_DELAYS = I915_READ(PCH_PP_ON_DELAYS); | 578 | dev_priv->savePP_ON_DELAYS = I915_READ(PCH_PP_ON_DELAYS); |
| 579 | dev_priv->savePP_OFF_DELAYS = I915_READ(PCH_PP_OFF_DELAYS); | 579 | dev_priv->savePP_OFF_DELAYS = I915_READ(PCH_PP_OFF_DELAYS); |
| 580 | dev_priv->savePP_DIVISOR = I915_READ(PCH_PP_DIVISOR); | 580 | dev_priv->savePP_DIVISOR = I915_READ(PCH_PP_DIVISOR); |
| @@ -614,7 +614,7 @@ void i915_save_display(struct drm_device *dev) | |||
| 614 | dev_priv->saveVGA0 = I915_READ(VGA0); | 614 | dev_priv->saveVGA0 = I915_READ(VGA0); |
| 615 | dev_priv->saveVGA1 = I915_READ(VGA1); | 615 | dev_priv->saveVGA1 = I915_READ(VGA1); |
| 616 | dev_priv->saveVGA_PD = I915_READ(VGA_PD); | 616 | dev_priv->saveVGA_PD = I915_READ(VGA_PD); |
| 617 | if (IS_IGDNG(dev)) | 617 | if (IS_IRONLAKE(dev)) |
| 618 | dev_priv->saveVGACNTRL = I915_READ(CPU_VGACNTRL); | 618 | dev_priv->saveVGACNTRL = I915_READ(CPU_VGACNTRL); |
| 619 | else | 619 | else |
| 620 | dev_priv->saveVGACNTRL = I915_READ(VGACNTRL); | 620 | dev_priv->saveVGACNTRL = I915_READ(VGACNTRL); |
| @@ -656,24 +656,24 @@ void i915_restore_display(struct drm_device *dev) | |||
| 656 | I915_WRITE(CURSIZE, dev_priv->saveCURSIZE); | 656 | I915_WRITE(CURSIZE, dev_priv->saveCURSIZE); |
| 657 | 657 | ||
| 658 | /* CRT state */ | 658 | /* CRT state */ |
| 659 | if (IS_IGDNG(dev)) | 659 | if (IS_IRONLAKE(dev)) |
| 660 | I915_WRITE(PCH_ADPA, dev_priv->saveADPA); | 660 | I915_WRITE(PCH_ADPA, dev_priv->saveADPA); |
| 661 | else | 661 | else |
| 662 | I915_WRITE(ADPA, dev_priv->saveADPA); | 662 | I915_WRITE(ADPA, dev_priv->saveADPA); |
| 663 | 663 | ||
| 664 | /* LVDS state */ | 664 | /* LVDS state */ |
| 665 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 665 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 666 | I915_WRITE(BLC_PWM_CTL2, dev_priv->saveBLC_PWM_CTL2); | 666 | I915_WRITE(BLC_PWM_CTL2, dev_priv->saveBLC_PWM_CTL2); |
| 667 | 667 | ||
| 668 | if (IS_IGDNG(dev)) { | 668 | if (IS_IRONLAKE(dev)) { |
| 669 | I915_WRITE(PCH_LVDS, dev_priv->saveLVDS); | 669 | I915_WRITE(PCH_LVDS, dev_priv->saveLVDS); |
| 670 | } else if (IS_MOBILE(dev) && !IS_I830(dev)) | 670 | } else if (IS_MOBILE(dev) && !IS_I830(dev)) |
| 671 | I915_WRITE(LVDS, dev_priv->saveLVDS); | 671 | I915_WRITE(LVDS, dev_priv->saveLVDS); |
| 672 | 672 | ||
| 673 | if (!IS_I830(dev) && !IS_845G(dev) && !IS_IGDNG(dev)) | 673 | if (!IS_I830(dev) && !IS_845G(dev) && !IS_IRONLAKE(dev)) |
| 674 | I915_WRITE(PFIT_CONTROL, dev_priv->savePFIT_CONTROL); | 674 | I915_WRITE(PFIT_CONTROL, dev_priv->savePFIT_CONTROL); |
| 675 | 675 | ||
| 676 | if (IS_IGDNG(dev)) { | 676 | if (IS_IRONLAKE(dev)) { |
| 677 | I915_WRITE(BLC_PWM_PCH_CTL1, dev_priv->saveBLC_PWM_CTL); | 677 | I915_WRITE(BLC_PWM_PCH_CTL1, dev_priv->saveBLC_PWM_CTL); |
| 678 | I915_WRITE(BLC_PWM_PCH_CTL2, dev_priv->saveBLC_PWM_CTL2); | 678 | I915_WRITE(BLC_PWM_PCH_CTL2, dev_priv->saveBLC_PWM_CTL2); |
| 679 | I915_WRITE(BLC_PWM_CPU_CTL, dev_priv->saveBLC_CPU_PWM_CTL); | 679 | I915_WRITE(BLC_PWM_CPU_CTL, dev_priv->saveBLC_CPU_PWM_CTL); |
| @@ -713,7 +713,7 @@ void i915_restore_display(struct drm_device *dev) | |||
| 713 | } | 713 | } |
| 714 | 714 | ||
| 715 | /* VGA state */ | 715 | /* VGA state */ |
| 716 | if (IS_IGDNG(dev)) | 716 | if (IS_IRONLAKE(dev)) |
| 717 | I915_WRITE(CPU_VGACNTRL, dev_priv->saveVGACNTRL); | 717 | I915_WRITE(CPU_VGACNTRL, dev_priv->saveVGACNTRL); |
| 718 | else | 718 | else |
| 719 | I915_WRITE(VGACNTRL, dev_priv->saveVGACNTRL); | 719 | I915_WRITE(VGACNTRL, dev_priv->saveVGACNTRL); |
| @@ -733,8 +733,10 @@ int i915_save_state(struct drm_device *dev) | |||
| 733 | pci_read_config_byte(dev->pdev, LBB, &dev_priv->saveLBB); | 733 | pci_read_config_byte(dev->pdev, LBB, &dev_priv->saveLBB); |
| 734 | 734 | ||
| 735 | /* Render Standby */ | 735 | /* Render Standby */ |
| 736 | if (IS_I965G(dev) && IS_MOBILE(dev)) | 736 | if (I915_HAS_RC6(dev)) { |
| 737 | dev_priv->saveRENDERSTANDBY = I915_READ(MCHBAR_RENDER_STANDBY); | 737 | dev_priv->saveRENDERSTANDBY = I915_READ(MCHBAR_RENDER_STANDBY); |
| 738 | dev_priv->savePWRCTXA = I915_READ(PWRCTXA); | ||
| 739 | } | ||
| 738 | 740 | ||
| 739 | /* Hardware status page */ | 741 | /* Hardware status page */ |
| 740 | dev_priv->saveHWS = I915_READ(HWS_PGA); | 742 | dev_priv->saveHWS = I915_READ(HWS_PGA); |
| @@ -742,7 +744,7 @@ int i915_save_state(struct drm_device *dev) | |||
| 742 | i915_save_display(dev); | 744 | i915_save_display(dev); |
| 743 | 745 | ||
| 744 | /* Interrupt state */ | 746 | /* Interrupt state */ |
| 745 | if (IS_IGDNG(dev)) { | 747 | if (IS_IRONLAKE(dev)) { |
| 746 | dev_priv->saveDEIER = I915_READ(DEIER); | 748 | dev_priv->saveDEIER = I915_READ(DEIER); |
| 747 | dev_priv->saveDEIMR = I915_READ(DEIMR); | 749 | dev_priv->saveDEIMR = I915_READ(DEIMR); |
| 748 | dev_priv->saveGTIER = I915_READ(GTIER); | 750 | dev_priv->saveGTIER = I915_READ(GTIER); |
| @@ -754,10 +756,6 @@ int i915_save_state(struct drm_device *dev) | |||
| 754 | dev_priv->saveIMR = I915_READ(IMR); | 756 | dev_priv->saveIMR = I915_READ(IMR); |
| 755 | } | 757 | } |
| 756 | 758 | ||
| 757 | /* Clock gating state */ | ||
| 758 | dev_priv->saveD_STATE = I915_READ(D_STATE); | ||
| 759 | dev_priv->saveDSPCLK_GATE_D = I915_READ(DSPCLK_GATE_D); /* Not sure about this */ | ||
| 760 | |||
| 761 | /* Cache mode state */ | 759 | /* Cache mode state */ |
| 762 | dev_priv->saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); | 760 | dev_priv->saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); |
| 763 | 761 | ||
| @@ -796,8 +794,10 @@ int i915_restore_state(struct drm_device *dev) | |||
| 796 | pci_write_config_byte(dev->pdev, LBB, dev_priv->saveLBB); | 794 | pci_write_config_byte(dev->pdev, LBB, dev_priv->saveLBB); |
| 797 | 795 | ||
| 798 | /* Render Standby */ | 796 | /* Render Standby */ |
| 799 | if (IS_I965G(dev) && IS_MOBILE(dev)) | 797 | if (I915_HAS_RC6(dev)) { |
| 800 | I915_WRITE(MCHBAR_RENDER_STANDBY, dev_priv->saveRENDERSTANDBY); | 798 | I915_WRITE(MCHBAR_RENDER_STANDBY, dev_priv->saveRENDERSTANDBY); |
| 799 | I915_WRITE(PWRCTXA, dev_priv->savePWRCTXA); | ||
| 800 | } | ||
| 801 | 801 | ||
| 802 | /* Hardware status page */ | 802 | /* Hardware status page */ |
| 803 | I915_WRITE(HWS_PGA, dev_priv->saveHWS); | 803 | I915_WRITE(HWS_PGA, dev_priv->saveHWS); |
| @@ -817,7 +817,7 @@ int i915_restore_state(struct drm_device *dev) | |||
| 817 | i915_restore_display(dev); | 817 | i915_restore_display(dev); |
| 818 | 818 | ||
| 819 | /* Interrupt state */ | 819 | /* Interrupt state */ |
| 820 | if (IS_IGDNG(dev)) { | 820 | if (IS_IRONLAKE(dev)) { |
| 821 | I915_WRITE(DEIER, dev_priv->saveDEIER); | 821 | I915_WRITE(DEIER, dev_priv->saveDEIER); |
| 822 | I915_WRITE(DEIMR, dev_priv->saveDEIMR); | 822 | I915_WRITE(DEIMR, dev_priv->saveDEIMR); |
| 823 | I915_WRITE(GTIER, dev_priv->saveGTIER); | 823 | I915_WRITE(GTIER, dev_priv->saveGTIER); |
| @@ -830,8 +830,7 @@ int i915_restore_state(struct drm_device *dev) | |||
| 830 | } | 830 | } |
| 831 | 831 | ||
| 832 | /* Clock gating state */ | 832 | /* Clock gating state */ |
| 833 | I915_WRITE (D_STATE, dev_priv->saveD_STATE); | 833 | intel_init_clock_gating(dev); |
| 834 | I915_WRITE (DSPCLK_GATE_D, dev_priv->saveDSPCLK_GATE_D); | ||
| 835 | 834 | ||
| 836 | /* Cache mode state */ | 835 | /* Cache mode state */ |
| 837 | I915_WRITE (CACHE_MODE_0, dev_priv->saveCACHE_MODE_0 | 0xffff0000); | 836 | I915_WRITE (CACHE_MODE_0, dev_priv->saveCACHE_MODE_0 | 0xffff0000); |
| @@ -846,6 +845,9 @@ int i915_restore_state(struct drm_device *dev) | |||
| 846 | for (i = 0; i < 3; i++) | 845 | for (i = 0; i < 3; i++) |
| 847 | I915_WRITE(SWF30 + (i << 2), dev_priv->saveSWF2[i]); | 846 | I915_WRITE(SWF30 + (i << 2), dev_priv->saveSWF2[i]); |
| 848 | 847 | ||
| 848 | /* I2C state */ | ||
| 849 | intel_i2c_reset_gmbus(dev); | ||
| 850 | |||
| 849 | return 0; | 851 | return 0; |
| 850 | } | 852 | } |
| 851 | 853 | ||
diff --git a/drivers/gpu/drm/i915/intel_bios.c b/drivers/gpu/drm/i915/intel_bios.c index 96cd256e60e6..f27567747580 100644 --- a/drivers/gpu/drm/i915/intel_bios.c +++ b/drivers/gpu/drm/i915/intel_bios.c | |||
| @@ -114,6 +114,8 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, | |||
| 114 | struct lvds_dvo_timing *dvo_timing; | 114 | struct lvds_dvo_timing *dvo_timing; |
| 115 | struct drm_display_mode *panel_fixed_mode; | 115 | struct drm_display_mode *panel_fixed_mode; |
| 116 | int lfp_data_size, dvo_timing_offset; | 116 | int lfp_data_size, dvo_timing_offset; |
| 117 | int i, temp_downclock; | ||
| 118 | struct drm_display_mode *temp_mode; | ||
| 117 | 119 | ||
| 118 | /* Defaults if we can't find VBT info */ | 120 | /* Defaults if we can't find VBT info */ |
| 119 | dev_priv->lvds_dither = 0; | 121 | dev_priv->lvds_dither = 0; |
| @@ -159,9 +161,49 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, | |||
| 159 | 161 | ||
| 160 | dev_priv->lfp_lvds_vbt_mode = panel_fixed_mode; | 162 | dev_priv->lfp_lvds_vbt_mode = panel_fixed_mode; |
| 161 | 163 | ||
| 162 | DRM_DEBUG("Found panel mode in BIOS VBT tables:\n"); | 164 | DRM_DEBUG_KMS("Found panel mode in BIOS VBT tables:\n"); |
| 163 | drm_mode_debug_printmodeline(panel_fixed_mode); | 165 | drm_mode_debug_printmodeline(panel_fixed_mode); |
| 164 | 166 | ||
| 167 | temp_mode = kzalloc(sizeof(*temp_mode), GFP_KERNEL); | ||
| 168 | temp_downclock = panel_fixed_mode->clock; | ||
| 169 | /* | ||
| 170 | * enumerate the LVDS panel timing info entry in VBT to check whether | ||
| 171 | * the LVDS downclock is found. | ||
| 172 | */ | ||
| 173 | for (i = 0; i < 16; i++) { | ||
| 174 | entry = (struct bdb_lvds_lfp_data_entry *) | ||
| 175 | ((uint8_t *)lvds_lfp_data->data + (lfp_data_size * i)); | ||
| 176 | dvo_timing = (struct lvds_dvo_timing *) | ||
| 177 | ((unsigned char *)entry + dvo_timing_offset); | ||
| 178 | |||
| 179 | fill_detail_timing_data(temp_mode, dvo_timing); | ||
| 180 | |||
| 181 | if (temp_mode->hdisplay == panel_fixed_mode->hdisplay && | ||
| 182 | temp_mode->hsync_start == panel_fixed_mode->hsync_start && | ||
| 183 | temp_mode->hsync_end == panel_fixed_mode->hsync_end && | ||
| 184 | temp_mode->htotal == panel_fixed_mode->htotal && | ||
| 185 | temp_mode->vdisplay == panel_fixed_mode->vdisplay && | ||
| 186 | temp_mode->vsync_start == panel_fixed_mode->vsync_start && | ||
| 187 | temp_mode->vsync_end == panel_fixed_mode->vsync_end && | ||
| 188 | temp_mode->vtotal == panel_fixed_mode->vtotal && | ||
| 189 | temp_mode->clock < temp_downclock) { | ||
| 190 | /* | ||
| 191 | * downclock is already found. But we expect | ||
| 192 | * to find the lower downclock. | ||
| 193 | */ | ||
| 194 | temp_downclock = temp_mode->clock; | ||
| 195 | } | ||
| 196 | /* clear it to zero */ | ||
| 197 | memset(temp_mode, 0, sizeof(*temp_mode)); | ||
| 198 | } | ||
| 199 | kfree(temp_mode); | ||
| 200 | if (temp_downclock < panel_fixed_mode->clock) { | ||
| 201 | dev_priv->lvds_downclock_avail = 1; | ||
| 202 | dev_priv->lvds_downclock = temp_downclock; | ||
| 203 | DRM_DEBUG_KMS("LVDS downclock is found in VBT. ", | ||
| 204 | "Normal Clock %dKHz, downclock %dKHz\n", | ||
| 205 | temp_downclock, panel_fixed_mode->clock); | ||
| 206 | } | ||
| 165 | return; | 207 | return; |
| 166 | } | 208 | } |
| 167 | 209 | ||
| @@ -217,7 +259,7 @@ parse_general_features(struct drm_i915_private *dev_priv, | |||
| 217 | if (IS_I85X(dev_priv->dev)) | 259 | if (IS_I85X(dev_priv->dev)) |
| 218 | dev_priv->lvds_ssc_freq = | 260 | dev_priv->lvds_ssc_freq = |
| 219 | general->ssc_freq ? 66 : 48; | 261 | general->ssc_freq ? 66 : 48; |
| 220 | else if (IS_IGDNG(dev_priv->dev)) | 262 | else if (IS_IRONLAKE(dev_priv->dev)) |
| 221 | dev_priv->lvds_ssc_freq = | 263 | dev_priv->lvds_ssc_freq = |
| 222 | general->ssc_freq ? 100 : 120; | 264 | general->ssc_freq ? 100 : 120; |
| 223 | else | 265 | else |
| @@ -241,22 +283,18 @@ parse_general_definitions(struct drm_i915_private *dev_priv, | |||
| 241 | GPIOF, | 283 | GPIOF, |
| 242 | }; | 284 | }; |
| 243 | 285 | ||
| 244 | /* Set sensible defaults in case we can't find the general block | ||
| 245 | or it is the wrong chipset */ | ||
| 246 | dev_priv->crt_ddc_bus = -1; | ||
| 247 | |||
| 248 | general = find_section(bdb, BDB_GENERAL_DEFINITIONS); | 286 | general = find_section(bdb, BDB_GENERAL_DEFINITIONS); |
| 249 | if (general) { | 287 | if (general) { |
| 250 | u16 block_size = get_blocksize(general); | 288 | u16 block_size = get_blocksize(general); |
| 251 | if (block_size >= sizeof(*general)) { | 289 | if (block_size >= sizeof(*general)) { |
| 252 | int bus_pin = general->crt_ddc_gmbus_pin; | 290 | int bus_pin = general->crt_ddc_gmbus_pin; |
| 253 | DRM_DEBUG("crt_ddc_bus_pin: %d\n", bus_pin); | 291 | DRM_DEBUG_KMS("crt_ddc_bus_pin: %d\n", bus_pin); |
| 254 | if ((bus_pin >= 1) && (bus_pin <= 6)) { | 292 | if ((bus_pin >= 1) && (bus_pin <= 6)) { |
| 255 | dev_priv->crt_ddc_bus = | 293 | dev_priv->crt_ddc_bus = |
| 256 | crt_bus_map_table[bus_pin-1]; | 294 | crt_bus_map_table[bus_pin-1]; |
| 257 | } | 295 | } |
| 258 | } else { | 296 | } else { |
| 259 | DRM_DEBUG("BDB_GD too small (%d). Invalid.\n", | 297 | DRM_DEBUG_KMS("BDB_GD too small (%d). Invalid.\n", |
| 260 | block_size); | 298 | block_size); |
| 261 | } | 299 | } |
| 262 | } | 300 | } |
| @@ -274,7 +312,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, | |||
| 274 | 312 | ||
| 275 | p_defs = find_section(bdb, BDB_GENERAL_DEFINITIONS); | 313 | p_defs = find_section(bdb, BDB_GENERAL_DEFINITIONS); |
| 276 | if (!p_defs) { | 314 | if (!p_defs) { |
| 277 | DRM_DEBUG("No general definition block is found\n"); | 315 | DRM_DEBUG_KMS("No general definition block is found\n"); |
| 278 | return; | 316 | return; |
| 279 | } | 317 | } |
| 280 | /* judge whether the size of child device meets the requirements. | 318 | /* judge whether the size of child device meets the requirements. |
| @@ -284,7 +322,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, | |||
| 284 | */ | 322 | */ |
| 285 | if (p_defs->child_dev_size != sizeof(*p_child)) { | 323 | if (p_defs->child_dev_size != sizeof(*p_child)) { |
| 286 | /* different child dev size . Ignore it */ | 324 | /* different child dev size . Ignore it */ |
| 287 | DRM_DEBUG("different child size is found. Invalid.\n"); | 325 | DRM_DEBUG_KMS("different child size is found. Invalid.\n"); |
| 288 | return; | 326 | return; |
| 289 | } | 327 | } |
| 290 | /* get the block size of general definitions */ | 328 | /* get the block size of general definitions */ |
| @@ -310,11 +348,11 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, | |||
| 310 | if (p_child->dvo_port != DEVICE_PORT_DVOB && | 348 | if (p_child->dvo_port != DEVICE_PORT_DVOB && |
| 311 | p_child->dvo_port != DEVICE_PORT_DVOC) { | 349 | p_child->dvo_port != DEVICE_PORT_DVOC) { |
| 312 | /* skip the incorrect SDVO port */ | 350 | /* skip the incorrect SDVO port */ |
| 313 | DRM_DEBUG("Incorrect SDVO port. Skip it \n"); | 351 | DRM_DEBUG_KMS("Incorrect SDVO port. Skip it \n"); |
| 314 | continue; | 352 | continue; |
| 315 | } | 353 | } |
| 316 | DRM_DEBUG("the SDVO device with slave addr %2x is found on " | 354 | DRM_DEBUG_KMS("the SDVO device with slave addr %2x is found on" |
| 317 | "%s port\n", | 355 | " %s port\n", |
| 318 | p_child->slave_addr, | 356 | p_child->slave_addr, |
| 319 | (p_child->dvo_port == DEVICE_PORT_DVOB) ? | 357 | (p_child->dvo_port == DEVICE_PORT_DVOB) ? |
| 320 | "SDVOB" : "SDVOC"); | 358 | "SDVOB" : "SDVOC"); |
| @@ -325,21 +363,21 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, | |||
| 325 | p_mapping->dvo_wiring = p_child->dvo_wiring; | 363 | p_mapping->dvo_wiring = p_child->dvo_wiring; |
| 326 | p_mapping->initialized = 1; | 364 | p_mapping->initialized = 1; |
| 327 | } else { | 365 | } else { |
| 328 | DRM_DEBUG("Maybe one SDVO port is shared by " | 366 | DRM_DEBUG_KMS("Maybe one SDVO port is shared by " |
| 329 | "two SDVO device.\n"); | 367 | "two SDVO device.\n"); |
| 330 | } | 368 | } |
| 331 | if (p_child->slave2_addr) { | 369 | if (p_child->slave2_addr) { |
| 332 | /* Maybe this is a SDVO device with multiple inputs */ | 370 | /* Maybe this is a SDVO device with multiple inputs */ |
| 333 | /* And the mapping info is not added */ | 371 | /* And the mapping info is not added */ |
| 334 | DRM_DEBUG("there exists the slave2_addr. Maybe this " | 372 | DRM_DEBUG_KMS("there exists the slave2_addr. Maybe this" |
| 335 | "is a SDVO device with multiple inputs.\n"); | 373 | " is a SDVO device with multiple inputs.\n"); |
| 336 | } | 374 | } |
| 337 | count++; | 375 | count++; |
| 338 | } | 376 | } |
| 339 | 377 | ||
| 340 | if (!count) { | 378 | if (!count) { |
| 341 | /* No SDVO device info is found */ | 379 | /* No SDVO device info is found */ |
| 342 | DRM_DEBUG("No SDVO device info is found in VBT\n"); | 380 | DRM_DEBUG_KMS("No SDVO device info is found in VBT\n"); |
| 343 | } | 381 | } |
| 344 | return; | 382 | return; |
| 345 | } | 383 | } |
| @@ -366,6 +404,70 @@ parse_driver_features(struct drm_i915_private *dev_priv, | |||
| 366 | dev_priv->render_reclock_avail = true; | 404 | dev_priv->render_reclock_avail = true; |
| 367 | } | 405 | } |
| 368 | 406 | ||
| 407 | static void | ||
| 408 | parse_device_mapping(struct drm_i915_private *dev_priv, | ||
| 409 | struct bdb_header *bdb) | ||
| 410 | { | ||
| 411 | struct bdb_general_definitions *p_defs; | ||
| 412 | struct child_device_config *p_child, *child_dev_ptr; | ||
| 413 | int i, child_device_num, count; | ||
| 414 | u16 block_size; | ||
| 415 | |||
| 416 | p_defs = find_section(bdb, BDB_GENERAL_DEFINITIONS); | ||
| 417 | if (!p_defs) { | ||
| 418 | DRM_DEBUG_KMS("No general definition block is found\n"); | ||
| 419 | return; | ||
| 420 | } | ||
| 421 | /* judge whether the size of child device meets the requirements. | ||
| 422 | * If the child device size obtained from general definition block | ||
| 423 | * is different with sizeof(struct child_device_config), skip the | ||
| 424 | * parsing of sdvo device info | ||
| 425 | */ | ||
| 426 | if (p_defs->child_dev_size != sizeof(*p_child)) { | ||
| 427 | /* different child dev size . Ignore it */ | ||
| 428 | DRM_DEBUG_KMS("different child size is found. Invalid.\n"); | ||
| 429 | return; | ||
| 430 | } | ||
| 431 | /* get the block size of general definitions */ | ||
| 432 | block_size = get_blocksize(p_defs); | ||
| 433 | /* get the number of child device */ | ||
| 434 | child_device_num = (block_size - sizeof(*p_defs)) / | ||
| 435 | sizeof(*p_child); | ||
| 436 | count = 0; | ||
| 437 | /* get the number of child device that is present */ | ||
| 438 | for (i = 0; i < child_device_num; i++) { | ||
| 439 | p_child = &(p_defs->devices[i]); | ||
| 440 | if (!p_child->device_type) { | ||
| 441 | /* skip the device block if device type is invalid */ | ||
| 442 | continue; | ||
| 443 | } | ||
| 444 | count++; | ||
| 445 | } | ||
| 446 | if (!count) { | ||
| 447 | DRM_DEBUG_KMS("no child dev is parsed from VBT \n"); | ||
| 448 | return; | ||
| 449 | } | ||
| 450 | dev_priv->child_dev = kzalloc(sizeof(*p_child) * count, GFP_KERNEL); | ||
| 451 | if (!dev_priv->child_dev) { | ||
| 452 | DRM_DEBUG_KMS("No memory space for child device\n"); | ||
| 453 | return; | ||
| 454 | } | ||
| 455 | |||
| 456 | dev_priv->child_dev_num = count; | ||
| 457 | count = 0; | ||
| 458 | for (i = 0; i < child_device_num; i++) { | ||
| 459 | p_child = &(p_defs->devices[i]); | ||
| 460 | if (!p_child->device_type) { | ||
| 461 | /* skip the device block if device type is invalid */ | ||
| 462 | continue; | ||
| 463 | } | ||
| 464 | child_dev_ptr = dev_priv->child_dev + count; | ||
| 465 | count++; | ||
| 466 | memcpy((void *)child_dev_ptr, (void *)p_child, | ||
| 467 | sizeof(*p_child)); | ||
| 468 | } | ||
| 469 | return; | ||
| 470 | } | ||
| 369 | /** | 471 | /** |
| 370 | * intel_init_bios - initialize VBIOS settings & find VBT | 472 | * intel_init_bios - initialize VBIOS settings & find VBT |
| 371 | * @dev: DRM device | 473 | * @dev: DRM device |
| @@ -417,6 +519,7 @@ intel_init_bios(struct drm_device *dev) | |||
| 417 | parse_lfp_panel_data(dev_priv, bdb); | 519 | parse_lfp_panel_data(dev_priv, bdb); |
| 418 | parse_sdvo_panel_data(dev_priv, bdb); | 520 | parse_sdvo_panel_data(dev_priv, bdb); |
| 419 | parse_sdvo_device_mapping(dev_priv, bdb); | 521 | parse_sdvo_device_mapping(dev_priv, bdb); |
| 522 | parse_device_mapping(dev_priv, bdb); | ||
| 420 | parse_driver_features(dev_priv, bdb); | 523 | parse_driver_features(dev_priv, bdb); |
| 421 | 524 | ||
| 422 | pci_unmap_rom(pdev, bios); | 525 | pci_unmap_rom(pdev, bios); |
diff --git a/drivers/gpu/drm/i915/intel_bios.h b/drivers/gpu/drm/i915/intel_bios.h index 0f8e5f69ac7a..425ac9d7f724 100644 --- a/drivers/gpu/drm/i915/intel_bios.h +++ b/drivers/gpu/drm/i915/intel_bios.h | |||
| @@ -549,4 +549,21 @@ bool intel_init_bios(struct drm_device *dev); | |||
| 549 | #define SWF14_APM_STANDBY 0x1 | 549 | #define SWF14_APM_STANDBY 0x1 |
| 550 | #define SWF14_APM_RESTORE 0x0 | 550 | #define SWF14_APM_RESTORE 0x0 |
| 551 | 551 | ||
| 552 | /* Add the device class for LFP, TV, HDMI */ | ||
| 553 | #define DEVICE_TYPE_INT_LFP 0x1022 | ||
| 554 | #define DEVICE_TYPE_INT_TV 0x1009 | ||
| 555 | #define DEVICE_TYPE_HDMI 0x60D2 | ||
| 556 | #define DEVICE_TYPE_DP 0x68C6 | ||
| 557 | #define DEVICE_TYPE_eDP 0x78C6 | ||
| 558 | |||
| 559 | /* define the DVO port for HDMI output type */ | ||
| 560 | #define DVO_B 1 | ||
| 561 | #define DVO_C 2 | ||
| 562 | #define DVO_D 3 | ||
| 563 | |||
| 564 | /* define the PORT for DP output type */ | ||
| 565 | #define PORT_IDPB 7 | ||
| 566 | #define PORT_IDPC 8 | ||
| 567 | #define PORT_IDPD 9 | ||
| 568 | |||
| 552 | #endif /* _I830_BIOS_H_ */ | 569 | #endif /* _I830_BIOS_H_ */ |
diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c index e5051446c48e..9f3d3e563414 100644 --- a/drivers/gpu/drm/i915/intel_crt.c +++ b/drivers/gpu/drm/i915/intel_crt.c | |||
| @@ -39,7 +39,7 @@ static void intel_crt_dpms(struct drm_encoder *encoder, int mode) | |||
| 39 | struct drm_i915_private *dev_priv = dev->dev_private; | 39 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 40 | u32 temp, reg; | 40 | u32 temp, reg; |
| 41 | 41 | ||
| 42 | if (IS_IGDNG(dev)) | 42 | if (IS_IRONLAKE(dev)) |
| 43 | reg = PCH_ADPA; | 43 | reg = PCH_ADPA; |
| 44 | else | 44 | else |
| 45 | reg = ADPA; | 45 | reg = ADPA; |
| @@ -64,34 +64,6 @@ static void intel_crt_dpms(struct drm_encoder *encoder, int mode) | |||
| 64 | } | 64 | } |
| 65 | 65 | ||
| 66 | I915_WRITE(reg, temp); | 66 | I915_WRITE(reg, temp); |
| 67 | |||
| 68 | if (IS_IGD(dev)) { | ||
| 69 | if (mode == DRM_MODE_DPMS_OFF) { | ||
| 70 | /* turn off DAC */ | ||
| 71 | temp = I915_READ(PORT_HOTPLUG_EN); | ||
| 72 | temp &= ~CRT_EOS_INT_EN; | ||
| 73 | I915_WRITE(PORT_HOTPLUG_EN, temp); | ||
| 74 | |||
| 75 | temp = I915_READ(PORT_HOTPLUG_STAT); | ||
| 76 | if (temp & CRT_EOS_INT_STATUS) | ||
| 77 | I915_WRITE(PORT_HOTPLUG_STAT, | ||
| 78 | CRT_EOS_INT_STATUS); | ||
| 79 | } else { | ||
| 80 | /* turn on DAC. EOS interrupt must be enabled after DAC | ||
| 81 | * is enabled, so it sounds not good to enable it in | ||
| 82 | * i915_driver_irq_postinstall() | ||
| 83 | * wait 12.5ms after DAC is enabled | ||
| 84 | */ | ||
| 85 | msleep(13); | ||
| 86 | temp = I915_READ(PORT_HOTPLUG_STAT); | ||
| 87 | if (temp & CRT_EOS_INT_STATUS) | ||
| 88 | I915_WRITE(PORT_HOTPLUG_STAT, | ||
| 89 | CRT_EOS_INT_STATUS); | ||
| 90 | temp = I915_READ(PORT_HOTPLUG_EN); | ||
| 91 | temp |= CRT_EOS_INT_EN; | ||
| 92 | I915_WRITE(PORT_HOTPLUG_EN, temp); | ||
| 93 | } | ||
| 94 | } | ||
| 95 | } | 67 | } |
| 96 | 68 | ||
| 97 | static int intel_crt_mode_valid(struct drm_connector *connector, | 69 | static int intel_crt_mode_valid(struct drm_connector *connector, |
| @@ -141,7 +113,7 @@ static void intel_crt_mode_set(struct drm_encoder *encoder, | |||
| 141 | else | 113 | else |
| 142 | dpll_md_reg = DPLL_B_MD; | 114 | dpll_md_reg = DPLL_B_MD; |
| 143 | 115 | ||
| 144 | if (IS_IGDNG(dev)) | 116 | if (IS_IRONLAKE(dev)) |
| 145 | adpa_reg = PCH_ADPA; | 117 | adpa_reg = PCH_ADPA; |
| 146 | else | 118 | else |
| 147 | adpa_reg = ADPA; | 119 | adpa_reg = ADPA; |
| @@ -150,7 +122,7 @@ static void intel_crt_mode_set(struct drm_encoder *encoder, | |||
| 150 | * Disable separate mode multiplier used when cloning SDVO to CRT | 122 | * Disable separate mode multiplier used when cloning SDVO to CRT |
| 151 | * XXX this needs to be adjusted when we really are cloning | 123 | * XXX this needs to be adjusted when we really are cloning |
| 152 | */ | 124 | */ |
| 153 | if (IS_I965G(dev) && !IS_IGDNG(dev)) { | 125 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) { |
| 154 | dpll_md = I915_READ(dpll_md_reg); | 126 | dpll_md = I915_READ(dpll_md_reg); |
| 155 | I915_WRITE(dpll_md_reg, | 127 | I915_WRITE(dpll_md_reg, |
| 156 | dpll_md & ~DPLL_MD_UDI_MULTIPLIER_MASK); | 128 | dpll_md & ~DPLL_MD_UDI_MULTIPLIER_MASK); |
| @@ -164,18 +136,18 @@ static void intel_crt_mode_set(struct drm_encoder *encoder, | |||
| 164 | 136 | ||
| 165 | if (intel_crtc->pipe == 0) { | 137 | if (intel_crtc->pipe == 0) { |
| 166 | adpa |= ADPA_PIPE_A_SELECT; | 138 | adpa |= ADPA_PIPE_A_SELECT; |
| 167 | if (!IS_IGDNG(dev)) | 139 | if (!IS_IRONLAKE(dev)) |
| 168 | I915_WRITE(BCLRPAT_A, 0); | 140 | I915_WRITE(BCLRPAT_A, 0); |
| 169 | } else { | 141 | } else { |
| 170 | adpa |= ADPA_PIPE_B_SELECT; | 142 | adpa |= ADPA_PIPE_B_SELECT; |
| 171 | if (!IS_IGDNG(dev)) | 143 | if (!IS_IRONLAKE(dev)) |
| 172 | I915_WRITE(BCLRPAT_B, 0); | 144 | I915_WRITE(BCLRPAT_B, 0); |
| 173 | } | 145 | } |
| 174 | 146 | ||
| 175 | I915_WRITE(adpa_reg, adpa); | 147 | I915_WRITE(adpa_reg, adpa); |
| 176 | } | 148 | } |
| 177 | 149 | ||
| 178 | static bool intel_igdng_crt_detect_hotplug(struct drm_connector *connector) | 150 | static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector) |
| 179 | { | 151 | { |
| 180 | struct drm_device *dev = connector->dev; | 152 | struct drm_device *dev = connector->dev; |
| 181 | struct drm_i915_private *dev_priv = dev->dev_private; | 153 | struct drm_i915_private *dev_priv = dev->dev_private; |
| @@ -194,7 +166,7 @@ static bool intel_igdng_crt_detect_hotplug(struct drm_connector *connector) | |||
| 194 | ADPA_CRT_HOTPLUG_ENABLE | | 166 | ADPA_CRT_HOTPLUG_ENABLE | |
| 195 | ADPA_CRT_HOTPLUG_FORCE_TRIGGER); | 167 | ADPA_CRT_HOTPLUG_FORCE_TRIGGER); |
| 196 | 168 | ||
| 197 | DRM_DEBUG("pch crt adpa 0x%x", adpa); | 169 | DRM_DEBUG_KMS("pch crt adpa 0x%x", adpa); |
| 198 | I915_WRITE(PCH_ADPA, adpa); | 170 | I915_WRITE(PCH_ADPA, adpa); |
| 199 | 171 | ||
| 200 | while ((I915_READ(PCH_ADPA) & ADPA_CRT_HOTPLUG_FORCE_TRIGGER) != 0) | 172 | while ((I915_READ(PCH_ADPA) & ADPA_CRT_HOTPLUG_FORCE_TRIGGER) != 0) |
| @@ -227,8 +199,8 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector) | |||
| 227 | u32 hotplug_en; | 199 | u32 hotplug_en; |
| 228 | int i, tries = 0; | 200 | int i, tries = 0; |
| 229 | 201 | ||
| 230 | if (IS_IGDNG(dev)) | 202 | if (IS_IRONLAKE(dev)) |
| 231 | return intel_igdng_crt_detect_hotplug(connector); | 203 | return intel_ironlake_crt_detect_hotplug(connector); |
| 232 | 204 | ||
| 233 | /* | 205 | /* |
| 234 | * On 4 series desktop, CRT detect sequence need to be done twice | 206 | * On 4 series desktop, CRT detect sequence need to be done twice |
| @@ -549,12 +521,12 @@ void intel_crt_init(struct drm_device *dev) | |||
| 549 | &intel_output->enc); | 521 | &intel_output->enc); |
| 550 | 522 | ||
| 551 | /* Set up the DDC bus. */ | 523 | /* Set up the DDC bus. */ |
| 552 | if (IS_IGDNG(dev)) | 524 | if (IS_IRONLAKE(dev)) |
| 553 | i2c_reg = PCH_GPIOA; | 525 | i2c_reg = PCH_GPIOA; |
| 554 | else { | 526 | else { |
| 555 | i2c_reg = GPIOA; | 527 | i2c_reg = GPIOA; |
| 556 | /* Use VBT information for CRT DDC if available */ | 528 | /* Use VBT information for CRT DDC if available */ |
| 557 | if (dev_priv->crt_ddc_bus != -1) | 529 | if (dev_priv->crt_ddc_bus != 0) |
| 558 | i2c_reg = dev_priv->crt_ddc_bus; | 530 | i2c_reg = dev_priv->crt_ddc_bus; |
| 559 | } | 531 | } |
| 560 | intel_output->ddc_bus = intel_i2c_create(dev, i2c_reg, "CRTDDC_A"); | 532 | intel_output->ddc_bus = intel_i2c_create(dev, i2c_reg, "CRTDDC_A"); |
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index 099f420de57a..52cd9b006da2 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
| @@ -32,7 +32,7 @@ | |||
| 32 | #include "intel_drv.h" | 32 | #include "intel_drv.h" |
| 33 | #include "i915_drm.h" | 33 | #include "i915_drm.h" |
| 34 | #include "i915_drv.h" | 34 | #include "i915_drv.h" |
| 35 | #include "intel_dp.h" | 35 | #include "drm_dp_helper.h" |
| 36 | 36 | ||
| 37 | #include "drm_crtc_helper.h" | 37 | #include "drm_crtc_helper.h" |
| 38 | 38 | ||
| @@ -102,32 +102,32 @@ struct intel_limit { | |||
| 102 | #define I9XX_DOT_MAX 400000 | 102 | #define I9XX_DOT_MAX 400000 |
| 103 | #define I9XX_VCO_MIN 1400000 | 103 | #define I9XX_VCO_MIN 1400000 |
| 104 | #define I9XX_VCO_MAX 2800000 | 104 | #define I9XX_VCO_MAX 2800000 |
| 105 | #define IGD_VCO_MIN 1700000 | 105 | #define PINEVIEW_VCO_MIN 1700000 |
| 106 | #define IGD_VCO_MAX 3500000 | 106 | #define PINEVIEW_VCO_MAX 3500000 |
| 107 | #define I9XX_N_MIN 1 | 107 | #define I9XX_N_MIN 1 |
| 108 | #define I9XX_N_MAX 6 | 108 | #define I9XX_N_MAX 6 |
| 109 | /* IGD's Ncounter is a ring counter */ | 109 | /* Pineview's Ncounter is a ring counter */ |
| 110 | #define IGD_N_MIN 3 | 110 | #define PINEVIEW_N_MIN 3 |
| 111 | #define IGD_N_MAX 6 | 111 | #define PINEVIEW_N_MAX 6 |
| 112 | #define I9XX_M_MIN 70 | 112 | #define I9XX_M_MIN 70 |
| 113 | #define I9XX_M_MAX 120 | 113 | #define I9XX_M_MAX 120 |
| 114 | #define IGD_M_MIN 2 | 114 | #define PINEVIEW_M_MIN 2 |
| 115 | #define IGD_M_MAX 256 | 115 | #define PINEVIEW_M_MAX 256 |
| 116 | #define I9XX_M1_MIN 10 | 116 | #define I9XX_M1_MIN 10 |
| 117 | #define I9XX_M1_MAX 22 | 117 | #define I9XX_M1_MAX 22 |
| 118 | #define I9XX_M2_MIN 5 | 118 | #define I9XX_M2_MIN 5 |
| 119 | #define I9XX_M2_MAX 9 | 119 | #define I9XX_M2_MAX 9 |
| 120 | /* IGD M1 is reserved, and must be 0 */ | 120 | /* Pineview M1 is reserved, and must be 0 */ |
| 121 | #define IGD_M1_MIN 0 | 121 | #define PINEVIEW_M1_MIN 0 |
| 122 | #define IGD_M1_MAX 0 | 122 | #define PINEVIEW_M1_MAX 0 |
| 123 | #define IGD_M2_MIN 0 | 123 | #define PINEVIEW_M2_MIN 0 |
| 124 | #define IGD_M2_MAX 254 | 124 | #define PINEVIEW_M2_MAX 254 |
| 125 | #define I9XX_P_SDVO_DAC_MIN 5 | 125 | #define I9XX_P_SDVO_DAC_MIN 5 |
| 126 | #define I9XX_P_SDVO_DAC_MAX 80 | 126 | #define I9XX_P_SDVO_DAC_MAX 80 |
| 127 | #define I9XX_P_LVDS_MIN 7 | 127 | #define I9XX_P_LVDS_MIN 7 |
| 128 | #define I9XX_P_LVDS_MAX 98 | 128 | #define I9XX_P_LVDS_MAX 98 |
| 129 | #define IGD_P_LVDS_MIN 7 | 129 | #define PINEVIEW_P_LVDS_MIN 7 |
| 130 | #define IGD_P_LVDS_MAX 112 | 130 | #define PINEVIEW_P_LVDS_MAX 112 |
| 131 | #define I9XX_P1_MIN 1 | 131 | #define I9XX_P1_MIN 1 |
| 132 | #define I9XX_P1_MAX 8 | 132 | #define I9XX_P1_MAX 8 |
| 133 | #define I9XX_P2_SDVO_DAC_SLOW 10 | 133 | #define I9XX_P2_SDVO_DAC_SLOW 10 |
| @@ -234,33 +234,33 @@ struct intel_limit { | |||
| 234 | #define G4X_P2_DISPLAY_PORT_FAST 10 | 234 | #define G4X_P2_DISPLAY_PORT_FAST 10 |
| 235 | #define G4X_P2_DISPLAY_PORT_LIMIT 0 | 235 | #define G4X_P2_DISPLAY_PORT_LIMIT 0 |
| 236 | 236 | ||
| 237 | /* IGDNG */ | 237 | /* Ironlake */ |
| 238 | /* as we calculate clock using (register_value + 2) for | 238 | /* as we calculate clock using (register_value + 2) for |
| 239 | N/M1/M2, so here the range value for them is (actual_value-2). | 239 | N/M1/M2, so here the range value for them is (actual_value-2). |
| 240 | */ | 240 | */ |
| 241 | #define IGDNG_DOT_MIN 25000 | 241 | #define IRONLAKE_DOT_MIN 25000 |
| 242 | #define IGDNG_DOT_MAX 350000 | 242 | #define IRONLAKE_DOT_MAX 350000 |
| 243 | #define IGDNG_VCO_MIN 1760000 | 243 | #define IRONLAKE_VCO_MIN 1760000 |
| 244 | #define IGDNG_VCO_MAX 3510000 | 244 | #define IRONLAKE_VCO_MAX 3510000 |
| 245 | #define IGDNG_N_MIN 1 | 245 | #define IRONLAKE_N_MIN 1 |
| 246 | #define IGDNG_N_MAX 5 | 246 | #define IRONLAKE_N_MAX 5 |
| 247 | #define IGDNG_M_MIN 79 | 247 | #define IRONLAKE_M_MIN 79 |
| 248 | #define IGDNG_M_MAX 118 | 248 | #define IRONLAKE_M_MAX 118 |
| 249 | #define IGDNG_M1_MIN 12 | 249 | #define IRONLAKE_M1_MIN 12 |
| 250 | #define IGDNG_M1_MAX 23 | 250 | #define IRONLAKE_M1_MAX 23 |
| 251 | #define IGDNG_M2_MIN 5 | 251 | #define IRONLAKE_M2_MIN 5 |
| 252 | #define IGDNG_M2_MAX 9 | 252 | #define IRONLAKE_M2_MAX 9 |
| 253 | #define IGDNG_P_SDVO_DAC_MIN 5 | 253 | #define IRONLAKE_P_SDVO_DAC_MIN 5 |
| 254 | #define IGDNG_P_SDVO_DAC_MAX 80 | 254 | #define IRONLAKE_P_SDVO_DAC_MAX 80 |
| 255 | #define IGDNG_P_LVDS_MIN 28 | 255 | #define IRONLAKE_P_LVDS_MIN 28 |
| 256 | #define IGDNG_P_LVDS_MAX 112 | 256 | #define IRONLAKE_P_LVDS_MAX 112 |
| 257 | #define IGDNG_P1_MIN 1 | 257 | #define IRONLAKE_P1_MIN 1 |
| 258 | #define IGDNG_P1_MAX 8 | 258 | #define IRONLAKE_P1_MAX 8 |
| 259 | #define IGDNG_P2_SDVO_DAC_SLOW 10 | 259 | #define IRONLAKE_P2_SDVO_DAC_SLOW 10 |
| 260 | #define IGDNG_P2_SDVO_DAC_FAST 5 | 260 | #define IRONLAKE_P2_SDVO_DAC_FAST 5 |
| 261 | #define IGDNG_P2_LVDS_SLOW 14 /* single channel */ | 261 | #define IRONLAKE_P2_LVDS_SLOW 14 /* single channel */ |
| 262 | #define IGDNG_P2_LVDS_FAST 7 /* double channel */ | 262 | #define IRONLAKE_P2_LVDS_FAST 7 /* double channel */ |
| 263 | #define IGDNG_P2_DOT_LIMIT 225000 /* 225Mhz */ | 263 | #define IRONLAKE_P2_DOT_LIMIT 225000 /* 225Mhz */ |
| 264 | 264 | ||
| 265 | static bool | 265 | static bool |
| 266 | intel_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | 266 | intel_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, |
| @@ -272,15 +272,15 @@ static bool | |||
| 272 | intel_g4x_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | 272 | intel_g4x_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, |
| 273 | int target, int refclk, intel_clock_t *best_clock); | 273 | int target, int refclk, intel_clock_t *best_clock); |
| 274 | static bool | 274 | static bool |
| 275 | intel_igdng_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | 275 | intel_ironlake_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, |
| 276 | int target, int refclk, intel_clock_t *best_clock); | 276 | int target, int refclk, intel_clock_t *best_clock); |
| 277 | 277 | ||
| 278 | static bool | 278 | static bool |
| 279 | intel_find_pll_g4x_dp(const intel_limit_t *, struct drm_crtc *crtc, | 279 | intel_find_pll_g4x_dp(const intel_limit_t *, struct drm_crtc *crtc, |
| 280 | int target, int refclk, intel_clock_t *best_clock); | 280 | int target, int refclk, intel_clock_t *best_clock); |
| 281 | static bool | 281 | static bool |
| 282 | intel_find_pll_igdng_dp(const intel_limit_t *, struct drm_crtc *crtc, | 282 | intel_find_pll_ironlake_dp(const intel_limit_t *, struct drm_crtc *crtc, |
| 283 | int target, int refclk, intel_clock_t *best_clock); | 283 | int target, int refclk, intel_clock_t *best_clock); |
| 284 | 284 | ||
| 285 | static const intel_limit_t intel_limits_i8xx_dvo = { | 285 | static const intel_limit_t intel_limits_i8xx_dvo = { |
| 286 | .dot = { .min = I8XX_DOT_MIN, .max = I8XX_DOT_MAX }, | 286 | .dot = { .min = I8XX_DOT_MIN, .max = I8XX_DOT_MAX }, |
| @@ -453,13 +453,13 @@ static const intel_limit_t intel_limits_g4x_display_port = { | |||
| 453 | .find_pll = intel_find_pll_g4x_dp, | 453 | .find_pll = intel_find_pll_g4x_dp, |
| 454 | }; | 454 | }; |
| 455 | 455 | ||
| 456 | static const intel_limit_t intel_limits_igd_sdvo = { | 456 | static const intel_limit_t intel_limits_pineview_sdvo = { |
| 457 | .dot = { .min = I9XX_DOT_MIN, .max = I9XX_DOT_MAX}, | 457 | .dot = { .min = I9XX_DOT_MIN, .max = I9XX_DOT_MAX}, |
| 458 | .vco = { .min = IGD_VCO_MIN, .max = IGD_VCO_MAX }, | 458 | .vco = { .min = PINEVIEW_VCO_MIN, .max = PINEVIEW_VCO_MAX }, |
| 459 | .n = { .min = IGD_N_MIN, .max = IGD_N_MAX }, | 459 | .n = { .min = PINEVIEW_N_MIN, .max = PINEVIEW_N_MAX }, |
| 460 | .m = { .min = IGD_M_MIN, .max = IGD_M_MAX }, | 460 | .m = { .min = PINEVIEW_M_MIN, .max = PINEVIEW_M_MAX }, |
| 461 | .m1 = { .min = IGD_M1_MIN, .max = IGD_M1_MAX }, | 461 | .m1 = { .min = PINEVIEW_M1_MIN, .max = PINEVIEW_M1_MAX }, |
| 462 | .m2 = { .min = IGD_M2_MIN, .max = IGD_M2_MAX }, | 462 | .m2 = { .min = PINEVIEW_M2_MIN, .max = PINEVIEW_M2_MAX }, |
| 463 | .p = { .min = I9XX_P_SDVO_DAC_MIN, .max = I9XX_P_SDVO_DAC_MAX }, | 463 | .p = { .min = I9XX_P_SDVO_DAC_MIN, .max = I9XX_P_SDVO_DAC_MAX }, |
| 464 | .p1 = { .min = I9XX_P1_MIN, .max = I9XX_P1_MAX }, | 464 | .p1 = { .min = I9XX_P1_MIN, .max = I9XX_P1_MAX }, |
| 465 | .p2 = { .dot_limit = I9XX_P2_SDVO_DAC_SLOW_LIMIT, | 465 | .p2 = { .dot_limit = I9XX_P2_SDVO_DAC_SLOW_LIMIT, |
| @@ -468,59 +468,59 @@ static const intel_limit_t intel_limits_igd_sdvo = { | |||
| 468 | .find_reduced_pll = intel_find_best_reduced_PLL, | 468 | .find_reduced_pll = intel_find_best_reduced_PLL, |
| 469 | }; | 469 | }; |
| 470 | 470 | ||
| 471 | static const intel_limit_t intel_limits_igd_lvds = { | 471 | static const intel_limit_t intel_limits_pineview_lvds = { |
| 472 | .dot = { .min = I9XX_DOT_MIN, .max = I9XX_DOT_MAX }, | 472 | .dot = { .min = I9XX_DOT_MIN, .max = I9XX_DOT_MAX }, |
| 473 | .vco = { .min = IGD_VCO_MIN, .max = IGD_VCO_MAX }, | 473 | .vco = { .min = PINEVIEW_VCO_MIN, .max = PINEVIEW_VCO_MAX }, |
| 474 | .n = { .min = IGD_N_MIN, .max = IGD_N_MAX }, | 474 | .n = { .min = PINEVIEW_N_MIN, .max = PINEVIEW_N_MAX }, |
| 475 | .m = { .min = IGD_M_MIN, .max = IGD_M_MAX }, | 475 | .m = { .min = PINEVIEW_M_MIN, .max = PINEVIEW_M_MAX }, |
| 476 | .m1 = { .min = IGD_M1_MIN, .max = IGD_M1_MAX }, | 476 | .m1 = { .min = PINEVIEW_M1_MIN, .max = PINEVIEW_M1_MAX }, |
| 477 | .m2 = { .min = IGD_M2_MIN, .max = IGD_M2_MAX }, | 477 | .m2 = { .min = PINEVIEW_M2_MIN, .max = PINEVIEW_M2_MAX }, |
| 478 | .p = { .min = IGD_P_LVDS_MIN, .max = IGD_P_LVDS_MAX }, | 478 | .p = { .min = PINEVIEW_P_LVDS_MIN, .max = PINEVIEW_P_LVDS_MAX }, |
| 479 | .p1 = { .min = I9XX_P1_MIN, .max = I9XX_P1_MAX }, | 479 | .p1 = { .min = I9XX_P1_MIN, .max = I9XX_P1_MAX }, |
| 480 | /* IGD only supports single-channel mode. */ | 480 | /* Pineview only supports single-channel mode. */ |
| 481 | .p2 = { .dot_limit = I9XX_P2_LVDS_SLOW_LIMIT, | 481 | .p2 = { .dot_limit = I9XX_P2_LVDS_SLOW_LIMIT, |
| 482 | .p2_slow = I9XX_P2_LVDS_SLOW, .p2_fast = I9XX_P2_LVDS_SLOW }, | 482 | .p2_slow = I9XX_P2_LVDS_SLOW, .p2_fast = I9XX_P2_LVDS_SLOW }, |
| 483 | .find_pll = intel_find_best_PLL, | 483 | .find_pll = intel_find_best_PLL, |
| 484 | .find_reduced_pll = intel_find_best_reduced_PLL, | 484 | .find_reduced_pll = intel_find_best_reduced_PLL, |
| 485 | }; | 485 | }; |
| 486 | 486 | ||
| 487 | static const intel_limit_t intel_limits_igdng_sdvo = { | 487 | static const intel_limit_t intel_limits_ironlake_sdvo = { |
| 488 | .dot = { .min = IGDNG_DOT_MIN, .max = IGDNG_DOT_MAX }, | 488 | .dot = { .min = IRONLAKE_DOT_MIN, .max = IRONLAKE_DOT_MAX }, |
| 489 | .vco = { .min = IGDNG_VCO_MIN, .max = IGDNG_VCO_MAX }, | 489 | .vco = { .min = IRONLAKE_VCO_MIN, .max = IRONLAKE_VCO_MAX }, |
| 490 | .n = { .min = IGDNG_N_MIN, .max = IGDNG_N_MAX }, | 490 | .n = { .min = IRONLAKE_N_MIN, .max = IRONLAKE_N_MAX }, |
| 491 | .m = { .min = IGDNG_M_MIN, .max = IGDNG_M_MAX }, | 491 | .m = { .min = IRONLAKE_M_MIN, .max = IRONLAKE_M_MAX }, |
| 492 | .m1 = { .min = IGDNG_M1_MIN, .max = IGDNG_M1_MAX }, | 492 | .m1 = { .min = IRONLAKE_M1_MIN, .max = IRONLAKE_M1_MAX }, |
| 493 | .m2 = { .min = IGDNG_M2_MIN, .max = IGDNG_M2_MAX }, | 493 | .m2 = { .min = IRONLAKE_M2_MIN, .max = IRONLAKE_M2_MAX }, |
| 494 | .p = { .min = IGDNG_P_SDVO_DAC_MIN, .max = IGDNG_P_SDVO_DAC_MAX }, | 494 | .p = { .min = IRONLAKE_P_SDVO_DAC_MIN, .max = IRONLAKE_P_SDVO_DAC_MAX }, |
| 495 | .p1 = { .min = IGDNG_P1_MIN, .max = IGDNG_P1_MAX }, | 495 | .p1 = { .min = IRONLAKE_P1_MIN, .max = IRONLAKE_P1_MAX }, |
| 496 | .p2 = { .dot_limit = IGDNG_P2_DOT_LIMIT, | 496 | .p2 = { .dot_limit = IRONLAKE_P2_DOT_LIMIT, |
| 497 | .p2_slow = IGDNG_P2_SDVO_DAC_SLOW, | 497 | .p2_slow = IRONLAKE_P2_SDVO_DAC_SLOW, |
| 498 | .p2_fast = IGDNG_P2_SDVO_DAC_FAST }, | 498 | .p2_fast = IRONLAKE_P2_SDVO_DAC_FAST }, |
| 499 | .find_pll = intel_igdng_find_best_PLL, | 499 | .find_pll = intel_ironlake_find_best_PLL, |
| 500 | }; | 500 | }; |
| 501 | 501 | ||
| 502 | static const intel_limit_t intel_limits_igdng_lvds = { | 502 | static const intel_limit_t intel_limits_ironlake_lvds = { |
| 503 | .dot = { .min = IGDNG_DOT_MIN, .max = IGDNG_DOT_MAX }, | 503 | .dot = { .min = IRONLAKE_DOT_MIN, .max = IRONLAKE_DOT_MAX }, |
| 504 | .vco = { .min = IGDNG_VCO_MIN, .max = IGDNG_VCO_MAX }, | 504 | .vco = { .min = IRONLAKE_VCO_MIN, .max = IRONLAKE_VCO_MAX }, |
| 505 | .n = { .min = IGDNG_N_MIN, .max = IGDNG_N_MAX }, | 505 | .n = { .min = IRONLAKE_N_MIN, .max = IRONLAKE_N_MAX }, |
| 506 | .m = { .min = IGDNG_M_MIN, .max = IGDNG_M_MAX }, | 506 | .m = { .min = IRONLAKE_M_MIN, .max = IRONLAKE_M_MAX }, |
| 507 | .m1 = { .min = IGDNG_M1_MIN, .max = IGDNG_M1_MAX }, | 507 | .m1 = { .min = IRONLAKE_M1_MIN, .max = IRONLAKE_M1_MAX }, |
| 508 | .m2 = { .min = IGDNG_M2_MIN, .max = IGDNG_M2_MAX }, | 508 | .m2 = { .min = IRONLAKE_M2_MIN, .max = IRONLAKE_M2_MAX }, |
| 509 | .p = { .min = IGDNG_P_LVDS_MIN, .max = IGDNG_P_LVDS_MAX }, | 509 | .p = { .min = IRONLAKE_P_LVDS_MIN, .max = IRONLAKE_P_LVDS_MAX }, |
| 510 | .p1 = { .min = IGDNG_P1_MIN, .max = IGDNG_P1_MAX }, | 510 | .p1 = { .min = IRONLAKE_P1_MIN, .max = IRONLAKE_P1_MAX }, |
| 511 | .p2 = { .dot_limit = IGDNG_P2_DOT_LIMIT, | 511 | .p2 = { .dot_limit = IRONLAKE_P2_DOT_LIMIT, |
| 512 | .p2_slow = IGDNG_P2_LVDS_SLOW, | 512 | .p2_slow = IRONLAKE_P2_LVDS_SLOW, |
| 513 | .p2_fast = IGDNG_P2_LVDS_FAST }, | 513 | .p2_fast = IRONLAKE_P2_LVDS_FAST }, |
| 514 | .find_pll = intel_igdng_find_best_PLL, | 514 | .find_pll = intel_ironlake_find_best_PLL, |
| 515 | }; | 515 | }; |
| 516 | 516 | ||
| 517 | static const intel_limit_t *intel_igdng_limit(struct drm_crtc *crtc) | 517 | static const intel_limit_t *intel_ironlake_limit(struct drm_crtc *crtc) |
| 518 | { | 518 | { |
| 519 | const intel_limit_t *limit; | 519 | const intel_limit_t *limit; |
| 520 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) | 520 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) |
| 521 | limit = &intel_limits_igdng_lvds; | 521 | limit = &intel_limits_ironlake_lvds; |
| 522 | else | 522 | else |
| 523 | limit = &intel_limits_igdng_sdvo; | 523 | limit = &intel_limits_ironlake_sdvo; |
| 524 | 524 | ||
| 525 | return limit; | 525 | return limit; |
| 526 | } | 526 | } |
| @@ -557,20 +557,20 @@ static const intel_limit_t *intel_limit(struct drm_crtc *crtc) | |||
| 557 | struct drm_device *dev = crtc->dev; | 557 | struct drm_device *dev = crtc->dev; |
| 558 | const intel_limit_t *limit; | 558 | const intel_limit_t *limit; |
| 559 | 559 | ||
| 560 | if (IS_IGDNG(dev)) | 560 | if (IS_IRONLAKE(dev)) |
| 561 | limit = intel_igdng_limit(crtc); | 561 | limit = intel_ironlake_limit(crtc); |
| 562 | else if (IS_G4X(dev)) { | 562 | else if (IS_G4X(dev)) { |
| 563 | limit = intel_g4x_limit(crtc); | 563 | limit = intel_g4x_limit(crtc); |
| 564 | } else if (IS_I9XX(dev) && !IS_IGD(dev)) { | 564 | } else if (IS_I9XX(dev) && !IS_PINEVIEW(dev)) { |
| 565 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) | 565 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) |
| 566 | limit = &intel_limits_i9xx_lvds; | 566 | limit = &intel_limits_i9xx_lvds; |
| 567 | else | 567 | else |
| 568 | limit = &intel_limits_i9xx_sdvo; | 568 | limit = &intel_limits_i9xx_sdvo; |
| 569 | } else if (IS_IGD(dev)) { | 569 | } else if (IS_PINEVIEW(dev)) { |
| 570 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) | 570 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) |
| 571 | limit = &intel_limits_igd_lvds; | 571 | limit = &intel_limits_pineview_lvds; |
| 572 | else | 572 | else |
| 573 | limit = &intel_limits_igd_sdvo; | 573 | limit = &intel_limits_pineview_sdvo; |
| 574 | } else { | 574 | } else { |
| 575 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) | 575 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) |
| 576 | limit = &intel_limits_i8xx_lvds; | 576 | limit = &intel_limits_i8xx_lvds; |
| @@ -580,8 +580,8 @@ static const intel_limit_t *intel_limit(struct drm_crtc *crtc) | |||
| 580 | return limit; | 580 | return limit; |
| 581 | } | 581 | } |
| 582 | 582 | ||
| 583 | /* m1 is reserved as 0 in IGD, n is a ring counter */ | 583 | /* m1 is reserved as 0 in Pineview, n is a ring counter */ |
| 584 | static void igd_clock(int refclk, intel_clock_t *clock) | 584 | static void pineview_clock(int refclk, intel_clock_t *clock) |
| 585 | { | 585 | { |
| 586 | clock->m = clock->m2 + 2; | 586 | clock->m = clock->m2 + 2; |
| 587 | clock->p = clock->p1 * clock->p2; | 587 | clock->p = clock->p1 * clock->p2; |
| @@ -591,8 +591,8 @@ static void igd_clock(int refclk, intel_clock_t *clock) | |||
| 591 | 591 | ||
| 592 | static void intel_clock(struct drm_device *dev, int refclk, intel_clock_t *clock) | 592 | static void intel_clock(struct drm_device *dev, int refclk, intel_clock_t *clock) |
| 593 | { | 593 | { |
| 594 | if (IS_IGD(dev)) { | 594 | if (IS_PINEVIEW(dev)) { |
| 595 | igd_clock(refclk, clock); | 595 | pineview_clock(refclk, clock); |
| 596 | return; | 596 | return; |
| 597 | } | 597 | } |
| 598 | clock->m = 5 * (clock->m1 + 2) + (clock->m2 + 2); | 598 | clock->m = 5 * (clock->m1 + 2) + (clock->m2 + 2); |
| @@ -657,7 +657,7 @@ static bool intel_PLL_is_valid(struct drm_crtc *crtc, intel_clock_t *clock) | |||
| 657 | INTELPllInvalid ("m2 out of range\n"); | 657 | INTELPllInvalid ("m2 out of range\n"); |
| 658 | if (clock->m1 < limit->m1.min || limit->m1.max < clock->m1) | 658 | if (clock->m1 < limit->m1.min || limit->m1.max < clock->m1) |
| 659 | INTELPllInvalid ("m1 out of range\n"); | 659 | INTELPllInvalid ("m1 out of range\n"); |
| 660 | if (clock->m1 <= clock->m2 && !IS_IGD(dev)) | 660 | if (clock->m1 <= clock->m2 && !IS_PINEVIEW(dev)) |
| 661 | INTELPllInvalid ("m1 <= m2\n"); | 661 | INTELPllInvalid ("m1 <= m2\n"); |
| 662 | if (clock->m < limit->m.min || limit->m.max < clock->m) | 662 | if (clock->m < limit->m.min || limit->m.max < clock->m) |
| 663 | INTELPllInvalid ("m out of range\n"); | 663 | INTELPllInvalid ("m out of range\n"); |
| @@ -706,16 +706,17 @@ intel_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | |||
| 706 | 706 | ||
| 707 | memset (best_clock, 0, sizeof (*best_clock)); | 707 | memset (best_clock, 0, sizeof (*best_clock)); |
| 708 | 708 | ||
| 709 | for (clock.p1 = limit->p1.max; clock.p1 >= limit->p1.min; clock.p1--) { | 709 | for (clock.m1 = limit->m1.min; clock.m1 <= limit->m1.max; |
| 710 | for (clock.m1 = limit->m1.min; clock.m1 <= limit->m1.max; | 710 | clock.m1++) { |
| 711 | clock.m1++) { | 711 | for (clock.m2 = limit->m2.min; |
| 712 | for (clock.m2 = limit->m2.min; | 712 | clock.m2 <= limit->m2.max; clock.m2++) { |
| 713 | clock.m2 <= limit->m2.max; clock.m2++) { | 713 | /* m1 is always 0 in Pineview */ |
| 714 | /* m1 is always 0 in IGD */ | 714 | if (clock.m2 >= clock.m1 && !IS_PINEVIEW(dev)) |
| 715 | if (clock.m2 >= clock.m1 && !IS_IGD(dev)) | 715 | break; |
| 716 | break; | 716 | for (clock.n = limit->n.min; |
| 717 | for (clock.n = limit->n.min; | 717 | clock.n <= limit->n.max; clock.n++) { |
| 718 | clock.n <= limit->n.max; clock.n++) { | 718 | for (clock.p1 = limit->p1.min; |
| 719 | clock.p1 <= limit->p1.max; clock.p1++) { | ||
| 719 | int this_err; | 720 | int this_err; |
| 720 | 721 | ||
| 721 | intel_clock(dev, refclk, &clock); | 722 | intel_clock(dev, refclk, &clock); |
| @@ -751,8 +752,8 @@ intel_find_best_reduced_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | |||
| 751 | 752 | ||
| 752 | for (clock.m1 = limit->m1.min; clock.m1 <= limit->m1.max; clock.m1++) { | 753 | for (clock.m1 = limit->m1.min; clock.m1 <= limit->m1.max; clock.m1++) { |
| 753 | for (clock.m2 = limit->m2.min; clock.m2 <= limit->m2.max; clock.m2++) { | 754 | for (clock.m2 = limit->m2.min; clock.m2 <= limit->m2.max; clock.m2++) { |
| 754 | /* m1 is always 0 in IGD */ | 755 | /* m1 is always 0 in Pineview */ |
| 755 | if (clock.m2 >= clock.m1 && !IS_IGD(dev)) | 756 | if (clock.m2 >= clock.m1 && !IS_PINEVIEW(dev)) |
| 756 | break; | 757 | break; |
| 757 | for (clock.n = limit->n.min; clock.n <= limit->n.max; | 758 | for (clock.n = limit->n.min; clock.n <= limit->n.max; |
| 758 | clock.n++) { | 759 | clock.n++) { |
| @@ -833,8 +834,8 @@ intel_g4x_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | |||
| 833 | } | 834 | } |
| 834 | 835 | ||
| 835 | static bool | 836 | static bool |
| 836 | intel_find_pll_igdng_dp(const intel_limit_t *limit, struct drm_crtc *crtc, | 837 | intel_find_pll_ironlake_dp(const intel_limit_t *limit, struct drm_crtc *crtc, |
| 837 | int target, int refclk, intel_clock_t *best_clock) | 838 | int target, int refclk, intel_clock_t *best_clock) |
| 838 | { | 839 | { |
| 839 | struct drm_device *dev = crtc->dev; | 840 | struct drm_device *dev = crtc->dev; |
| 840 | intel_clock_t clock; | 841 | intel_clock_t clock; |
| @@ -857,8 +858,8 @@ intel_find_pll_igdng_dp(const intel_limit_t *limit, struct drm_crtc *crtc, | |||
| 857 | } | 858 | } |
| 858 | 859 | ||
| 859 | static bool | 860 | static bool |
| 860 | intel_igdng_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | 861 | intel_ironlake_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, |
| 861 | int target, int refclk, intel_clock_t *best_clock) | 862 | int target, int refclk, intel_clock_t *best_clock) |
| 862 | { | 863 | { |
| 863 | struct drm_device *dev = crtc->dev; | 864 | struct drm_device *dev = crtc->dev; |
| 864 | struct drm_i915_private *dev_priv = dev->dev_private; | 865 | struct drm_i915_private *dev_priv = dev->dev_private; |
| @@ -871,7 +872,7 @@ intel_igdng_find_best_PLL(const intel_limit_t *limit, struct drm_crtc *crtc, | |||
| 871 | return true; | 872 | return true; |
| 872 | 873 | ||
| 873 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_DISPLAYPORT)) | 874 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_DISPLAYPORT)) |
| 874 | return intel_find_pll_igdng_dp(limit, crtc, target, | 875 | return intel_find_pll_ironlake_dp(limit, crtc, target, |
| 875 | refclk, best_clock); | 876 | refclk, best_clock); |
| 876 | 877 | ||
| 877 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) { | 878 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) { |
| @@ -949,7 +950,7 @@ void | |||
| 949 | intel_wait_for_vblank(struct drm_device *dev) | 950 | intel_wait_for_vblank(struct drm_device *dev) |
| 950 | { | 951 | { |
| 951 | /* Wait for 20ms, i.e. one cycle at 50hz. */ | 952 | /* Wait for 20ms, i.e. one cycle at 50hz. */ |
| 952 | mdelay(20); | 953 | msleep(20); |
| 953 | } | 954 | } |
| 954 | 955 | ||
| 955 | /* Parameters have changed, update FBC info */ | 956 | /* Parameters have changed, update FBC info */ |
| @@ -994,7 +995,7 @@ static void i8xx_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
| 994 | fbc_ctl |= dev_priv->cfb_fence; | 995 | fbc_ctl |= dev_priv->cfb_fence; |
| 995 | I915_WRITE(FBC_CONTROL, fbc_ctl); | 996 | I915_WRITE(FBC_CONTROL, fbc_ctl); |
| 996 | 997 | ||
| 997 | DRM_DEBUG("enabled FBC, pitch %ld, yoff %d, plane %d, ", | 998 | DRM_DEBUG_KMS("enabled FBC, pitch %ld, yoff %d, plane %d, ", |
| 998 | dev_priv->cfb_pitch, crtc->y, dev_priv->cfb_plane); | 999 | dev_priv->cfb_pitch, crtc->y, dev_priv->cfb_plane); |
| 999 | } | 1000 | } |
| 1000 | 1001 | ||
| @@ -1017,7 +1018,7 @@ void i8xx_disable_fbc(struct drm_device *dev) | |||
| 1017 | 1018 | ||
| 1018 | intel_wait_for_vblank(dev); | 1019 | intel_wait_for_vblank(dev); |
| 1019 | 1020 | ||
| 1020 | DRM_DEBUG("disabled FBC\n"); | 1021 | DRM_DEBUG_KMS("disabled FBC\n"); |
| 1021 | } | 1022 | } |
| 1022 | 1023 | ||
| 1023 | static bool i8xx_fbc_enabled(struct drm_crtc *crtc) | 1024 | static bool i8xx_fbc_enabled(struct drm_crtc *crtc) |
| @@ -1062,7 +1063,7 @@ static void g4x_enable_fbc(struct drm_crtc *crtc, unsigned long interval) | |||
| 1062 | /* enable it... */ | 1063 | /* enable it... */ |
| 1063 | I915_WRITE(DPFC_CONTROL, I915_READ(DPFC_CONTROL) | DPFC_CTL_EN); | 1064 | I915_WRITE(DPFC_CONTROL, I915_READ(DPFC_CONTROL) | DPFC_CTL_EN); |
| 1064 | 1065 | ||
| 1065 | DRM_DEBUG("enabled fbc on plane %d\n", intel_crtc->plane); | 1066 | DRM_DEBUG_KMS("enabled fbc on plane %d\n", intel_crtc->plane); |
| 1066 | } | 1067 | } |
| 1067 | 1068 | ||
| 1068 | void g4x_disable_fbc(struct drm_device *dev) | 1069 | void g4x_disable_fbc(struct drm_device *dev) |
| @@ -1076,7 +1077,7 @@ void g4x_disable_fbc(struct drm_device *dev) | |||
| 1076 | I915_WRITE(DPFC_CONTROL, dpfc_ctl); | 1077 | I915_WRITE(DPFC_CONTROL, dpfc_ctl); |
| 1077 | intel_wait_for_vblank(dev); | 1078 | intel_wait_for_vblank(dev); |
| 1078 | 1079 | ||
| 1079 | DRM_DEBUG("disabled FBC\n"); | 1080 | DRM_DEBUG_KMS("disabled FBC\n"); |
| 1080 | } | 1081 | } |
| 1081 | 1082 | ||
| 1082 | static bool g4x_fbc_enabled(struct drm_crtc *crtc) | 1083 | static bool g4x_fbc_enabled(struct drm_crtc *crtc) |
| @@ -1141,25 +1142,27 @@ static void intel_update_fbc(struct drm_crtc *crtc, | |||
| 1141 | * - going to an unsupported config (interlace, pixel multiply, etc.) | 1142 | * - going to an unsupported config (interlace, pixel multiply, etc.) |
| 1142 | */ | 1143 | */ |
| 1143 | if (intel_fb->obj->size > dev_priv->cfb_size) { | 1144 | if (intel_fb->obj->size > dev_priv->cfb_size) { |
| 1144 | DRM_DEBUG("framebuffer too large, disabling compression\n"); | 1145 | DRM_DEBUG_KMS("framebuffer too large, disabling " |
| 1146 | "compression\n"); | ||
| 1145 | goto out_disable; | 1147 | goto out_disable; |
| 1146 | } | 1148 | } |
| 1147 | if ((mode->flags & DRM_MODE_FLAG_INTERLACE) || | 1149 | if ((mode->flags & DRM_MODE_FLAG_INTERLACE) || |
| 1148 | (mode->flags & DRM_MODE_FLAG_DBLSCAN)) { | 1150 | (mode->flags & DRM_MODE_FLAG_DBLSCAN)) { |
| 1149 | DRM_DEBUG("mode incompatible with compression, disabling\n"); | 1151 | DRM_DEBUG_KMS("mode incompatible with compression, " |
| 1152 | "disabling\n"); | ||
| 1150 | goto out_disable; | 1153 | goto out_disable; |
| 1151 | } | 1154 | } |
| 1152 | if ((mode->hdisplay > 2048) || | 1155 | if ((mode->hdisplay > 2048) || |
| 1153 | (mode->vdisplay > 1536)) { | 1156 | (mode->vdisplay > 1536)) { |
| 1154 | DRM_DEBUG("mode too large for compression, disabling\n"); | 1157 | DRM_DEBUG_KMS("mode too large for compression, disabling\n"); |
| 1155 | goto out_disable; | 1158 | goto out_disable; |
| 1156 | } | 1159 | } |
| 1157 | if ((IS_I915GM(dev) || IS_I945GM(dev)) && plane != 0) { | 1160 | if ((IS_I915GM(dev) || IS_I945GM(dev)) && plane != 0) { |
| 1158 | DRM_DEBUG("plane not 0, disabling compression\n"); | 1161 | DRM_DEBUG_KMS("plane not 0, disabling compression\n"); |
| 1159 | goto out_disable; | 1162 | goto out_disable; |
| 1160 | } | 1163 | } |
| 1161 | if (obj_priv->tiling_mode != I915_TILING_X) { | 1164 | if (obj_priv->tiling_mode != I915_TILING_X) { |
| 1162 | DRM_DEBUG("framebuffer not tiled, disabling compression\n"); | 1165 | DRM_DEBUG_KMS("framebuffer not tiled, disabling compression\n"); |
| 1163 | goto out_disable; | 1166 | goto out_disable; |
| 1164 | } | 1167 | } |
| 1165 | 1168 | ||
| @@ -1181,13 +1184,57 @@ static void intel_update_fbc(struct drm_crtc *crtc, | |||
| 1181 | return; | 1184 | return; |
| 1182 | 1185 | ||
| 1183 | out_disable: | 1186 | out_disable: |
| 1184 | DRM_DEBUG("unsupported config, disabling FBC\n"); | 1187 | DRM_DEBUG_KMS("unsupported config, disabling FBC\n"); |
| 1185 | /* Multiple disables should be harmless */ | 1188 | /* Multiple disables should be harmless */ |
| 1186 | if (dev_priv->display.fbc_enabled(crtc)) | 1189 | if (dev_priv->display.fbc_enabled(crtc)) |
| 1187 | dev_priv->display.disable_fbc(dev); | 1190 | dev_priv->display.disable_fbc(dev); |
| 1188 | } | 1191 | } |
| 1189 | 1192 | ||
| 1190 | static int | 1193 | static int |
| 1194 | intel_pin_and_fence_fb_obj(struct drm_device *dev, struct drm_gem_object *obj) | ||
| 1195 | { | ||
| 1196 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | ||
| 1197 | u32 alignment; | ||
| 1198 | int ret; | ||
| 1199 | |||
| 1200 | switch (obj_priv->tiling_mode) { | ||
| 1201 | case I915_TILING_NONE: | ||
| 1202 | alignment = 64 * 1024; | ||
| 1203 | break; | ||
| 1204 | case I915_TILING_X: | ||
| 1205 | /* pin() will align the object as required by fence */ | ||
| 1206 | alignment = 0; | ||
| 1207 | break; | ||
| 1208 | case I915_TILING_Y: | ||
| 1209 | /* FIXME: Is this true? */ | ||
| 1210 | DRM_ERROR("Y tiled not allowed for scan out buffers\n"); | ||
| 1211 | return -EINVAL; | ||
| 1212 | default: | ||
| 1213 | BUG(); | ||
| 1214 | } | ||
| 1215 | |||
| 1216 | ret = i915_gem_object_pin(obj, alignment); | ||
| 1217 | if (ret != 0) | ||
| 1218 | return ret; | ||
| 1219 | |||
| 1220 | /* Install a fence for tiled scan-out. Pre-i965 always needs a | ||
| 1221 | * fence, whereas 965+ only requires a fence if using | ||
| 1222 | * framebuffer compression. For simplicity, we always install | ||
| 1223 | * a fence as the cost is not that onerous. | ||
| 1224 | */ | ||
| 1225 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE && | ||
| 1226 | obj_priv->tiling_mode != I915_TILING_NONE) { | ||
| 1227 | ret = i915_gem_object_get_fence_reg(obj); | ||
| 1228 | if (ret != 0) { | ||
| 1229 | i915_gem_object_unpin(obj); | ||
| 1230 | return ret; | ||
| 1231 | } | ||
| 1232 | } | ||
| 1233 | |||
| 1234 | return 0; | ||
| 1235 | } | ||
| 1236 | |||
| 1237 | static int | ||
| 1191 | intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | 1238 | intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, |
| 1192 | struct drm_framebuffer *old_fb) | 1239 | struct drm_framebuffer *old_fb) |
| 1193 | { | 1240 | { |
| @@ -1206,12 +1253,12 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1206 | int dspstride = (plane == 0) ? DSPASTRIDE : DSPBSTRIDE; | 1253 | int dspstride = (plane == 0) ? DSPASTRIDE : DSPBSTRIDE; |
| 1207 | int dsptileoff = (plane == 0 ? DSPATILEOFF : DSPBTILEOFF); | 1254 | int dsptileoff = (plane == 0 ? DSPATILEOFF : DSPBTILEOFF); |
| 1208 | int dspcntr_reg = (plane == 0) ? DSPACNTR : DSPBCNTR; | 1255 | int dspcntr_reg = (plane == 0) ? DSPACNTR : DSPBCNTR; |
| 1209 | u32 dspcntr, alignment; | 1256 | u32 dspcntr; |
| 1210 | int ret; | 1257 | int ret; |
| 1211 | 1258 | ||
| 1212 | /* no fb bound */ | 1259 | /* no fb bound */ |
| 1213 | if (!crtc->fb) { | 1260 | if (!crtc->fb) { |
| 1214 | DRM_DEBUG("No FB bound\n"); | 1261 | DRM_DEBUG_KMS("No FB bound\n"); |
| 1215 | return 0; | 1262 | return 0; |
| 1216 | } | 1263 | } |
| 1217 | 1264 | ||
| @@ -1228,24 +1275,8 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1228 | obj = intel_fb->obj; | 1275 | obj = intel_fb->obj; |
| 1229 | obj_priv = obj->driver_private; | 1276 | obj_priv = obj->driver_private; |
| 1230 | 1277 | ||
| 1231 | switch (obj_priv->tiling_mode) { | ||
| 1232 | case I915_TILING_NONE: | ||
| 1233 | alignment = 64 * 1024; | ||
| 1234 | break; | ||
| 1235 | case I915_TILING_X: | ||
| 1236 | /* pin() will align the object as required by fence */ | ||
| 1237 | alignment = 0; | ||
| 1238 | break; | ||
| 1239 | case I915_TILING_Y: | ||
| 1240 | /* FIXME: Is this true? */ | ||
| 1241 | DRM_ERROR("Y tiled not allowed for scan out buffers\n"); | ||
| 1242 | return -EINVAL; | ||
| 1243 | default: | ||
| 1244 | BUG(); | ||
| 1245 | } | ||
| 1246 | |||
| 1247 | mutex_lock(&dev->struct_mutex); | 1278 | mutex_lock(&dev->struct_mutex); |
| 1248 | ret = i915_gem_object_pin(obj, alignment); | 1279 | ret = intel_pin_and_fence_fb_obj(dev, obj); |
| 1249 | if (ret != 0) { | 1280 | if (ret != 0) { |
| 1250 | mutex_unlock(&dev->struct_mutex); | 1281 | mutex_unlock(&dev->struct_mutex); |
| 1251 | return ret; | 1282 | return ret; |
| @@ -1258,20 +1289,6 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1258 | return ret; | 1289 | return ret; |
| 1259 | } | 1290 | } |
| 1260 | 1291 | ||
| 1261 | /* Install a fence for tiled scan-out. Pre-i965 always needs a fence, | ||
| 1262 | * whereas 965+ only requires a fence if using framebuffer compression. | ||
| 1263 | * For simplicity, we always install a fence as the cost is not that onerous. | ||
| 1264 | */ | ||
| 1265 | if (obj_priv->fence_reg == I915_FENCE_REG_NONE && | ||
| 1266 | obj_priv->tiling_mode != I915_TILING_NONE) { | ||
| 1267 | ret = i915_gem_object_get_fence_reg(obj); | ||
| 1268 | if (ret != 0) { | ||
| 1269 | i915_gem_object_unpin(obj); | ||
| 1270 | mutex_unlock(&dev->struct_mutex); | ||
| 1271 | return ret; | ||
| 1272 | } | ||
| 1273 | } | ||
| 1274 | |||
| 1275 | dspcntr = I915_READ(dspcntr_reg); | 1292 | dspcntr = I915_READ(dspcntr_reg); |
| 1276 | /* Mask out pixel format bits in case we change it */ | 1293 | /* Mask out pixel format bits in case we change it */ |
| 1277 | dspcntr &= ~DISPPLANE_PIXFORMAT_MASK; | 1294 | dspcntr &= ~DISPPLANE_PIXFORMAT_MASK; |
| @@ -1287,7 +1304,10 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1287 | break; | 1304 | break; |
| 1288 | case 24: | 1305 | case 24: |
| 1289 | case 32: | 1306 | case 32: |
| 1290 | dspcntr |= DISPPLANE_32BPP_NO_ALPHA; | 1307 | if (crtc->fb->depth == 30) |
| 1308 | dspcntr |= DISPPLANE_32BPP_30BIT_NO_ALPHA; | ||
| 1309 | else | ||
| 1310 | dspcntr |= DISPPLANE_32BPP_NO_ALPHA; | ||
| 1291 | break; | 1311 | break; |
| 1292 | default: | 1312 | default: |
| 1293 | DRM_ERROR("Unknown color depth\n"); | 1313 | DRM_ERROR("Unknown color depth\n"); |
| @@ -1302,7 +1322,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1302 | dspcntr &= ~DISPPLANE_TILED; | 1322 | dspcntr &= ~DISPPLANE_TILED; |
| 1303 | } | 1323 | } |
| 1304 | 1324 | ||
| 1305 | if (IS_IGDNG(dev)) | 1325 | if (IS_IRONLAKE(dev)) |
| 1306 | /* must disable */ | 1326 | /* must disable */ |
| 1307 | dspcntr |= DISPPLANE_TRICKLE_FEED_DISABLE; | 1327 | dspcntr |= DISPPLANE_TRICKLE_FEED_DISABLE; |
| 1308 | 1328 | ||
| @@ -1311,7 +1331,7 @@ intel_pipe_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 1311 | Start = obj_priv->gtt_offset; | 1331 | Start = obj_priv->gtt_offset; |
| 1312 | Offset = y * crtc->fb->pitch + x * (crtc->fb->bits_per_pixel / 8); | 1332 | Offset = y * crtc->fb->pitch + x * (crtc->fb->bits_per_pixel / 8); |
| 1313 | 1333 | ||
| 1314 | DRM_DEBUG("Writing base %08lX %08lX %d %d\n", Start, Offset, x, y); | 1334 | DRM_DEBUG_KMS("Writing base %08lX %08lX %d %d\n", Start, Offset, x, y); |
| 1315 | I915_WRITE(dspstride, crtc->fb->pitch); | 1335 | I915_WRITE(dspstride, crtc->fb->pitch); |
| 1316 | if (IS_I965G(dev)) { | 1336 | if (IS_I965G(dev)) { |
| 1317 | I915_WRITE(dspbase, Offset); | 1337 | I915_WRITE(dspbase, Offset); |
| @@ -1363,7 +1383,7 @@ static void i915_disable_vga (struct drm_device *dev) | |||
| 1363 | u8 sr1; | 1383 | u8 sr1; |
| 1364 | u32 vga_reg; | 1384 | u32 vga_reg; |
| 1365 | 1385 | ||
| 1366 | if (IS_IGDNG(dev)) | 1386 | if (IS_IRONLAKE(dev)) |
| 1367 | vga_reg = CPU_VGACNTRL; | 1387 | vga_reg = CPU_VGACNTRL; |
| 1368 | else | 1388 | else |
| 1369 | vga_reg = VGACNTRL; | 1389 | vga_reg = VGACNTRL; |
| @@ -1379,19 +1399,19 @@ static void i915_disable_vga (struct drm_device *dev) | |||
| 1379 | I915_WRITE(vga_reg, VGA_DISP_DISABLE); | 1399 | I915_WRITE(vga_reg, VGA_DISP_DISABLE); |
| 1380 | } | 1400 | } |
| 1381 | 1401 | ||
| 1382 | static void igdng_disable_pll_edp (struct drm_crtc *crtc) | 1402 | static void ironlake_disable_pll_edp (struct drm_crtc *crtc) |
| 1383 | { | 1403 | { |
| 1384 | struct drm_device *dev = crtc->dev; | 1404 | struct drm_device *dev = crtc->dev; |
| 1385 | struct drm_i915_private *dev_priv = dev->dev_private; | 1405 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1386 | u32 dpa_ctl; | 1406 | u32 dpa_ctl; |
| 1387 | 1407 | ||
| 1388 | DRM_DEBUG("\n"); | 1408 | DRM_DEBUG_KMS("\n"); |
| 1389 | dpa_ctl = I915_READ(DP_A); | 1409 | dpa_ctl = I915_READ(DP_A); |
| 1390 | dpa_ctl &= ~DP_PLL_ENABLE; | 1410 | dpa_ctl &= ~DP_PLL_ENABLE; |
| 1391 | I915_WRITE(DP_A, dpa_ctl); | 1411 | I915_WRITE(DP_A, dpa_ctl); |
| 1392 | } | 1412 | } |
| 1393 | 1413 | ||
| 1394 | static void igdng_enable_pll_edp (struct drm_crtc *crtc) | 1414 | static void ironlake_enable_pll_edp (struct drm_crtc *crtc) |
| 1395 | { | 1415 | { |
| 1396 | struct drm_device *dev = crtc->dev; | 1416 | struct drm_device *dev = crtc->dev; |
| 1397 | struct drm_i915_private *dev_priv = dev->dev_private; | 1417 | struct drm_i915_private *dev_priv = dev->dev_private; |
| @@ -1404,13 +1424,13 @@ static void igdng_enable_pll_edp (struct drm_crtc *crtc) | |||
| 1404 | } | 1424 | } |
| 1405 | 1425 | ||
| 1406 | 1426 | ||
| 1407 | static void igdng_set_pll_edp (struct drm_crtc *crtc, int clock) | 1427 | static void ironlake_set_pll_edp (struct drm_crtc *crtc, int clock) |
| 1408 | { | 1428 | { |
| 1409 | struct drm_device *dev = crtc->dev; | 1429 | struct drm_device *dev = crtc->dev; |
| 1410 | struct drm_i915_private *dev_priv = dev->dev_private; | 1430 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1411 | u32 dpa_ctl; | 1431 | u32 dpa_ctl; |
| 1412 | 1432 | ||
| 1413 | DRM_DEBUG("eDP PLL enable for clock %d\n", clock); | 1433 | DRM_DEBUG_KMS("eDP PLL enable for clock %d\n", clock); |
| 1414 | dpa_ctl = I915_READ(DP_A); | 1434 | dpa_ctl = I915_READ(DP_A); |
| 1415 | dpa_ctl &= ~DP_PLL_FREQ_MASK; | 1435 | dpa_ctl &= ~DP_PLL_FREQ_MASK; |
| 1416 | 1436 | ||
| @@ -1440,7 +1460,7 @@ static void igdng_set_pll_edp (struct drm_crtc *crtc, int clock) | |||
| 1440 | udelay(500); | 1460 | udelay(500); |
| 1441 | } | 1461 | } |
| 1442 | 1462 | ||
| 1443 | static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | 1463 | static void ironlake_crtc_dpms(struct drm_crtc *crtc, int mode) |
| 1444 | { | 1464 | { |
| 1445 | struct drm_device *dev = crtc->dev; | 1465 | struct drm_device *dev = crtc->dev; |
| 1446 | struct drm_i915_private *dev_priv = dev->dev_private; | 1466 | struct drm_i915_private *dev_priv = dev->dev_private; |
| @@ -1481,10 +1501,19 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1481 | case DRM_MODE_DPMS_ON: | 1501 | case DRM_MODE_DPMS_ON: |
| 1482 | case DRM_MODE_DPMS_STANDBY: | 1502 | case DRM_MODE_DPMS_STANDBY: |
| 1483 | case DRM_MODE_DPMS_SUSPEND: | 1503 | case DRM_MODE_DPMS_SUSPEND: |
| 1484 | DRM_DEBUG("crtc %d dpms on\n", pipe); | 1504 | DRM_DEBUG_KMS("crtc %d dpms on\n", pipe); |
| 1505 | |||
| 1506 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) { | ||
| 1507 | temp = I915_READ(PCH_LVDS); | ||
| 1508 | if ((temp & LVDS_PORT_EN) == 0) { | ||
| 1509 | I915_WRITE(PCH_LVDS, temp | LVDS_PORT_EN); | ||
| 1510 | POSTING_READ(PCH_LVDS); | ||
| 1511 | } | ||
| 1512 | } | ||
| 1513 | |||
| 1485 | if (HAS_eDP) { | 1514 | if (HAS_eDP) { |
| 1486 | /* enable eDP PLL */ | 1515 | /* enable eDP PLL */ |
| 1487 | igdng_enable_pll_edp(crtc); | 1516 | ironlake_enable_pll_edp(crtc); |
| 1488 | } else { | 1517 | } else { |
| 1489 | /* enable PCH DPLL */ | 1518 | /* enable PCH DPLL */ |
| 1490 | temp = I915_READ(pch_dpll_reg); | 1519 | temp = I915_READ(pch_dpll_reg); |
| @@ -1501,7 +1530,7 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1501 | I915_READ(fdi_rx_reg); | 1530 | I915_READ(fdi_rx_reg); |
| 1502 | udelay(200); | 1531 | udelay(200); |
| 1503 | 1532 | ||
| 1504 | /* Enable CPU FDI TX PLL, always on for IGDNG */ | 1533 | /* Enable CPU FDI TX PLL, always on for Ironlake */ |
| 1505 | temp = I915_READ(fdi_tx_reg); | 1534 | temp = I915_READ(fdi_tx_reg); |
| 1506 | if ((temp & FDI_TX_PLL_ENABLE) == 0) { | 1535 | if ((temp & FDI_TX_PLL_ENABLE) == 0) { |
| 1507 | I915_WRITE(fdi_tx_reg, temp | FDI_TX_PLL_ENABLE); | 1536 | I915_WRITE(fdi_tx_reg, temp | FDI_TX_PLL_ENABLE); |
| @@ -1568,12 +1597,13 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1568 | udelay(150); | 1597 | udelay(150); |
| 1569 | 1598 | ||
| 1570 | temp = I915_READ(fdi_rx_iir_reg); | 1599 | temp = I915_READ(fdi_rx_iir_reg); |
| 1571 | DRM_DEBUG("FDI_RX_IIR 0x%x\n", temp); | 1600 | DRM_DEBUG_KMS("FDI_RX_IIR 0x%x\n", temp); |
| 1572 | 1601 | ||
| 1573 | if ((temp & FDI_RX_BIT_LOCK) == 0) { | 1602 | if ((temp & FDI_RX_BIT_LOCK) == 0) { |
| 1574 | for (j = 0; j < tries; j++) { | 1603 | for (j = 0; j < tries; j++) { |
| 1575 | temp = I915_READ(fdi_rx_iir_reg); | 1604 | temp = I915_READ(fdi_rx_iir_reg); |
| 1576 | DRM_DEBUG("FDI_RX_IIR 0x%x\n", temp); | 1605 | DRM_DEBUG_KMS("FDI_RX_IIR 0x%x\n", |
| 1606 | temp); | ||
| 1577 | if (temp & FDI_RX_BIT_LOCK) | 1607 | if (temp & FDI_RX_BIT_LOCK) |
| 1578 | break; | 1608 | break; |
| 1579 | udelay(200); | 1609 | udelay(200); |
| @@ -1582,11 +1612,11 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1582 | I915_WRITE(fdi_rx_iir_reg, | 1612 | I915_WRITE(fdi_rx_iir_reg, |
| 1583 | temp | FDI_RX_BIT_LOCK); | 1613 | temp | FDI_RX_BIT_LOCK); |
| 1584 | else | 1614 | else |
| 1585 | DRM_DEBUG("train 1 fail\n"); | 1615 | DRM_DEBUG_KMS("train 1 fail\n"); |
| 1586 | } else { | 1616 | } else { |
| 1587 | I915_WRITE(fdi_rx_iir_reg, | 1617 | I915_WRITE(fdi_rx_iir_reg, |
| 1588 | temp | FDI_RX_BIT_LOCK); | 1618 | temp | FDI_RX_BIT_LOCK); |
| 1589 | DRM_DEBUG("train 1 ok 2!\n"); | 1619 | DRM_DEBUG_KMS("train 1 ok 2!\n"); |
| 1590 | } | 1620 | } |
| 1591 | temp = I915_READ(fdi_tx_reg); | 1621 | temp = I915_READ(fdi_tx_reg); |
| 1592 | temp &= ~FDI_LINK_TRAIN_NONE; | 1622 | temp &= ~FDI_LINK_TRAIN_NONE; |
| @@ -1601,12 +1631,13 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1601 | udelay(150); | 1631 | udelay(150); |
| 1602 | 1632 | ||
| 1603 | temp = I915_READ(fdi_rx_iir_reg); | 1633 | temp = I915_READ(fdi_rx_iir_reg); |
| 1604 | DRM_DEBUG("FDI_RX_IIR 0x%x\n", temp); | 1634 | DRM_DEBUG_KMS("FDI_RX_IIR 0x%x\n", temp); |
| 1605 | 1635 | ||
| 1606 | if ((temp & FDI_RX_SYMBOL_LOCK) == 0) { | 1636 | if ((temp & FDI_RX_SYMBOL_LOCK) == 0) { |
| 1607 | for (j = 0; j < tries; j++) { | 1637 | for (j = 0; j < tries; j++) { |
| 1608 | temp = I915_READ(fdi_rx_iir_reg); | 1638 | temp = I915_READ(fdi_rx_iir_reg); |
| 1609 | DRM_DEBUG("FDI_RX_IIR 0x%x\n", temp); | 1639 | DRM_DEBUG_KMS("FDI_RX_IIR 0x%x\n", |
| 1640 | temp); | ||
| 1610 | if (temp & FDI_RX_SYMBOL_LOCK) | 1641 | if (temp & FDI_RX_SYMBOL_LOCK) |
| 1611 | break; | 1642 | break; |
| 1612 | udelay(200); | 1643 | udelay(200); |
| @@ -1614,15 +1645,15 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1614 | if (j != tries) { | 1645 | if (j != tries) { |
| 1615 | I915_WRITE(fdi_rx_iir_reg, | 1646 | I915_WRITE(fdi_rx_iir_reg, |
| 1616 | temp | FDI_RX_SYMBOL_LOCK); | 1647 | temp | FDI_RX_SYMBOL_LOCK); |
| 1617 | DRM_DEBUG("train 2 ok 1!\n"); | 1648 | DRM_DEBUG_KMS("train 2 ok 1!\n"); |
| 1618 | } else | 1649 | } else |
| 1619 | DRM_DEBUG("train 2 fail\n"); | 1650 | DRM_DEBUG_KMS("train 2 fail\n"); |
| 1620 | } else { | 1651 | } else { |
| 1621 | I915_WRITE(fdi_rx_iir_reg, | 1652 | I915_WRITE(fdi_rx_iir_reg, |
| 1622 | temp | FDI_RX_SYMBOL_LOCK); | 1653 | temp | FDI_RX_SYMBOL_LOCK); |
| 1623 | DRM_DEBUG("train 2 ok 2!\n"); | 1654 | DRM_DEBUG_KMS("train 2 ok 2!\n"); |
| 1624 | } | 1655 | } |
| 1625 | DRM_DEBUG("train done\n"); | 1656 | DRM_DEBUG_KMS("train done\n"); |
| 1626 | 1657 | ||
| 1627 | /* set transcoder timing */ | 1658 | /* set transcoder timing */ |
| 1628 | I915_WRITE(trans_htot_reg, I915_READ(cpu_htot_reg)); | 1659 | I915_WRITE(trans_htot_reg, I915_READ(cpu_htot_reg)); |
| @@ -1664,9 +1695,7 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1664 | 1695 | ||
| 1665 | break; | 1696 | break; |
| 1666 | case DRM_MODE_DPMS_OFF: | 1697 | case DRM_MODE_DPMS_OFF: |
| 1667 | DRM_DEBUG("crtc %d dpms off\n", pipe); | 1698 | DRM_DEBUG_KMS("crtc %d dpms off\n", pipe); |
| 1668 | |||
| 1669 | i915_disable_vga(dev); | ||
| 1670 | 1699 | ||
| 1671 | /* Disable display plane */ | 1700 | /* Disable display plane */ |
| 1672 | temp = I915_READ(dspcntr_reg); | 1701 | temp = I915_READ(dspcntr_reg); |
| @@ -1677,6 +1706,8 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1677 | I915_READ(dspbase_reg); | 1706 | I915_READ(dspbase_reg); |
| 1678 | } | 1707 | } |
| 1679 | 1708 | ||
| 1709 | i915_disable_vga(dev); | ||
| 1710 | |||
| 1680 | /* disable cpu pipe, disable after all planes disabled */ | 1711 | /* disable cpu pipe, disable after all planes disabled */ |
| 1681 | temp = I915_READ(pipeconf_reg); | 1712 | temp = I915_READ(pipeconf_reg); |
| 1682 | if ((temp & PIPEACONF_ENABLE) != 0) { | 1713 | if ((temp & PIPEACONF_ENABLE) != 0) { |
| @@ -1690,16 +1721,23 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1690 | udelay(500); | 1721 | udelay(500); |
| 1691 | continue; | 1722 | continue; |
| 1692 | } else { | 1723 | } else { |
| 1693 | DRM_DEBUG("pipe %d off delay\n", pipe); | 1724 | DRM_DEBUG_KMS("pipe %d off delay\n", |
| 1725 | pipe); | ||
| 1694 | break; | 1726 | break; |
| 1695 | } | 1727 | } |
| 1696 | } | 1728 | } |
| 1697 | } else | 1729 | } else |
| 1698 | DRM_DEBUG("crtc %d is disabled\n", pipe); | 1730 | DRM_DEBUG_KMS("crtc %d is disabled\n", pipe); |
| 1699 | 1731 | ||
| 1700 | if (HAS_eDP) { | 1732 | udelay(100); |
| 1701 | igdng_disable_pll_edp(crtc); | 1733 | |
| 1734 | /* Disable PF */ | ||
| 1735 | temp = I915_READ(pf_ctl_reg); | ||
| 1736 | if ((temp & PF_ENABLE) != 0) { | ||
| 1737 | I915_WRITE(pf_ctl_reg, temp & ~PF_ENABLE); | ||
| 1738 | I915_READ(pf_ctl_reg); | ||
| 1702 | } | 1739 | } |
| 1740 | I915_WRITE(pf_win_size, 0); | ||
| 1703 | 1741 | ||
| 1704 | /* disable CPU FDI tx and PCH FDI rx */ | 1742 | /* disable CPU FDI tx and PCH FDI rx */ |
| 1705 | temp = I915_READ(fdi_tx_reg); | 1743 | temp = I915_READ(fdi_tx_reg); |
| @@ -1725,6 +1763,13 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1725 | 1763 | ||
| 1726 | udelay(100); | 1764 | udelay(100); |
| 1727 | 1765 | ||
| 1766 | if (intel_pipe_has_type(crtc, INTEL_OUTPUT_LVDS)) { | ||
| 1767 | temp = I915_READ(PCH_LVDS); | ||
| 1768 | I915_WRITE(PCH_LVDS, temp & ~LVDS_PORT_EN); | ||
| 1769 | I915_READ(PCH_LVDS); | ||
| 1770 | udelay(100); | ||
| 1771 | } | ||
| 1772 | |||
| 1728 | /* disable PCH transcoder */ | 1773 | /* disable PCH transcoder */ |
| 1729 | temp = I915_READ(transconf_reg); | 1774 | temp = I915_READ(transconf_reg); |
| 1730 | if ((temp & TRANS_ENABLE) != 0) { | 1775 | if ((temp & TRANS_ENABLE) != 0) { |
| @@ -1738,12 +1783,15 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1738 | udelay(500); | 1783 | udelay(500); |
| 1739 | continue; | 1784 | continue; |
| 1740 | } else { | 1785 | } else { |
| 1741 | DRM_DEBUG("transcoder %d off delay\n", pipe); | 1786 | DRM_DEBUG_KMS("transcoder %d off " |
| 1787 | "delay\n", pipe); | ||
| 1742 | break; | 1788 | break; |
| 1743 | } | 1789 | } |
| 1744 | } | 1790 | } |
| 1745 | } | 1791 | } |
| 1746 | 1792 | ||
| 1793 | udelay(100); | ||
| 1794 | |||
| 1747 | /* disable PCH DPLL */ | 1795 | /* disable PCH DPLL */ |
| 1748 | temp = I915_READ(pch_dpll_reg); | 1796 | temp = I915_READ(pch_dpll_reg); |
| 1749 | if ((temp & DPLL_VCO_ENABLE) != 0) { | 1797 | if ((temp & DPLL_VCO_ENABLE) != 0) { |
| @@ -1751,14 +1799,20 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1751 | I915_READ(pch_dpll_reg); | 1799 | I915_READ(pch_dpll_reg); |
| 1752 | } | 1800 | } |
| 1753 | 1801 | ||
| 1754 | temp = I915_READ(fdi_rx_reg); | 1802 | if (HAS_eDP) { |
| 1755 | if ((temp & FDI_RX_PLL_ENABLE) != 0) { | 1803 | ironlake_disable_pll_edp(crtc); |
| 1756 | temp &= ~FDI_SEL_PCDCLK; | ||
| 1757 | temp &= ~FDI_RX_PLL_ENABLE; | ||
| 1758 | I915_WRITE(fdi_rx_reg, temp); | ||
| 1759 | I915_READ(fdi_rx_reg); | ||
| 1760 | } | 1804 | } |
| 1761 | 1805 | ||
| 1806 | temp = I915_READ(fdi_rx_reg); | ||
| 1807 | temp &= ~FDI_SEL_PCDCLK; | ||
| 1808 | I915_WRITE(fdi_rx_reg, temp); | ||
| 1809 | I915_READ(fdi_rx_reg); | ||
| 1810 | |||
| 1811 | temp = I915_READ(fdi_rx_reg); | ||
| 1812 | temp &= ~FDI_RX_PLL_ENABLE; | ||
| 1813 | I915_WRITE(fdi_rx_reg, temp); | ||
| 1814 | I915_READ(fdi_rx_reg); | ||
| 1815 | |||
| 1762 | /* Disable CPU FDI TX PLL */ | 1816 | /* Disable CPU FDI TX PLL */ |
| 1763 | temp = I915_READ(fdi_tx_reg); | 1817 | temp = I915_READ(fdi_tx_reg); |
| 1764 | if ((temp & FDI_TX_PLL_ENABLE) != 0) { | 1818 | if ((temp & FDI_TX_PLL_ENABLE) != 0) { |
| @@ -1767,20 +1821,43 @@ static void igdng_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1767 | udelay(100); | 1821 | udelay(100); |
| 1768 | } | 1822 | } |
| 1769 | 1823 | ||
| 1770 | /* Disable PF */ | ||
| 1771 | temp = I915_READ(pf_ctl_reg); | ||
| 1772 | if ((temp & PF_ENABLE) != 0) { | ||
| 1773 | I915_WRITE(pf_ctl_reg, temp & ~PF_ENABLE); | ||
| 1774 | I915_READ(pf_ctl_reg); | ||
| 1775 | } | ||
| 1776 | I915_WRITE(pf_win_size, 0); | ||
| 1777 | |||
| 1778 | /* Wait for the clocks to turn off. */ | 1824 | /* Wait for the clocks to turn off. */ |
| 1779 | udelay(150); | 1825 | udelay(100); |
| 1780 | break; | 1826 | break; |
| 1781 | } | 1827 | } |
| 1782 | } | 1828 | } |
| 1783 | 1829 | ||
| 1830 | static void intel_crtc_dpms_overlay(struct intel_crtc *intel_crtc, bool enable) | ||
| 1831 | { | ||
| 1832 | struct intel_overlay *overlay; | ||
| 1833 | int ret; | ||
| 1834 | |||
| 1835 | if (!enable && intel_crtc->overlay) { | ||
| 1836 | overlay = intel_crtc->overlay; | ||
| 1837 | mutex_lock(&overlay->dev->struct_mutex); | ||
| 1838 | for (;;) { | ||
| 1839 | ret = intel_overlay_switch_off(overlay); | ||
| 1840 | if (ret == 0) | ||
| 1841 | break; | ||
| 1842 | |||
| 1843 | ret = intel_overlay_recover_from_interrupt(overlay, 0); | ||
| 1844 | if (ret != 0) { | ||
| 1845 | /* overlay doesn't react anymore. Usually | ||
| 1846 | * results in a black screen and an unkillable | ||
| 1847 | * X server. */ | ||
| 1848 | BUG(); | ||
| 1849 | overlay->hw_wedged = HW_WEDGED; | ||
| 1850 | break; | ||
| 1851 | } | ||
| 1852 | } | ||
| 1853 | mutex_unlock(&overlay->dev->struct_mutex); | ||
| 1854 | } | ||
| 1855 | /* Let userspace switch the overlay on again. In most cases userspace | ||
| 1856 | * has to recompute where to put it anyway. */ | ||
| 1857 | |||
| 1858 | return; | ||
| 1859 | } | ||
| 1860 | |||
| 1784 | static void i9xx_crtc_dpms(struct drm_crtc *crtc, int mode) | 1861 | static void i9xx_crtc_dpms(struct drm_crtc *crtc, int mode) |
| 1785 | { | 1862 | { |
| 1786 | struct drm_device *dev = crtc->dev; | 1863 | struct drm_device *dev = crtc->dev; |
| @@ -1839,12 +1916,14 @@ static void i9xx_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 1839 | intel_update_fbc(crtc, &crtc->mode); | 1916 | intel_update_fbc(crtc, &crtc->mode); |
| 1840 | 1917 | ||
| 1841 | /* Give the overlay scaler a chance to enable if it's on this pipe */ | 1918 | /* Give the overlay scaler a chance to enable if it's on this pipe */ |
| 1842 | //intel_crtc_dpms_video(crtc, true); TODO | 1919 | intel_crtc_dpms_overlay(intel_crtc, true); |
| 1843 | break; | 1920 | break; |
| 1844 | case DRM_MODE_DPMS_OFF: | 1921 | case DRM_MODE_DPMS_OFF: |
| 1845 | intel_update_watermarks(dev); | 1922 | intel_update_watermarks(dev); |
| 1923 | |||
| 1846 | /* Give the overlay scaler a chance to disable if it's on this pipe */ | 1924 | /* Give the overlay scaler a chance to disable if it's on this pipe */ |
| 1847 | //intel_crtc_dpms_video(crtc, FALSE); TODO | 1925 | intel_crtc_dpms_overlay(intel_crtc, false); |
| 1926 | drm_vblank_off(dev, pipe); | ||
| 1848 | 1927 | ||
| 1849 | if (dev_priv->cfb_plane == plane && | 1928 | if (dev_priv->cfb_plane == plane && |
| 1850 | dev_priv->display.disable_fbc) | 1929 | dev_priv->display.disable_fbc) |
| @@ -1963,7 +2042,7 @@ static bool intel_crtc_mode_fixup(struct drm_crtc *crtc, | |||
| 1963 | struct drm_display_mode *adjusted_mode) | 2042 | struct drm_display_mode *adjusted_mode) |
| 1964 | { | 2043 | { |
| 1965 | struct drm_device *dev = crtc->dev; | 2044 | struct drm_device *dev = crtc->dev; |
| 1966 | if (IS_IGDNG(dev)) { | 2045 | if (IS_IRONLAKE(dev)) { |
| 1967 | /* FDI link clock is fixed at 2.7G */ | 2046 | /* FDI link clock is fixed at 2.7G */ |
| 1968 | if (mode->clock * 3 > 27000 * 4) | 2047 | if (mode->clock * 3 > 27000 * 4) |
| 1969 | return MODE_CLOCK_HIGH; | 2048 | return MODE_CLOCK_HIGH; |
| @@ -2039,7 +2118,7 @@ static int i830_get_display_clock_speed(struct drm_device *dev) | |||
| 2039 | * Return the pipe currently connected to the panel fitter, | 2118 | * Return the pipe currently connected to the panel fitter, |
| 2040 | * or -1 if the panel fitter is not present or not in use | 2119 | * or -1 if the panel fitter is not present or not in use |
| 2041 | */ | 2120 | */ |
| 2042 | static int intel_panel_fitter_pipe (struct drm_device *dev) | 2121 | int intel_panel_fitter_pipe (struct drm_device *dev) |
| 2043 | { | 2122 | { |
| 2044 | struct drm_i915_private *dev_priv = dev->dev_private; | 2123 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 2045 | u32 pfit_control; | 2124 | u32 pfit_control; |
| @@ -2083,9 +2162,8 @@ fdi_reduce_ratio(u32 *num, u32 *den) | |||
| 2083 | #define LINK_N 0x80000 | 2162 | #define LINK_N 0x80000 |
| 2084 | 2163 | ||
| 2085 | static void | 2164 | static void |
| 2086 | igdng_compute_m_n(int bits_per_pixel, int nlanes, | 2165 | ironlake_compute_m_n(int bits_per_pixel, int nlanes, int pixel_clock, |
| 2087 | int pixel_clock, int link_clock, | 2166 | int link_clock, struct fdi_m_n *m_n) |
| 2088 | struct fdi_m_n *m_n) | ||
| 2089 | { | 2167 | { |
| 2090 | u64 temp; | 2168 | u64 temp; |
| 2091 | 2169 | ||
| @@ -2113,34 +2191,34 @@ struct intel_watermark_params { | |||
| 2113 | unsigned long cacheline_size; | 2191 | unsigned long cacheline_size; |
| 2114 | }; | 2192 | }; |
| 2115 | 2193 | ||
| 2116 | /* IGD has different values for various configs */ | 2194 | /* Pineview has different values for various configs */ |
| 2117 | static struct intel_watermark_params igd_display_wm = { | 2195 | static struct intel_watermark_params pineview_display_wm = { |
| 2118 | IGD_DISPLAY_FIFO, | 2196 | PINEVIEW_DISPLAY_FIFO, |
| 2119 | IGD_MAX_WM, | 2197 | PINEVIEW_MAX_WM, |
| 2120 | IGD_DFT_WM, | 2198 | PINEVIEW_DFT_WM, |
| 2121 | IGD_GUARD_WM, | 2199 | PINEVIEW_GUARD_WM, |
| 2122 | IGD_FIFO_LINE_SIZE | 2200 | PINEVIEW_FIFO_LINE_SIZE |
| 2123 | }; | 2201 | }; |
| 2124 | static struct intel_watermark_params igd_display_hplloff_wm = { | 2202 | static struct intel_watermark_params pineview_display_hplloff_wm = { |
| 2125 | IGD_DISPLAY_FIFO, | 2203 | PINEVIEW_DISPLAY_FIFO, |
| 2126 | IGD_MAX_WM, | 2204 | PINEVIEW_MAX_WM, |
| 2127 | IGD_DFT_HPLLOFF_WM, | 2205 | PINEVIEW_DFT_HPLLOFF_WM, |
| 2128 | IGD_GUARD_WM, | 2206 | PINEVIEW_GUARD_WM, |
| 2129 | IGD_FIFO_LINE_SIZE | 2207 | PINEVIEW_FIFO_LINE_SIZE |
| 2130 | }; | 2208 | }; |
| 2131 | static struct intel_watermark_params igd_cursor_wm = { | 2209 | static struct intel_watermark_params pineview_cursor_wm = { |
| 2132 | IGD_CURSOR_FIFO, | 2210 | PINEVIEW_CURSOR_FIFO, |
| 2133 | IGD_CURSOR_MAX_WM, | 2211 | PINEVIEW_CURSOR_MAX_WM, |
| 2134 | IGD_CURSOR_DFT_WM, | 2212 | PINEVIEW_CURSOR_DFT_WM, |
| 2135 | IGD_CURSOR_GUARD_WM, | 2213 | PINEVIEW_CURSOR_GUARD_WM, |
| 2136 | IGD_FIFO_LINE_SIZE, | 2214 | PINEVIEW_FIFO_LINE_SIZE, |
| 2137 | }; | 2215 | }; |
| 2138 | static struct intel_watermark_params igd_cursor_hplloff_wm = { | 2216 | static struct intel_watermark_params pineview_cursor_hplloff_wm = { |
| 2139 | IGD_CURSOR_FIFO, | 2217 | PINEVIEW_CURSOR_FIFO, |
| 2140 | IGD_CURSOR_MAX_WM, | 2218 | PINEVIEW_CURSOR_MAX_WM, |
| 2141 | IGD_CURSOR_DFT_WM, | 2219 | PINEVIEW_CURSOR_DFT_WM, |
| 2142 | IGD_CURSOR_GUARD_WM, | 2220 | PINEVIEW_CURSOR_GUARD_WM, |
| 2143 | IGD_FIFO_LINE_SIZE | 2221 | PINEVIEW_FIFO_LINE_SIZE |
| 2144 | }; | 2222 | }; |
| 2145 | static struct intel_watermark_params g4x_wm_info = { | 2223 | static struct intel_watermark_params g4x_wm_info = { |
| 2146 | G4X_FIFO_SIZE, | 2224 | G4X_FIFO_SIZE, |
| @@ -2213,11 +2291,11 @@ static unsigned long intel_calculate_wm(unsigned long clock_in_khz, | |||
| 2213 | 1000; | 2291 | 1000; |
| 2214 | entries_required /= wm->cacheline_size; | 2292 | entries_required /= wm->cacheline_size; |
| 2215 | 2293 | ||
| 2216 | DRM_DEBUG("FIFO entries required for mode: %d\n", entries_required); | 2294 | DRM_DEBUG_KMS("FIFO entries required for mode: %d\n", entries_required); |
| 2217 | 2295 | ||
| 2218 | wm_size = wm->fifo_size - (entries_required + wm->guard_size); | 2296 | wm_size = wm->fifo_size - (entries_required + wm->guard_size); |
| 2219 | 2297 | ||
| 2220 | DRM_DEBUG("FIFO watermark level: %d\n", wm_size); | 2298 | DRM_DEBUG_KMS("FIFO watermark level: %d\n", wm_size); |
| 2221 | 2299 | ||
| 2222 | /* Don't promote wm_size to unsigned... */ | 2300 | /* Don't promote wm_size to unsigned... */ |
| 2223 | if (wm_size > (long)wm->max_wm) | 2301 | if (wm_size > (long)wm->max_wm) |
| @@ -2279,50 +2357,50 @@ static struct cxsr_latency *intel_get_cxsr_latency(int is_desktop, int fsb, | |||
| 2279 | return latency; | 2357 | return latency; |
| 2280 | } | 2358 | } |
| 2281 | 2359 | ||
| 2282 | DRM_DEBUG("Unknown FSB/MEM found, disable CxSR\n"); | 2360 | DRM_DEBUG_KMS("Unknown FSB/MEM found, disable CxSR\n"); |
| 2283 | 2361 | ||
| 2284 | return NULL; | 2362 | return NULL; |
| 2285 | } | 2363 | } |
| 2286 | 2364 | ||
| 2287 | static void igd_disable_cxsr(struct drm_device *dev) | 2365 | static void pineview_disable_cxsr(struct drm_device *dev) |
| 2288 | { | 2366 | { |
| 2289 | struct drm_i915_private *dev_priv = dev->dev_private; | 2367 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 2290 | u32 reg; | 2368 | u32 reg; |
| 2291 | 2369 | ||
| 2292 | /* deactivate cxsr */ | 2370 | /* deactivate cxsr */ |
| 2293 | reg = I915_READ(DSPFW3); | 2371 | reg = I915_READ(DSPFW3); |
| 2294 | reg &= ~(IGD_SELF_REFRESH_EN); | 2372 | reg &= ~(PINEVIEW_SELF_REFRESH_EN); |
| 2295 | I915_WRITE(DSPFW3, reg); | 2373 | I915_WRITE(DSPFW3, reg); |
| 2296 | DRM_INFO("Big FIFO is disabled\n"); | 2374 | DRM_INFO("Big FIFO is disabled\n"); |
| 2297 | } | 2375 | } |
| 2298 | 2376 | ||
| 2299 | static void igd_enable_cxsr(struct drm_device *dev, unsigned long clock, | 2377 | static void pineview_enable_cxsr(struct drm_device *dev, unsigned long clock, |
| 2300 | int pixel_size) | 2378 | int pixel_size) |
| 2301 | { | 2379 | { |
| 2302 | struct drm_i915_private *dev_priv = dev->dev_private; | 2380 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 2303 | u32 reg; | 2381 | u32 reg; |
| 2304 | unsigned long wm; | 2382 | unsigned long wm; |
| 2305 | struct cxsr_latency *latency; | 2383 | struct cxsr_latency *latency; |
| 2306 | 2384 | ||
| 2307 | latency = intel_get_cxsr_latency(IS_IGDG(dev), dev_priv->fsb_freq, | 2385 | latency = intel_get_cxsr_latency(IS_PINEVIEW_G(dev), dev_priv->fsb_freq, |
| 2308 | dev_priv->mem_freq); | 2386 | dev_priv->mem_freq); |
| 2309 | if (!latency) { | 2387 | if (!latency) { |
| 2310 | DRM_DEBUG("Unknown FSB/MEM found, disable CxSR\n"); | 2388 | DRM_DEBUG_KMS("Unknown FSB/MEM found, disable CxSR\n"); |
| 2311 | igd_disable_cxsr(dev); | 2389 | pineview_disable_cxsr(dev); |
| 2312 | return; | 2390 | return; |
| 2313 | } | 2391 | } |
| 2314 | 2392 | ||
| 2315 | /* Display SR */ | 2393 | /* Display SR */ |
| 2316 | wm = intel_calculate_wm(clock, &igd_display_wm, pixel_size, | 2394 | wm = intel_calculate_wm(clock, &pineview_display_wm, pixel_size, |
| 2317 | latency->display_sr); | 2395 | latency->display_sr); |
| 2318 | reg = I915_READ(DSPFW1); | 2396 | reg = I915_READ(DSPFW1); |
| 2319 | reg &= 0x7fffff; | 2397 | reg &= 0x7fffff; |
| 2320 | reg |= wm << 23; | 2398 | reg |= wm << 23; |
| 2321 | I915_WRITE(DSPFW1, reg); | 2399 | I915_WRITE(DSPFW1, reg); |
| 2322 | DRM_DEBUG("DSPFW1 register is %x\n", reg); | 2400 | DRM_DEBUG_KMS("DSPFW1 register is %x\n", reg); |
| 2323 | 2401 | ||
| 2324 | /* cursor SR */ | 2402 | /* cursor SR */ |
| 2325 | wm = intel_calculate_wm(clock, &igd_cursor_wm, pixel_size, | 2403 | wm = intel_calculate_wm(clock, &pineview_cursor_wm, pixel_size, |
| 2326 | latency->cursor_sr); | 2404 | latency->cursor_sr); |
| 2327 | reg = I915_READ(DSPFW3); | 2405 | reg = I915_READ(DSPFW3); |
| 2328 | reg &= ~(0x3f << 24); | 2406 | reg &= ~(0x3f << 24); |
| @@ -2330,7 +2408,7 @@ static void igd_enable_cxsr(struct drm_device *dev, unsigned long clock, | |||
| 2330 | I915_WRITE(DSPFW3, reg); | 2408 | I915_WRITE(DSPFW3, reg); |
| 2331 | 2409 | ||
| 2332 | /* Display HPLL off SR */ | 2410 | /* Display HPLL off SR */ |
| 2333 | wm = intel_calculate_wm(clock, &igd_display_hplloff_wm, | 2411 | wm = intel_calculate_wm(clock, &pineview_display_hplloff_wm, |
| 2334 | latency->display_hpll_disable, I915_FIFO_LINE_SIZE); | 2412 | latency->display_hpll_disable, I915_FIFO_LINE_SIZE); |
| 2335 | reg = I915_READ(DSPFW3); | 2413 | reg = I915_READ(DSPFW3); |
| 2336 | reg &= 0xfffffe00; | 2414 | reg &= 0xfffffe00; |
| @@ -2338,17 +2416,17 @@ static void igd_enable_cxsr(struct drm_device *dev, unsigned long clock, | |||
| 2338 | I915_WRITE(DSPFW3, reg); | 2416 | I915_WRITE(DSPFW3, reg); |
| 2339 | 2417 | ||
| 2340 | /* cursor HPLL off SR */ | 2418 | /* cursor HPLL off SR */ |
| 2341 | wm = intel_calculate_wm(clock, &igd_cursor_hplloff_wm, pixel_size, | 2419 | wm = intel_calculate_wm(clock, &pineview_cursor_hplloff_wm, pixel_size, |
| 2342 | latency->cursor_hpll_disable); | 2420 | latency->cursor_hpll_disable); |
| 2343 | reg = I915_READ(DSPFW3); | 2421 | reg = I915_READ(DSPFW3); |
| 2344 | reg &= ~(0x3f << 16); | 2422 | reg &= ~(0x3f << 16); |
| 2345 | reg |= (wm & 0x3f) << 16; | 2423 | reg |= (wm & 0x3f) << 16; |
| 2346 | I915_WRITE(DSPFW3, reg); | 2424 | I915_WRITE(DSPFW3, reg); |
| 2347 | DRM_DEBUG("DSPFW3 register is %x\n", reg); | 2425 | DRM_DEBUG_KMS("DSPFW3 register is %x\n", reg); |
| 2348 | 2426 | ||
| 2349 | /* activate cxsr */ | 2427 | /* activate cxsr */ |
| 2350 | reg = I915_READ(DSPFW3); | 2428 | reg = I915_READ(DSPFW3); |
| 2351 | reg |= IGD_SELF_REFRESH_EN; | 2429 | reg |= PINEVIEW_SELF_REFRESH_EN; |
| 2352 | I915_WRITE(DSPFW3, reg); | 2430 | I915_WRITE(DSPFW3, reg); |
| 2353 | 2431 | ||
| 2354 | DRM_INFO("Big FIFO is enabled\n"); | 2432 | DRM_INFO("Big FIFO is enabled\n"); |
| @@ -2384,8 +2462,8 @@ static int i9xx_get_fifo_size(struct drm_device *dev, int plane) | |||
| 2384 | size = ((dsparb >> DSPARB_CSTART_SHIFT) & 0x7f) - | 2462 | size = ((dsparb >> DSPARB_CSTART_SHIFT) & 0x7f) - |
| 2385 | (dsparb & 0x7f); | 2463 | (dsparb & 0x7f); |
| 2386 | 2464 | ||
| 2387 | DRM_DEBUG("FIFO size - (0x%08x) %s: %d\n", dsparb, plane ? "B" : "A", | 2465 | DRM_DEBUG_KMS("FIFO size - (0x%08x) %s: %d\n", dsparb, |
| 2388 | size); | 2466 | plane ? "B" : "A", size); |
| 2389 | 2467 | ||
| 2390 | return size; | 2468 | return size; |
| 2391 | } | 2469 | } |
| @@ -2403,8 +2481,8 @@ static int i85x_get_fifo_size(struct drm_device *dev, int plane) | |||
| 2403 | (dsparb & 0x1ff); | 2481 | (dsparb & 0x1ff); |
| 2404 | size >>= 1; /* Convert to cachelines */ | 2482 | size >>= 1; /* Convert to cachelines */ |
| 2405 | 2483 | ||
| 2406 | DRM_DEBUG("FIFO size - (0x%08x) %s: %d\n", dsparb, plane ? "B" : "A", | 2484 | DRM_DEBUG_KMS("FIFO size - (0x%08x) %s: %d\n", dsparb, |
| 2407 | size); | 2485 | plane ? "B" : "A", size); |
| 2408 | 2486 | ||
| 2409 | return size; | 2487 | return size; |
| 2410 | } | 2488 | } |
| @@ -2418,7 +2496,8 @@ static int i845_get_fifo_size(struct drm_device *dev, int plane) | |||
| 2418 | size = dsparb & 0x7f; | 2496 | size = dsparb & 0x7f; |
| 2419 | size >>= 2; /* Convert to cachelines */ | 2497 | size >>= 2; /* Convert to cachelines */ |
| 2420 | 2498 | ||
| 2421 | DRM_DEBUG("FIFO size - (0x%08x) %s: %d\n", dsparb, plane ? "B" : "A", | 2499 | DRM_DEBUG_KMS("FIFO size - (0x%08x) %s: %d\n", dsparb, |
| 2500 | plane ? "B" : "A", | ||
| 2422 | size); | 2501 | size); |
| 2423 | 2502 | ||
| 2424 | return size; | 2503 | return size; |
| @@ -2433,8 +2512,8 @@ static int i830_get_fifo_size(struct drm_device *dev, int plane) | |||
| 2433 | size = dsparb & 0x7f; | 2512 | size = dsparb & 0x7f; |
| 2434 | size >>= 1; /* Convert to cachelines */ | 2513 | size >>= 1; /* Convert to cachelines */ |
| 2435 | 2514 | ||
| 2436 | DRM_DEBUG("FIFO size - (0x%08x) %s: %d\n", dsparb, plane ? "B" : "A", | 2515 | DRM_DEBUG_KMS("FIFO size - (0x%08x) %s: %d\n", dsparb, |
| 2437 | size); | 2516 | plane ? "B" : "A", size); |
| 2438 | 2517 | ||
| 2439 | return size; | 2518 | return size; |
| 2440 | } | 2519 | } |
| @@ -2509,15 +2588,39 @@ static void g4x_update_wm(struct drm_device *dev, int planea_clock, | |||
| 2509 | (cursor_sr << DSPFW_CURSOR_SR_SHIFT)); | 2588 | (cursor_sr << DSPFW_CURSOR_SR_SHIFT)); |
| 2510 | } | 2589 | } |
| 2511 | 2590 | ||
| 2512 | static void i965_update_wm(struct drm_device *dev, int unused, int unused2, | 2591 | static void i965_update_wm(struct drm_device *dev, int planea_clock, |
| 2513 | int unused3, int unused4) | 2592 | int planeb_clock, int sr_hdisplay, int pixel_size) |
| 2514 | { | 2593 | { |
| 2515 | struct drm_i915_private *dev_priv = dev->dev_private; | 2594 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 2595 | unsigned long line_time_us; | ||
| 2596 | int sr_clock, sr_entries, srwm = 1; | ||
| 2597 | |||
| 2598 | /* Calc sr entries for one plane configs */ | ||
| 2599 | if (sr_hdisplay && (!planea_clock || !planeb_clock)) { | ||
| 2600 | /* self-refresh has much higher latency */ | ||
| 2601 | const static int sr_latency_ns = 12000; | ||
| 2602 | |||
| 2603 | sr_clock = planea_clock ? planea_clock : planeb_clock; | ||
| 2604 | line_time_us = ((sr_hdisplay * 1000) / sr_clock); | ||
| 2605 | |||
| 2606 | /* Use ns/us then divide to preserve precision */ | ||
| 2607 | sr_entries = (((sr_latency_ns / line_time_us) + 1) * | ||
| 2608 | pixel_size * sr_hdisplay) / 1000; | ||
| 2609 | sr_entries = roundup(sr_entries / I915_FIFO_LINE_SIZE, 1); | ||
| 2610 | DRM_DEBUG("self-refresh entries: %d\n", sr_entries); | ||
| 2611 | srwm = I945_FIFO_SIZE - sr_entries; | ||
| 2612 | if (srwm < 0) | ||
| 2613 | srwm = 1; | ||
| 2614 | srwm &= 0x3f; | ||
| 2615 | I915_WRITE(FW_BLC_SELF, FW_BLC_SELF_EN); | ||
| 2616 | } | ||
| 2516 | 2617 | ||
| 2517 | DRM_DEBUG("Setting FIFO watermarks - A: 8, B: 8, C: 8, SR 8\n"); | 2618 | DRM_DEBUG_KMS("Setting FIFO watermarks - A: 8, B: 8, C: 8, SR %d\n", |
| 2619 | srwm); | ||
| 2518 | 2620 | ||
| 2519 | /* 965 has limitations... */ | 2621 | /* 965 has limitations... */ |
| 2520 | I915_WRITE(DSPFW1, (8 << 16) | (8 << 8) | (8 << 0)); | 2622 | I915_WRITE(DSPFW1, (srwm << DSPFW_SR_SHIFT) | (8 << 16) | (8 << 8) | |
| 2623 | (8 << 0)); | ||
| 2521 | I915_WRITE(DSPFW2, (8 << 8) | (8 << 0)); | 2624 | I915_WRITE(DSPFW2, (8 << 8) | (8 << 0)); |
| 2522 | } | 2625 | } |
| 2523 | 2626 | ||
| @@ -2553,7 +2656,7 @@ static void i9xx_update_wm(struct drm_device *dev, int planea_clock, | |||
| 2553 | pixel_size, latency_ns); | 2656 | pixel_size, latency_ns); |
| 2554 | planeb_wm = intel_calculate_wm(planeb_clock, &planeb_params, | 2657 | planeb_wm = intel_calculate_wm(planeb_clock, &planeb_params, |
| 2555 | pixel_size, latency_ns); | 2658 | pixel_size, latency_ns); |
| 2556 | DRM_DEBUG("FIFO watermarks - A: %d, B: %d\n", planea_wm, planeb_wm); | 2659 | DRM_DEBUG_KMS("FIFO watermarks - A: %d, B: %d\n", planea_wm, planeb_wm); |
| 2557 | 2660 | ||
| 2558 | /* | 2661 | /* |
| 2559 | * Overlay gets an aggressive default since video jitter is bad. | 2662 | * Overlay gets an aggressive default since video jitter is bad. |
| @@ -2573,14 +2676,14 @@ static void i9xx_update_wm(struct drm_device *dev, int planea_clock, | |||
| 2573 | sr_entries = (((sr_latency_ns / line_time_us) + 1) * | 2676 | sr_entries = (((sr_latency_ns / line_time_us) + 1) * |
| 2574 | pixel_size * sr_hdisplay) / 1000; | 2677 | pixel_size * sr_hdisplay) / 1000; |
| 2575 | sr_entries = roundup(sr_entries / cacheline_size, 1); | 2678 | sr_entries = roundup(sr_entries / cacheline_size, 1); |
| 2576 | DRM_DEBUG("self-refresh entries: %d\n", sr_entries); | 2679 | DRM_DEBUG_KMS("self-refresh entries: %d\n", sr_entries); |
| 2577 | srwm = total_size - sr_entries; | 2680 | srwm = total_size - sr_entries; |
| 2578 | if (srwm < 0) | 2681 | if (srwm < 0) |
| 2579 | srwm = 1; | 2682 | srwm = 1; |
| 2580 | I915_WRITE(FW_BLC_SELF, FW_BLC_SELF_EN | (srwm & 0x3f)); | 2683 | I915_WRITE(FW_BLC_SELF, FW_BLC_SELF_EN | (srwm & 0x3f)); |
| 2581 | } | 2684 | } |
| 2582 | 2685 | ||
| 2583 | DRM_DEBUG("Setting FIFO watermarks - A: %d, B: %d, C: %d, SR %d\n", | 2686 | DRM_DEBUG_KMS("Setting FIFO watermarks - A: %d, B: %d, C: %d, SR %d\n", |
| 2584 | planea_wm, planeb_wm, cwm, srwm); | 2687 | planea_wm, planeb_wm, cwm, srwm); |
| 2585 | 2688 | ||
| 2586 | fwater_lo = ((planeb_wm & 0x3f) << 16) | (planea_wm & 0x3f); | 2689 | fwater_lo = ((planeb_wm & 0x3f) << 16) | (planea_wm & 0x3f); |
| @@ -2607,7 +2710,7 @@ static void i830_update_wm(struct drm_device *dev, int planea_clock, int unused, | |||
| 2607 | pixel_size, latency_ns); | 2710 | pixel_size, latency_ns); |
| 2608 | fwater_lo |= (3<<8) | planea_wm; | 2711 | fwater_lo |= (3<<8) | planea_wm; |
| 2609 | 2712 | ||
| 2610 | DRM_DEBUG("Setting FIFO watermarks - A: %d\n", planea_wm); | 2713 | DRM_DEBUG_KMS("Setting FIFO watermarks - A: %d\n", planea_wm); |
| 2611 | 2714 | ||
| 2612 | I915_WRITE(FW_BLC, fwater_lo); | 2715 | I915_WRITE(FW_BLC, fwater_lo); |
| 2613 | } | 2716 | } |
| @@ -2661,11 +2764,11 @@ static void intel_update_watermarks(struct drm_device *dev) | |||
| 2661 | if (crtc->enabled) { | 2764 | if (crtc->enabled) { |
| 2662 | enabled++; | 2765 | enabled++; |
| 2663 | if (intel_crtc->plane == 0) { | 2766 | if (intel_crtc->plane == 0) { |
| 2664 | DRM_DEBUG("plane A (pipe %d) clock: %d\n", | 2767 | DRM_DEBUG_KMS("plane A (pipe %d) clock: %d\n", |
| 2665 | intel_crtc->pipe, crtc->mode.clock); | 2768 | intel_crtc->pipe, crtc->mode.clock); |
| 2666 | planea_clock = crtc->mode.clock; | 2769 | planea_clock = crtc->mode.clock; |
| 2667 | } else { | 2770 | } else { |
| 2668 | DRM_DEBUG("plane B (pipe %d) clock: %d\n", | 2771 | DRM_DEBUG_KMS("plane B (pipe %d) clock: %d\n", |
| 2669 | intel_crtc->pipe, crtc->mode.clock); | 2772 | intel_crtc->pipe, crtc->mode.clock); |
| 2670 | planeb_clock = crtc->mode.clock; | 2773 | planeb_clock = crtc->mode.clock; |
| 2671 | } | 2774 | } |
| @@ -2682,10 +2785,10 @@ static void intel_update_watermarks(struct drm_device *dev) | |||
| 2682 | return; | 2785 | return; |
| 2683 | 2786 | ||
| 2684 | /* Single plane configs can enable self refresh */ | 2787 | /* Single plane configs can enable self refresh */ |
| 2685 | if (enabled == 1 && IS_IGD(dev)) | 2788 | if (enabled == 1 && IS_PINEVIEW(dev)) |
| 2686 | igd_enable_cxsr(dev, sr_clock, pixel_size); | 2789 | pineview_enable_cxsr(dev, sr_clock, pixel_size); |
| 2687 | else if (IS_IGD(dev)) | 2790 | else if (IS_PINEVIEW(dev)) |
| 2688 | igd_disable_cxsr(dev); | 2791 | pineview_disable_cxsr(dev); |
| 2689 | 2792 | ||
| 2690 | dev_priv->display.update_wm(dev, planea_clock, planeb_clock, | 2793 | dev_priv->display.update_wm(dev, planea_clock, planeb_clock, |
| 2691 | sr_hdisplay, pixel_size); | 2794 | sr_hdisplay, pixel_size); |
| @@ -2779,10 +2882,11 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2779 | 2882 | ||
| 2780 | if (is_lvds && dev_priv->lvds_use_ssc && num_outputs < 2) { | 2883 | if (is_lvds && dev_priv->lvds_use_ssc && num_outputs < 2) { |
| 2781 | refclk = dev_priv->lvds_ssc_freq * 1000; | 2884 | refclk = dev_priv->lvds_ssc_freq * 1000; |
| 2782 | DRM_DEBUG("using SSC reference clock of %d MHz\n", refclk / 1000); | 2885 | DRM_DEBUG_KMS("using SSC reference clock of %d MHz\n", |
| 2886 | refclk / 1000); | ||
| 2783 | } else if (IS_I9XX(dev)) { | 2887 | } else if (IS_I9XX(dev)) { |
| 2784 | refclk = 96000; | 2888 | refclk = 96000; |
| 2785 | if (IS_IGDNG(dev)) | 2889 | if (IS_IRONLAKE(dev)) |
| 2786 | refclk = 120000; /* 120Mhz refclk */ | 2890 | refclk = 120000; /* 120Mhz refclk */ |
| 2787 | } else { | 2891 | } else { |
| 2788 | refclk = 48000; | 2892 | refclk = 48000; |
| @@ -2802,14 +2906,25 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2802 | return -EINVAL; | 2906 | return -EINVAL; |
| 2803 | } | 2907 | } |
| 2804 | 2908 | ||
| 2805 | if (limit->find_reduced_pll && dev_priv->lvds_downclock_avail) { | 2909 | if (is_lvds && limit->find_reduced_pll && |
| 2910 | dev_priv->lvds_downclock_avail) { | ||
| 2806 | memcpy(&reduced_clock, &clock, sizeof(intel_clock_t)); | 2911 | memcpy(&reduced_clock, &clock, sizeof(intel_clock_t)); |
| 2807 | has_reduced_clock = limit->find_reduced_pll(limit, crtc, | 2912 | has_reduced_clock = limit->find_reduced_pll(limit, crtc, |
| 2808 | (adjusted_mode->clock*3/4), | 2913 | dev_priv->lvds_downclock, |
| 2809 | refclk, | 2914 | refclk, |
| 2810 | &reduced_clock); | 2915 | &reduced_clock); |
| 2916 | if (has_reduced_clock && (clock.p != reduced_clock.p)) { | ||
| 2917 | /* | ||
| 2918 | * If the different P is found, it means that we can't | ||
| 2919 | * switch the display clock by using the FP0/FP1. | ||
| 2920 | * In such case we will disable the LVDS downclock | ||
| 2921 | * feature. | ||
| 2922 | */ | ||
| 2923 | DRM_DEBUG_KMS("Different P is found for " | ||
| 2924 | "LVDS clock/downclock\n"); | ||
| 2925 | has_reduced_clock = 0; | ||
| 2926 | } | ||
| 2811 | } | 2927 | } |
| 2812 | |||
| 2813 | /* SDVO TV has fixed PLL values depend on its clock range, | 2928 | /* SDVO TV has fixed PLL values depend on its clock range, |
| 2814 | this mirrors vbios setting. */ | 2929 | this mirrors vbios setting. */ |
| 2815 | if (is_sdvo && is_tv) { | 2930 | if (is_sdvo && is_tv) { |
| @@ -2831,7 +2946,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2831 | } | 2946 | } |
| 2832 | 2947 | ||
| 2833 | /* FDI link */ | 2948 | /* FDI link */ |
| 2834 | if (IS_IGDNG(dev)) { | 2949 | if (IS_IRONLAKE(dev)) { |
| 2835 | int lane, link_bw, bpp; | 2950 | int lane, link_bw, bpp; |
| 2836 | /* eDP doesn't require FDI link, so just set DP M/N | 2951 | /* eDP doesn't require FDI link, so just set DP M/N |
| 2837 | according to current link config */ | 2952 | according to current link config */ |
| @@ -2873,8 +2988,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2873 | bpp = 24; | 2988 | bpp = 24; |
| 2874 | } | 2989 | } |
| 2875 | 2990 | ||
| 2876 | igdng_compute_m_n(bpp, lane, target_clock, | 2991 | ironlake_compute_m_n(bpp, lane, target_clock, link_bw, &m_n); |
| 2877 | link_bw, &m_n); | ||
| 2878 | } | 2992 | } |
| 2879 | 2993 | ||
| 2880 | /* Ironlake: try to setup display ref clock before DPLL | 2994 | /* Ironlake: try to setup display ref clock before DPLL |
| @@ -2882,7 +2996,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2882 | * PCH B stepping, previous chipset stepping should be | 2996 | * PCH B stepping, previous chipset stepping should be |
| 2883 | * ignoring this setting. | 2997 | * ignoring this setting. |
| 2884 | */ | 2998 | */ |
| 2885 | if (IS_IGDNG(dev)) { | 2999 | if (IS_IRONLAKE(dev)) { |
| 2886 | temp = I915_READ(PCH_DREF_CONTROL); | 3000 | temp = I915_READ(PCH_DREF_CONTROL); |
| 2887 | /* Always enable nonspread source */ | 3001 | /* Always enable nonspread source */ |
| 2888 | temp &= ~DREF_NONSPREAD_SOURCE_MASK; | 3002 | temp &= ~DREF_NONSPREAD_SOURCE_MASK; |
| @@ -2917,7 +3031,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2917 | } | 3031 | } |
| 2918 | } | 3032 | } |
| 2919 | 3033 | ||
| 2920 | if (IS_IGD(dev)) { | 3034 | if (IS_PINEVIEW(dev)) { |
| 2921 | fp = (1 << clock.n) << 16 | clock.m1 << 8 | clock.m2; | 3035 | fp = (1 << clock.n) << 16 | clock.m1 << 8 | clock.m2; |
| 2922 | if (has_reduced_clock) | 3036 | if (has_reduced_clock) |
| 2923 | fp2 = (1 << reduced_clock.n) << 16 | | 3037 | fp2 = (1 << reduced_clock.n) << 16 | |
| @@ -2929,7 +3043,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2929 | reduced_clock.m2; | 3043 | reduced_clock.m2; |
| 2930 | } | 3044 | } |
| 2931 | 3045 | ||
| 2932 | if (!IS_IGDNG(dev)) | 3046 | if (!IS_IRONLAKE(dev)) |
| 2933 | dpll = DPLL_VGA_MODE_DIS; | 3047 | dpll = DPLL_VGA_MODE_DIS; |
| 2934 | 3048 | ||
| 2935 | if (IS_I9XX(dev)) { | 3049 | if (IS_I9XX(dev)) { |
| @@ -2942,19 +3056,19 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2942 | sdvo_pixel_multiply = adjusted_mode->clock / mode->clock; | 3056 | sdvo_pixel_multiply = adjusted_mode->clock / mode->clock; |
| 2943 | if (IS_I945G(dev) || IS_I945GM(dev) || IS_G33(dev)) | 3057 | if (IS_I945G(dev) || IS_I945GM(dev) || IS_G33(dev)) |
| 2944 | dpll |= (sdvo_pixel_multiply - 1) << SDVO_MULTIPLIER_SHIFT_HIRES; | 3058 | dpll |= (sdvo_pixel_multiply - 1) << SDVO_MULTIPLIER_SHIFT_HIRES; |
| 2945 | else if (IS_IGDNG(dev)) | 3059 | else if (IS_IRONLAKE(dev)) |
| 2946 | dpll |= (sdvo_pixel_multiply - 1) << PLL_REF_SDVO_HDMI_MULTIPLIER_SHIFT; | 3060 | dpll |= (sdvo_pixel_multiply - 1) << PLL_REF_SDVO_HDMI_MULTIPLIER_SHIFT; |
| 2947 | } | 3061 | } |
| 2948 | if (is_dp) | 3062 | if (is_dp) |
| 2949 | dpll |= DPLL_DVO_HIGH_SPEED; | 3063 | dpll |= DPLL_DVO_HIGH_SPEED; |
| 2950 | 3064 | ||
| 2951 | /* compute bitmask from p1 value */ | 3065 | /* compute bitmask from p1 value */ |
| 2952 | if (IS_IGD(dev)) | 3066 | if (IS_PINEVIEW(dev)) |
| 2953 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT_IGD; | 3067 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW; |
| 2954 | else { | 3068 | else { |
| 2955 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT; | 3069 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT; |
| 2956 | /* also FPA1 */ | 3070 | /* also FPA1 */ |
| 2957 | if (IS_IGDNG(dev)) | 3071 | if (IS_IRONLAKE(dev)) |
| 2958 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT; | 3072 | dpll |= (1 << (clock.p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT; |
| 2959 | if (IS_G4X(dev) && has_reduced_clock) | 3073 | if (IS_G4X(dev) && has_reduced_clock) |
| 2960 | dpll |= (1 << (reduced_clock.p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT; | 3074 | dpll |= (1 << (reduced_clock.p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT; |
| @@ -2973,7 +3087,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 2973 | dpll |= DPLLB_LVDS_P2_CLOCK_DIV_14; | 3087 | dpll |= DPLLB_LVDS_P2_CLOCK_DIV_14; |
| 2974 | break; | 3088 | break; |
| 2975 | } | 3089 | } |
| 2976 | if (IS_I965G(dev) && !IS_IGDNG(dev)) | 3090 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) |
| 2977 | dpll |= (6 << PLL_LOAD_PULSE_PHASE_SHIFT); | 3091 | dpll |= (6 << PLL_LOAD_PULSE_PHASE_SHIFT); |
| 2978 | } else { | 3092 | } else { |
| 2979 | if (is_lvds) { | 3093 | if (is_lvds) { |
| @@ -3005,9 +3119,9 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3005 | /* Set up the display plane register */ | 3119 | /* Set up the display plane register */ |
| 3006 | dspcntr = DISPPLANE_GAMMA_ENABLE; | 3120 | dspcntr = DISPPLANE_GAMMA_ENABLE; |
| 3007 | 3121 | ||
| 3008 | /* IGDNG's plane is forced to pipe, bit 24 is to | 3122 | /* Ironlake's plane is forced to pipe, bit 24 is to |
| 3009 | enable color space conversion */ | 3123 | enable color space conversion */ |
| 3010 | if (!IS_IGDNG(dev)) { | 3124 | if (!IS_IRONLAKE(dev)) { |
| 3011 | if (pipe == 0) | 3125 | if (pipe == 0) |
| 3012 | dspcntr &= ~DISPPLANE_SEL_PIPE_MASK; | 3126 | dspcntr &= ~DISPPLANE_SEL_PIPE_MASK; |
| 3013 | else | 3127 | else |
| @@ -3034,20 +3148,20 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3034 | 3148 | ||
| 3035 | 3149 | ||
| 3036 | /* Disable the panel fitter if it was on our pipe */ | 3150 | /* Disable the panel fitter if it was on our pipe */ |
| 3037 | if (!IS_IGDNG(dev) && intel_panel_fitter_pipe(dev) == pipe) | 3151 | if (!IS_IRONLAKE(dev) && intel_panel_fitter_pipe(dev) == pipe) |
| 3038 | I915_WRITE(PFIT_CONTROL, 0); | 3152 | I915_WRITE(PFIT_CONTROL, 0); |
| 3039 | 3153 | ||
| 3040 | DRM_DEBUG("Mode for pipe %c:\n", pipe == 0 ? 'A' : 'B'); | 3154 | DRM_DEBUG_KMS("Mode for pipe %c:\n", pipe == 0 ? 'A' : 'B'); |
| 3041 | drm_mode_debug_printmodeline(mode); | 3155 | drm_mode_debug_printmodeline(mode); |
| 3042 | 3156 | ||
| 3043 | /* assign to IGDNG registers */ | 3157 | /* assign to Ironlake registers */ |
| 3044 | if (IS_IGDNG(dev)) { | 3158 | if (IS_IRONLAKE(dev)) { |
| 3045 | fp_reg = pch_fp_reg; | 3159 | fp_reg = pch_fp_reg; |
| 3046 | dpll_reg = pch_dpll_reg; | 3160 | dpll_reg = pch_dpll_reg; |
| 3047 | } | 3161 | } |
| 3048 | 3162 | ||
| 3049 | if (is_edp) { | 3163 | if (is_edp) { |
| 3050 | igdng_disable_pll_edp(crtc); | 3164 | ironlake_disable_pll_edp(crtc); |
| 3051 | } else if ((dpll & DPLL_VCO_ENABLE)) { | 3165 | } else if ((dpll & DPLL_VCO_ENABLE)) { |
| 3052 | I915_WRITE(fp_reg, fp); | 3166 | I915_WRITE(fp_reg, fp); |
| 3053 | I915_WRITE(dpll_reg, dpll & ~DPLL_VCO_ENABLE); | 3167 | I915_WRITE(dpll_reg, dpll & ~DPLL_VCO_ENABLE); |
| @@ -3062,7 +3176,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3062 | if (is_lvds) { | 3176 | if (is_lvds) { |
| 3063 | u32 lvds; | 3177 | u32 lvds; |
| 3064 | 3178 | ||
| 3065 | if (IS_IGDNG(dev)) | 3179 | if (IS_IRONLAKE(dev)) |
| 3066 | lvds_reg = PCH_LVDS; | 3180 | lvds_reg = PCH_LVDS; |
| 3067 | 3181 | ||
| 3068 | lvds = I915_READ(lvds_reg); | 3182 | lvds = I915_READ(lvds_reg); |
| @@ -3095,7 +3209,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3095 | /* Wait for the clocks to stabilize. */ | 3209 | /* Wait for the clocks to stabilize. */ |
| 3096 | udelay(150); | 3210 | udelay(150); |
| 3097 | 3211 | ||
| 3098 | if (IS_I965G(dev) && !IS_IGDNG(dev)) { | 3212 | if (IS_I965G(dev) && !IS_IRONLAKE(dev)) { |
| 3099 | if (is_sdvo) { | 3213 | if (is_sdvo) { |
| 3100 | sdvo_pixel_multiply = adjusted_mode->clock / mode->clock; | 3214 | sdvo_pixel_multiply = adjusted_mode->clock / mode->clock; |
| 3101 | I915_WRITE(dpll_md_reg, (0 << DPLL_MD_UDI_DIVIDER_SHIFT) | | 3215 | I915_WRITE(dpll_md_reg, (0 << DPLL_MD_UDI_DIVIDER_SHIFT) | |
| @@ -3115,14 +3229,14 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3115 | I915_WRITE(fp_reg + 4, fp2); | 3229 | I915_WRITE(fp_reg + 4, fp2); |
| 3116 | intel_crtc->lowfreq_avail = true; | 3230 | intel_crtc->lowfreq_avail = true; |
| 3117 | if (HAS_PIPE_CXSR(dev)) { | 3231 | if (HAS_PIPE_CXSR(dev)) { |
| 3118 | DRM_DEBUG("enabling CxSR downclocking\n"); | 3232 | DRM_DEBUG_KMS("enabling CxSR downclocking\n"); |
| 3119 | pipeconf |= PIPECONF_CXSR_DOWNCLOCK; | 3233 | pipeconf |= PIPECONF_CXSR_DOWNCLOCK; |
| 3120 | } | 3234 | } |
| 3121 | } else { | 3235 | } else { |
| 3122 | I915_WRITE(fp_reg + 4, fp); | 3236 | I915_WRITE(fp_reg + 4, fp); |
| 3123 | intel_crtc->lowfreq_avail = false; | 3237 | intel_crtc->lowfreq_avail = false; |
| 3124 | if (HAS_PIPE_CXSR(dev)) { | 3238 | if (HAS_PIPE_CXSR(dev)) { |
| 3125 | DRM_DEBUG("disabling CxSR downclocking\n"); | 3239 | DRM_DEBUG_KMS("disabling CxSR downclocking\n"); |
| 3126 | pipeconf &= ~PIPECONF_CXSR_DOWNCLOCK; | 3240 | pipeconf &= ~PIPECONF_CXSR_DOWNCLOCK; |
| 3127 | } | 3241 | } |
| 3128 | } | 3242 | } |
| @@ -3142,21 +3256,21 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3142 | /* pipesrc and dspsize control the size that is scaled from, which should | 3256 | /* pipesrc and dspsize control the size that is scaled from, which should |
| 3143 | * always be the user's requested size. | 3257 | * always be the user's requested size. |
| 3144 | */ | 3258 | */ |
| 3145 | if (!IS_IGDNG(dev)) { | 3259 | if (!IS_IRONLAKE(dev)) { |
| 3146 | I915_WRITE(dspsize_reg, ((mode->vdisplay - 1) << 16) | | 3260 | I915_WRITE(dspsize_reg, ((mode->vdisplay - 1) << 16) | |
| 3147 | (mode->hdisplay - 1)); | 3261 | (mode->hdisplay - 1)); |
| 3148 | I915_WRITE(dsppos_reg, 0); | 3262 | I915_WRITE(dsppos_reg, 0); |
| 3149 | } | 3263 | } |
| 3150 | I915_WRITE(pipesrc_reg, ((mode->hdisplay - 1) << 16) | (mode->vdisplay - 1)); | 3264 | I915_WRITE(pipesrc_reg, ((mode->hdisplay - 1) << 16) | (mode->vdisplay - 1)); |
| 3151 | 3265 | ||
| 3152 | if (IS_IGDNG(dev)) { | 3266 | if (IS_IRONLAKE(dev)) { |
| 3153 | I915_WRITE(data_m1_reg, TU_SIZE(m_n.tu) | m_n.gmch_m); | 3267 | I915_WRITE(data_m1_reg, TU_SIZE(m_n.tu) | m_n.gmch_m); |
| 3154 | I915_WRITE(data_n1_reg, TU_SIZE(m_n.tu) | m_n.gmch_n); | 3268 | I915_WRITE(data_n1_reg, TU_SIZE(m_n.tu) | m_n.gmch_n); |
| 3155 | I915_WRITE(link_m1_reg, m_n.link_m); | 3269 | I915_WRITE(link_m1_reg, m_n.link_m); |
| 3156 | I915_WRITE(link_n1_reg, m_n.link_n); | 3270 | I915_WRITE(link_n1_reg, m_n.link_n); |
| 3157 | 3271 | ||
| 3158 | if (is_edp) { | 3272 | if (is_edp) { |
| 3159 | igdng_set_pll_edp(crtc, adjusted_mode->clock); | 3273 | ironlake_set_pll_edp(crtc, adjusted_mode->clock); |
| 3160 | } else { | 3274 | } else { |
| 3161 | /* enable FDI RX PLL too */ | 3275 | /* enable FDI RX PLL too */ |
| 3162 | temp = I915_READ(fdi_rx_reg); | 3276 | temp = I915_READ(fdi_rx_reg); |
| @@ -3170,7 +3284,7 @@ static int intel_crtc_mode_set(struct drm_crtc *crtc, | |||
| 3170 | 3284 | ||
| 3171 | intel_wait_for_vblank(dev); | 3285 | intel_wait_for_vblank(dev); |
| 3172 | 3286 | ||
| 3173 | if (IS_IGDNG(dev)) { | 3287 | if (IS_IRONLAKE(dev)) { |
| 3174 | /* enable address swizzle for tiling buffer */ | 3288 | /* enable address swizzle for tiling buffer */ |
| 3175 | temp = I915_READ(DISP_ARB_CTL); | 3289 | temp = I915_READ(DISP_ARB_CTL); |
| 3176 | I915_WRITE(DISP_ARB_CTL, temp | DISP_TILE_SURFACE_SWIZZLING); | 3290 | I915_WRITE(DISP_ARB_CTL, temp | DISP_TILE_SURFACE_SWIZZLING); |
| @@ -3204,8 +3318,8 @@ void intel_crtc_load_lut(struct drm_crtc *crtc) | |||
| 3204 | if (!crtc->enabled) | 3318 | if (!crtc->enabled) |
| 3205 | return; | 3319 | return; |
| 3206 | 3320 | ||
| 3207 | /* use legacy palette for IGDNG */ | 3321 | /* use legacy palette for Ironlake */ |
| 3208 | if (IS_IGDNG(dev)) | 3322 | if (IS_IRONLAKE(dev)) |
| 3209 | palreg = (intel_crtc->pipe == 0) ? LGC_PALETTE_A : | 3323 | palreg = (intel_crtc->pipe == 0) ? LGC_PALETTE_A : |
| 3210 | LGC_PALETTE_B; | 3324 | LGC_PALETTE_B; |
| 3211 | 3325 | ||
| @@ -3234,11 +3348,11 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc, | |||
| 3234 | size_t addr; | 3348 | size_t addr; |
| 3235 | int ret; | 3349 | int ret; |
| 3236 | 3350 | ||
| 3237 | DRM_DEBUG("\n"); | 3351 | DRM_DEBUG_KMS("\n"); |
| 3238 | 3352 | ||
| 3239 | /* if we want to turn off the cursor ignore width and height */ | 3353 | /* if we want to turn off the cursor ignore width and height */ |
| 3240 | if (!handle) { | 3354 | if (!handle) { |
| 3241 | DRM_DEBUG("cursor off\n"); | 3355 | DRM_DEBUG_KMS("cursor off\n"); |
| 3242 | if (IS_MOBILE(dev) || IS_I9XX(dev)) { | 3356 | if (IS_MOBILE(dev) || IS_I9XX(dev)) { |
| 3243 | temp &= ~(CURSOR_MODE | MCURSOR_GAMMA_ENABLE); | 3357 | temp &= ~(CURSOR_MODE | MCURSOR_GAMMA_ENABLE); |
| 3244 | temp |= CURSOR_MODE_DISABLE; | 3358 | temp |= CURSOR_MODE_DISABLE; |
| @@ -3546,18 +3660,18 @@ static int intel_crtc_clock_get(struct drm_device *dev, struct drm_crtc *crtc) | |||
| 3546 | fp = I915_READ((pipe == 0) ? FPA1 : FPB1); | 3660 | fp = I915_READ((pipe == 0) ? FPA1 : FPB1); |
| 3547 | 3661 | ||
| 3548 | clock.m1 = (fp & FP_M1_DIV_MASK) >> FP_M1_DIV_SHIFT; | 3662 | clock.m1 = (fp & FP_M1_DIV_MASK) >> FP_M1_DIV_SHIFT; |
| 3549 | if (IS_IGD(dev)) { | 3663 | if (IS_PINEVIEW(dev)) { |
| 3550 | clock.n = ffs((fp & FP_N_IGD_DIV_MASK) >> FP_N_DIV_SHIFT) - 1; | 3664 | clock.n = ffs((fp & FP_N_PINEVIEW_DIV_MASK) >> FP_N_DIV_SHIFT) - 1; |
| 3551 | clock.m2 = (fp & FP_M2_IGD_DIV_MASK) >> FP_M2_DIV_SHIFT; | 3665 | clock.m2 = (fp & FP_M2_PINEVIEW_DIV_MASK) >> FP_M2_DIV_SHIFT; |
| 3552 | } else { | 3666 | } else { |
| 3553 | clock.n = (fp & FP_N_DIV_MASK) >> FP_N_DIV_SHIFT; | 3667 | clock.n = (fp & FP_N_DIV_MASK) >> FP_N_DIV_SHIFT; |
| 3554 | clock.m2 = (fp & FP_M2_DIV_MASK) >> FP_M2_DIV_SHIFT; | 3668 | clock.m2 = (fp & FP_M2_DIV_MASK) >> FP_M2_DIV_SHIFT; |
| 3555 | } | 3669 | } |
| 3556 | 3670 | ||
| 3557 | if (IS_I9XX(dev)) { | 3671 | if (IS_I9XX(dev)) { |
| 3558 | if (IS_IGD(dev)) | 3672 | if (IS_PINEVIEW(dev)) |
| 3559 | clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK_IGD) >> | 3673 | clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK_PINEVIEW) >> |
| 3560 | DPLL_FPA01_P1_POST_DIV_SHIFT_IGD); | 3674 | DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW); |
| 3561 | else | 3675 | else |
| 3562 | clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK) >> | 3676 | clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK) >> |
| 3563 | DPLL_FPA01_P1_POST_DIV_SHIFT); | 3677 | DPLL_FPA01_P1_POST_DIV_SHIFT); |
| @@ -3572,7 +3686,7 @@ static int intel_crtc_clock_get(struct drm_device *dev, struct drm_crtc *crtc) | |||
| 3572 | 7 : 14; | 3686 | 7 : 14; |
| 3573 | break; | 3687 | break; |
| 3574 | default: | 3688 | default: |
| 3575 | DRM_DEBUG("Unknown DPLL mode %08x in programmed " | 3689 | DRM_DEBUG_KMS("Unknown DPLL mode %08x in programmed " |
| 3576 | "mode\n", (int)(dpll & DPLL_MODE_MASK)); | 3690 | "mode\n", (int)(dpll & DPLL_MODE_MASK)); |
| 3577 | return 0; | 3691 | return 0; |
| 3578 | } | 3692 | } |
| @@ -3658,7 +3772,7 @@ static void intel_gpu_idle_timer(unsigned long arg) | |||
| 3658 | struct drm_device *dev = (struct drm_device *)arg; | 3772 | struct drm_device *dev = (struct drm_device *)arg; |
| 3659 | drm_i915_private_t *dev_priv = dev->dev_private; | 3773 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 3660 | 3774 | ||
| 3661 | DRM_DEBUG("idle timer fired, downclocking\n"); | 3775 | DRM_DEBUG_DRIVER("idle timer fired, downclocking\n"); |
| 3662 | 3776 | ||
| 3663 | dev_priv->busy = false; | 3777 | dev_priv->busy = false; |
| 3664 | 3778 | ||
| @@ -3669,11 +3783,11 @@ void intel_increase_renderclock(struct drm_device *dev, bool schedule) | |||
| 3669 | { | 3783 | { |
| 3670 | drm_i915_private_t *dev_priv = dev->dev_private; | 3784 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 3671 | 3785 | ||
| 3672 | if (IS_IGDNG(dev)) | 3786 | if (IS_IRONLAKE(dev)) |
| 3673 | return; | 3787 | return; |
| 3674 | 3788 | ||
| 3675 | if (!dev_priv->render_reclock_avail) { | 3789 | if (!dev_priv->render_reclock_avail) { |
| 3676 | DRM_DEBUG("not reclocking render clock\n"); | 3790 | DRM_DEBUG_DRIVER("not reclocking render clock\n"); |
| 3677 | return; | 3791 | return; |
| 3678 | } | 3792 | } |
| 3679 | 3793 | ||
| @@ -3682,7 +3796,7 @@ void intel_increase_renderclock(struct drm_device *dev, bool schedule) | |||
| 3682 | pci_write_config_word(dev->pdev, GCFGC, dev_priv->orig_clock); | 3796 | pci_write_config_word(dev->pdev, GCFGC, dev_priv->orig_clock); |
| 3683 | else if (IS_I85X(dev)) | 3797 | else if (IS_I85X(dev)) |
| 3684 | pci_write_config_word(dev->pdev, HPLLCC, dev_priv->orig_clock); | 3798 | pci_write_config_word(dev->pdev, HPLLCC, dev_priv->orig_clock); |
| 3685 | DRM_DEBUG("increasing render clock frequency\n"); | 3799 | DRM_DEBUG_DRIVER("increasing render clock frequency\n"); |
| 3686 | 3800 | ||
| 3687 | /* Schedule downclock */ | 3801 | /* Schedule downclock */ |
| 3688 | if (schedule) | 3802 | if (schedule) |
| @@ -3694,11 +3808,11 @@ void intel_decrease_renderclock(struct drm_device *dev) | |||
| 3694 | { | 3808 | { |
| 3695 | drm_i915_private_t *dev_priv = dev->dev_private; | 3809 | drm_i915_private_t *dev_priv = dev->dev_private; |
| 3696 | 3810 | ||
| 3697 | if (IS_IGDNG(dev)) | 3811 | if (IS_IRONLAKE(dev)) |
| 3698 | return; | 3812 | return; |
| 3699 | 3813 | ||
| 3700 | if (!dev_priv->render_reclock_avail) { | 3814 | if (!dev_priv->render_reclock_avail) { |
| 3701 | DRM_DEBUG("not reclocking render clock\n"); | 3815 | DRM_DEBUG_DRIVER("not reclocking render clock\n"); |
| 3702 | return; | 3816 | return; |
| 3703 | } | 3817 | } |
| 3704 | 3818 | ||
| @@ -3758,7 +3872,7 @@ void intel_decrease_renderclock(struct drm_device *dev) | |||
| 3758 | 3872 | ||
| 3759 | pci_write_config_word(dev->pdev, HPLLCC, hpllcc); | 3873 | pci_write_config_word(dev->pdev, HPLLCC, hpllcc); |
| 3760 | } | 3874 | } |
| 3761 | DRM_DEBUG("decreasing render clock frequency\n"); | 3875 | DRM_DEBUG_DRIVER("decreasing render clock frequency\n"); |
| 3762 | } | 3876 | } |
| 3763 | 3877 | ||
| 3764 | /* Note that no increase function is needed for this - increase_renderclock() | 3878 | /* Note that no increase function is needed for this - increase_renderclock() |
| @@ -3766,7 +3880,7 @@ void intel_decrease_renderclock(struct drm_device *dev) | |||
| 3766 | */ | 3880 | */ |
| 3767 | void intel_decrease_displayclock(struct drm_device *dev) | 3881 | void intel_decrease_displayclock(struct drm_device *dev) |
| 3768 | { | 3882 | { |
| 3769 | if (IS_IGDNG(dev)) | 3883 | if (IS_IRONLAKE(dev)) |
| 3770 | return; | 3884 | return; |
| 3771 | 3885 | ||
| 3772 | if (IS_I945G(dev) || IS_I945GM(dev) || IS_I915G(dev) || | 3886 | if (IS_I945G(dev) || IS_I945GM(dev) || IS_I915G(dev) || |
| @@ -3792,7 +3906,7 @@ static void intel_crtc_idle_timer(unsigned long arg) | |||
| 3792 | struct drm_crtc *crtc = &intel_crtc->base; | 3906 | struct drm_crtc *crtc = &intel_crtc->base; |
| 3793 | drm_i915_private_t *dev_priv = crtc->dev->dev_private; | 3907 | drm_i915_private_t *dev_priv = crtc->dev->dev_private; |
| 3794 | 3908 | ||
| 3795 | DRM_DEBUG("idle timer fired, downclocking\n"); | 3909 | DRM_DEBUG_DRIVER("idle timer fired, downclocking\n"); |
| 3796 | 3910 | ||
| 3797 | intel_crtc->busy = false; | 3911 | intel_crtc->busy = false; |
| 3798 | 3912 | ||
| @@ -3808,14 +3922,14 @@ static void intel_increase_pllclock(struct drm_crtc *crtc, bool schedule) | |||
| 3808 | int dpll_reg = (pipe == 0) ? DPLL_A : DPLL_B; | 3922 | int dpll_reg = (pipe == 0) ? DPLL_A : DPLL_B; |
| 3809 | int dpll = I915_READ(dpll_reg); | 3923 | int dpll = I915_READ(dpll_reg); |
| 3810 | 3924 | ||
| 3811 | if (IS_IGDNG(dev)) | 3925 | if (IS_IRONLAKE(dev)) |
| 3812 | return; | 3926 | return; |
| 3813 | 3927 | ||
| 3814 | if (!dev_priv->lvds_downclock_avail) | 3928 | if (!dev_priv->lvds_downclock_avail) |
| 3815 | return; | 3929 | return; |
| 3816 | 3930 | ||
| 3817 | if (!HAS_PIPE_CXSR(dev) && (dpll & DISPLAY_RATE_SELECT_FPA1)) { | 3931 | if (!HAS_PIPE_CXSR(dev) && (dpll & DISPLAY_RATE_SELECT_FPA1)) { |
| 3818 | DRM_DEBUG("upclocking LVDS\n"); | 3932 | DRM_DEBUG_DRIVER("upclocking LVDS\n"); |
| 3819 | 3933 | ||
| 3820 | /* Unlock panel regs */ | 3934 | /* Unlock panel regs */ |
| 3821 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) | (0xabcd << 16)); | 3935 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) | (0xabcd << 16)); |
| @@ -3826,7 +3940,7 @@ static void intel_increase_pllclock(struct drm_crtc *crtc, bool schedule) | |||
| 3826 | intel_wait_for_vblank(dev); | 3940 | intel_wait_for_vblank(dev); |
| 3827 | dpll = I915_READ(dpll_reg); | 3941 | dpll = I915_READ(dpll_reg); |
| 3828 | if (dpll & DISPLAY_RATE_SELECT_FPA1) | 3942 | if (dpll & DISPLAY_RATE_SELECT_FPA1) |
| 3829 | DRM_DEBUG("failed to upclock LVDS!\n"); | 3943 | DRM_DEBUG_DRIVER("failed to upclock LVDS!\n"); |
| 3830 | 3944 | ||
| 3831 | /* ...and lock them again */ | 3945 | /* ...and lock them again */ |
| 3832 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) & 0x3); | 3946 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) & 0x3); |
| @@ -3847,7 +3961,7 @@ static void intel_decrease_pllclock(struct drm_crtc *crtc) | |||
| 3847 | int dpll_reg = (pipe == 0) ? DPLL_A : DPLL_B; | 3961 | int dpll_reg = (pipe == 0) ? DPLL_A : DPLL_B; |
| 3848 | int dpll = I915_READ(dpll_reg); | 3962 | int dpll = I915_READ(dpll_reg); |
| 3849 | 3963 | ||
| 3850 | if (IS_IGDNG(dev)) | 3964 | if (IS_IRONLAKE(dev)) |
| 3851 | return; | 3965 | return; |
| 3852 | 3966 | ||
| 3853 | if (!dev_priv->lvds_downclock_avail) | 3967 | if (!dev_priv->lvds_downclock_avail) |
| @@ -3858,7 +3972,7 @@ static void intel_decrease_pllclock(struct drm_crtc *crtc) | |||
| 3858 | * the manual case. | 3972 | * the manual case. |
| 3859 | */ | 3973 | */ |
| 3860 | if (!HAS_PIPE_CXSR(dev) && intel_crtc->lowfreq_avail) { | 3974 | if (!HAS_PIPE_CXSR(dev) && intel_crtc->lowfreq_avail) { |
| 3861 | DRM_DEBUG("downclocking LVDS\n"); | 3975 | DRM_DEBUG_DRIVER("downclocking LVDS\n"); |
| 3862 | 3976 | ||
| 3863 | /* Unlock panel regs */ | 3977 | /* Unlock panel regs */ |
| 3864 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) | (0xabcd << 16)); | 3978 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) | (0xabcd << 16)); |
| @@ -3869,7 +3983,7 @@ static void intel_decrease_pllclock(struct drm_crtc *crtc) | |||
| 3869 | intel_wait_for_vblank(dev); | 3983 | intel_wait_for_vblank(dev); |
| 3870 | dpll = I915_READ(dpll_reg); | 3984 | dpll = I915_READ(dpll_reg); |
| 3871 | if (!(dpll & DISPLAY_RATE_SELECT_FPA1)) | 3985 | if (!(dpll & DISPLAY_RATE_SELECT_FPA1)) |
| 3872 | DRM_DEBUG("failed to downclock LVDS!\n"); | 3986 | DRM_DEBUG_DRIVER("failed to downclock LVDS!\n"); |
| 3873 | 3987 | ||
| 3874 | /* ...and lock them again */ | 3988 | /* ...and lock them again */ |
| 3875 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) & 0x3); | 3989 | I915_WRITE(PP_CONTROL, I915_READ(PP_CONTROL) & 0x3); |
| @@ -3936,8 +4050,13 @@ void intel_mark_busy(struct drm_device *dev, struct drm_gem_object *obj) | |||
| 3936 | if (!drm_core_check_feature(dev, DRIVER_MODESET)) | 4050 | if (!drm_core_check_feature(dev, DRIVER_MODESET)) |
| 3937 | return; | 4051 | return; |
| 3938 | 4052 | ||
| 3939 | dev_priv->busy = true; | 4053 | if (!dev_priv->busy) { |
| 3940 | intel_increase_renderclock(dev, true); | 4054 | dev_priv->busy = true; |
| 4055 | intel_increase_renderclock(dev, true); | ||
| 4056 | } else { | ||
| 4057 | mod_timer(&dev_priv->idle_timer, jiffies + | ||
| 4058 | msecs_to_jiffies(GPU_IDLE_TIMEOUT)); | ||
| 4059 | } | ||
| 3941 | 4060 | ||
| 3942 | list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { | 4061 | list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { |
| 3943 | if (!crtc->fb) | 4062 | if (!crtc->fb) |
| @@ -3967,6 +4086,158 @@ static void intel_crtc_destroy(struct drm_crtc *crtc) | |||
| 3967 | kfree(intel_crtc); | 4086 | kfree(intel_crtc); |
| 3968 | } | 4087 | } |
| 3969 | 4088 | ||
| 4089 | struct intel_unpin_work { | ||
| 4090 | struct work_struct work; | ||
| 4091 | struct drm_device *dev; | ||
| 4092 | struct drm_gem_object *obj; | ||
| 4093 | struct drm_pending_vblank_event *event; | ||
| 4094 | int pending; | ||
| 4095 | }; | ||
| 4096 | |||
| 4097 | static void intel_unpin_work_fn(struct work_struct *__work) | ||
| 4098 | { | ||
| 4099 | struct intel_unpin_work *work = | ||
| 4100 | container_of(__work, struct intel_unpin_work, work); | ||
| 4101 | |||
| 4102 | mutex_lock(&work->dev->struct_mutex); | ||
| 4103 | i915_gem_object_unpin(work->obj); | ||
| 4104 | drm_gem_object_unreference(work->obj); | ||
| 4105 | mutex_unlock(&work->dev->struct_mutex); | ||
| 4106 | kfree(work); | ||
| 4107 | } | ||
| 4108 | |||
| 4109 | void intel_finish_page_flip(struct drm_device *dev, int pipe) | ||
| 4110 | { | ||
| 4111 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 4112 | struct drm_crtc *crtc = dev_priv->pipe_to_crtc_mapping[pipe]; | ||
| 4113 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | ||
| 4114 | struct intel_unpin_work *work; | ||
| 4115 | struct drm_i915_gem_object *obj_priv; | ||
| 4116 | struct drm_pending_vblank_event *e; | ||
| 4117 | struct timeval now; | ||
| 4118 | unsigned long flags; | ||
| 4119 | |||
| 4120 | /* Ignore early vblank irqs */ | ||
| 4121 | if (intel_crtc == NULL) | ||
| 4122 | return; | ||
| 4123 | |||
| 4124 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 4125 | work = intel_crtc->unpin_work; | ||
| 4126 | if (work == NULL || !work->pending) { | ||
| 4127 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 4128 | return; | ||
| 4129 | } | ||
| 4130 | |||
| 4131 | intel_crtc->unpin_work = NULL; | ||
| 4132 | drm_vblank_put(dev, intel_crtc->pipe); | ||
| 4133 | |||
| 4134 | if (work->event) { | ||
| 4135 | e = work->event; | ||
| 4136 | do_gettimeofday(&now); | ||
| 4137 | e->event.sequence = drm_vblank_count(dev, intel_crtc->pipe); | ||
| 4138 | e->event.tv_sec = now.tv_sec; | ||
| 4139 | e->event.tv_usec = now.tv_usec; | ||
| 4140 | list_add_tail(&e->base.link, | ||
| 4141 | &e->base.file_priv->event_list); | ||
| 4142 | wake_up_interruptible(&e->base.file_priv->event_wait); | ||
| 4143 | } | ||
| 4144 | |||
| 4145 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 4146 | |||
| 4147 | obj_priv = work->obj->driver_private; | ||
| 4148 | if (atomic_dec_and_test(&obj_priv->pending_flip)) | ||
| 4149 | DRM_WAKEUP(&dev_priv->pending_flip_queue); | ||
| 4150 | schedule_work(&work->work); | ||
| 4151 | } | ||
| 4152 | |||
| 4153 | void intel_prepare_page_flip(struct drm_device *dev, int plane) | ||
| 4154 | { | ||
| 4155 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 4156 | struct intel_crtc *intel_crtc = | ||
| 4157 | to_intel_crtc(dev_priv->plane_to_crtc_mapping[plane]); | ||
| 4158 | unsigned long flags; | ||
| 4159 | |||
| 4160 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 4161 | if (intel_crtc->unpin_work) | ||
| 4162 | intel_crtc->unpin_work->pending = 1; | ||
| 4163 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 4164 | } | ||
| 4165 | |||
| 4166 | static int intel_crtc_page_flip(struct drm_crtc *crtc, | ||
| 4167 | struct drm_framebuffer *fb, | ||
| 4168 | struct drm_pending_vblank_event *event) | ||
| 4169 | { | ||
| 4170 | struct drm_device *dev = crtc->dev; | ||
| 4171 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 4172 | struct intel_framebuffer *intel_fb; | ||
| 4173 | struct drm_i915_gem_object *obj_priv; | ||
| 4174 | struct drm_gem_object *obj; | ||
| 4175 | struct intel_crtc *intel_crtc = to_intel_crtc(crtc); | ||
| 4176 | struct intel_unpin_work *work; | ||
| 4177 | unsigned long flags; | ||
| 4178 | int ret; | ||
| 4179 | RING_LOCALS; | ||
| 4180 | |||
| 4181 | work = kzalloc(sizeof *work, GFP_KERNEL); | ||
| 4182 | if (work == NULL) | ||
| 4183 | return -ENOMEM; | ||
| 4184 | |||
| 4185 | mutex_lock(&dev->struct_mutex); | ||
| 4186 | |||
| 4187 | work->event = event; | ||
| 4188 | work->dev = crtc->dev; | ||
| 4189 | intel_fb = to_intel_framebuffer(crtc->fb); | ||
| 4190 | work->obj = intel_fb->obj; | ||
| 4191 | INIT_WORK(&work->work, intel_unpin_work_fn); | ||
| 4192 | |||
| 4193 | /* We borrow the event spin lock for protecting unpin_work */ | ||
| 4194 | spin_lock_irqsave(&dev->event_lock, flags); | ||
| 4195 | if (intel_crtc->unpin_work) { | ||
| 4196 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 4197 | kfree(work); | ||
| 4198 | mutex_unlock(&dev->struct_mutex); | ||
| 4199 | return -EBUSY; | ||
| 4200 | } | ||
| 4201 | intel_crtc->unpin_work = work; | ||
| 4202 | spin_unlock_irqrestore(&dev->event_lock, flags); | ||
| 4203 | |||
| 4204 | intel_fb = to_intel_framebuffer(fb); | ||
| 4205 | obj = intel_fb->obj; | ||
| 4206 | |||
| 4207 | ret = intel_pin_and_fence_fb_obj(dev, obj); | ||
| 4208 | if (ret != 0) { | ||
| 4209 | kfree(work); | ||
| 4210 | mutex_unlock(&dev->struct_mutex); | ||
| 4211 | return ret; | ||
| 4212 | } | ||
| 4213 | |||
| 4214 | /* Reference the old fb object for the scheduled work. */ | ||
| 4215 | drm_gem_object_reference(work->obj); | ||
| 4216 | |||
| 4217 | crtc->fb = fb; | ||
| 4218 | i915_gem_object_flush_write_domain(obj); | ||
| 4219 | drm_vblank_get(dev, intel_crtc->pipe); | ||
| 4220 | obj_priv = obj->driver_private; | ||
| 4221 | atomic_inc(&obj_priv->pending_flip); | ||
| 4222 | |||
| 4223 | BEGIN_LP_RING(4); | ||
| 4224 | OUT_RING(MI_DISPLAY_FLIP | | ||
| 4225 | MI_DISPLAY_FLIP_PLANE(intel_crtc->plane)); | ||
| 4226 | OUT_RING(fb->pitch); | ||
| 4227 | if (IS_I965G(dev)) { | ||
| 4228 | OUT_RING(obj_priv->gtt_offset | obj_priv->tiling_mode); | ||
| 4229 | OUT_RING((fb->width << 16) | fb->height); | ||
| 4230 | } else { | ||
| 4231 | OUT_RING(obj_priv->gtt_offset); | ||
| 4232 | OUT_RING(MI_NOOP); | ||
| 4233 | } | ||
| 4234 | ADVANCE_LP_RING(); | ||
| 4235 | |||
| 4236 | mutex_unlock(&dev->struct_mutex); | ||
| 4237 | |||
| 4238 | return 0; | ||
| 4239 | } | ||
| 4240 | |||
| 3970 | static const struct drm_crtc_helper_funcs intel_helper_funcs = { | 4241 | static const struct drm_crtc_helper_funcs intel_helper_funcs = { |
| 3971 | .dpms = intel_crtc_dpms, | 4242 | .dpms = intel_crtc_dpms, |
| 3972 | .mode_fixup = intel_crtc_mode_fixup, | 4243 | .mode_fixup = intel_crtc_mode_fixup, |
| @@ -3983,11 +4254,13 @@ static const struct drm_crtc_funcs intel_crtc_funcs = { | |||
| 3983 | .gamma_set = intel_crtc_gamma_set, | 4254 | .gamma_set = intel_crtc_gamma_set, |
| 3984 | .set_config = drm_crtc_helper_set_config, | 4255 | .set_config = drm_crtc_helper_set_config, |
| 3985 | .destroy = intel_crtc_destroy, | 4256 | .destroy = intel_crtc_destroy, |
| 4257 | .page_flip = intel_crtc_page_flip, | ||
| 3986 | }; | 4258 | }; |
| 3987 | 4259 | ||
| 3988 | 4260 | ||
| 3989 | static void intel_crtc_init(struct drm_device *dev, int pipe) | 4261 | static void intel_crtc_init(struct drm_device *dev, int pipe) |
| 3990 | { | 4262 | { |
| 4263 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 3991 | struct intel_crtc *intel_crtc; | 4264 | struct intel_crtc *intel_crtc; |
| 3992 | int i; | 4265 | int i; |
| 3993 | 4266 | ||
| @@ -4010,10 +4283,15 @@ static void intel_crtc_init(struct drm_device *dev, int pipe) | |||
| 4010 | intel_crtc->pipe = pipe; | 4283 | intel_crtc->pipe = pipe; |
| 4011 | intel_crtc->plane = pipe; | 4284 | intel_crtc->plane = pipe; |
| 4012 | if (IS_MOBILE(dev) && (IS_I9XX(dev) && !IS_I965G(dev))) { | 4285 | if (IS_MOBILE(dev) && (IS_I9XX(dev) && !IS_I965G(dev))) { |
| 4013 | DRM_DEBUG("swapping pipes & planes for FBC\n"); | 4286 | DRM_DEBUG_KMS("swapping pipes & planes for FBC\n"); |
| 4014 | intel_crtc->plane = ((pipe == 0) ? 1 : 0); | 4287 | intel_crtc->plane = ((pipe == 0) ? 1 : 0); |
| 4015 | } | 4288 | } |
| 4016 | 4289 | ||
| 4290 | BUG_ON(pipe >= ARRAY_SIZE(dev_priv->plane_to_crtc_mapping) || | ||
| 4291 | dev_priv->plane_to_crtc_mapping[intel_crtc->plane] != NULL); | ||
| 4292 | dev_priv->plane_to_crtc_mapping[intel_crtc->plane] = &intel_crtc->base; | ||
| 4293 | dev_priv->pipe_to_crtc_mapping[intel_crtc->pipe] = &intel_crtc->base; | ||
| 4294 | |||
| 4017 | intel_crtc->cursor_addr = 0; | 4295 | intel_crtc->cursor_addr = 0; |
| 4018 | intel_crtc->dpms_mode = DRM_MODE_DPMS_OFF; | 4296 | intel_crtc->dpms_mode = DRM_MODE_DPMS_OFF; |
| 4019 | drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs); | 4297 | drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs); |
| @@ -4090,7 +4368,7 @@ static void intel_setup_outputs(struct drm_device *dev) | |||
| 4090 | if (IS_MOBILE(dev) && !IS_I830(dev)) | 4368 | if (IS_MOBILE(dev) && !IS_I830(dev)) |
| 4091 | intel_lvds_init(dev); | 4369 | intel_lvds_init(dev); |
| 4092 | 4370 | ||
| 4093 | if (IS_IGDNG(dev)) { | 4371 | if (IS_IRONLAKE(dev)) { |
| 4094 | int found; | 4372 | int found; |
| 4095 | 4373 | ||
| 4096 | if (IS_MOBILE(dev) && (I915_READ(DP_A) & DP_DETECTED)) | 4374 | if (IS_MOBILE(dev) && (I915_READ(DP_A) & DP_DETECTED)) |
| @@ -4118,7 +4396,7 @@ static void intel_setup_outputs(struct drm_device *dev) | |||
| 4118 | if (I915_READ(PCH_DP_D) & DP_DETECTED) | 4396 | if (I915_READ(PCH_DP_D) & DP_DETECTED) |
| 4119 | intel_dp_init(dev, PCH_DP_D); | 4397 | intel_dp_init(dev, PCH_DP_D); |
| 4120 | 4398 | ||
| 4121 | } else if (IS_I9XX(dev)) { | 4399 | } else if (SUPPORTS_DIGITAL_OUTPUTS(dev)) { |
| 4122 | bool found = false; | 4400 | bool found = false; |
| 4123 | 4401 | ||
| 4124 | if (I915_READ(SDVOB) & SDVO_DETECTED) { | 4402 | if (I915_READ(SDVOB) & SDVO_DETECTED) { |
| @@ -4145,10 +4423,10 @@ static void intel_setup_outputs(struct drm_device *dev) | |||
| 4145 | 4423 | ||
| 4146 | if (SUPPORTS_INTEGRATED_DP(dev) && (I915_READ(DP_D) & DP_DETECTED)) | 4424 | if (SUPPORTS_INTEGRATED_DP(dev) && (I915_READ(DP_D) & DP_DETECTED)) |
| 4147 | intel_dp_init(dev, DP_D); | 4425 | intel_dp_init(dev, DP_D); |
| 4148 | } else | 4426 | } else if (IS_I8XX(dev)) |
| 4149 | intel_dvo_init(dev); | 4427 | intel_dvo_init(dev); |
| 4150 | 4428 | ||
| 4151 | if (IS_I9XX(dev) && IS_MOBILE(dev) && !IS_IGDNG(dev)) | 4429 | if (SUPPORTS_TV(dev)) |
| 4152 | intel_tv_init(dev); | 4430 | intel_tv_init(dev); |
| 4153 | 4431 | ||
| 4154 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | 4432 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { |
| @@ -4257,7 +4535,7 @@ void intel_init_clock_gating(struct drm_device *dev) | |||
| 4257 | * Disable clock gating reported to work incorrectly according to the | 4535 | * Disable clock gating reported to work incorrectly according to the |
| 4258 | * specs, but enable as much else as we can. | 4536 | * specs, but enable as much else as we can. |
| 4259 | */ | 4537 | */ |
| 4260 | if (IS_IGDNG(dev)) { | 4538 | if (IS_IRONLAKE(dev)) { |
| 4261 | return; | 4539 | return; |
| 4262 | } else if (IS_G4X(dev)) { | 4540 | } else if (IS_G4X(dev)) { |
| 4263 | uint32_t dspclk_gate; | 4541 | uint32_t dspclk_gate; |
| @@ -4291,11 +4569,52 @@ void intel_init_clock_gating(struct drm_device *dev) | |||
| 4291 | dstate |= DSTATE_PLL_D3_OFF | DSTATE_GFX_CLOCK_GATING | | 4569 | dstate |= DSTATE_PLL_D3_OFF | DSTATE_GFX_CLOCK_GATING | |
| 4292 | DSTATE_DOT_CLOCK_GATING; | 4570 | DSTATE_DOT_CLOCK_GATING; |
| 4293 | I915_WRITE(D_STATE, dstate); | 4571 | I915_WRITE(D_STATE, dstate); |
| 4294 | } else if (IS_I855(dev) || IS_I865G(dev)) { | 4572 | } else if (IS_I85X(dev) || IS_I865G(dev)) { |
| 4295 | I915_WRITE(RENCLK_GATE_D1, SV_CLOCK_GATE_DISABLE); | 4573 | I915_WRITE(RENCLK_GATE_D1, SV_CLOCK_GATE_DISABLE); |
| 4296 | } else if (IS_I830(dev)) { | 4574 | } else if (IS_I830(dev)) { |
| 4297 | I915_WRITE(DSPCLK_GATE_D, OVRUNIT_CLOCK_GATE_DISABLE); | 4575 | I915_WRITE(DSPCLK_GATE_D, OVRUNIT_CLOCK_GATE_DISABLE); |
| 4298 | } | 4576 | } |
| 4577 | |||
| 4578 | /* | ||
| 4579 | * GPU can automatically power down the render unit if given a page | ||
| 4580 | * to save state. | ||
| 4581 | */ | ||
| 4582 | if (I915_HAS_RC6(dev)) { | ||
| 4583 | struct drm_gem_object *pwrctx; | ||
| 4584 | struct drm_i915_gem_object *obj_priv; | ||
| 4585 | int ret; | ||
| 4586 | |||
| 4587 | if (dev_priv->pwrctx) { | ||
| 4588 | obj_priv = dev_priv->pwrctx->driver_private; | ||
| 4589 | } else { | ||
| 4590 | pwrctx = drm_gem_object_alloc(dev, 4096); | ||
| 4591 | if (!pwrctx) { | ||
| 4592 | DRM_DEBUG("failed to alloc power context, " | ||
| 4593 | "RC6 disabled\n"); | ||
| 4594 | goto out; | ||
| 4595 | } | ||
| 4596 | |||
| 4597 | ret = i915_gem_object_pin(pwrctx, 4096); | ||
| 4598 | if (ret) { | ||
| 4599 | DRM_ERROR("failed to pin power context: %d\n", | ||
| 4600 | ret); | ||
| 4601 | drm_gem_object_unreference(pwrctx); | ||
| 4602 | goto out; | ||
| 4603 | } | ||
| 4604 | |||
| 4605 | i915_gem_object_set_to_gtt_domain(pwrctx, 1); | ||
| 4606 | |||
| 4607 | dev_priv->pwrctx = pwrctx; | ||
| 4608 | obj_priv = pwrctx->driver_private; | ||
| 4609 | } | ||
| 4610 | |||
| 4611 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset | PWRCTX_EN); | ||
| 4612 | I915_WRITE(MCHBAR_RENDER_STANDBY, | ||
| 4613 | I915_READ(MCHBAR_RENDER_STANDBY) & ~RCX_SW_EXIT); | ||
| 4614 | } | ||
| 4615 | |||
| 4616 | out: | ||
| 4617 | return; | ||
| 4299 | } | 4618 | } |
| 4300 | 4619 | ||
| 4301 | /* Set up chip specific display functions */ | 4620 | /* Set up chip specific display functions */ |
| @@ -4304,8 +4623,8 @@ static void intel_init_display(struct drm_device *dev) | |||
| 4304 | struct drm_i915_private *dev_priv = dev->dev_private; | 4623 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 4305 | 4624 | ||
| 4306 | /* We always want a DPMS function */ | 4625 | /* We always want a DPMS function */ |
| 4307 | if (IS_IGDNG(dev)) | 4626 | if (IS_IRONLAKE(dev)) |
| 4308 | dev_priv->display.dpms = igdng_crtc_dpms; | 4627 | dev_priv->display.dpms = ironlake_crtc_dpms; |
| 4309 | else | 4628 | else |
| 4310 | dev_priv->display.dpms = i9xx_crtc_dpms; | 4629 | dev_priv->display.dpms = i9xx_crtc_dpms; |
| 4311 | 4630 | ||
| @@ -4324,13 +4643,13 @@ static void intel_init_display(struct drm_device *dev) | |||
| 4324 | } | 4643 | } |
| 4325 | 4644 | ||
| 4326 | /* Returns the core display clock speed */ | 4645 | /* Returns the core display clock speed */ |
| 4327 | if (IS_I945G(dev)) | 4646 | if (IS_I945G(dev) || (IS_G33(dev) && ! IS_PINEVIEW_M(dev))) |
| 4328 | dev_priv->display.get_display_clock_speed = | 4647 | dev_priv->display.get_display_clock_speed = |
| 4329 | i945_get_display_clock_speed; | 4648 | i945_get_display_clock_speed; |
| 4330 | else if (IS_I915G(dev)) | 4649 | else if (IS_I915G(dev)) |
| 4331 | dev_priv->display.get_display_clock_speed = | 4650 | dev_priv->display.get_display_clock_speed = |
| 4332 | i915_get_display_clock_speed; | 4651 | i915_get_display_clock_speed; |
| 4333 | else if (IS_I945GM(dev) || IS_845G(dev) || IS_IGDGM(dev)) | 4652 | else if (IS_I945GM(dev) || IS_845G(dev) || IS_PINEVIEW_M(dev)) |
| 4334 | dev_priv->display.get_display_clock_speed = | 4653 | dev_priv->display.get_display_clock_speed = |
| 4335 | i9xx_misc_get_display_clock_speed; | 4654 | i9xx_misc_get_display_clock_speed; |
| 4336 | else if (IS_I915GM(dev)) | 4655 | else if (IS_I915GM(dev)) |
| @@ -4339,7 +4658,7 @@ static void intel_init_display(struct drm_device *dev) | |||
| 4339 | else if (IS_I865G(dev)) | 4658 | else if (IS_I865G(dev)) |
| 4340 | dev_priv->display.get_display_clock_speed = | 4659 | dev_priv->display.get_display_clock_speed = |
| 4341 | i865_get_display_clock_speed; | 4660 | i865_get_display_clock_speed; |
| 4342 | else if (IS_I855(dev)) | 4661 | else if (IS_I85X(dev)) |
| 4343 | dev_priv->display.get_display_clock_speed = | 4662 | dev_priv->display.get_display_clock_speed = |
| 4344 | i855_get_display_clock_speed; | 4663 | i855_get_display_clock_speed; |
| 4345 | else /* 852, 830 */ | 4664 | else /* 852, 830 */ |
| @@ -4347,7 +4666,7 @@ static void intel_init_display(struct drm_device *dev) | |||
| 4347 | i830_get_display_clock_speed; | 4666 | i830_get_display_clock_speed; |
| 4348 | 4667 | ||
| 4349 | /* For FIFO watermark updates */ | 4668 | /* For FIFO watermark updates */ |
| 4350 | if (IS_IGDNG(dev)) | 4669 | if (IS_IRONLAKE(dev)) |
| 4351 | dev_priv->display.update_wm = NULL; | 4670 | dev_priv->display.update_wm = NULL; |
| 4352 | else if (IS_G4X(dev)) | 4671 | else if (IS_G4X(dev)) |
| 4353 | dev_priv->display.update_wm = g4x_update_wm; | 4672 | dev_priv->display.update_wm = g4x_update_wm; |
| @@ -4403,7 +4722,7 @@ void intel_modeset_init(struct drm_device *dev) | |||
| 4403 | num_pipe = 2; | 4722 | num_pipe = 2; |
| 4404 | else | 4723 | else |
| 4405 | num_pipe = 1; | 4724 | num_pipe = 1; |
| 4406 | DRM_DEBUG("%d display pipe%s available.\n", | 4725 | DRM_DEBUG_KMS("%d display pipe%s available.\n", |
| 4407 | num_pipe, num_pipe > 1 ? "s" : ""); | 4726 | num_pipe, num_pipe > 1 ? "s" : ""); |
| 4408 | 4727 | ||
| 4409 | if (IS_I85X(dev)) | 4728 | if (IS_I85X(dev)) |
| @@ -4422,6 +4741,15 @@ void intel_modeset_init(struct drm_device *dev) | |||
| 4422 | INIT_WORK(&dev_priv->idle_work, intel_idle_update); | 4741 | INIT_WORK(&dev_priv->idle_work, intel_idle_update); |
| 4423 | setup_timer(&dev_priv->idle_timer, intel_gpu_idle_timer, | 4742 | setup_timer(&dev_priv->idle_timer, intel_gpu_idle_timer, |
| 4424 | (unsigned long)dev); | 4743 | (unsigned long)dev); |
| 4744 | |||
| 4745 | intel_setup_overlay(dev); | ||
| 4746 | |||
| 4747 | if (IS_PINEVIEW(dev) && !intel_get_cxsr_latency(IS_PINEVIEW_G(dev), | ||
| 4748 | dev_priv->fsb_freq, | ||
| 4749 | dev_priv->mem_freq)) | ||
| 4750 | DRM_INFO("failed to find known CxSR latency " | ||
| 4751 | "(found fsb freq %d, mem freq %d), disabling CxSR\n", | ||
| 4752 | dev_priv->fsb_freq, dev_priv->mem_freq); | ||
| 4425 | } | 4753 | } |
| 4426 | 4754 | ||
| 4427 | void intel_modeset_cleanup(struct drm_device *dev) | 4755 | void intel_modeset_cleanup(struct drm_device *dev) |
| @@ -4445,11 +4773,21 @@ void intel_modeset_cleanup(struct drm_device *dev) | |||
| 4445 | intel_increase_renderclock(dev, false); | 4773 | intel_increase_renderclock(dev, false); |
| 4446 | del_timer_sync(&dev_priv->idle_timer); | 4774 | del_timer_sync(&dev_priv->idle_timer); |
| 4447 | 4775 | ||
| 4448 | mutex_unlock(&dev->struct_mutex); | ||
| 4449 | |||
| 4450 | if (dev_priv->display.disable_fbc) | 4776 | if (dev_priv->display.disable_fbc) |
| 4451 | dev_priv->display.disable_fbc(dev); | 4777 | dev_priv->display.disable_fbc(dev); |
| 4452 | 4778 | ||
| 4779 | if (dev_priv->pwrctx) { | ||
| 4780 | struct drm_i915_gem_object *obj_priv; | ||
| 4781 | |||
| 4782 | obj_priv = dev_priv->pwrctx->driver_private; | ||
| 4783 | I915_WRITE(PWRCTXA, obj_priv->gtt_offset &~ PWRCTX_EN); | ||
| 4784 | I915_READ(PWRCTXA); | ||
| 4785 | i915_gem_object_unpin(dev_priv->pwrctx); | ||
| 4786 | drm_gem_object_unreference(dev_priv->pwrctx); | ||
| 4787 | } | ||
| 4788 | |||
| 4789 | mutex_unlock(&dev->struct_mutex); | ||
| 4790 | |||
| 4453 | drm_mode_config_cleanup(dev); | 4791 | drm_mode_config_cleanup(dev); |
| 4454 | } | 4792 | } |
| 4455 | 4793 | ||
diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c index d83447557f9b..4e7aa8b7b938 100644 --- a/drivers/gpu/drm/i915/intel_dp.c +++ b/drivers/gpu/drm/i915/intel_dp.c | |||
| @@ -33,7 +33,8 @@ | |||
| 33 | #include "intel_drv.h" | 33 | #include "intel_drv.h" |
| 34 | #include "i915_drm.h" | 34 | #include "i915_drm.h" |
| 35 | #include "i915_drv.h" | 35 | #include "i915_drv.h" |
| 36 | #include "intel_dp.h" | 36 | #include "drm_dp_helper.h" |
| 37 | |||
| 37 | 38 | ||
| 38 | #define DP_LINK_STATUS_SIZE 6 | 39 | #define DP_LINK_STATUS_SIZE 6 |
| 39 | #define DP_LINK_CHECK_TIMEOUT (10 * 1000) | 40 | #define DP_LINK_CHECK_TIMEOUT (10 * 1000) |
| @@ -223,8 +224,8 @@ intel_dp_aux_ch(struct intel_output *intel_output, | |||
| 223 | */ | 224 | */ |
| 224 | if (IS_eDP(intel_output)) | 225 | if (IS_eDP(intel_output)) |
| 225 | aux_clock_divider = 225; /* eDP input clock at 450Mhz */ | 226 | aux_clock_divider = 225; /* eDP input clock at 450Mhz */ |
| 226 | else if (IS_IGDNG(dev)) | 227 | else if (IS_IRONLAKE(dev)) |
| 227 | aux_clock_divider = 62; /* IGDNG: input clock fixed at 125Mhz */ | 228 | aux_clock_divider = 62; /* IRL input clock fixed at 125Mhz */ |
| 228 | else | 229 | else |
| 229 | aux_clock_divider = intel_hrawclk(dev) / 2; | 230 | aux_clock_divider = intel_hrawclk(dev) / 2; |
| 230 | 231 | ||
| @@ -282,7 +283,7 @@ intel_dp_aux_ch(struct intel_output *intel_output, | |||
| 282 | /* Timeouts occur when the device isn't connected, so they're | 283 | /* Timeouts occur when the device isn't connected, so they're |
| 283 | * "normal" -- don't fill the kernel log with these */ | 284 | * "normal" -- don't fill the kernel log with these */ |
| 284 | if (status & DP_AUX_CH_CTL_TIME_OUT_ERROR) { | 285 | if (status & DP_AUX_CH_CTL_TIME_OUT_ERROR) { |
| 285 | DRM_DEBUG("dp_aux_ch timeout status 0x%08x\n", status); | 286 | DRM_DEBUG_KMS("dp_aux_ch timeout status 0x%08x\n", status); |
| 286 | return -ETIMEDOUT; | 287 | return -ETIMEDOUT; |
| 287 | } | 288 | } |
| 288 | 289 | ||
| @@ -382,17 +383,77 @@ intel_dp_aux_native_read(struct intel_output *intel_output, | |||
| 382 | } | 383 | } |
| 383 | 384 | ||
| 384 | static int | 385 | static int |
| 385 | intel_dp_i2c_aux_ch(struct i2c_adapter *adapter, | 386 | intel_dp_i2c_aux_ch(struct i2c_adapter *adapter, int mode, |
| 386 | uint8_t *send, int send_bytes, | 387 | uint8_t write_byte, uint8_t *read_byte) |
| 387 | uint8_t *recv, int recv_bytes) | ||
| 388 | { | 388 | { |
| 389 | struct i2c_algo_dp_aux_data *algo_data = adapter->algo_data; | ||
| 389 | struct intel_dp_priv *dp_priv = container_of(adapter, | 390 | struct intel_dp_priv *dp_priv = container_of(adapter, |
| 390 | struct intel_dp_priv, | 391 | struct intel_dp_priv, |
| 391 | adapter); | 392 | adapter); |
| 392 | struct intel_output *intel_output = dp_priv->intel_output; | 393 | struct intel_output *intel_output = dp_priv->intel_output; |
| 394 | uint16_t address = algo_data->address; | ||
| 395 | uint8_t msg[5]; | ||
| 396 | uint8_t reply[2]; | ||
| 397 | int msg_bytes; | ||
| 398 | int reply_bytes; | ||
| 399 | int ret; | ||
| 400 | |||
| 401 | /* Set up the command byte */ | ||
| 402 | if (mode & MODE_I2C_READ) | ||
| 403 | msg[0] = AUX_I2C_READ << 4; | ||
| 404 | else | ||
| 405 | msg[0] = AUX_I2C_WRITE << 4; | ||
| 406 | |||
| 407 | if (!(mode & MODE_I2C_STOP)) | ||
| 408 | msg[0] |= AUX_I2C_MOT << 4; | ||
| 393 | 409 | ||
| 394 | return intel_dp_aux_ch(intel_output, | 410 | msg[1] = address >> 8; |
| 395 | send, send_bytes, recv, recv_bytes); | 411 | msg[2] = address; |
| 412 | |||
| 413 | switch (mode) { | ||
| 414 | case MODE_I2C_WRITE: | ||
| 415 | msg[3] = 0; | ||
| 416 | msg[4] = write_byte; | ||
| 417 | msg_bytes = 5; | ||
| 418 | reply_bytes = 1; | ||
| 419 | break; | ||
| 420 | case MODE_I2C_READ: | ||
| 421 | msg[3] = 0; | ||
| 422 | msg_bytes = 4; | ||
| 423 | reply_bytes = 2; | ||
| 424 | break; | ||
| 425 | default: | ||
| 426 | msg_bytes = 3; | ||
| 427 | reply_bytes = 1; | ||
| 428 | break; | ||
| 429 | } | ||
| 430 | |||
| 431 | for (;;) { | ||
| 432 | ret = intel_dp_aux_ch(intel_output, | ||
| 433 | msg, msg_bytes, | ||
| 434 | reply, reply_bytes); | ||
| 435 | if (ret < 0) { | ||
| 436 | DRM_DEBUG_KMS("aux_ch failed %d\n", ret); | ||
| 437 | return ret; | ||
| 438 | } | ||
| 439 | switch (reply[0] & AUX_I2C_REPLY_MASK) { | ||
| 440 | case AUX_I2C_REPLY_ACK: | ||
| 441 | if (mode == MODE_I2C_READ) { | ||
| 442 | *read_byte = reply[1]; | ||
| 443 | } | ||
| 444 | return reply_bytes - 1; | ||
| 445 | case AUX_I2C_REPLY_NACK: | ||
| 446 | DRM_DEBUG_KMS("aux_ch nack\n"); | ||
| 447 | return -EREMOTEIO; | ||
| 448 | case AUX_I2C_REPLY_DEFER: | ||
| 449 | DRM_DEBUG_KMS("aux_ch defer\n"); | ||
| 450 | udelay(100); | ||
| 451 | break; | ||
| 452 | default: | ||
| 453 | DRM_ERROR("aux_ch invalid reply 0x%02x\n", reply[0]); | ||
| 454 | return -EREMOTEIO; | ||
| 455 | } | ||
| 456 | } | ||
| 396 | } | 457 | } |
| 397 | 458 | ||
| 398 | static int | 459 | static int |
| @@ -435,7 +496,8 @@ intel_dp_mode_fixup(struct drm_encoder *encoder, struct drm_display_mode *mode, | |||
| 435 | dp_priv->link_bw = bws[clock]; | 496 | dp_priv->link_bw = bws[clock]; |
| 436 | dp_priv->lane_count = lane_count; | 497 | dp_priv->lane_count = lane_count; |
| 437 | adjusted_mode->clock = intel_dp_link_clock(dp_priv->link_bw); | 498 | adjusted_mode->clock = intel_dp_link_clock(dp_priv->link_bw); |
| 438 | DRM_DEBUG("Display port link bw %02x lane count %d clock %d\n", | 499 | DRM_DEBUG_KMS("Display port link bw %02x lane " |
| 500 | "count %d clock %d\n", | ||
| 439 | dp_priv->link_bw, dp_priv->lane_count, | 501 | dp_priv->link_bw, dp_priv->lane_count, |
| 440 | adjusted_mode->clock); | 502 | adjusted_mode->clock); |
| 441 | return true; | 503 | return true; |
| @@ -514,7 +576,7 @@ intel_dp_set_m_n(struct drm_crtc *crtc, struct drm_display_mode *mode, | |||
| 514 | intel_dp_compute_m_n(3, lane_count, | 576 | intel_dp_compute_m_n(3, lane_count, |
| 515 | mode->clock, adjusted_mode->clock, &m_n); | 577 | mode->clock, adjusted_mode->clock, &m_n); |
| 516 | 578 | ||
| 517 | if (IS_IGDNG(dev)) { | 579 | if (IS_IRONLAKE(dev)) { |
| 518 | if (intel_crtc->pipe == 0) { | 580 | if (intel_crtc->pipe == 0) { |
| 519 | I915_WRITE(TRANSA_DATA_M1, | 581 | I915_WRITE(TRANSA_DATA_M1, |
| 520 | ((m_n.tu - 1) << PIPE_GMCH_DATA_M_TU_SIZE_SHIFT) | | 582 | ((m_n.tu - 1) << PIPE_GMCH_DATA_M_TU_SIZE_SHIFT) | |
| @@ -606,23 +668,23 @@ intel_dp_mode_set(struct drm_encoder *encoder, struct drm_display_mode *mode, | |||
| 606 | } | 668 | } |
| 607 | } | 669 | } |
| 608 | 670 | ||
| 609 | static void igdng_edp_backlight_on (struct drm_device *dev) | 671 | static void ironlake_edp_backlight_on (struct drm_device *dev) |
| 610 | { | 672 | { |
| 611 | struct drm_i915_private *dev_priv = dev->dev_private; | 673 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 612 | u32 pp; | 674 | u32 pp; |
| 613 | 675 | ||
| 614 | DRM_DEBUG("\n"); | 676 | DRM_DEBUG_KMS("\n"); |
| 615 | pp = I915_READ(PCH_PP_CONTROL); | 677 | pp = I915_READ(PCH_PP_CONTROL); |
| 616 | pp |= EDP_BLC_ENABLE; | 678 | pp |= EDP_BLC_ENABLE; |
| 617 | I915_WRITE(PCH_PP_CONTROL, pp); | 679 | I915_WRITE(PCH_PP_CONTROL, pp); |
| 618 | } | 680 | } |
| 619 | 681 | ||
| 620 | static void igdng_edp_backlight_off (struct drm_device *dev) | 682 | static void ironlake_edp_backlight_off (struct drm_device *dev) |
| 621 | { | 683 | { |
| 622 | struct drm_i915_private *dev_priv = dev->dev_private; | 684 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 623 | u32 pp; | 685 | u32 pp; |
| 624 | 686 | ||
| 625 | DRM_DEBUG("\n"); | 687 | DRM_DEBUG_KMS("\n"); |
| 626 | pp = I915_READ(PCH_PP_CONTROL); | 688 | pp = I915_READ(PCH_PP_CONTROL); |
| 627 | pp &= ~EDP_BLC_ENABLE; | 689 | pp &= ~EDP_BLC_ENABLE; |
| 628 | I915_WRITE(PCH_PP_CONTROL, pp); | 690 | I915_WRITE(PCH_PP_CONTROL, pp); |
| @@ -641,13 +703,13 @@ intel_dp_dpms(struct drm_encoder *encoder, int mode) | |||
| 641 | if (dp_reg & DP_PORT_EN) { | 703 | if (dp_reg & DP_PORT_EN) { |
| 642 | intel_dp_link_down(intel_output, dp_priv->DP); | 704 | intel_dp_link_down(intel_output, dp_priv->DP); |
| 643 | if (IS_eDP(intel_output)) | 705 | if (IS_eDP(intel_output)) |
| 644 | igdng_edp_backlight_off(dev); | 706 | ironlake_edp_backlight_off(dev); |
| 645 | } | 707 | } |
| 646 | } else { | 708 | } else { |
| 647 | if (!(dp_reg & DP_PORT_EN)) { | 709 | if (!(dp_reg & DP_PORT_EN)) { |
| 648 | intel_dp_link_train(intel_output, dp_priv->DP, dp_priv->link_configuration); | 710 | intel_dp_link_train(intel_output, dp_priv->DP, dp_priv->link_configuration); |
| 649 | if (IS_eDP(intel_output)) | 711 | if (IS_eDP(intel_output)) |
| 650 | igdng_edp_backlight_on(dev); | 712 | ironlake_edp_backlight_on(dev); |
| 651 | } | 713 | } |
| 652 | } | 714 | } |
| 653 | dp_priv->dpms_mode = mode; | 715 | dp_priv->dpms_mode = mode; |
| @@ -1010,7 +1072,7 @@ intel_dp_link_down(struct intel_output *intel_output, uint32_t DP) | |||
| 1010 | struct drm_i915_private *dev_priv = dev->dev_private; | 1072 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 1011 | struct intel_dp_priv *dp_priv = intel_output->dev_priv; | 1073 | struct intel_dp_priv *dp_priv = intel_output->dev_priv; |
| 1012 | 1074 | ||
| 1013 | DRM_DEBUG("\n"); | 1075 | DRM_DEBUG_KMS("\n"); |
| 1014 | 1076 | ||
| 1015 | if (IS_eDP(intel_output)) { | 1077 | if (IS_eDP(intel_output)) { |
| 1016 | DP &= ~DP_PLL_ENABLE; | 1078 | DP &= ~DP_PLL_ENABLE; |
| @@ -1071,7 +1133,7 @@ intel_dp_check_link_status(struct intel_output *intel_output) | |||
| 1071 | } | 1133 | } |
| 1072 | 1134 | ||
| 1073 | static enum drm_connector_status | 1135 | static enum drm_connector_status |
| 1074 | igdng_dp_detect(struct drm_connector *connector) | 1136 | ironlake_dp_detect(struct drm_connector *connector) |
| 1075 | { | 1137 | { |
| 1076 | struct intel_output *intel_output = to_intel_output(connector); | 1138 | struct intel_output *intel_output = to_intel_output(connector); |
| 1077 | struct intel_dp_priv *dp_priv = intel_output->dev_priv; | 1139 | struct intel_dp_priv *dp_priv = intel_output->dev_priv; |
| @@ -1106,8 +1168,8 @@ intel_dp_detect(struct drm_connector *connector) | |||
| 1106 | 1168 | ||
| 1107 | dp_priv->has_audio = false; | 1169 | dp_priv->has_audio = false; |
| 1108 | 1170 | ||
| 1109 | if (IS_IGDNG(dev)) | 1171 | if (IS_IRONLAKE(dev)) |
| 1110 | return igdng_dp_detect(connector); | 1172 | return ironlake_dp_detect(connector); |
| 1111 | 1173 | ||
| 1112 | temp = I915_READ(PORT_HOTPLUG_EN); | 1174 | temp = I915_READ(PORT_HOTPLUG_EN); |
| 1113 | 1175 | ||
| @@ -1227,7 +1289,53 @@ intel_dp_hot_plug(struct intel_output *intel_output) | |||
| 1227 | if (dp_priv->dpms_mode == DRM_MODE_DPMS_ON) | 1289 | if (dp_priv->dpms_mode == DRM_MODE_DPMS_ON) |
| 1228 | intel_dp_check_link_status(intel_output); | 1290 | intel_dp_check_link_status(intel_output); |
| 1229 | } | 1291 | } |
| 1230 | 1292 | /* | |
| 1293 | * Enumerate the child dev array parsed from VBT to check whether | ||
| 1294 | * the given DP is present. | ||
| 1295 | * If it is present, return 1. | ||
| 1296 | * If it is not present, return false. | ||
| 1297 | * If no child dev is parsed from VBT, it is assumed that the given | ||
| 1298 | * DP is present. | ||
| 1299 | */ | ||
| 1300 | static int dp_is_present_in_vbt(struct drm_device *dev, int dp_reg) | ||
| 1301 | { | ||
| 1302 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 1303 | struct child_device_config *p_child; | ||
| 1304 | int i, dp_port, ret; | ||
| 1305 | |||
| 1306 | if (!dev_priv->child_dev_num) | ||
| 1307 | return 1; | ||
| 1308 | |||
| 1309 | dp_port = 0; | ||
| 1310 | if (dp_reg == DP_B || dp_reg == PCH_DP_B) | ||
| 1311 | dp_port = PORT_IDPB; | ||
| 1312 | else if (dp_reg == DP_C || dp_reg == PCH_DP_C) | ||
| 1313 | dp_port = PORT_IDPC; | ||
| 1314 | else if (dp_reg == DP_D || dp_reg == PCH_DP_D) | ||
| 1315 | dp_port = PORT_IDPD; | ||
| 1316 | |||
| 1317 | ret = 0; | ||
| 1318 | for (i = 0; i < dev_priv->child_dev_num; i++) { | ||
| 1319 | p_child = dev_priv->child_dev + i; | ||
| 1320 | /* | ||
| 1321 | * If the device type is not DP, continue. | ||
| 1322 | */ | ||
| 1323 | if (p_child->device_type != DEVICE_TYPE_DP && | ||
| 1324 | p_child->device_type != DEVICE_TYPE_eDP) | ||
| 1325 | continue; | ||
| 1326 | /* Find the eDP port */ | ||
| 1327 | if (dp_reg == DP_A && p_child->device_type == DEVICE_TYPE_eDP) { | ||
| 1328 | ret = 1; | ||
| 1329 | break; | ||
| 1330 | } | ||
| 1331 | /* Find the DP port */ | ||
| 1332 | if (p_child->dvo_port == dp_port) { | ||
| 1333 | ret = 1; | ||
| 1334 | break; | ||
| 1335 | } | ||
| 1336 | } | ||
| 1337 | return ret; | ||
| 1338 | } | ||
| 1231 | void | 1339 | void |
| 1232 | intel_dp_init(struct drm_device *dev, int output_reg) | 1340 | intel_dp_init(struct drm_device *dev, int output_reg) |
| 1233 | { | 1341 | { |
| @@ -1237,6 +1345,10 @@ intel_dp_init(struct drm_device *dev, int output_reg) | |||
| 1237 | struct intel_dp_priv *dp_priv; | 1345 | struct intel_dp_priv *dp_priv; |
| 1238 | const char *name = NULL; | 1346 | const char *name = NULL; |
| 1239 | 1347 | ||
| 1348 | if (!dp_is_present_in_vbt(dev, output_reg)) { | ||
| 1349 | DRM_DEBUG_KMS("DP is not present. Ignore it\n"); | ||
| 1350 | return; | ||
| 1351 | } | ||
| 1240 | intel_output = kcalloc(sizeof(struct intel_output) + | 1352 | intel_output = kcalloc(sizeof(struct intel_output) + |
| 1241 | sizeof(struct intel_dp_priv), 1, GFP_KERNEL); | 1353 | sizeof(struct intel_dp_priv), 1, GFP_KERNEL); |
| 1242 | if (!intel_output) | 1354 | if (!intel_output) |
| @@ -1254,11 +1366,11 @@ intel_dp_init(struct drm_device *dev, int output_reg) | |||
| 1254 | else | 1366 | else |
| 1255 | intel_output->type = INTEL_OUTPUT_DISPLAYPORT; | 1367 | intel_output->type = INTEL_OUTPUT_DISPLAYPORT; |
| 1256 | 1368 | ||
| 1257 | if (output_reg == DP_B) | 1369 | if (output_reg == DP_B || output_reg == PCH_DP_B) |
| 1258 | intel_output->clone_mask = (1 << INTEL_DP_B_CLONE_BIT); | 1370 | intel_output->clone_mask = (1 << INTEL_DP_B_CLONE_BIT); |
| 1259 | else if (output_reg == DP_C) | 1371 | else if (output_reg == DP_C || output_reg == PCH_DP_C) |
| 1260 | intel_output->clone_mask = (1 << INTEL_DP_C_CLONE_BIT); | 1372 | intel_output->clone_mask = (1 << INTEL_DP_C_CLONE_BIT); |
| 1261 | else if (output_reg == DP_D) | 1373 | else if (output_reg == DP_D || output_reg == PCH_DP_D) |
| 1262 | intel_output->clone_mask = (1 << INTEL_DP_D_CLONE_BIT); | 1374 | intel_output->clone_mask = (1 << INTEL_DP_D_CLONE_BIT); |
| 1263 | 1375 | ||
| 1264 | if (IS_eDP(intel_output)) { | 1376 | if (IS_eDP(intel_output)) { |
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h index ef61fe9507e2..a51573da1ff6 100644 --- a/drivers/gpu/drm/i915/intel_drv.h +++ b/drivers/gpu/drm/i915/intel_drv.h | |||
| @@ -110,6 +110,32 @@ struct intel_output { | |||
| 110 | int clone_mask; | 110 | int clone_mask; |
| 111 | }; | 111 | }; |
| 112 | 112 | ||
| 113 | struct intel_crtc; | ||
| 114 | struct intel_overlay { | ||
| 115 | struct drm_device *dev; | ||
| 116 | struct intel_crtc *crtc; | ||
| 117 | struct drm_i915_gem_object *vid_bo; | ||
| 118 | struct drm_i915_gem_object *old_vid_bo; | ||
| 119 | int active; | ||
| 120 | int pfit_active; | ||
| 121 | u32 pfit_vscale_ratio; /* shifted-point number, (1<<12) == 1.0 */ | ||
| 122 | u32 color_key; | ||
| 123 | u32 brightness, contrast, saturation; | ||
| 124 | u32 old_xscale, old_yscale; | ||
| 125 | /* register access */ | ||
| 126 | u32 flip_addr; | ||
| 127 | struct drm_i915_gem_object *reg_bo; | ||
| 128 | void *virt_addr; | ||
| 129 | /* flip handling */ | ||
| 130 | uint32_t last_flip_req; | ||
| 131 | int hw_wedged; | ||
| 132 | #define HW_WEDGED 1 | ||
| 133 | #define NEEDS_WAIT_FOR_FLIP 2 | ||
| 134 | #define RELEASE_OLD_VID 3 | ||
| 135 | #define SWITCH_OFF_STAGE_1 4 | ||
| 136 | #define SWITCH_OFF_STAGE_2 5 | ||
| 137 | }; | ||
| 138 | |||
| 113 | struct intel_crtc { | 139 | struct intel_crtc { |
| 114 | struct drm_crtc base; | 140 | struct drm_crtc base; |
| 115 | enum pipe pipe; | 141 | enum pipe pipe; |
| @@ -121,6 +147,8 @@ struct intel_crtc { | |||
| 121 | bool busy; /* is scanout buffer being updated frequently? */ | 147 | bool busy; /* is scanout buffer being updated frequently? */ |
| 122 | struct timer_list idle_timer; | 148 | struct timer_list idle_timer; |
| 123 | bool lowfreq_avail; | 149 | bool lowfreq_avail; |
| 150 | struct intel_overlay *overlay; | ||
| 151 | struct intel_unpin_work *unpin_work; | ||
| 124 | }; | 152 | }; |
| 125 | 153 | ||
| 126 | #define to_intel_crtc(x) container_of(x, struct intel_crtc, base) | 154 | #define to_intel_crtc(x) container_of(x, struct intel_crtc, base) |
| @@ -134,6 +162,8 @@ void intel_i2c_destroy(struct i2c_adapter *adapter); | |||
| 134 | int intel_ddc_get_modes(struct intel_output *intel_output); | 162 | int intel_ddc_get_modes(struct intel_output *intel_output); |
| 135 | extern bool intel_ddc_probe(struct intel_output *intel_output); | 163 | extern bool intel_ddc_probe(struct intel_output *intel_output); |
| 136 | void intel_i2c_quirk_set(struct drm_device *dev, bool enable); | 164 | void intel_i2c_quirk_set(struct drm_device *dev, bool enable); |
| 165 | void intel_i2c_reset_gmbus(struct drm_device *dev); | ||
| 166 | |||
| 137 | extern void intel_crt_init(struct drm_device *dev); | 167 | extern void intel_crt_init(struct drm_device *dev); |
| 138 | extern void intel_hdmi_init(struct drm_device *dev, int sdvox_reg); | 168 | extern void intel_hdmi_init(struct drm_device *dev, int sdvox_reg); |
| 139 | extern bool intel_sdvo_init(struct drm_device *dev, int output_device); | 169 | extern bool intel_sdvo_init(struct drm_device *dev, int output_device); |
| @@ -148,6 +178,7 @@ intel_dp_set_m_n(struct drm_crtc *crtc, struct drm_display_mode *mode, | |||
| 148 | extern void intel_edp_link_config (struct intel_output *, int *, int *); | 178 | extern void intel_edp_link_config (struct intel_output *, int *, int *); |
| 149 | 179 | ||
| 150 | 180 | ||
| 181 | extern int intel_panel_fitter_pipe (struct drm_device *dev); | ||
| 151 | extern void intel_crtc_load_lut(struct drm_crtc *crtc); | 182 | extern void intel_crtc_load_lut(struct drm_crtc *crtc); |
| 152 | extern void intel_encoder_prepare (struct drm_encoder *encoder); | 183 | extern void intel_encoder_prepare (struct drm_encoder *encoder); |
| 153 | extern void intel_encoder_commit (struct drm_encoder *encoder); | 184 | extern void intel_encoder_commit (struct drm_encoder *encoder); |
| @@ -177,10 +208,23 @@ extern void intel_crtc_fb_gamma_set(struct drm_crtc *crtc, u16 red, u16 green, | |||
| 177 | u16 blue, int regno); | 208 | u16 blue, int regno); |
| 178 | extern void intel_crtc_fb_gamma_get(struct drm_crtc *crtc, u16 *red, u16 *green, | 209 | extern void intel_crtc_fb_gamma_get(struct drm_crtc *crtc, u16 *red, u16 *green, |
| 179 | u16 *blue, int regno); | 210 | u16 *blue, int regno); |
| 211 | extern void intel_init_clock_gating(struct drm_device *dev); | ||
| 180 | 212 | ||
| 181 | extern int intel_framebuffer_create(struct drm_device *dev, | 213 | extern int intel_framebuffer_create(struct drm_device *dev, |
| 182 | struct drm_mode_fb_cmd *mode_cmd, | 214 | struct drm_mode_fb_cmd *mode_cmd, |
| 183 | struct drm_framebuffer **fb, | 215 | struct drm_framebuffer **fb, |
| 184 | struct drm_gem_object *obj); | 216 | struct drm_gem_object *obj); |
| 185 | 217 | ||
| 218 | extern void intel_prepare_page_flip(struct drm_device *dev, int plane); | ||
| 219 | extern void intel_finish_page_flip(struct drm_device *dev, int pipe); | ||
| 220 | |||
| 221 | extern void intel_setup_overlay(struct drm_device *dev); | ||
| 222 | extern void intel_cleanup_overlay(struct drm_device *dev); | ||
| 223 | extern int intel_overlay_switch_off(struct intel_overlay *overlay); | ||
| 224 | extern int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay, | ||
| 225 | int interruptible); | ||
| 226 | extern int intel_overlay_put_image(struct drm_device *dev, void *data, | ||
| 227 | struct drm_file *file_priv); | ||
| 228 | extern int intel_overlay_attrs(struct drm_device *dev, void *data, | ||
| 229 | struct drm_file *file_priv); | ||
| 186 | #endif /* __INTEL_DRV_H__ */ | 230 | #endif /* __INTEL_DRV_H__ */ |
diff --git a/drivers/gpu/drm/i915/intel_fb.c b/drivers/gpu/drm/i915/intel_fb.c index 40fcf6fdef38..371d753e362b 100644 --- a/drivers/gpu/drm/i915/intel_fb.c +++ b/drivers/gpu/drm/i915/intel_fb.c | |||
| @@ -230,8 +230,9 @@ static int intelfb_create(struct drm_device *dev, uint32_t fb_width, | |||
| 230 | par->intel_fb = intel_fb; | 230 | par->intel_fb = intel_fb; |
| 231 | 231 | ||
| 232 | /* To allow resizeing without swapping buffers */ | 232 | /* To allow resizeing without swapping buffers */ |
| 233 | DRM_DEBUG("allocated %dx%d fb: 0x%08x, bo %p\n", intel_fb->base.width, | 233 | DRM_DEBUG_KMS("allocated %dx%d fb: 0x%08x, bo %p\n", |
| 234 | intel_fb->base.height, obj_priv->gtt_offset, fbo); | 234 | intel_fb->base.width, intel_fb->base.height, |
| 235 | obj_priv->gtt_offset, fbo); | ||
| 235 | 236 | ||
| 236 | mutex_unlock(&dev->struct_mutex); | 237 | mutex_unlock(&dev->struct_mutex); |
| 237 | return 0; | 238 | return 0; |
| @@ -249,7 +250,7 @@ int intelfb_probe(struct drm_device *dev) | |||
| 249 | { | 250 | { |
| 250 | int ret; | 251 | int ret; |
| 251 | 252 | ||
| 252 | DRM_DEBUG("\n"); | 253 | DRM_DEBUG_KMS("\n"); |
| 253 | ret = drm_fb_helper_single_fb_probe(dev, 32, intelfb_create); | 254 | ret = drm_fb_helper_single_fb_probe(dev, 32, intelfb_create); |
| 254 | return ret; | 255 | return ret; |
| 255 | } | 256 | } |
diff --git a/drivers/gpu/drm/i915/intel_hdmi.c b/drivers/gpu/drm/i915/intel_hdmi.c index c33451aec1bd..f04dbbe7d400 100644 --- a/drivers/gpu/drm/i915/intel_hdmi.c +++ b/drivers/gpu/drm/i915/intel_hdmi.c | |||
| @@ -82,7 +82,7 @@ static void intel_hdmi_dpms(struct drm_encoder *encoder, int mode) | |||
| 82 | /* HW workaround, need to toggle enable bit off and on for 12bpc, but | 82 | /* HW workaround, need to toggle enable bit off and on for 12bpc, but |
| 83 | * we do this anyway which shows more stable in testing. | 83 | * we do this anyway which shows more stable in testing. |
| 84 | */ | 84 | */ |
| 85 | if (IS_IGDNG(dev)) { | 85 | if (IS_IRONLAKE(dev)) { |
| 86 | I915_WRITE(hdmi_priv->sdvox_reg, temp & ~SDVO_ENABLE); | 86 | I915_WRITE(hdmi_priv->sdvox_reg, temp & ~SDVO_ENABLE); |
| 87 | POSTING_READ(hdmi_priv->sdvox_reg); | 87 | POSTING_READ(hdmi_priv->sdvox_reg); |
| 88 | } | 88 | } |
| @@ -99,7 +99,7 @@ static void intel_hdmi_dpms(struct drm_encoder *encoder, int mode) | |||
| 99 | /* HW workaround, need to write this twice for issue that may result | 99 | /* HW workaround, need to write this twice for issue that may result |
| 100 | * in first write getting masked. | 100 | * in first write getting masked. |
| 101 | */ | 101 | */ |
| 102 | if (IS_IGDNG(dev)) { | 102 | if (IS_IRONLAKE(dev)) { |
| 103 | I915_WRITE(hdmi_priv->sdvox_reg, temp); | 103 | I915_WRITE(hdmi_priv->sdvox_reg, temp); |
| 104 | POSTING_READ(hdmi_priv->sdvox_reg); | 104 | POSTING_READ(hdmi_priv->sdvox_reg); |
| 105 | } | 105 | } |
| @@ -225,7 +225,52 @@ static const struct drm_encoder_funcs intel_hdmi_enc_funcs = { | |||
| 225 | .destroy = intel_hdmi_enc_destroy, | 225 | .destroy = intel_hdmi_enc_destroy, |
| 226 | }; | 226 | }; |
| 227 | 227 | ||
| 228 | 228 | /* | |
| 229 | * Enumerate the child dev array parsed from VBT to check whether | ||
| 230 | * the given HDMI is present. | ||
| 231 | * If it is present, return 1. | ||
| 232 | * If it is not present, return false. | ||
| 233 | * If no child dev is parsed from VBT, it assumes that the given | ||
| 234 | * HDMI is present. | ||
| 235 | */ | ||
| 236 | static int hdmi_is_present_in_vbt(struct drm_device *dev, int hdmi_reg) | ||
| 237 | { | ||
| 238 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 239 | struct child_device_config *p_child; | ||
| 240 | int i, hdmi_port, ret; | ||
| 241 | |||
| 242 | if (!dev_priv->child_dev_num) | ||
| 243 | return 1; | ||
| 244 | |||
| 245 | if (hdmi_reg == SDVOB) | ||
| 246 | hdmi_port = DVO_B; | ||
| 247 | else if (hdmi_reg == SDVOC) | ||
| 248 | hdmi_port = DVO_C; | ||
| 249 | else if (hdmi_reg == HDMIB) | ||
| 250 | hdmi_port = DVO_B; | ||
| 251 | else if (hdmi_reg == HDMIC) | ||
| 252 | hdmi_port = DVO_C; | ||
| 253 | else if (hdmi_reg == HDMID) | ||
| 254 | hdmi_port = DVO_D; | ||
| 255 | else | ||
| 256 | return 0; | ||
| 257 | |||
| 258 | ret = 0; | ||
| 259 | for (i = 0; i < dev_priv->child_dev_num; i++) { | ||
| 260 | p_child = dev_priv->child_dev + i; | ||
| 261 | /* | ||
| 262 | * If the device type is not HDMI, continue. | ||
| 263 | */ | ||
| 264 | if (p_child->device_type != DEVICE_TYPE_HDMI) | ||
| 265 | continue; | ||
| 266 | /* Find the HDMI port */ | ||
| 267 | if (p_child->dvo_port == hdmi_port) { | ||
| 268 | ret = 1; | ||
| 269 | break; | ||
| 270 | } | ||
| 271 | } | ||
| 272 | return ret; | ||
| 273 | } | ||
| 229 | void intel_hdmi_init(struct drm_device *dev, int sdvox_reg) | 274 | void intel_hdmi_init(struct drm_device *dev, int sdvox_reg) |
| 230 | { | 275 | { |
| 231 | struct drm_i915_private *dev_priv = dev->dev_private; | 276 | struct drm_i915_private *dev_priv = dev->dev_private; |
| @@ -233,6 +278,10 @@ void intel_hdmi_init(struct drm_device *dev, int sdvox_reg) | |||
| 233 | struct intel_output *intel_output; | 278 | struct intel_output *intel_output; |
| 234 | struct intel_hdmi_priv *hdmi_priv; | 279 | struct intel_hdmi_priv *hdmi_priv; |
| 235 | 280 | ||
| 281 | if (!hdmi_is_present_in_vbt(dev, sdvox_reg)) { | ||
| 282 | DRM_DEBUG_KMS("HDMI is not present. Ignored it \n"); | ||
| 283 | return; | ||
| 284 | } | ||
| 236 | intel_output = kcalloc(sizeof(struct intel_output) + | 285 | intel_output = kcalloc(sizeof(struct intel_output) + |
| 237 | sizeof(struct intel_hdmi_priv), 1, GFP_KERNEL); | 286 | sizeof(struct intel_hdmi_priv), 1, GFP_KERNEL); |
| 238 | if (!intel_output) | 287 | if (!intel_output) |
diff --git a/drivers/gpu/drm/i915/intel_i2c.c b/drivers/gpu/drm/i915/intel_i2c.c index c7eab724c418..8673c735b8ab 100644 --- a/drivers/gpu/drm/i915/intel_i2c.c +++ b/drivers/gpu/drm/i915/intel_i2c.c | |||
| @@ -39,7 +39,7 @@ void intel_i2c_quirk_set(struct drm_device *dev, bool enable) | |||
| 39 | struct drm_i915_private *dev_priv = dev->dev_private; | 39 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 40 | 40 | ||
| 41 | /* When using bit bashing for I2C, this bit needs to be set to 1 */ | 41 | /* When using bit bashing for I2C, this bit needs to be set to 1 */ |
| 42 | if (!IS_IGD(dev)) | 42 | if (!IS_PINEVIEW(dev)) |
| 43 | return; | 43 | return; |
| 44 | if (enable) | 44 | if (enable) |
| 45 | I915_WRITE(DSPCLK_GATE_D, | 45 | I915_WRITE(DSPCLK_GATE_D, |
| @@ -118,6 +118,23 @@ static void set_data(void *data, int state_high) | |||
| 118 | udelay(I2C_RISEFALL_TIME); /* wait for the line to change state */ | 118 | udelay(I2C_RISEFALL_TIME); /* wait for the line to change state */ |
| 119 | } | 119 | } |
| 120 | 120 | ||
| 121 | /* Clears the GMBUS setup. Our driver doesn't make use of the GMBUS I2C | ||
| 122 | * engine, but if the BIOS leaves it enabled, then that can break our use | ||
| 123 | * of the bit-banging I2C interfaces. This is notably the case with the | ||
| 124 | * Mac Mini in EFI mode. | ||
| 125 | */ | ||
| 126 | void | ||
| 127 | intel_i2c_reset_gmbus(struct drm_device *dev) | ||
| 128 | { | ||
| 129 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 130 | |||
| 131 | if (IS_IRONLAKE(dev)) { | ||
| 132 | I915_WRITE(PCH_GMBUS0, 0); | ||
| 133 | } else { | ||
| 134 | I915_WRITE(GMBUS0, 0); | ||
| 135 | } | ||
| 136 | } | ||
| 137 | |||
| 121 | /** | 138 | /** |
| 122 | * intel_i2c_create - instantiate an Intel i2c bus using the specified GPIO reg | 139 | * intel_i2c_create - instantiate an Intel i2c bus using the specified GPIO reg |
| 123 | * @dev: DRM device | 140 | * @dev: DRM device |
| @@ -168,6 +185,8 @@ struct i2c_adapter *intel_i2c_create(struct drm_device *dev, const u32 reg, | |||
| 168 | if(i2c_bit_add_bus(&chan->adapter)) | 185 | if(i2c_bit_add_bus(&chan->adapter)) |
| 169 | goto out_free; | 186 | goto out_free; |
| 170 | 187 | ||
| 188 | intel_i2c_reset_gmbus(dev); | ||
| 189 | |||
| 171 | /* JJJ: raise SCL and SDA? */ | 190 | /* JJJ: raise SCL and SDA? */ |
| 172 | intel_i2c_quirk_set(dev, true); | 191 | intel_i2c_quirk_set(dev, true); |
| 173 | set_data(chan, 1); | 192 | set_data(chan, 1); |
diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c index eb365021bb5a..3118ce274e67 100644 --- a/drivers/gpu/drm/i915/intel_lvds.c +++ b/drivers/gpu/drm/i915/intel_lvds.c | |||
| @@ -56,7 +56,7 @@ static void intel_lvds_set_backlight(struct drm_device *dev, int level) | |||
| 56 | struct drm_i915_private *dev_priv = dev->dev_private; | 56 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 57 | u32 blc_pwm_ctl, reg; | 57 | u32 blc_pwm_ctl, reg; |
| 58 | 58 | ||
| 59 | if (IS_IGDNG(dev)) | 59 | if (IS_IRONLAKE(dev)) |
| 60 | reg = BLC_PWM_CPU_CTL; | 60 | reg = BLC_PWM_CPU_CTL; |
| 61 | else | 61 | else |
| 62 | reg = BLC_PWM_CTL; | 62 | reg = BLC_PWM_CTL; |
| @@ -74,7 +74,7 @@ static u32 intel_lvds_get_max_backlight(struct drm_device *dev) | |||
| 74 | struct drm_i915_private *dev_priv = dev->dev_private; | 74 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 75 | u32 reg; | 75 | u32 reg; |
| 76 | 76 | ||
| 77 | if (IS_IGDNG(dev)) | 77 | if (IS_IRONLAKE(dev)) |
| 78 | reg = BLC_PWM_PCH_CTL2; | 78 | reg = BLC_PWM_PCH_CTL2; |
| 79 | else | 79 | else |
| 80 | reg = BLC_PWM_CTL; | 80 | reg = BLC_PWM_CTL; |
| @@ -91,7 +91,7 @@ static void intel_lvds_set_power(struct drm_device *dev, bool on) | |||
| 91 | struct drm_i915_private *dev_priv = dev->dev_private; | 91 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 92 | u32 pp_status, ctl_reg, status_reg; | 92 | u32 pp_status, ctl_reg, status_reg; |
| 93 | 93 | ||
| 94 | if (IS_IGDNG(dev)) { | 94 | if (IS_IRONLAKE(dev)) { |
| 95 | ctl_reg = PCH_PP_CONTROL; | 95 | ctl_reg = PCH_PP_CONTROL; |
| 96 | status_reg = PCH_PP_STATUS; | 96 | status_reg = PCH_PP_STATUS; |
| 97 | } else { | 97 | } else { |
| @@ -137,7 +137,7 @@ static void intel_lvds_save(struct drm_connector *connector) | |||
| 137 | u32 pp_on_reg, pp_off_reg, pp_ctl_reg, pp_div_reg; | 137 | u32 pp_on_reg, pp_off_reg, pp_ctl_reg, pp_div_reg; |
| 138 | u32 pwm_ctl_reg; | 138 | u32 pwm_ctl_reg; |
| 139 | 139 | ||
| 140 | if (IS_IGDNG(dev)) { | 140 | if (IS_IRONLAKE(dev)) { |
| 141 | pp_on_reg = PCH_PP_ON_DELAYS; | 141 | pp_on_reg = PCH_PP_ON_DELAYS; |
| 142 | pp_off_reg = PCH_PP_OFF_DELAYS; | 142 | pp_off_reg = PCH_PP_OFF_DELAYS; |
| 143 | pp_ctl_reg = PCH_PP_CONTROL; | 143 | pp_ctl_reg = PCH_PP_CONTROL; |
| @@ -174,7 +174,7 @@ static void intel_lvds_restore(struct drm_connector *connector) | |||
| 174 | u32 pp_on_reg, pp_off_reg, pp_ctl_reg, pp_div_reg; | 174 | u32 pp_on_reg, pp_off_reg, pp_ctl_reg, pp_div_reg; |
| 175 | u32 pwm_ctl_reg; | 175 | u32 pwm_ctl_reg; |
| 176 | 176 | ||
| 177 | if (IS_IGDNG(dev)) { | 177 | if (IS_IRONLAKE(dev)) { |
| 178 | pp_on_reg = PCH_PP_ON_DELAYS; | 178 | pp_on_reg = PCH_PP_ON_DELAYS; |
| 179 | pp_off_reg = PCH_PP_OFF_DELAYS; | 179 | pp_off_reg = PCH_PP_OFF_DELAYS; |
| 180 | pp_ctl_reg = PCH_PP_CONTROL; | 180 | pp_ctl_reg = PCH_PP_CONTROL; |
| @@ -297,7 +297,7 @@ static bool intel_lvds_mode_fixup(struct drm_encoder *encoder, | |||
| 297 | } | 297 | } |
| 298 | 298 | ||
| 299 | /* full screen scale for now */ | 299 | /* full screen scale for now */ |
| 300 | if (IS_IGDNG(dev)) | 300 | if (IS_IRONLAKE(dev)) |
| 301 | goto out; | 301 | goto out; |
| 302 | 302 | ||
| 303 | /* 965+ wants fuzzy fitting */ | 303 | /* 965+ wants fuzzy fitting */ |
| @@ -327,7 +327,7 @@ static bool intel_lvds_mode_fixup(struct drm_encoder *encoder, | |||
| 327 | * to register description and PRM. | 327 | * to register description and PRM. |
| 328 | * Change the value here to see the borders for debugging | 328 | * Change the value here to see the borders for debugging |
| 329 | */ | 329 | */ |
| 330 | if (!IS_IGDNG(dev)) { | 330 | if (!IS_IRONLAKE(dev)) { |
| 331 | I915_WRITE(BCLRPAT_A, 0); | 331 | I915_WRITE(BCLRPAT_A, 0); |
| 332 | I915_WRITE(BCLRPAT_B, 0); | 332 | I915_WRITE(BCLRPAT_B, 0); |
| 333 | } | 333 | } |
| @@ -548,7 +548,7 @@ static void intel_lvds_prepare(struct drm_encoder *encoder) | |||
| 548 | struct drm_i915_private *dev_priv = dev->dev_private; | 548 | struct drm_i915_private *dev_priv = dev->dev_private; |
| 549 | u32 reg; | 549 | u32 reg; |
| 550 | 550 | ||
| 551 | if (IS_IGDNG(dev)) | 551 | if (IS_IRONLAKE(dev)) |
| 552 | reg = BLC_PWM_CPU_CTL; | 552 | reg = BLC_PWM_CPU_CTL; |
| 553 | else | 553 | else |
| 554 | reg = BLC_PWM_CTL; | 554 | reg = BLC_PWM_CTL; |
| @@ -587,7 +587,7 @@ static void intel_lvds_mode_set(struct drm_encoder *encoder, | |||
| 587 | * settings. | 587 | * settings. |
| 588 | */ | 588 | */ |
| 589 | 589 | ||
| 590 | if (IS_IGDNG(dev)) | 590 | if (IS_IRONLAKE(dev)) |
| 591 | return; | 591 | return; |
| 592 | 592 | ||
| 593 | /* | 593 | /* |
| @@ -914,6 +914,101 @@ static int intel_lid_present(void) | |||
| 914 | #endif | 914 | #endif |
| 915 | 915 | ||
| 916 | /** | 916 | /** |
| 917 | * intel_find_lvds_downclock - find the reduced downclock for LVDS in EDID | ||
| 918 | * @dev: drm device | ||
| 919 | * @connector: LVDS connector | ||
| 920 | * | ||
| 921 | * Find the reduced downclock for LVDS in EDID. | ||
| 922 | */ | ||
| 923 | static void intel_find_lvds_downclock(struct drm_device *dev, | ||
| 924 | struct drm_connector *connector) | ||
| 925 | { | ||
| 926 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 927 | struct drm_display_mode *scan, *panel_fixed_mode; | ||
| 928 | int temp_downclock; | ||
| 929 | |||
| 930 | panel_fixed_mode = dev_priv->panel_fixed_mode; | ||
| 931 | temp_downclock = panel_fixed_mode->clock; | ||
| 932 | |||
| 933 | mutex_lock(&dev->mode_config.mutex); | ||
| 934 | list_for_each_entry(scan, &connector->probed_modes, head) { | ||
| 935 | /* | ||
| 936 | * If one mode has the same resolution with the fixed_panel | ||
| 937 | * mode while they have the different refresh rate, it means | ||
| 938 | * that the reduced downclock is found for the LVDS. In such | ||
| 939 | * case we can set the different FPx0/1 to dynamically select | ||
| 940 | * between low and high frequency. | ||
| 941 | */ | ||
| 942 | if (scan->hdisplay == panel_fixed_mode->hdisplay && | ||
| 943 | scan->hsync_start == panel_fixed_mode->hsync_start && | ||
| 944 | scan->hsync_end == panel_fixed_mode->hsync_end && | ||
| 945 | scan->htotal == panel_fixed_mode->htotal && | ||
| 946 | scan->vdisplay == panel_fixed_mode->vdisplay && | ||
| 947 | scan->vsync_start == panel_fixed_mode->vsync_start && | ||
| 948 | scan->vsync_end == panel_fixed_mode->vsync_end && | ||
| 949 | scan->vtotal == panel_fixed_mode->vtotal) { | ||
| 950 | if (scan->clock < temp_downclock) { | ||
| 951 | /* | ||
| 952 | * The downclock is already found. But we | ||
| 953 | * expect to find the lower downclock. | ||
| 954 | */ | ||
| 955 | temp_downclock = scan->clock; | ||
| 956 | } | ||
| 957 | } | ||
| 958 | } | ||
| 959 | mutex_unlock(&dev->mode_config.mutex); | ||
| 960 | if (temp_downclock < panel_fixed_mode->clock) { | ||
| 961 | /* We found the downclock for LVDS. */ | ||
| 962 | dev_priv->lvds_downclock_avail = 1; | ||
| 963 | dev_priv->lvds_downclock = temp_downclock; | ||
| 964 | DRM_DEBUG_KMS("LVDS downclock is found in EDID. " | ||
| 965 | "Normal clock %dKhz, downclock %dKhz\n", | ||
| 966 | panel_fixed_mode->clock, temp_downclock); | ||
| 967 | } | ||
| 968 | return; | ||
| 969 | } | ||
| 970 | |||
| 971 | /* | ||
| 972 | * Enumerate the child dev array parsed from VBT to check whether | ||
| 973 | * the LVDS is present. | ||
| 974 | * If it is present, return 1. | ||
| 975 | * If it is not present, return false. | ||
| 976 | * If no child dev is parsed from VBT, it assumes that the LVDS is present. | ||
| 977 | * Note: The addin_offset should also be checked for LVDS panel. | ||
| 978 | * Only when it is non-zero, it is assumed that it is present. | ||
| 979 | */ | ||
| 980 | static int lvds_is_present_in_vbt(struct drm_device *dev) | ||
| 981 | { | ||
| 982 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 983 | struct child_device_config *p_child; | ||
| 984 | int i, ret; | ||
| 985 | |||
| 986 | if (!dev_priv->child_dev_num) | ||
| 987 | return 1; | ||
| 988 | |||
| 989 | ret = 0; | ||
| 990 | for (i = 0; i < dev_priv->child_dev_num; i++) { | ||
| 991 | p_child = dev_priv->child_dev + i; | ||
| 992 | /* | ||
| 993 | * If the device type is not LFP, continue. | ||
| 994 | * If the device type is 0x22, it is also regarded as LFP. | ||
| 995 | */ | ||
| 996 | if (p_child->device_type != DEVICE_TYPE_INT_LFP && | ||
| 997 | p_child->device_type != DEVICE_TYPE_LFP) | ||
| 998 | continue; | ||
| 999 | |||
| 1000 | /* The addin_offset should be checked. Only when it is | ||
| 1001 | * non-zero, it is regarded as present. | ||
| 1002 | */ | ||
| 1003 | if (p_child->addin_offset) { | ||
| 1004 | ret = 1; | ||
| 1005 | break; | ||
| 1006 | } | ||
| 1007 | } | ||
| 1008 | return ret; | ||
| 1009 | } | ||
| 1010 | |||
| 1011 | /** | ||
| 917 | * intel_lvds_init - setup LVDS connectors on this device | 1012 | * intel_lvds_init - setup LVDS connectors on this device |
| 918 | * @dev: drm device | 1013 | * @dev: drm device |
| 919 | * | 1014 | * |
| @@ -936,21 +1031,20 @@ void intel_lvds_init(struct drm_device *dev) | |||
| 936 | if (dmi_check_system(intel_no_lvds)) | 1031 | if (dmi_check_system(intel_no_lvds)) |
| 937 | return; | 1032 | return; |
| 938 | 1033 | ||
| 939 | /* Assume that any device without an ACPI LID device also doesn't | 1034 | /* |
| 940 | * have an integrated LVDS. We would be better off parsing the BIOS | 1035 | * Assume LVDS is present if there's an ACPI lid device or if the |
| 941 | * to get a reliable indicator, but that code isn't written yet. | 1036 | * device is present in the VBT. |
| 942 | * | ||
| 943 | * In the case of all-in-one desktops using LVDS that we've seen, | ||
| 944 | * they're using SDVO LVDS. | ||
| 945 | */ | 1037 | */ |
| 946 | if (!intel_lid_present()) | 1038 | if (!lvds_is_present_in_vbt(dev) && !intel_lid_present()) { |
| 1039 | DRM_DEBUG_KMS("LVDS is not present in VBT and no lid detected\n"); | ||
| 947 | return; | 1040 | return; |
| 1041 | } | ||
| 948 | 1042 | ||
| 949 | if (IS_IGDNG(dev)) { | 1043 | if (IS_IRONLAKE(dev)) { |
| 950 | if ((I915_READ(PCH_LVDS) & LVDS_DETECTED) == 0) | 1044 | if ((I915_READ(PCH_LVDS) & LVDS_DETECTED) == 0) |
| 951 | return; | 1045 | return; |
| 952 | if (dev_priv->edp_support) { | 1046 | if (dev_priv->edp_support) { |
| 953 | DRM_DEBUG("disable LVDS for eDP support\n"); | 1047 | DRM_DEBUG_KMS("disable LVDS for eDP support\n"); |
| 954 | return; | 1048 | return; |
| 955 | } | 1049 | } |
| 956 | gpio = PCH_GPIOC; | 1050 | gpio = PCH_GPIOC; |
| @@ -1023,6 +1117,7 @@ void intel_lvds_init(struct drm_device *dev) | |||
| 1023 | dev_priv->panel_fixed_mode = | 1117 | dev_priv->panel_fixed_mode = |
| 1024 | drm_mode_duplicate(dev, scan); | 1118 | drm_mode_duplicate(dev, scan); |
| 1025 | mutex_unlock(&dev->mode_config.mutex); | 1119 | mutex_unlock(&dev->mode_config.mutex); |
| 1120 | intel_find_lvds_downclock(dev, connector); | ||
| 1026 | goto out; | 1121 | goto out; |
| 1027 | } | 1122 | } |
| 1028 | mutex_unlock(&dev->mode_config.mutex); | 1123 | mutex_unlock(&dev->mode_config.mutex); |
| @@ -1047,8 +1142,8 @@ void intel_lvds_init(struct drm_device *dev) | |||
| 1047 | * correct mode. | 1142 | * correct mode. |
| 1048 | */ | 1143 | */ |
| 1049 | 1144 | ||
| 1050 | /* IGDNG: FIXME if still fail, not try pipe mode now */ | 1145 | /* Ironlake: FIXME if still fail, not try pipe mode now */ |
| 1051 | if (IS_IGDNG(dev)) | 1146 | if (IS_IRONLAKE(dev)) |
| 1052 | goto failed; | 1147 | goto failed; |
| 1053 | 1148 | ||
| 1054 | lvds = I915_READ(LVDS); | 1149 | lvds = I915_READ(LVDS); |
| @@ -1069,7 +1164,7 @@ void intel_lvds_init(struct drm_device *dev) | |||
| 1069 | goto failed; | 1164 | goto failed; |
| 1070 | 1165 | ||
| 1071 | out: | 1166 | out: |
| 1072 | if (IS_IGDNG(dev)) { | 1167 | if (IS_IRONLAKE(dev)) { |
| 1073 | u32 pwm; | 1168 | u32 pwm; |
| 1074 | /* make sure PWM is enabled */ | 1169 | /* make sure PWM is enabled */ |
| 1075 | pwm = I915_READ(BLC_PWM_CPU_CTL2); | 1170 | pwm = I915_READ(BLC_PWM_CPU_CTL2); |
| @@ -1082,7 +1177,7 @@ out: | |||
| 1082 | } | 1177 | } |
| 1083 | dev_priv->lid_notifier.notifier_call = intel_lid_notify; | 1178 | dev_priv->lid_notifier.notifier_call = intel_lid_notify; |
| 1084 | if (acpi_lid_notifier_register(&dev_priv->lid_notifier)) { | 1179 | if (acpi_lid_notifier_register(&dev_priv->lid_notifier)) { |
| 1085 | DRM_DEBUG("lid notifier registration failed\n"); | 1180 | DRM_DEBUG_KMS("lid notifier registration failed\n"); |
| 1086 | dev_priv->lid_notifier.notifier_call = NULL; | 1181 | dev_priv->lid_notifier.notifier_call = NULL; |
| 1087 | } | 1182 | } |
| 1088 | drm_sysfs_connector_add(connector); | 1183 | drm_sysfs_connector_add(connector); |
| @@ -1093,5 +1188,6 @@ failed: | |||
| 1093 | if (intel_output->ddc_bus) | 1188 | if (intel_output->ddc_bus) |
| 1094 | intel_i2c_destroy(intel_output->ddc_bus); | 1189 | intel_i2c_destroy(intel_output->ddc_bus); |
| 1095 | drm_connector_cleanup(connector); | 1190 | drm_connector_cleanup(connector); |
| 1191 | drm_encoder_cleanup(encoder); | ||
| 1096 | kfree(intel_output); | 1192 | kfree(intel_output); |
| 1097 | } | 1193 | } |
diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c new file mode 100644 index 000000000000..2639591c72e9 --- /dev/null +++ b/drivers/gpu/drm/i915/intel_overlay.c | |||
| @@ -0,0 +1,1416 @@ | |||
| 1 | /* | ||
| 2 | * Copyright © 2009 | ||
| 3 | * | ||
| 4 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 5 | * copy of this software and associated documentation files (the "Software"), | ||
| 6 | * to deal in the Software without restriction, including without limitation | ||
| 7 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | ||
| 8 | * and/or sell copies of the Software, and to permit persons to whom the | ||
| 9 | * Software is furnished to do so, subject to the following conditions: | ||
| 10 | * | ||
| 11 | * The above copyright notice and this permission notice (including the next | ||
| 12 | * paragraph) shall be included in all copies or substantial portions of the | ||
| 13 | * Software. | ||
| 14 | * | ||
| 15 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 16 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 17 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 18 | * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 19 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
| 20 | * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||
| 21 | * SOFTWARE. | ||
| 22 | * | ||
| 23 | * Authors: | ||
| 24 | * Daniel Vetter <daniel@ffwll.ch> | ||
| 25 | * | ||
| 26 | * Derived from Xorg ddx, xf86-video-intel, src/i830_video.c | ||
| 27 | */ | ||
| 28 | #include "drmP.h" | ||
| 29 | #include "drm.h" | ||
| 30 | #include "i915_drm.h" | ||
| 31 | #include "i915_drv.h" | ||
| 32 | #include "i915_reg.h" | ||
| 33 | #include "intel_drv.h" | ||
| 34 | |||
| 35 | /* Limits for overlay size. According to intel doc, the real limits are: | ||
| 36 | * Y width: 4095, UV width (planar): 2047, Y height: 2047, | ||
| 37 | * UV width (planar): * 1023. But the xorg thinks 2048 for height and width. Use | ||
| 38 | * the mininum of both. */ | ||
| 39 | #define IMAGE_MAX_WIDTH 2048 | ||
| 40 | #define IMAGE_MAX_HEIGHT 2046 /* 2 * 1023 */ | ||
| 41 | /* on 830 and 845 these large limits result in the card hanging */ | ||
| 42 | #define IMAGE_MAX_WIDTH_LEGACY 1024 | ||
| 43 | #define IMAGE_MAX_HEIGHT_LEGACY 1088 | ||
| 44 | |||
| 45 | /* overlay register definitions */ | ||
| 46 | /* OCMD register */ | ||
| 47 | #define OCMD_TILED_SURFACE (0x1<<19) | ||
| 48 | #define OCMD_MIRROR_MASK (0x3<<17) | ||
| 49 | #define OCMD_MIRROR_MODE (0x3<<17) | ||
| 50 | #define OCMD_MIRROR_HORIZONTAL (0x1<<17) | ||
| 51 | #define OCMD_MIRROR_VERTICAL (0x2<<17) | ||
| 52 | #define OCMD_MIRROR_BOTH (0x3<<17) | ||
| 53 | #define OCMD_BYTEORDER_MASK (0x3<<14) /* zero for YUYV or FOURCC YUY2 */ | ||
| 54 | #define OCMD_UV_SWAP (0x1<<14) /* YVYU */ | ||
| 55 | #define OCMD_Y_SWAP (0x2<<14) /* UYVY or FOURCC UYVY */ | ||
| 56 | #define OCMD_Y_AND_UV_SWAP (0x3<<14) /* VYUY */ | ||
| 57 | #define OCMD_SOURCE_FORMAT_MASK (0xf<<10) | ||
| 58 | #define OCMD_RGB_888 (0x1<<10) /* not in i965 Intel docs */ | ||
| 59 | #define OCMD_RGB_555 (0x2<<10) /* not in i965 Intel docs */ | ||
| 60 | #define OCMD_RGB_565 (0x3<<10) /* not in i965 Intel docs */ | ||
| 61 | #define OCMD_YUV_422_PACKED (0x8<<10) | ||
| 62 | #define OCMD_YUV_411_PACKED (0x9<<10) /* not in i965 Intel docs */ | ||
| 63 | #define OCMD_YUV_420_PLANAR (0xc<<10) | ||
| 64 | #define OCMD_YUV_422_PLANAR (0xd<<10) | ||
| 65 | #define OCMD_YUV_410_PLANAR (0xe<<10) /* also 411 */ | ||
| 66 | #define OCMD_TVSYNCFLIP_PARITY (0x1<<9) | ||
| 67 | #define OCMD_TVSYNCFLIP_ENABLE (0x1<<7) | ||
| 68 | #define OCMD_BUF_TYPE_MASK (Ox1<<5) | ||
| 69 | #define OCMD_BUF_TYPE_FRAME (0x0<<5) | ||
| 70 | #define OCMD_BUF_TYPE_FIELD (0x1<<5) | ||
| 71 | #define OCMD_TEST_MODE (0x1<<4) | ||
| 72 | #define OCMD_BUFFER_SELECT (0x3<<2) | ||
| 73 | #define OCMD_BUFFER0 (0x0<<2) | ||
| 74 | #define OCMD_BUFFER1 (0x1<<2) | ||
| 75 | #define OCMD_FIELD_SELECT (0x1<<2) | ||
| 76 | #define OCMD_FIELD0 (0x0<<1) | ||
| 77 | #define OCMD_FIELD1 (0x1<<1) | ||
| 78 | #define OCMD_ENABLE (0x1<<0) | ||
| 79 | |||
| 80 | /* OCONFIG register */ | ||
| 81 | #define OCONF_PIPE_MASK (0x1<<18) | ||
| 82 | #define OCONF_PIPE_A (0x0<<18) | ||
| 83 | #define OCONF_PIPE_B (0x1<<18) | ||
| 84 | #define OCONF_GAMMA2_ENABLE (0x1<<16) | ||
| 85 | #define OCONF_CSC_MODE_BT601 (0x0<<5) | ||
| 86 | #define OCONF_CSC_MODE_BT709 (0x1<<5) | ||
| 87 | #define OCONF_CSC_BYPASS (0x1<<4) | ||
| 88 | #define OCONF_CC_OUT_8BIT (0x1<<3) | ||
| 89 | #define OCONF_TEST_MODE (0x1<<2) | ||
| 90 | #define OCONF_THREE_LINE_BUFFER (0x1<<0) | ||
| 91 | #define OCONF_TWO_LINE_BUFFER (0x0<<0) | ||
| 92 | |||
| 93 | /* DCLRKM (dst-key) register */ | ||
| 94 | #define DST_KEY_ENABLE (0x1<<31) | ||
| 95 | #define CLK_RGB24_MASK 0x0 | ||
| 96 | #define CLK_RGB16_MASK 0x070307 | ||
| 97 | #define CLK_RGB15_MASK 0x070707 | ||
| 98 | #define CLK_RGB8I_MASK 0xffffff | ||
| 99 | |||
| 100 | #define RGB16_TO_COLORKEY(c) \ | ||
| 101 | (((c & 0xF800) << 8) | ((c & 0x07E0) << 5) | ((c & 0x001F) << 3)) | ||
| 102 | #define RGB15_TO_COLORKEY(c) \ | ||
| 103 | (((c & 0x7c00) << 9) | ((c & 0x03E0) << 6) | ((c & 0x001F) << 3)) | ||
| 104 | |||
| 105 | /* overlay flip addr flag */ | ||
| 106 | #define OFC_UPDATE 0x1 | ||
| 107 | |||
| 108 | /* polyphase filter coefficients */ | ||
| 109 | #define N_HORIZ_Y_TAPS 5 | ||
| 110 | #define N_VERT_Y_TAPS 3 | ||
| 111 | #define N_HORIZ_UV_TAPS 3 | ||
| 112 | #define N_VERT_UV_TAPS 3 | ||
| 113 | #define N_PHASES 17 | ||
| 114 | #define MAX_TAPS 5 | ||
| 115 | |||
| 116 | /* memory bufferd overlay registers */ | ||
| 117 | struct overlay_registers { | ||
| 118 | u32 OBUF_0Y; | ||
| 119 | u32 OBUF_1Y; | ||
| 120 | u32 OBUF_0U; | ||
| 121 | u32 OBUF_0V; | ||
| 122 | u32 OBUF_1U; | ||
| 123 | u32 OBUF_1V; | ||
| 124 | u32 OSTRIDE; | ||
| 125 | u32 YRGB_VPH; | ||
| 126 | u32 UV_VPH; | ||
| 127 | u32 HORZ_PH; | ||
| 128 | u32 INIT_PHS; | ||
| 129 | u32 DWINPOS; | ||
| 130 | u32 DWINSZ; | ||
| 131 | u32 SWIDTH; | ||
| 132 | u32 SWIDTHSW; | ||
| 133 | u32 SHEIGHT; | ||
| 134 | u32 YRGBSCALE; | ||
| 135 | u32 UVSCALE; | ||
| 136 | u32 OCLRC0; | ||
| 137 | u32 OCLRC1; | ||
| 138 | u32 DCLRKV; | ||
| 139 | u32 DCLRKM; | ||
| 140 | u32 SCLRKVH; | ||
| 141 | u32 SCLRKVL; | ||
| 142 | u32 SCLRKEN; | ||
| 143 | u32 OCONFIG; | ||
| 144 | u32 OCMD; | ||
| 145 | u32 RESERVED1; /* 0x6C */ | ||
| 146 | u32 OSTART_0Y; | ||
| 147 | u32 OSTART_1Y; | ||
| 148 | u32 OSTART_0U; | ||
| 149 | u32 OSTART_0V; | ||
| 150 | u32 OSTART_1U; | ||
| 151 | u32 OSTART_1V; | ||
| 152 | u32 OTILEOFF_0Y; | ||
| 153 | u32 OTILEOFF_1Y; | ||
| 154 | u32 OTILEOFF_0U; | ||
| 155 | u32 OTILEOFF_0V; | ||
| 156 | u32 OTILEOFF_1U; | ||
| 157 | u32 OTILEOFF_1V; | ||
| 158 | u32 FASTHSCALE; /* 0xA0 */ | ||
| 159 | u32 UVSCALEV; /* 0xA4 */ | ||
| 160 | u32 RESERVEDC[(0x200 - 0xA8) / 4]; /* 0xA8 - 0x1FC */ | ||
| 161 | u16 Y_VCOEFS[N_VERT_Y_TAPS * N_PHASES]; /* 0x200 */ | ||
| 162 | u16 RESERVEDD[0x100 / 2 - N_VERT_Y_TAPS * N_PHASES]; | ||
| 163 | u16 Y_HCOEFS[N_HORIZ_Y_TAPS * N_PHASES]; /* 0x300 */ | ||
| 164 | u16 RESERVEDE[0x200 / 2 - N_HORIZ_Y_TAPS * N_PHASES]; | ||
| 165 | u16 UV_VCOEFS[N_VERT_UV_TAPS * N_PHASES]; /* 0x500 */ | ||
| 166 | u16 RESERVEDF[0x100 / 2 - N_VERT_UV_TAPS * N_PHASES]; | ||
| 167 | u16 UV_HCOEFS[N_HORIZ_UV_TAPS * N_PHASES]; /* 0x600 */ | ||
| 168 | u16 RESERVEDG[0x100 / 2 - N_HORIZ_UV_TAPS * N_PHASES]; | ||
| 169 | }; | ||
| 170 | |||
| 171 | /* overlay flip addr flag */ | ||
| 172 | #define OFC_UPDATE 0x1 | ||
| 173 | |||
| 174 | #define OVERLAY_NONPHYSICAL(dev) (IS_G33(dev) || IS_I965G(dev)) | ||
| 175 | #define OVERLAY_EXISTS(dev) (!IS_G4X(dev) && !IS_IRONLAKE(dev)) | ||
| 176 | |||
| 177 | |||
| 178 | static struct overlay_registers *intel_overlay_map_regs_atomic(struct intel_overlay *overlay) | ||
| 179 | { | ||
| 180 | drm_i915_private_t *dev_priv = overlay->dev->dev_private; | ||
| 181 | struct overlay_registers *regs; | ||
| 182 | |||
| 183 | /* no recursive mappings */ | ||
| 184 | BUG_ON(overlay->virt_addr); | ||
| 185 | |||
| 186 | if (OVERLAY_NONPHYSICAL(overlay->dev)) { | ||
| 187 | regs = io_mapping_map_atomic_wc(dev_priv->mm.gtt_mapping, | ||
| 188 | overlay->reg_bo->gtt_offset); | ||
| 189 | |||
| 190 | if (!regs) { | ||
| 191 | DRM_ERROR("failed to map overlay regs in GTT\n"); | ||
| 192 | return NULL; | ||
| 193 | } | ||
| 194 | } else | ||
| 195 | regs = overlay->reg_bo->phys_obj->handle->vaddr; | ||
| 196 | |||
| 197 | return overlay->virt_addr = regs; | ||
| 198 | } | ||
| 199 | |||
| 200 | static void intel_overlay_unmap_regs_atomic(struct intel_overlay *overlay) | ||
| 201 | { | ||
| 202 | struct drm_device *dev = overlay->dev; | ||
| 203 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 204 | |||
| 205 | if (OVERLAY_NONPHYSICAL(overlay->dev)) | ||
| 206 | io_mapping_unmap_atomic(overlay->virt_addr); | ||
| 207 | |||
| 208 | overlay->virt_addr = NULL; | ||
| 209 | |||
| 210 | I915_READ(OVADD); /* flush wc cashes */ | ||
| 211 | |||
| 212 | return; | ||
| 213 | } | ||
| 214 | |||
| 215 | /* overlay needs to be disable in OCMD reg */ | ||
| 216 | static int intel_overlay_on(struct intel_overlay *overlay) | ||
| 217 | { | ||
| 218 | struct drm_device *dev = overlay->dev; | ||
| 219 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 220 | int ret; | ||
| 221 | RING_LOCALS; | ||
| 222 | |||
| 223 | BUG_ON(overlay->active); | ||
| 224 | |||
| 225 | overlay->active = 1; | ||
| 226 | overlay->hw_wedged = NEEDS_WAIT_FOR_FLIP; | ||
| 227 | |||
| 228 | BEGIN_LP_RING(6); | ||
| 229 | OUT_RING(MI_FLUSH); | ||
| 230 | OUT_RING(MI_NOOP); | ||
| 231 | OUT_RING(MI_OVERLAY_FLIP | MI_OVERLAY_ON); | ||
| 232 | OUT_RING(overlay->flip_addr | OFC_UPDATE); | ||
| 233 | OUT_RING(MI_WAIT_FOR_EVENT | MI_WAIT_FOR_OVERLAY_FLIP); | ||
| 234 | OUT_RING(MI_NOOP); | ||
| 235 | ADVANCE_LP_RING(); | ||
| 236 | |||
| 237 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 238 | if (overlay->last_flip_req == 0) | ||
| 239 | return -ENOMEM; | ||
| 240 | |||
| 241 | ret = i915_do_wait_request(dev, overlay->last_flip_req, 1); | ||
| 242 | if (ret != 0) | ||
| 243 | return ret; | ||
| 244 | |||
| 245 | overlay->hw_wedged = 0; | ||
| 246 | overlay->last_flip_req = 0; | ||
| 247 | return 0; | ||
| 248 | } | ||
| 249 | |||
| 250 | /* overlay needs to be enabled in OCMD reg */ | ||
| 251 | static void intel_overlay_continue(struct intel_overlay *overlay, | ||
| 252 | bool load_polyphase_filter) | ||
| 253 | { | ||
| 254 | struct drm_device *dev = overlay->dev; | ||
| 255 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 256 | u32 flip_addr = overlay->flip_addr; | ||
| 257 | u32 tmp; | ||
| 258 | RING_LOCALS; | ||
| 259 | |||
| 260 | BUG_ON(!overlay->active); | ||
| 261 | |||
| 262 | if (load_polyphase_filter) | ||
| 263 | flip_addr |= OFC_UPDATE; | ||
| 264 | |||
| 265 | /* check for underruns */ | ||
| 266 | tmp = I915_READ(DOVSTA); | ||
| 267 | if (tmp & (1 << 17)) | ||
| 268 | DRM_DEBUG("overlay underrun, DOVSTA: %x\n", tmp); | ||
| 269 | |||
| 270 | BEGIN_LP_RING(4); | ||
| 271 | OUT_RING(MI_FLUSH); | ||
| 272 | OUT_RING(MI_NOOP); | ||
| 273 | OUT_RING(MI_OVERLAY_FLIP | MI_OVERLAY_CONTINUE); | ||
| 274 | OUT_RING(flip_addr); | ||
| 275 | ADVANCE_LP_RING(); | ||
| 276 | |||
| 277 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 278 | } | ||
| 279 | |||
| 280 | static int intel_overlay_wait_flip(struct intel_overlay *overlay) | ||
| 281 | { | ||
| 282 | struct drm_device *dev = overlay->dev; | ||
| 283 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 284 | int ret; | ||
| 285 | u32 tmp; | ||
| 286 | RING_LOCALS; | ||
| 287 | |||
| 288 | if (overlay->last_flip_req != 0) { | ||
| 289 | ret = i915_do_wait_request(dev, overlay->last_flip_req, 1); | ||
| 290 | if (ret == 0) { | ||
| 291 | overlay->last_flip_req = 0; | ||
| 292 | |||
| 293 | tmp = I915_READ(ISR); | ||
| 294 | |||
| 295 | if (!(tmp & I915_OVERLAY_PLANE_FLIP_PENDING_INTERRUPT)) | ||
| 296 | return 0; | ||
| 297 | } | ||
| 298 | } | ||
| 299 | |||
| 300 | /* synchronous slowpath */ | ||
| 301 | overlay->hw_wedged = RELEASE_OLD_VID; | ||
| 302 | |||
| 303 | BEGIN_LP_RING(2); | ||
| 304 | OUT_RING(MI_WAIT_FOR_EVENT | MI_WAIT_FOR_OVERLAY_FLIP); | ||
| 305 | OUT_RING(MI_NOOP); | ||
| 306 | ADVANCE_LP_RING(); | ||
| 307 | |||
| 308 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 309 | if (overlay->last_flip_req == 0) | ||
| 310 | return -ENOMEM; | ||
| 311 | |||
| 312 | ret = i915_do_wait_request(dev, overlay->last_flip_req, 1); | ||
| 313 | if (ret != 0) | ||
| 314 | return ret; | ||
| 315 | |||
| 316 | overlay->hw_wedged = 0; | ||
| 317 | overlay->last_flip_req = 0; | ||
| 318 | return 0; | ||
| 319 | } | ||
| 320 | |||
| 321 | /* overlay needs to be disabled in OCMD reg */ | ||
| 322 | static int intel_overlay_off(struct intel_overlay *overlay) | ||
| 323 | { | ||
| 324 | u32 flip_addr = overlay->flip_addr; | ||
| 325 | struct drm_device *dev = overlay->dev; | ||
| 326 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 327 | int ret; | ||
| 328 | RING_LOCALS; | ||
| 329 | |||
| 330 | BUG_ON(!overlay->active); | ||
| 331 | |||
| 332 | /* According to intel docs the overlay hw may hang (when switching | ||
| 333 | * off) without loading the filter coeffs. It is however unclear whether | ||
| 334 | * this applies to the disabling of the overlay or to the switching off | ||
| 335 | * of the hw. Do it in both cases */ | ||
| 336 | flip_addr |= OFC_UPDATE; | ||
| 337 | |||
| 338 | /* wait for overlay to go idle */ | ||
| 339 | overlay->hw_wedged = SWITCH_OFF_STAGE_1; | ||
| 340 | |||
| 341 | BEGIN_LP_RING(6); | ||
| 342 | OUT_RING(MI_FLUSH); | ||
| 343 | OUT_RING(MI_NOOP); | ||
| 344 | OUT_RING(MI_OVERLAY_FLIP | MI_OVERLAY_CONTINUE); | ||
| 345 | OUT_RING(flip_addr); | ||
| 346 | OUT_RING(MI_WAIT_FOR_EVENT | MI_WAIT_FOR_OVERLAY_FLIP); | ||
| 347 | OUT_RING(MI_NOOP); | ||
| 348 | ADVANCE_LP_RING(); | ||
| 349 | |||
| 350 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 351 | if (overlay->last_flip_req == 0) | ||
| 352 | return -ENOMEM; | ||
| 353 | |||
| 354 | ret = i915_do_wait_request(dev, overlay->last_flip_req, 1); | ||
| 355 | if (ret != 0) | ||
| 356 | return ret; | ||
| 357 | |||
| 358 | /* turn overlay off */ | ||
| 359 | overlay->hw_wedged = SWITCH_OFF_STAGE_2; | ||
| 360 | |||
| 361 | BEGIN_LP_RING(6); | ||
| 362 | OUT_RING(MI_FLUSH); | ||
| 363 | OUT_RING(MI_NOOP); | ||
| 364 | OUT_RING(MI_OVERLAY_FLIP | MI_OVERLAY_OFF); | ||
| 365 | OUT_RING(flip_addr); | ||
| 366 | OUT_RING(MI_WAIT_FOR_EVENT | MI_WAIT_FOR_OVERLAY_FLIP); | ||
| 367 | OUT_RING(MI_NOOP); | ||
| 368 | ADVANCE_LP_RING(); | ||
| 369 | |||
| 370 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 371 | if (overlay->last_flip_req == 0) | ||
| 372 | return -ENOMEM; | ||
| 373 | |||
| 374 | ret = i915_do_wait_request(dev, overlay->last_flip_req, 1); | ||
| 375 | if (ret != 0) | ||
| 376 | return ret; | ||
| 377 | |||
| 378 | overlay->hw_wedged = 0; | ||
| 379 | overlay->last_flip_req = 0; | ||
| 380 | return ret; | ||
| 381 | } | ||
| 382 | |||
| 383 | static void intel_overlay_off_tail(struct intel_overlay *overlay) | ||
| 384 | { | ||
| 385 | struct drm_gem_object *obj; | ||
| 386 | |||
| 387 | /* never have the overlay hw on without showing a frame */ | ||
| 388 | BUG_ON(!overlay->vid_bo); | ||
| 389 | obj = overlay->vid_bo->obj; | ||
| 390 | |||
| 391 | i915_gem_object_unpin(obj); | ||
| 392 | drm_gem_object_unreference(obj); | ||
| 393 | overlay->vid_bo = NULL; | ||
| 394 | |||
| 395 | overlay->crtc->overlay = NULL; | ||
| 396 | overlay->crtc = NULL; | ||
| 397 | overlay->active = 0; | ||
| 398 | } | ||
| 399 | |||
| 400 | /* recover from an interruption due to a signal | ||
| 401 | * We have to be careful not to repeat work forever an make forward progess. */ | ||
| 402 | int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay, | ||
| 403 | int interruptible) | ||
| 404 | { | ||
| 405 | struct drm_device *dev = overlay->dev; | ||
| 406 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 407 | struct drm_gem_object *obj; | ||
| 408 | u32 flip_addr; | ||
| 409 | int ret; | ||
| 410 | RING_LOCALS; | ||
| 411 | |||
| 412 | if (overlay->hw_wedged == HW_WEDGED) | ||
| 413 | return -EIO; | ||
| 414 | |||
| 415 | if (overlay->last_flip_req == 0) { | ||
| 416 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 417 | if (overlay->last_flip_req == 0) | ||
| 418 | return -ENOMEM; | ||
| 419 | } | ||
| 420 | |||
| 421 | ret = i915_do_wait_request(dev, overlay->last_flip_req, interruptible); | ||
| 422 | if (ret != 0) | ||
| 423 | return ret; | ||
| 424 | |||
| 425 | switch (overlay->hw_wedged) { | ||
| 426 | case RELEASE_OLD_VID: | ||
| 427 | obj = overlay->old_vid_bo->obj; | ||
| 428 | i915_gem_object_unpin(obj); | ||
| 429 | drm_gem_object_unreference(obj); | ||
| 430 | overlay->old_vid_bo = NULL; | ||
| 431 | break; | ||
| 432 | case SWITCH_OFF_STAGE_1: | ||
| 433 | flip_addr = overlay->flip_addr; | ||
| 434 | flip_addr |= OFC_UPDATE; | ||
| 435 | |||
| 436 | overlay->hw_wedged = SWITCH_OFF_STAGE_2; | ||
| 437 | |||
| 438 | BEGIN_LP_RING(6); | ||
| 439 | OUT_RING(MI_FLUSH); | ||
| 440 | OUT_RING(MI_NOOP); | ||
| 441 | OUT_RING(MI_OVERLAY_FLIP | MI_OVERLAY_OFF); | ||
| 442 | OUT_RING(flip_addr); | ||
| 443 | OUT_RING(MI_WAIT_FOR_EVENT | MI_WAIT_FOR_OVERLAY_FLIP); | ||
| 444 | OUT_RING(MI_NOOP); | ||
| 445 | ADVANCE_LP_RING(); | ||
| 446 | |||
| 447 | overlay->last_flip_req = i915_add_request(dev, NULL, 0); | ||
| 448 | if (overlay->last_flip_req == 0) | ||
| 449 | return -ENOMEM; | ||
| 450 | |||
| 451 | ret = i915_do_wait_request(dev, overlay->last_flip_req, | ||
| 452 | interruptible); | ||
| 453 | if (ret != 0) | ||
| 454 | return ret; | ||
| 455 | |||
| 456 | case SWITCH_OFF_STAGE_2: | ||
| 457 | intel_overlay_off_tail(overlay); | ||
| 458 | break; | ||
| 459 | default: | ||
| 460 | BUG_ON(overlay->hw_wedged != NEEDS_WAIT_FOR_FLIP); | ||
| 461 | } | ||
| 462 | |||
| 463 | overlay->hw_wedged = 0; | ||
| 464 | overlay->last_flip_req = 0; | ||
| 465 | return 0; | ||
| 466 | } | ||
| 467 | |||
| 468 | /* Wait for pending overlay flip and release old frame. | ||
| 469 | * Needs to be called before the overlay register are changed | ||
| 470 | * via intel_overlay_(un)map_regs_atomic */ | ||
| 471 | static int intel_overlay_release_old_vid(struct intel_overlay *overlay) | ||
| 472 | { | ||
| 473 | int ret; | ||
| 474 | struct drm_gem_object *obj; | ||
| 475 | |||
| 476 | /* only wait if there is actually an old frame to release to | ||
| 477 | * guarantee forward progress */ | ||
| 478 | if (!overlay->old_vid_bo) | ||
| 479 | return 0; | ||
| 480 | |||
| 481 | ret = intel_overlay_wait_flip(overlay); | ||
| 482 | if (ret != 0) | ||
| 483 | return ret; | ||
| 484 | |||
| 485 | obj = overlay->old_vid_bo->obj; | ||
| 486 | i915_gem_object_unpin(obj); | ||
| 487 | drm_gem_object_unreference(obj); | ||
| 488 | overlay->old_vid_bo = NULL; | ||
| 489 | |||
| 490 | return 0; | ||
| 491 | } | ||
| 492 | |||
| 493 | struct put_image_params { | ||
| 494 | int format; | ||
| 495 | short dst_x; | ||
| 496 | short dst_y; | ||
| 497 | short dst_w; | ||
| 498 | short dst_h; | ||
| 499 | short src_w; | ||
| 500 | short src_scan_h; | ||
| 501 | short src_scan_w; | ||
| 502 | short src_h; | ||
| 503 | short stride_Y; | ||
| 504 | short stride_UV; | ||
| 505 | int offset_Y; | ||
| 506 | int offset_U; | ||
| 507 | int offset_V; | ||
| 508 | }; | ||
| 509 | |||
| 510 | static int packed_depth_bytes(u32 format) | ||
| 511 | { | ||
| 512 | switch (format & I915_OVERLAY_DEPTH_MASK) { | ||
| 513 | case I915_OVERLAY_YUV422: | ||
| 514 | return 4; | ||
| 515 | case I915_OVERLAY_YUV411: | ||
| 516 | /* return 6; not implemented */ | ||
| 517 | default: | ||
| 518 | return -EINVAL; | ||
| 519 | } | ||
| 520 | } | ||
| 521 | |||
| 522 | static int packed_width_bytes(u32 format, short width) | ||
| 523 | { | ||
| 524 | switch (format & I915_OVERLAY_DEPTH_MASK) { | ||
| 525 | case I915_OVERLAY_YUV422: | ||
| 526 | return width << 1; | ||
| 527 | default: | ||
| 528 | return -EINVAL; | ||
| 529 | } | ||
| 530 | } | ||
| 531 | |||
| 532 | static int uv_hsubsampling(u32 format) | ||
| 533 | { | ||
| 534 | switch (format & I915_OVERLAY_DEPTH_MASK) { | ||
| 535 | case I915_OVERLAY_YUV422: | ||
| 536 | case I915_OVERLAY_YUV420: | ||
| 537 | return 2; | ||
| 538 | case I915_OVERLAY_YUV411: | ||
| 539 | case I915_OVERLAY_YUV410: | ||
| 540 | return 4; | ||
| 541 | default: | ||
| 542 | return -EINVAL; | ||
| 543 | } | ||
| 544 | } | ||
| 545 | |||
| 546 | static int uv_vsubsampling(u32 format) | ||
| 547 | { | ||
| 548 | switch (format & I915_OVERLAY_DEPTH_MASK) { | ||
| 549 | case I915_OVERLAY_YUV420: | ||
| 550 | case I915_OVERLAY_YUV410: | ||
| 551 | return 2; | ||
| 552 | case I915_OVERLAY_YUV422: | ||
| 553 | case I915_OVERLAY_YUV411: | ||
| 554 | return 1; | ||
| 555 | default: | ||
| 556 | return -EINVAL; | ||
| 557 | } | ||
| 558 | } | ||
| 559 | |||
| 560 | static u32 calc_swidthsw(struct drm_device *dev, u32 offset, u32 width) | ||
| 561 | { | ||
| 562 | u32 mask, shift, ret; | ||
| 563 | if (IS_I9XX(dev)) { | ||
| 564 | mask = 0x3f; | ||
| 565 | shift = 6; | ||
| 566 | } else { | ||
| 567 | mask = 0x1f; | ||
| 568 | shift = 5; | ||
| 569 | } | ||
| 570 | ret = ((offset + width + mask) >> shift) - (offset >> shift); | ||
| 571 | if (IS_I9XX(dev)) | ||
| 572 | ret <<= 1; | ||
| 573 | ret -=1; | ||
| 574 | return ret << 2; | ||
| 575 | } | ||
| 576 | |||
| 577 | static const u16 y_static_hcoeffs[N_HORIZ_Y_TAPS * N_PHASES] = { | ||
| 578 | 0x3000, 0xb4a0, 0x1930, 0x1920, 0xb4a0, | ||
| 579 | 0x3000, 0xb500, 0x19d0, 0x1880, 0xb440, | ||
| 580 | 0x3000, 0xb540, 0x1a88, 0x2f80, 0xb3e0, | ||
| 581 | 0x3000, 0xb580, 0x1b30, 0x2e20, 0xb380, | ||
| 582 | 0x3000, 0xb5c0, 0x1bd8, 0x2cc0, 0xb320, | ||
| 583 | 0x3020, 0xb5e0, 0x1c60, 0x2b80, 0xb2c0, | ||
| 584 | 0x3020, 0xb5e0, 0x1cf8, 0x2a20, 0xb260, | ||
| 585 | 0x3020, 0xb5e0, 0x1d80, 0x28e0, 0xb200, | ||
| 586 | 0x3020, 0xb5c0, 0x1e08, 0x3f40, 0xb1c0, | ||
| 587 | 0x3020, 0xb580, 0x1e78, 0x3ce0, 0xb160, | ||
| 588 | 0x3040, 0xb520, 0x1ed8, 0x3aa0, 0xb120, | ||
| 589 | 0x3040, 0xb4a0, 0x1f30, 0x3880, 0xb0e0, | ||
| 590 | 0x3040, 0xb400, 0x1f78, 0x3680, 0xb0a0, | ||
| 591 | 0x3020, 0xb340, 0x1fb8, 0x34a0, 0xb060, | ||
| 592 | 0x3020, 0xb240, 0x1fe0, 0x32e0, 0xb040, | ||
| 593 | 0x3020, 0xb140, 0x1ff8, 0x3160, 0xb020, | ||
| 594 | 0xb000, 0x3000, 0x0800, 0x3000, 0xb000}; | ||
| 595 | static const u16 uv_static_hcoeffs[N_HORIZ_UV_TAPS * N_PHASES] = { | ||
| 596 | 0x3000, 0x1800, 0x1800, 0xb000, 0x18d0, 0x2e60, | ||
| 597 | 0xb000, 0x1990, 0x2ce0, 0xb020, 0x1a68, 0x2b40, | ||
| 598 | 0xb040, 0x1b20, 0x29e0, 0xb060, 0x1bd8, 0x2880, | ||
| 599 | 0xb080, 0x1c88, 0x3e60, 0xb0a0, 0x1d28, 0x3c00, | ||
| 600 | 0xb0c0, 0x1db8, 0x39e0, 0xb0e0, 0x1e40, 0x37e0, | ||
| 601 | 0xb100, 0x1eb8, 0x3620, 0xb100, 0x1f18, 0x34a0, | ||
| 602 | 0xb100, 0x1f68, 0x3360, 0xb0e0, 0x1fa8, 0x3240, | ||
| 603 | 0xb0c0, 0x1fe0, 0x3140, 0xb060, 0x1ff0, 0x30a0, | ||
| 604 | 0x3000, 0x0800, 0x3000}; | ||
| 605 | |||
| 606 | static void update_polyphase_filter(struct overlay_registers *regs) | ||
| 607 | { | ||
| 608 | memcpy(regs->Y_HCOEFS, y_static_hcoeffs, sizeof(y_static_hcoeffs)); | ||
| 609 | memcpy(regs->UV_HCOEFS, uv_static_hcoeffs, sizeof(uv_static_hcoeffs)); | ||
| 610 | } | ||
| 611 | |||
| 612 | static bool update_scaling_factors(struct intel_overlay *overlay, | ||
| 613 | struct overlay_registers *regs, | ||
| 614 | struct put_image_params *params) | ||
| 615 | { | ||
| 616 | /* fixed point with a 12 bit shift */ | ||
| 617 | u32 xscale, yscale, xscale_UV, yscale_UV; | ||
| 618 | #define FP_SHIFT 12 | ||
| 619 | #define FRACT_MASK 0xfff | ||
| 620 | bool scale_changed = false; | ||
| 621 | int uv_hscale = uv_hsubsampling(params->format); | ||
| 622 | int uv_vscale = uv_vsubsampling(params->format); | ||
| 623 | |||
| 624 | if (params->dst_w > 1) | ||
| 625 | xscale = ((params->src_scan_w - 1) << FP_SHIFT) | ||
| 626 | /(params->dst_w); | ||
| 627 | else | ||
| 628 | xscale = 1 << FP_SHIFT; | ||
| 629 | |||
| 630 | if (params->dst_h > 1) | ||
| 631 | yscale = ((params->src_scan_h - 1) << FP_SHIFT) | ||
| 632 | /(params->dst_h); | ||
| 633 | else | ||
| 634 | yscale = 1 << FP_SHIFT; | ||
| 635 | |||
| 636 | /*if (params->format & I915_OVERLAY_YUV_PLANAR) {*/ | ||
| 637 | xscale_UV = xscale/uv_hscale; | ||
| 638 | yscale_UV = yscale/uv_vscale; | ||
| 639 | /* make the Y scale to UV scale ratio an exact multiply */ | ||
| 640 | xscale = xscale_UV * uv_hscale; | ||
| 641 | yscale = yscale_UV * uv_vscale; | ||
| 642 | /*} else { | ||
| 643 | xscale_UV = 0; | ||
| 644 | yscale_UV = 0; | ||
| 645 | }*/ | ||
| 646 | |||
| 647 | if (xscale != overlay->old_xscale || yscale != overlay->old_yscale) | ||
| 648 | scale_changed = true; | ||
| 649 | overlay->old_xscale = xscale; | ||
| 650 | overlay->old_yscale = yscale; | ||
| 651 | |||
| 652 | regs->YRGBSCALE = ((yscale & FRACT_MASK) << 20) | ||
| 653 | | ((xscale >> FP_SHIFT) << 16) | ||
| 654 | | ((xscale & FRACT_MASK) << 3); | ||
| 655 | regs->UVSCALE = ((yscale_UV & FRACT_MASK) << 20) | ||
| 656 | | ((xscale_UV >> FP_SHIFT) << 16) | ||
| 657 | | ((xscale_UV & FRACT_MASK) << 3); | ||
| 658 | regs->UVSCALEV = ((yscale >> FP_SHIFT) << 16) | ||
| 659 | | ((yscale_UV >> FP_SHIFT) << 0); | ||
| 660 | |||
| 661 | if (scale_changed) | ||
| 662 | update_polyphase_filter(regs); | ||
| 663 | |||
| 664 | return scale_changed; | ||
| 665 | } | ||
| 666 | |||
| 667 | static void update_colorkey(struct intel_overlay *overlay, | ||
| 668 | struct overlay_registers *regs) | ||
| 669 | { | ||
| 670 | u32 key = overlay->color_key; | ||
| 671 | switch (overlay->crtc->base.fb->bits_per_pixel) { | ||
| 672 | case 8: | ||
| 673 | regs->DCLRKV = 0; | ||
| 674 | regs->DCLRKM = CLK_RGB8I_MASK | DST_KEY_ENABLE; | ||
| 675 | case 16: | ||
| 676 | if (overlay->crtc->base.fb->depth == 15) { | ||
| 677 | regs->DCLRKV = RGB15_TO_COLORKEY(key); | ||
| 678 | regs->DCLRKM = CLK_RGB15_MASK | DST_KEY_ENABLE; | ||
| 679 | } else { | ||
| 680 | regs->DCLRKV = RGB16_TO_COLORKEY(key); | ||
| 681 | regs->DCLRKM = CLK_RGB16_MASK | DST_KEY_ENABLE; | ||
| 682 | } | ||
| 683 | case 24: | ||
| 684 | case 32: | ||
| 685 | regs->DCLRKV = key; | ||
| 686 | regs->DCLRKM = CLK_RGB24_MASK | DST_KEY_ENABLE; | ||
| 687 | } | ||
| 688 | } | ||
| 689 | |||
| 690 | static u32 overlay_cmd_reg(struct put_image_params *params) | ||
| 691 | { | ||
| 692 | u32 cmd = OCMD_ENABLE | OCMD_BUF_TYPE_FRAME | OCMD_BUFFER0; | ||
| 693 | |||
| 694 | if (params->format & I915_OVERLAY_YUV_PLANAR) { | ||
| 695 | switch (params->format & I915_OVERLAY_DEPTH_MASK) { | ||
| 696 | case I915_OVERLAY_YUV422: | ||
| 697 | cmd |= OCMD_YUV_422_PLANAR; | ||
| 698 | break; | ||
| 699 | case I915_OVERLAY_YUV420: | ||
| 700 | cmd |= OCMD_YUV_420_PLANAR; | ||
| 701 | break; | ||
| 702 | case I915_OVERLAY_YUV411: | ||
| 703 | case I915_OVERLAY_YUV410: | ||
| 704 | cmd |= OCMD_YUV_410_PLANAR; | ||
| 705 | break; | ||
| 706 | } | ||
| 707 | } else { /* YUV packed */ | ||
| 708 | switch (params->format & I915_OVERLAY_DEPTH_MASK) { | ||
| 709 | case I915_OVERLAY_YUV422: | ||
| 710 | cmd |= OCMD_YUV_422_PACKED; | ||
| 711 | break; | ||
| 712 | case I915_OVERLAY_YUV411: | ||
| 713 | cmd |= OCMD_YUV_411_PACKED; | ||
| 714 | break; | ||
| 715 | } | ||
| 716 | |||
| 717 | switch (params->format & I915_OVERLAY_SWAP_MASK) { | ||
| 718 | case I915_OVERLAY_NO_SWAP: | ||
| 719 | break; | ||
| 720 | case I915_OVERLAY_UV_SWAP: | ||
| 721 | cmd |= OCMD_UV_SWAP; | ||
| 722 | break; | ||
| 723 | case I915_OVERLAY_Y_SWAP: | ||
| 724 | cmd |= OCMD_Y_SWAP; | ||
| 725 | break; | ||
| 726 | case I915_OVERLAY_Y_AND_UV_SWAP: | ||
| 727 | cmd |= OCMD_Y_AND_UV_SWAP; | ||
| 728 | break; | ||
| 729 | } | ||
| 730 | } | ||
| 731 | |||
| 732 | return cmd; | ||
| 733 | } | ||
| 734 | |||
| 735 | int intel_overlay_do_put_image(struct intel_overlay *overlay, | ||
| 736 | struct drm_gem_object *new_bo, | ||
| 737 | struct put_image_params *params) | ||
| 738 | { | ||
| 739 | int ret, tmp_width; | ||
| 740 | struct overlay_registers *regs; | ||
| 741 | bool scale_changed = false; | ||
| 742 | struct drm_i915_gem_object *bo_priv = new_bo->driver_private; | ||
| 743 | struct drm_device *dev = overlay->dev; | ||
| 744 | |||
| 745 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); | ||
| 746 | BUG_ON(!mutex_is_locked(&dev->mode_config.mutex)); | ||
| 747 | BUG_ON(!overlay); | ||
| 748 | |||
| 749 | ret = intel_overlay_release_old_vid(overlay); | ||
| 750 | if (ret != 0) | ||
| 751 | return ret; | ||
| 752 | |||
| 753 | ret = i915_gem_object_pin(new_bo, PAGE_SIZE); | ||
| 754 | if (ret != 0) | ||
| 755 | return ret; | ||
| 756 | |||
| 757 | ret = i915_gem_object_set_to_gtt_domain(new_bo, 0); | ||
| 758 | if (ret != 0) | ||
| 759 | goto out_unpin; | ||
| 760 | |||
| 761 | if (!overlay->active) { | ||
| 762 | regs = intel_overlay_map_regs_atomic(overlay); | ||
| 763 | if (!regs) { | ||
| 764 | ret = -ENOMEM; | ||
| 765 | goto out_unpin; | ||
| 766 | } | ||
| 767 | regs->OCONFIG = OCONF_CC_OUT_8BIT; | ||
| 768 | if (IS_I965GM(overlay->dev)) | ||
| 769 | regs->OCONFIG |= OCONF_CSC_MODE_BT709; | ||
| 770 | regs->OCONFIG |= overlay->crtc->pipe == 0 ? | ||
| 771 | OCONF_PIPE_A : OCONF_PIPE_B; | ||
| 772 | intel_overlay_unmap_regs_atomic(overlay); | ||
| 773 | |||
| 774 | ret = intel_overlay_on(overlay); | ||
| 775 | if (ret != 0) | ||
| 776 | goto out_unpin; | ||
| 777 | } | ||
| 778 | |||
| 779 | regs = intel_overlay_map_regs_atomic(overlay); | ||
| 780 | if (!regs) { | ||
| 781 | ret = -ENOMEM; | ||
| 782 | goto out_unpin; | ||
| 783 | } | ||
| 784 | |||
| 785 | regs->DWINPOS = (params->dst_y << 16) | params->dst_x; | ||
| 786 | regs->DWINSZ = (params->dst_h << 16) | params->dst_w; | ||
| 787 | |||
| 788 | if (params->format & I915_OVERLAY_YUV_PACKED) | ||
| 789 | tmp_width = packed_width_bytes(params->format, params->src_w); | ||
| 790 | else | ||
| 791 | tmp_width = params->src_w; | ||
| 792 | |||
| 793 | regs->SWIDTH = params->src_w; | ||
| 794 | regs->SWIDTHSW = calc_swidthsw(overlay->dev, | ||
| 795 | params->offset_Y, tmp_width); | ||
| 796 | regs->SHEIGHT = params->src_h; | ||
| 797 | regs->OBUF_0Y = bo_priv->gtt_offset + params-> offset_Y; | ||
| 798 | regs->OSTRIDE = params->stride_Y; | ||
| 799 | |||
| 800 | if (params->format & I915_OVERLAY_YUV_PLANAR) { | ||
| 801 | int uv_hscale = uv_hsubsampling(params->format); | ||
| 802 | int uv_vscale = uv_vsubsampling(params->format); | ||
| 803 | u32 tmp_U, tmp_V; | ||
| 804 | regs->SWIDTH |= (params->src_w/uv_hscale) << 16; | ||
| 805 | tmp_U = calc_swidthsw(overlay->dev, params->offset_U, | ||
| 806 | params->src_w/uv_hscale); | ||
| 807 | tmp_V = calc_swidthsw(overlay->dev, params->offset_V, | ||
| 808 | params->src_w/uv_hscale); | ||
| 809 | regs->SWIDTHSW |= max_t(u32, tmp_U, tmp_V) << 16; | ||
| 810 | regs->SHEIGHT |= (params->src_h/uv_vscale) << 16; | ||
| 811 | regs->OBUF_0U = bo_priv->gtt_offset + params->offset_U; | ||
| 812 | regs->OBUF_0V = bo_priv->gtt_offset + params->offset_V; | ||
| 813 | regs->OSTRIDE |= params->stride_UV << 16; | ||
| 814 | } | ||
| 815 | |||
| 816 | scale_changed = update_scaling_factors(overlay, regs, params); | ||
| 817 | |||
| 818 | update_colorkey(overlay, regs); | ||
| 819 | |||
| 820 | regs->OCMD = overlay_cmd_reg(params); | ||
| 821 | |||
| 822 | intel_overlay_unmap_regs_atomic(overlay); | ||
| 823 | |||
| 824 | intel_overlay_continue(overlay, scale_changed); | ||
| 825 | |||
| 826 | overlay->old_vid_bo = overlay->vid_bo; | ||
| 827 | overlay->vid_bo = new_bo->driver_private; | ||
| 828 | |||
| 829 | return 0; | ||
| 830 | |||
| 831 | out_unpin: | ||
| 832 | i915_gem_object_unpin(new_bo); | ||
| 833 | return ret; | ||
| 834 | } | ||
| 835 | |||
| 836 | int intel_overlay_switch_off(struct intel_overlay *overlay) | ||
| 837 | { | ||
| 838 | int ret; | ||
| 839 | struct overlay_registers *regs; | ||
| 840 | struct drm_device *dev = overlay->dev; | ||
| 841 | |||
| 842 | BUG_ON(!mutex_is_locked(&dev->struct_mutex)); | ||
| 843 | BUG_ON(!mutex_is_locked(&dev->mode_config.mutex)); | ||
| 844 | |||
| 845 | if (overlay->hw_wedged) { | ||
| 846 | ret = intel_overlay_recover_from_interrupt(overlay, 1); | ||
| 847 | if (ret != 0) | ||
| 848 | return ret; | ||
| 849 | } | ||
| 850 | |||
| 851 | if (!overlay->active) | ||
| 852 | return 0; | ||
| 853 | |||
| 854 | ret = intel_overlay_release_old_vid(overlay); | ||
| 855 | if (ret != 0) | ||
| 856 | return ret; | ||
| 857 | |||
| 858 | regs = intel_overlay_map_regs_atomic(overlay); | ||
| 859 | regs->OCMD = 0; | ||
| 860 | intel_overlay_unmap_regs_atomic(overlay); | ||
| 861 | |||
| 862 | ret = intel_overlay_off(overlay); | ||
| 863 | if (ret != 0) | ||
| 864 | return ret; | ||
| 865 | |||
| 866 | intel_overlay_off_tail(overlay); | ||
| 867 | |||
| 868 | return 0; | ||
| 869 | } | ||
| 870 | |||
| 871 | static int check_overlay_possible_on_crtc(struct intel_overlay *overlay, | ||
| 872 | struct intel_crtc *crtc) | ||
| 873 | { | ||
| 874 | drm_i915_private_t *dev_priv = overlay->dev->dev_private; | ||
| 875 | u32 pipeconf; | ||
| 876 | int pipeconf_reg = (crtc->pipe == 0) ? PIPEACONF : PIPEBCONF; | ||
| 877 | |||
| 878 | if (!crtc->base.enabled || crtc->dpms_mode != DRM_MODE_DPMS_ON) | ||
| 879 | return -EINVAL; | ||
| 880 | |||
| 881 | pipeconf = I915_READ(pipeconf_reg); | ||
| 882 | |||
| 883 | /* can't use the overlay with double wide pipe */ | ||
| 884 | if (!IS_I965G(overlay->dev) && pipeconf & PIPEACONF_DOUBLE_WIDE) | ||
| 885 | return -EINVAL; | ||
| 886 | |||
| 887 | return 0; | ||
| 888 | } | ||
| 889 | |||
| 890 | static void update_pfit_vscale_ratio(struct intel_overlay *overlay) | ||
| 891 | { | ||
| 892 | struct drm_device *dev = overlay->dev; | ||
| 893 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 894 | u32 ratio; | ||
| 895 | u32 pfit_control = I915_READ(PFIT_CONTROL); | ||
| 896 | |||
| 897 | /* XXX: This is not the same logic as in the xorg driver, but more in | ||
| 898 | * line with the intel documentation for the i965 */ | ||
| 899 | if (!IS_I965G(dev) && (pfit_control & VERT_AUTO_SCALE)) { | ||
| 900 | ratio = I915_READ(PFIT_AUTO_RATIOS) >> PFIT_VERT_SCALE_SHIFT; | ||
| 901 | } else { /* on i965 use the PGM reg to read out the autoscaler values */ | ||
| 902 | ratio = I915_READ(PFIT_PGM_RATIOS); | ||
| 903 | if (IS_I965G(dev)) | ||
| 904 | ratio >>= PFIT_VERT_SCALE_SHIFT_965; | ||
| 905 | else | ||
| 906 | ratio >>= PFIT_VERT_SCALE_SHIFT; | ||
| 907 | } | ||
| 908 | |||
| 909 | overlay->pfit_vscale_ratio = ratio; | ||
| 910 | } | ||
| 911 | |||
| 912 | static int check_overlay_dst(struct intel_overlay *overlay, | ||
| 913 | struct drm_intel_overlay_put_image *rec) | ||
| 914 | { | ||
| 915 | struct drm_display_mode *mode = &overlay->crtc->base.mode; | ||
| 916 | |||
| 917 | if ((rec->dst_x < mode->crtc_hdisplay) | ||
| 918 | && (rec->dst_x + rec->dst_width | ||
| 919 | <= mode->crtc_hdisplay) | ||
| 920 | && (rec->dst_y < mode->crtc_vdisplay) | ||
| 921 | && (rec->dst_y + rec->dst_height | ||
| 922 | <= mode->crtc_vdisplay)) | ||
| 923 | return 0; | ||
| 924 | else | ||
| 925 | return -EINVAL; | ||
| 926 | } | ||
| 927 | |||
| 928 | static int check_overlay_scaling(struct put_image_params *rec) | ||
| 929 | { | ||
| 930 | u32 tmp; | ||
| 931 | |||
| 932 | /* downscaling limit is 8.0 */ | ||
| 933 | tmp = ((rec->src_scan_h << 16) / rec->dst_h) >> 16; | ||
| 934 | if (tmp > 7) | ||
| 935 | return -EINVAL; | ||
| 936 | tmp = ((rec->src_scan_w << 16) / rec->dst_w) >> 16; | ||
| 937 | if (tmp > 7) | ||
| 938 | return -EINVAL; | ||
| 939 | |||
| 940 | return 0; | ||
| 941 | } | ||
| 942 | |||
| 943 | static int check_overlay_src(struct drm_device *dev, | ||
| 944 | struct drm_intel_overlay_put_image *rec, | ||
| 945 | struct drm_gem_object *new_bo) | ||
| 946 | { | ||
| 947 | u32 stride_mask; | ||
| 948 | int depth; | ||
| 949 | int uv_hscale = uv_hsubsampling(rec->flags); | ||
| 950 | int uv_vscale = uv_vsubsampling(rec->flags); | ||
| 951 | size_t tmp; | ||
| 952 | |||
| 953 | /* check src dimensions */ | ||
| 954 | if (IS_845G(dev) || IS_I830(dev)) { | ||
| 955 | if (rec->src_height > IMAGE_MAX_HEIGHT_LEGACY | ||
| 956 | || rec->src_width > IMAGE_MAX_WIDTH_LEGACY) | ||
| 957 | return -EINVAL; | ||
| 958 | } else { | ||
| 959 | if (rec->src_height > IMAGE_MAX_HEIGHT | ||
| 960 | || rec->src_width > IMAGE_MAX_WIDTH) | ||
| 961 | return -EINVAL; | ||
| 962 | } | ||
| 963 | /* better safe than sorry, use 4 as the maximal subsampling ratio */ | ||
| 964 | if (rec->src_height < N_VERT_Y_TAPS*4 | ||
| 965 | || rec->src_width < N_HORIZ_Y_TAPS*4) | ||
| 966 | return -EINVAL; | ||
| 967 | |||
| 968 | /* check alingment constrains */ | ||
| 969 | switch (rec->flags & I915_OVERLAY_TYPE_MASK) { | ||
| 970 | case I915_OVERLAY_RGB: | ||
| 971 | /* not implemented */ | ||
| 972 | return -EINVAL; | ||
| 973 | case I915_OVERLAY_YUV_PACKED: | ||
| 974 | depth = packed_depth_bytes(rec->flags); | ||
| 975 | if (uv_vscale != 1) | ||
| 976 | return -EINVAL; | ||
| 977 | if (depth < 0) | ||
| 978 | return depth; | ||
| 979 | /* ignore UV planes */ | ||
| 980 | rec->stride_UV = 0; | ||
| 981 | rec->offset_U = 0; | ||
| 982 | rec->offset_V = 0; | ||
| 983 | /* check pixel alignment */ | ||
| 984 | if (rec->offset_Y % depth) | ||
| 985 | return -EINVAL; | ||
| 986 | break; | ||
| 987 | case I915_OVERLAY_YUV_PLANAR: | ||
| 988 | if (uv_vscale < 0 || uv_hscale < 0) | ||
| 989 | return -EINVAL; | ||
| 990 | /* no offset restrictions for planar formats */ | ||
| 991 | break; | ||
| 992 | default: | ||
| 993 | return -EINVAL; | ||
| 994 | } | ||
| 995 | |||
| 996 | if (rec->src_width % uv_hscale) | ||
| 997 | return -EINVAL; | ||
| 998 | |||
| 999 | /* stride checking */ | ||
| 1000 | stride_mask = 63; | ||
| 1001 | |||
| 1002 | if (rec->stride_Y & stride_mask || rec->stride_UV & stride_mask) | ||
| 1003 | return -EINVAL; | ||
| 1004 | if (IS_I965G(dev) && rec->stride_Y < 512) | ||
| 1005 | return -EINVAL; | ||
| 1006 | |||
| 1007 | tmp = (rec->flags & I915_OVERLAY_TYPE_MASK) == I915_OVERLAY_YUV_PLANAR ? | ||
| 1008 | 4 : 8; | ||
| 1009 | if (rec->stride_Y > tmp*1024 || rec->stride_UV > 2*1024) | ||
| 1010 | return -EINVAL; | ||
| 1011 | |||
| 1012 | /* check buffer dimensions */ | ||
| 1013 | switch (rec->flags & I915_OVERLAY_TYPE_MASK) { | ||
| 1014 | case I915_OVERLAY_RGB: | ||
| 1015 | case I915_OVERLAY_YUV_PACKED: | ||
| 1016 | /* always 4 Y values per depth pixels */ | ||
| 1017 | if (packed_width_bytes(rec->flags, rec->src_width) | ||
| 1018 | > rec->stride_Y) | ||
| 1019 | return -EINVAL; | ||
| 1020 | |||
| 1021 | tmp = rec->stride_Y*rec->src_height; | ||
| 1022 | if (rec->offset_Y + tmp > new_bo->size) | ||
| 1023 | return -EINVAL; | ||
| 1024 | break; | ||
| 1025 | case I915_OVERLAY_YUV_PLANAR: | ||
| 1026 | if (rec->src_width > rec->stride_Y) | ||
| 1027 | return -EINVAL; | ||
| 1028 | if (rec->src_width/uv_hscale > rec->stride_UV) | ||
| 1029 | return -EINVAL; | ||
| 1030 | |||
| 1031 | tmp = rec->stride_Y*rec->src_height; | ||
| 1032 | if (rec->offset_Y + tmp > new_bo->size) | ||
| 1033 | return -EINVAL; | ||
| 1034 | tmp = rec->stride_UV*rec->src_height; | ||
| 1035 | tmp /= uv_vscale; | ||
| 1036 | if (rec->offset_U + tmp > new_bo->size | ||
| 1037 | || rec->offset_V + tmp > new_bo->size) | ||
| 1038 | return -EINVAL; | ||
| 1039 | break; | ||
| 1040 | } | ||
| 1041 | |||
| 1042 | return 0; | ||
| 1043 | } | ||
| 1044 | |||
| 1045 | int intel_overlay_put_image(struct drm_device *dev, void *data, | ||
| 1046 | struct drm_file *file_priv) | ||
| 1047 | { | ||
| 1048 | struct drm_intel_overlay_put_image *put_image_rec = data; | ||
| 1049 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 1050 | struct intel_overlay *overlay; | ||
| 1051 | struct drm_mode_object *drmmode_obj; | ||
| 1052 | struct intel_crtc *crtc; | ||
| 1053 | struct drm_gem_object *new_bo; | ||
| 1054 | struct put_image_params *params; | ||
| 1055 | int ret; | ||
| 1056 | |||
| 1057 | if (!dev_priv) { | ||
| 1058 | DRM_ERROR("called with no initialization\n"); | ||
| 1059 | return -EINVAL; | ||
| 1060 | } | ||
| 1061 | |||
| 1062 | overlay = dev_priv->overlay; | ||
| 1063 | if (!overlay) { | ||
| 1064 | DRM_DEBUG("userspace bug: no overlay\n"); | ||
| 1065 | return -ENODEV; | ||
| 1066 | } | ||
| 1067 | |||
| 1068 | if (!(put_image_rec->flags & I915_OVERLAY_ENABLE)) { | ||
| 1069 | mutex_lock(&dev->mode_config.mutex); | ||
| 1070 | mutex_lock(&dev->struct_mutex); | ||
| 1071 | |||
| 1072 | ret = intel_overlay_switch_off(overlay); | ||
| 1073 | |||
| 1074 | mutex_unlock(&dev->struct_mutex); | ||
| 1075 | mutex_unlock(&dev->mode_config.mutex); | ||
| 1076 | |||
| 1077 | return ret; | ||
| 1078 | } | ||
| 1079 | |||
| 1080 | params = kmalloc(sizeof(struct put_image_params), GFP_KERNEL); | ||
| 1081 | if (!params) | ||
| 1082 | return -ENOMEM; | ||
| 1083 | |||
| 1084 | drmmode_obj = drm_mode_object_find(dev, put_image_rec->crtc_id, | ||
| 1085 | DRM_MODE_OBJECT_CRTC); | ||
| 1086 | if (!drmmode_obj) | ||
| 1087 | return -ENOENT; | ||
| 1088 | crtc = to_intel_crtc(obj_to_crtc(drmmode_obj)); | ||
| 1089 | |||
| 1090 | new_bo = drm_gem_object_lookup(dev, file_priv, | ||
| 1091 | put_image_rec->bo_handle); | ||
| 1092 | if (!new_bo) | ||
| 1093 | return -ENOENT; | ||
| 1094 | |||
| 1095 | mutex_lock(&dev->mode_config.mutex); | ||
| 1096 | mutex_lock(&dev->struct_mutex); | ||
| 1097 | |||
| 1098 | if (overlay->hw_wedged) { | ||
| 1099 | ret = intel_overlay_recover_from_interrupt(overlay, 1); | ||
| 1100 | if (ret != 0) | ||
| 1101 | goto out_unlock; | ||
| 1102 | } | ||
| 1103 | |||
| 1104 | if (overlay->crtc != crtc) { | ||
| 1105 | struct drm_display_mode *mode = &crtc->base.mode; | ||
| 1106 | ret = intel_overlay_switch_off(overlay); | ||
| 1107 | if (ret != 0) | ||
| 1108 | goto out_unlock; | ||
| 1109 | |||
| 1110 | ret = check_overlay_possible_on_crtc(overlay, crtc); | ||
| 1111 | if (ret != 0) | ||
| 1112 | goto out_unlock; | ||
| 1113 | |||
| 1114 | overlay->crtc = crtc; | ||
| 1115 | crtc->overlay = overlay; | ||
| 1116 | |||
| 1117 | if (intel_panel_fitter_pipe(dev) == crtc->pipe | ||
| 1118 | /* and line to wide, i.e. one-line-mode */ | ||
| 1119 | && mode->hdisplay > 1024) { | ||
| 1120 | overlay->pfit_active = 1; | ||
| 1121 | update_pfit_vscale_ratio(overlay); | ||
| 1122 | } else | ||
| 1123 | overlay->pfit_active = 0; | ||
| 1124 | } | ||
| 1125 | |||
| 1126 | ret = check_overlay_dst(overlay, put_image_rec); | ||
| 1127 | if (ret != 0) | ||
| 1128 | goto out_unlock; | ||
| 1129 | |||
| 1130 | if (overlay->pfit_active) { | ||
| 1131 | params->dst_y = ((((u32)put_image_rec->dst_y) << 12) / | ||
| 1132 | overlay->pfit_vscale_ratio); | ||
| 1133 | /* shifting right rounds downwards, so add 1 */ | ||
| 1134 | params->dst_h = ((((u32)put_image_rec->dst_height) << 12) / | ||
| 1135 | overlay->pfit_vscale_ratio) + 1; | ||
| 1136 | } else { | ||
| 1137 | params->dst_y = put_image_rec->dst_y; | ||
| 1138 | params->dst_h = put_image_rec->dst_height; | ||
| 1139 | } | ||
| 1140 | params->dst_x = put_image_rec->dst_x; | ||
| 1141 | params->dst_w = put_image_rec->dst_width; | ||
| 1142 | |||
| 1143 | params->src_w = put_image_rec->src_width; | ||
| 1144 | params->src_h = put_image_rec->src_height; | ||
| 1145 | params->src_scan_w = put_image_rec->src_scan_width; | ||
| 1146 | params->src_scan_h = put_image_rec->src_scan_height; | ||
| 1147 | if (params->src_scan_h > params->src_h | ||
| 1148 | || params->src_scan_w > params->src_w) { | ||
| 1149 | ret = -EINVAL; | ||
| 1150 | goto out_unlock; | ||
| 1151 | } | ||
| 1152 | |||
| 1153 | ret = check_overlay_src(dev, put_image_rec, new_bo); | ||
| 1154 | if (ret != 0) | ||
| 1155 | goto out_unlock; | ||
| 1156 | params->format = put_image_rec->flags & ~I915_OVERLAY_FLAGS_MASK; | ||
| 1157 | params->stride_Y = put_image_rec->stride_Y; | ||
| 1158 | params->stride_UV = put_image_rec->stride_UV; | ||
| 1159 | params->offset_Y = put_image_rec->offset_Y; | ||
| 1160 | params->offset_U = put_image_rec->offset_U; | ||
| 1161 | params->offset_V = put_image_rec->offset_V; | ||
| 1162 | |||
| 1163 | /* Check scaling after src size to prevent a divide-by-zero. */ | ||
| 1164 | ret = check_overlay_scaling(params); | ||
| 1165 | if (ret != 0) | ||
| 1166 | goto out_unlock; | ||
| 1167 | |||
| 1168 | ret = intel_overlay_do_put_image(overlay, new_bo, params); | ||
| 1169 | if (ret != 0) | ||
| 1170 | goto out_unlock; | ||
| 1171 | |||
| 1172 | mutex_unlock(&dev->struct_mutex); | ||
| 1173 | mutex_unlock(&dev->mode_config.mutex); | ||
| 1174 | |||
| 1175 | kfree(params); | ||
| 1176 | |||
| 1177 | return 0; | ||
| 1178 | |||
| 1179 | out_unlock: | ||
| 1180 | mutex_unlock(&dev->struct_mutex); | ||
| 1181 | mutex_unlock(&dev->mode_config.mutex); | ||
| 1182 | drm_gem_object_unreference(new_bo); | ||
| 1183 | kfree(params); | ||
| 1184 | |||
| 1185 | return ret; | ||
| 1186 | } | ||
| 1187 | |||
| 1188 | static void update_reg_attrs(struct intel_overlay *overlay, | ||
| 1189 | struct overlay_registers *regs) | ||
| 1190 | { | ||
| 1191 | regs->OCLRC0 = (overlay->contrast << 18) | (overlay->brightness & 0xff); | ||
| 1192 | regs->OCLRC1 = overlay->saturation; | ||
| 1193 | } | ||
| 1194 | |||
| 1195 | static bool check_gamma_bounds(u32 gamma1, u32 gamma2) | ||
| 1196 | { | ||
| 1197 | int i; | ||
| 1198 | |||
| 1199 | if (gamma1 & 0xff000000 || gamma2 & 0xff000000) | ||
| 1200 | return false; | ||
| 1201 | |||
| 1202 | for (i = 0; i < 3; i++) { | ||
| 1203 | if (((gamma1 >> i * 8) & 0xff) >= ((gamma2 >> i*8) & 0xff)) | ||
| 1204 | return false; | ||
| 1205 | } | ||
| 1206 | |||
| 1207 | return true; | ||
| 1208 | } | ||
| 1209 | |||
| 1210 | static bool check_gamma5_errata(u32 gamma5) | ||
| 1211 | { | ||
| 1212 | int i; | ||
| 1213 | |||
| 1214 | for (i = 0; i < 3; i++) { | ||
| 1215 | if (((gamma5 >> i*8) & 0xff) == 0x80) | ||
| 1216 | return false; | ||
| 1217 | } | ||
| 1218 | |||
| 1219 | return true; | ||
| 1220 | } | ||
| 1221 | |||
| 1222 | static int check_gamma(struct drm_intel_overlay_attrs *attrs) | ||
| 1223 | { | ||
| 1224 | if (!check_gamma_bounds(0, attrs->gamma0) | ||
| 1225 | || !check_gamma_bounds(attrs->gamma0, attrs->gamma1) | ||
| 1226 | || !check_gamma_bounds(attrs->gamma1, attrs->gamma2) | ||
| 1227 | || !check_gamma_bounds(attrs->gamma2, attrs->gamma3) | ||
| 1228 | || !check_gamma_bounds(attrs->gamma3, attrs->gamma4) | ||
| 1229 | || !check_gamma_bounds(attrs->gamma4, attrs->gamma5) | ||
| 1230 | || !check_gamma_bounds(attrs->gamma5, 0x00ffffff)) | ||
| 1231 | return -EINVAL; | ||
| 1232 | if (!check_gamma5_errata(attrs->gamma5)) | ||
| 1233 | return -EINVAL; | ||
| 1234 | return 0; | ||
| 1235 | } | ||
| 1236 | |||
| 1237 | int intel_overlay_attrs(struct drm_device *dev, void *data, | ||
| 1238 | struct drm_file *file_priv) | ||
| 1239 | { | ||
| 1240 | struct drm_intel_overlay_attrs *attrs = data; | ||
| 1241 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 1242 | struct intel_overlay *overlay; | ||
| 1243 | struct overlay_registers *regs; | ||
| 1244 | int ret; | ||
| 1245 | |||
| 1246 | if (!dev_priv) { | ||
| 1247 | DRM_ERROR("called with no initialization\n"); | ||
| 1248 | return -EINVAL; | ||
| 1249 | } | ||
| 1250 | |||
| 1251 | overlay = dev_priv->overlay; | ||
| 1252 | if (!overlay) { | ||
| 1253 | DRM_DEBUG("userspace bug: no overlay\n"); | ||
| 1254 | return -ENODEV; | ||
| 1255 | } | ||
| 1256 | |||
| 1257 | mutex_lock(&dev->mode_config.mutex); | ||
| 1258 | mutex_lock(&dev->struct_mutex); | ||
| 1259 | |||
| 1260 | if (!(attrs->flags & I915_OVERLAY_UPDATE_ATTRS)) { | ||
| 1261 | attrs->color_key = overlay->color_key; | ||
| 1262 | attrs->brightness = overlay->brightness; | ||
| 1263 | attrs->contrast = overlay->contrast; | ||
| 1264 | attrs->saturation = overlay->saturation; | ||
| 1265 | |||
| 1266 | if (IS_I9XX(dev)) { | ||
| 1267 | attrs->gamma0 = I915_READ(OGAMC0); | ||
| 1268 | attrs->gamma1 = I915_READ(OGAMC1); | ||
| 1269 | attrs->gamma2 = I915_READ(OGAMC2); | ||
| 1270 | attrs->gamma3 = I915_READ(OGAMC3); | ||
| 1271 | attrs->gamma4 = I915_READ(OGAMC4); | ||
| 1272 | attrs->gamma5 = I915_READ(OGAMC5); | ||
| 1273 | } | ||
| 1274 | ret = 0; | ||
| 1275 | } else { | ||
| 1276 | overlay->color_key = attrs->color_key; | ||
| 1277 | if (attrs->brightness >= -128 && attrs->brightness <= 127) { | ||
| 1278 | overlay->brightness = attrs->brightness; | ||
| 1279 | } else { | ||
| 1280 | ret = -EINVAL; | ||
| 1281 | goto out_unlock; | ||
| 1282 | } | ||
| 1283 | if (attrs->contrast <= 255) { | ||
| 1284 | overlay->contrast = attrs->contrast; | ||
| 1285 | } else { | ||
| 1286 | ret = -EINVAL; | ||
| 1287 | goto out_unlock; | ||
| 1288 | } | ||
| 1289 | if (attrs->saturation <= 1023) { | ||
| 1290 | overlay->saturation = attrs->saturation; | ||
| 1291 | } else { | ||
| 1292 | ret = -EINVAL; | ||
| 1293 | goto out_unlock; | ||
| 1294 | } | ||
| 1295 | |||
| 1296 | regs = intel_overlay_map_regs_atomic(overlay); | ||
| 1297 | if (!regs) { | ||
| 1298 | ret = -ENOMEM; | ||
| 1299 | goto out_unlock; | ||
| 1300 | } | ||
| 1301 | |||
| 1302 | update_reg_attrs(overlay, regs); | ||
| 1303 | |||
| 1304 | intel_overlay_unmap_regs_atomic(overlay); | ||
| 1305 | |||
| 1306 | if (attrs->flags & I915_OVERLAY_UPDATE_GAMMA) { | ||
| 1307 | if (!IS_I9XX(dev)) { | ||
| 1308 | ret = -EINVAL; | ||
| 1309 | goto out_unlock; | ||
| 1310 | } | ||
| 1311 | |||
| 1312 | if (overlay->active) { | ||
| 1313 | ret = -EBUSY; | ||
| 1314 | goto out_unlock; | ||
| 1315 | } | ||
| 1316 | |||
| 1317 | ret = check_gamma(attrs); | ||
| 1318 | if (ret != 0) | ||
| 1319 | goto out_unlock; | ||
| 1320 | |||
| 1321 | I915_WRITE(OGAMC0, attrs->gamma0); | ||
| 1322 | I915_WRITE(OGAMC1, attrs->gamma1); | ||
| 1323 | I915_WRITE(OGAMC2, attrs->gamma2); | ||
| 1324 | I915_WRITE(OGAMC3, attrs->gamma3); | ||
| 1325 | I915_WRITE(OGAMC4, attrs->gamma4); | ||
| 1326 | I915_WRITE(OGAMC5, attrs->gamma5); | ||
| 1327 | } | ||
| 1328 | ret = 0; | ||
| 1329 | } | ||
| 1330 | |||
| 1331 | out_unlock: | ||
| 1332 | mutex_unlock(&dev->struct_mutex); | ||
| 1333 | mutex_unlock(&dev->mode_config.mutex); | ||
| 1334 | |||
| 1335 | return ret; | ||
| 1336 | } | ||
| 1337 | |||
| 1338 | void intel_setup_overlay(struct drm_device *dev) | ||
| 1339 | { | ||
| 1340 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 1341 | struct intel_overlay *overlay; | ||
| 1342 | struct drm_gem_object *reg_bo; | ||
| 1343 | struct overlay_registers *regs; | ||
| 1344 | int ret; | ||
| 1345 | |||
| 1346 | if (!OVERLAY_EXISTS(dev)) | ||
| 1347 | return; | ||
| 1348 | |||
| 1349 | overlay = kzalloc(sizeof(struct intel_overlay), GFP_KERNEL); | ||
| 1350 | if (!overlay) | ||
| 1351 | return; | ||
| 1352 | overlay->dev = dev; | ||
| 1353 | |||
| 1354 | reg_bo = drm_gem_object_alloc(dev, PAGE_SIZE); | ||
| 1355 | if (!reg_bo) | ||
| 1356 | goto out_free; | ||
| 1357 | overlay->reg_bo = reg_bo->driver_private; | ||
| 1358 | |||
| 1359 | if (OVERLAY_NONPHYSICAL(dev)) { | ||
| 1360 | ret = i915_gem_object_pin(reg_bo, PAGE_SIZE); | ||
| 1361 | if (ret) { | ||
| 1362 | DRM_ERROR("failed to pin overlay register bo\n"); | ||
| 1363 | goto out_free_bo; | ||
| 1364 | } | ||
| 1365 | overlay->flip_addr = overlay->reg_bo->gtt_offset; | ||
| 1366 | } else { | ||
| 1367 | ret = i915_gem_attach_phys_object(dev, reg_bo, | ||
| 1368 | I915_GEM_PHYS_OVERLAY_REGS); | ||
| 1369 | if (ret) { | ||
| 1370 | DRM_ERROR("failed to attach phys overlay regs\n"); | ||
| 1371 | goto out_free_bo; | ||
| 1372 | } | ||
| 1373 | overlay->flip_addr = overlay->reg_bo->phys_obj->handle->busaddr; | ||
| 1374 | } | ||
| 1375 | |||
| 1376 | /* init all values */ | ||
| 1377 | overlay->color_key = 0x0101fe; | ||
| 1378 | overlay->brightness = -19; | ||
| 1379 | overlay->contrast = 75; | ||
| 1380 | overlay->saturation = 146; | ||
| 1381 | |||
| 1382 | regs = intel_overlay_map_regs_atomic(overlay); | ||
| 1383 | if (!regs) | ||
| 1384 | goto out_free_bo; | ||
| 1385 | |||
| 1386 | memset(regs, 0, sizeof(struct overlay_registers)); | ||
| 1387 | update_polyphase_filter(regs); | ||
| 1388 | |||
| 1389 | update_reg_attrs(overlay, regs); | ||
| 1390 | |||
| 1391 | intel_overlay_unmap_regs_atomic(overlay); | ||
| 1392 | |||
| 1393 | dev_priv->overlay = overlay; | ||
| 1394 | DRM_INFO("initialized overlay support\n"); | ||
| 1395 | return; | ||
| 1396 | |||
| 1397 | out_free_bo: | ||
| 1398 | drm_gem_object_unreference(reg_bo); | ||
| 1399 | out_free: | ||
| 1400 | kfree(overlay); | ||
| 1401 | return; | ||
| 1402 | } | ||
| 1403 | |||
| 1404 | void intel_cleanup_overlay(struct drm_device *dev) | ||
| 1405 | { | ||
| 1406 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
| 1407 | |||
| 1408 | if (dev_priv->overlay) { | ||
| 1409 | /* The bo's should be free'd by the generic code already. | ||
| 1410 | * Furthermore modesetting teardown happens beforehand so the | ||
| 1411 | * hardware should be off already */ | ||
| 1412 | BUG_ON(dev_priv->overlay->active); | ||
| 1413 | |||
| 1414 | kfree(dev_priv->overlay); | ||
| 1415 | } | ||
| 1416 | } | ||
diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c index e7fa3279e2f8..24a3dc99716c 100644 --- a/drivers/gpu/drm/i915/intel_sdvo.c +++ b/drivers/gpu/drm/i915/intel_sdvo.c | |||
| @@ -36,8 +36,6 @@ | |||
| 36 | #include "i915_drv.h" | 36 | #include "i915_drv.h" |
| 37 | #include "intel_sdvo_regs.h" | 37 | #include "intel_sdvo_regs.h" |
| 38 | 38 | ||
| 39 | #undef SDVO_DEBUG | ||
| 40 | |||
| 41 | static char *tv_format_names[] = { | 39 | static char *tv_format_names[] = { |
| 42 | "NTSC_M" , "NTSC_J" , "NTSC_443", | 40 | "NTSC_M" , "NTSC_J" , "NTSC_443", |
| 43 | "PAL_B" , "PAL_D" , "PAL_G" , | 41 | "PAL_B" , "PAL_D" , "PAL_G" , |
| @@ -356,7 +354,6 @@ static const struct _sdvo_cmd_name { | |||
| 356 | #define SDVO_NAME(dev_priv) ((dev_priv)->output_device == SDVOB ? "SDVOB" : "SDVOC") | 354 | #define SDVO_NAME(dev_priv) ((dev_priv)->output_device == SDVOB ? "SDVOB" : "SDVOC") |
| 357 | #define SDVO_PRIV(output) ((struct intel_sdvo_priv *) (output)->dev_priv) | 355 | #define SDVO_PRIV(output) ((struct intel_sdvo_priv *) (output)->dev_priv) |
| 358 | 356 | ||
| 359 | #ifdef SDVO_DEBUG | ||
| 360 | static void intel_sdvo_debug_write(struct intel_output *intel_output, u8 cmd, | 357 | static void intel_sdvo_debug_write(struct intel_output *intel_output, u8 cmd, |
| 361 | void *args, int args_len) | 358 | void *args, int args_len) |
| 362 | { | 359 | { |
| @@ -379,9 +376,6 @@ static void intel_sdvo_debug_write(struct intel_output *intel_output, u8 cmd, | |||
| 379 | DRM_LOG_KMS("(%02X)", cmd); | 376 | DRM_LOG_KMS("(%02X)", cmd); |
| 380 | DRM_LOG_KMS("\n"); | 377 | DRM_LOG_KMS("\n"); |
| 381 | } | 378 | } |
| 382 | #else | ||
| 383 | #define intel_sdvo_debug_write(o, c, a, l) | ||
| 384 | #endif | ||
| 385 | 379 | ||
| 386 | static void intel_sdvo_write_cmd(struct intel_output *intel_output, u8 cmd, | 380 | static void intel_sdvo_write_cmd(struct intel_output *intel_output, u8 cmd, |
| 387 | void *args, int args_len) | 381 | void *args, int args_len) |
| @@ -398,7 +392,6 @@ static void intel_sdvo_write_cmd(struct intel_output *intel_output, u8 cmd, | |||
| 398 | intel_sdvo_write_byte(intel_output, SDVO_I2C_OPCODE, cmd); | 392 | intel_sdvo_write_byte(intel_output, SDVO_I2C_OPCODE, cmd); |
| 399 | } | 393 | } |
| 400 | 394 | ||
| 401 | #ifdef SDVO_DEBUG | ||
| 402 | static const char *cmd_status_names[] = { | 395 | static const char *cmd_status_names[] = { |
| 403 | "Power on", | 396 | "Power on", |
| 404 | "Success", | 397 | "Success", |
| @@ -427,9 +420,6 @@ static void intel_sdvo_debug_response(struct intel_output *intel_output, | |||
| 427 | DRM_LOG_KMS("(??? %d)", status); | 420 | DRM_LOG_KMS("(??? %d)", status); |
| 428 | DRM_LOG_KMS("\n"); | 421 | DRM_LOG_KMS("\n"); |
| 429 | } | 422 | } |
| 430 | #else | ||
| 431 | #define intel_sdvo_debug_response(o, r, l, s) | ||
| 432 | #endif | ||
| 433 | 423 | ||
| 434 | static u8 intel_sdvo_read_response(struct intel_output *intel_output, | 424 | static u8 intel_sdvo_read_response(struct intel_output *intel_output, |
| 435 | void *response, int response_len) | 425 | void *response, int response_len) |
| @@ -1627,6 +1617,10 @@ static enum drm_connector_status intel_sdvo_detect(struct drm_connector *connect | |||
| 1627 | 1617 | ||
| 1628 | intel_sdvo_write_cmd(intel_output, | 1618 | intel_sdvo_write_cmd(intel_output, |
| 1629 | SDVO_CMD_GET_ATTACHED_DISPLAYS, NULL, 0); | 1619 | SDVO_CMD_GET_ATTACHED_DISPLAYS, NULL, 0); |
| 1620 | if (sdvo_priv->is_tv) { | ||
| 1621 | /* add 30ms delay when the output type is SDVO-TV */ | ||
| 1622 | mdelay(30); | ||
| 1623 | } | ||
| 1630 | status = intel_sdvo_read_response(intel_output, &response, 2); | 1624 | status = intel_sdvo_read_response(intel_output, &response, 2); |
| 1631 | 1625 | ||
| 1632 | DRM_DEBUG_KMS("SDVO response %d %d\n", response & 0xff, response >> 8); | 1626 | DRM_DEBUG_KMS("SDVO response %d %d\n", response & 0xff, response >> 8); |
diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c index 9ca917931afb..552ec110b741 100644 --- a/drivers/gpu/drm/i915/intel_tv.c +++ b/drivers/gpu/drm/i915/intel_tv.c | |||
| @@ -1213,20 +1213,17 @@ intel_tv_mode_set(struct drm_encoder *encoder, struct drm_display_mode *mode, | |||
| 1213 | tv_ctl |= TV_TRILEVEL_SYNC; | 1213 | tv_ctl |= TV_TRILEVEL_SYNC; |
| 1214 | if (tv_mode->pal_burst) | 1214 | if (tv_mode->pal_burst) |
| 1215 | tv_ctl |= TV_PAL_BURST; | 1215 | tv_ctl |= TV_PAL_BURST; |
| 1216 | |||
| 1216 | scctl1 = 0; | 1217 | scctl1 = 0; |
| 1217 | /* dda1 implies valid video levels */ | 1218 | if (tv_mode->dda1_inc) |
| 1218 | if (tv_mode->dda1_inc) { | ||
| 1219 | scctl1 |= TV_SC_DDA1_EN; | 1219 | scctl1 |= TV_SC_DDA1_EN; |
| 1220 | } | ||
| 1221 | |||
| 1222 | if (tv_mode->dda2_inc) | 1220 | if (tv_mode->dda2_inc) |
| 1223 | scctl1 |= TV_SC_DDA2_EN; | 1221 | scctl1 |= TV_SC_DDA2_EN; |
| 1224 | |||
| 1225 | if (tv_mode->dda3_inc) | 1222 | if (tv_mode->dda3_inc) |
| 1226 | scctl1 |= TV_SC_DDA3_EN; | 1223 | scctl1 |= TV_SC_DDA3_EN; |
| 1227 | |||
| 1228 | scctl1 |= tv_mode->sc_reset; | 1224 | scctl1 |= tv_mode->sc_reset; |
| 1229 | scctl1 |= video_levels->burst << TV_BURST_LEVEL_SHIFT; | 1225 | if (video_levels) |
| 1226 | scctl1 |= video_levels->burst << TV_BURST_LEVEL_SHIFT; | ||
| 1230 | scctl1 |= tv_mode->dda1_inc << TV_SCDDA1_INC_SHIFT; | 1227 | scctl1 |= tv_mode->dda1_inc << TV_SCDDA1_INC_SHIFT; |
| 1231 | 1228 | ||
| 1232 | scctl2 = tv_mode->dda2_size << TV_SCDDA2_SIZE_SHIFT | | 1229 | scctl2 = tv_mode->dda2_size << TV_SCDDA2_SIZE_SHIFT | |
| @@ -1416,16 +1413,16 @@ intel_tv_detect_type (struct drm_crtc *crtc, struct intel_output *intel_output) | |||
| 1416 | * 0 0 0 Component | 1413 | * 0 0 0 Component |
| 1417 | */ | 1414 | */ |
| 1418 | if ((tv_dac & TVDAC_SENSE_MASK) == (TVDAC_B_SENSE | TVDAC_C_SENSE)) { | 1415 | if ((tv_dac & TVDAC_SENSE_MASK) == (TVDAC_B_SENSE | TVDAC_C_SENSE)) { |
| 1419 | DRM_DEBUG("Detected Composite TV connection\n"); | 1416 | DRM_DEBUG_KMS("Detected Composite TV connection\n"); |
| 1420 | type = DRM_MODE_CONNECTOR_Composite; | 1417 | type = DRM_MODE_CONNECTOR_Composite; |
| 1421 | } else if ((tv_dac & (TVDAC_A_SENSE|TVDAC_B_SENSE)) == TVDAC_A_SENSE) { | 1418 | } else if ((tv_dac & (TVDAC_A_SENSE|TVDAC_B_SENSE)) == TVDAC_A_SENSE) { |
| 1422 | DRM_DEBUG("Detected S-Video TV connection\n"); | 1419 | DRM_DEBUG_KMS("Detected S-Video TV connection\n"); |
| 1423 | type = DRM_MODE_CONNECTOR_SVIDEO; | 1420 | type = DRM_MODE_CONNECTOR_SVIDEO; |
| 1424 | } else if ((tv_dac & TVDAC_SENSE_MASK) == 0) { | 1421 | } else if ((tv_dac & TVDAC_SENSE_MASK) == 0) { |
| 1425 | DRM_DEBUG("Detected Component TV connection\n"); | 1422 | DRM_DEBUG_KMS("Detected Component TV connection\n"); |
| 1426 | type = DRM_MODE_CONNECTOR_Component; | 1423 | type = DRM_MODE_CONNECTOR_Component; |
| 1427 | } else { | 1424 | } else { |
| 1428 | DRM_DEBUG("No TV connection detected\n"); | 1425 | DRM_DEBUG_KMS("No TV connection detected\n"); |
| 1429 | type = -1; | 1426 | type = -1; |
| 1430 | } | 1427 | } |
| 1431 | 1428 | ||
| @@ -1702,6 +1699,41 @@ static const struct drm_encoder_funcs intel_tv_enc_funcs = { | |||
| 1702 | .destroy = intel_tv_enc_destroy, | 1699 | .destroy = intel_tv_enc_destroy, |
| 1703 | }; | 1700 | }; |
| 1704 | 1701 | ||
| 1702 | /* | ||
| 1703 | * Enumerate the child dev array parsed from VBT to check whether | ||
| 1704 | * the integrated TV is present. | ||
| 1705 | * If it is present, return 1. | ||
| 1706 | * If it is not present, return false. | ||
| 1707 | * If no child dev is parsed from VBT, it assumes that the TV is present. | ||
| 1708 | */ | ||
| 1709 | static int tv_is_present_in_vbt(struct drm_device *dev) | ||
| 1710 | { | ||
| 1711 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
| 1712 | struct child_device_config *p_child; | ||
| 1713 | int i, ret; | ||
| 1714 | |||
| 1715 | if (!dev_priv->child_dev_num) | ||
| 1716 | return 1; | ||
| 1717 | |||
| 1718 | ret = 0; | ||
| 1719 | for (i = 0; i < dev_priv->child_dev_num; i++) { | ||
| 1720 | p_child = dev_priv->child_dev + i; | ||
| 1721 | /* | ||
| 1722 | * If the device type is not TV, continue. | ||
| 1723 | */ | ||
| 1724 | if (p_child->device_type != DEVICE_TYPE_INT_TV && | ||
| 1725 | p_child->device_type != DEVICE_TYPE_TV) | ||
| 1726 | continue; | ||
| 1727 | /* Only when the addin_offset is non-zero, it is regarded | ||
| 1728 | * as present. | ||
| 1729 | */ | ||
| 1730 | if (p_child->addin_offset) { | ||
| 1731 | ret = 1; | ||
| 1732 | break; | ||
| 1733 | } | ||
| 1734 | } | ||
| 1735 | return ret; | ||
| 1736 | } | ||
| 1705 | 1737 | ||
| 1706 | void | 1738 | void |
| 1707 | intel_tv_init(struct drm_device *dev) | 1739 | intel_tv_init(struct drm_device *dev) |
| @@ -1717,6 +1749,10 @@ intel_tv_init(struct drm_device *dev) | |||
| 1717 | if ((I915_READ(TV_CTL) & TV_FUSE_STATE_MASK) == TV_FUSE_STATE_DISABLED) | 1749 | if ((I915_READ(TV_CTL) & TV_FUSE_STATE_MASK) == TV_FUSE_STATE_DISABLED) |
| 1718 | return; | 1750 | return; |
| 1719 | 1751 | ||
| 1752 | if (!tv_is_present_in_vbt(dev)) { | ||
| 1753 | DRM_DEBUG_KMS("Integrated TV is not present.\n"); | ||
| 1754 | return; | ||
| 1755 | } | ||
| 1720 | /* Even if we have an encoder we may not have a connector */ | 1756 | /* Even if we have an encoder we may not have a connector */ |
| 1721 | if (!dev_priv->int_tv_support) | 1757 | if (!dev_priv->int_tv_support) |
| 1722 | return; | 1758 | return; |
diff --git a/drivers/gpu/drm/radeon/Makefile b/drivers/gpu/drm/radeon/Makefile index b5713eedd6e1..feb52eee4314 100644 --- a/drivers/gpu/drm/radeon/Makefile +++ b/drivers/gpu/drm/radeon/Makefile | |||
| @@ -49,7 +49,7 @@ radeon-y += radeon_device.o radeon_kms.o \ | |||
| 49 | radeon_cs.o radeon_bios.o radeon_benchmark.o r100.o r300.o r420.o \ | 49 | radeon_cs.o radeon_bios.o radeon_benchmark.o r100.o r300.o r420.o \ |
| 50 | rs400.o rs600.o rs690.o rv515.o r520.o r600.o rv770.o radeon_test.o \ | 50 | rs400.o rs600.o rs690.o rv515.o r520.o r600.o rv770.o radeon_test.o \ |
| 51 | r200.o radeon_legacy_tv.o r600_cs.o r600_blit.o r600_blit_shaders.o \ | 51 | r200.o radeon_legacy_tv.o r600_cs.o r600_blit.o r600_blit_shaders.o \ |
| 52 | r600_blit_kms.o radeon_pm.o | 52 | r600_blit_kms.o radeon_pm.o atombios_dp.o |
| 53 | 53 | ||
| 54 | radeon-$(CONFIG_COMPAT) += radeon_ioc32.o | 54 | radeon-$(CONFIG_COMPAT) += radeon_ioc32.o |
| 55 | 55 | ||
diff --git a/drivers/gpu/drm/radeon/atom.c b/drivers/gpu/drm/radeon/atom.c index d67c42555ab9..6578d19dff93 100644 --- a/drivers/gpu/drm/radeon/atom.c +++ b/drivers/gpu/drm/radeon/atom.c | |||
| @@ -263,10 +263,10 @@ static uint32_t atom_get_src_int(atom_exec_context *ctx, uint8_t attr, | |||
| 263 | case ATOM_ARG_FB: | 263 | case ATOM_ARG_FB: |
| 264 | idx = U8(*ptr); | 264 | idx = U8(*ptr); |
| 265 | (*ptr)++; | 265 | (*ptr)++; |
| 266 | val = gctx->scratch[((gctx->fb_base + idx) / 4)]; | ||
| 266 | if (print) | 267 | if (print) |
| 267 | DEBUG("FB[0x%02X]", idx); | 268 | DEBUG("FB[0x%02X]", idx); |
| 268 | printk(KERN_INFO "FB access is not implemented.\n"); | 269 | break; |
| 269 | return 0; | ||
| 270 | case ATOM_ARG_IMM: | 270 | case ATOM_ARG_IMM: |
| 271 | switch (align) { | 271 | switch (align) { |
| 272 | case ATOM_SRC_DWORD: | 272 | case ATOM_SRC_DWORD: |
| @@ -488,9 +488,9 @@ static void atom_put_dst(atom_exec_context *ctx, int arg, uint8_t attr, | |||
| 488 | case ATOM_ARG_FB: | 488 | case ATOM_ARG_FB: |
| 489 | idx = U8(*ptr); | 489 | idx = U8(*ptr); |
| 490 | (*ptr)++; | 490 | (*ptr)++; |
| 491 | gctx->scratch[((gctx->fb_base + idx) / 4)] = val; | ||
| 491 | DEBUG("FB[0x%02X]", idx); | 492 | DEBUG("FB[0x%02X]", idx); |
| 492 | printk(KERN_INFO "FB access is not implemented.\n"); | 493 | break; |
| 493 | return; | ||
| 494 | case ATOM_ARG_PLL: | 494 | case ATOM_ARG_PLL: |
| 495 | idx = U8(*ptr); | 495 | idx = U8(*ptr); |
| 496 | (*ptr)++; | 496 | (*ptr)++; |
| @@ -1214,3 +1214,28 @@ void atom_parse_cmd_header(struct atom_context *ctx, int index, uint8_t * frev, | |||
| 1214 | *crev = CU8(idx + 3); | 1214 | *crev = CU8(idx + 3); |
| 1215 | return; | 1215 | return; |
| 1216 | } | 1216 | } |
| 1217 | |||
| 1218 | int atom_allocate_fb_scratch(struct atom_context *ctx) | ||
| 1219 | { | ||
| 1220 | int index = GetIndexIntoMasterTable(DATA, VRAM_UsageByFirmware); | ||
| 1221 | uint16_t data_offset; | ||
| 1222 | int usage_bytes; | ||
| 1223 | struct _ATOM_VRAM_USAGE_BY_FIRMWARE *firmware_usage; | ||
| 1224 | |||
| 1225 | atom_parse_data_header(ctx, index, NULL, NULL, NULL, &data_offset); | ||
| 1226 | |||
| 1227 | firmware_usage = (struct _ATOM_VRAM_USAGE_BY_FIRMWARE *)(ctx->bios + data_offset); | ||
| 1228 | |||
| 1229 | DRM_DEBUG("atom firmware requested %08x %dkb\n", | ||
| 1230 | firmware_usage->asFirmwareVramReserveInfo[0].ulStartAddrUsedByFirmware, | ||
| 1231 | firmware_usage->asFirmwareVramReserveInfo[0].usFirmwareUseInKb); | ||
| 1232 | |||
| 1233 | usage_bytes = firmware_usage->asFirmwareVramReserveInfo[0].usFirmwareUseInKb * 1024; | ||
| 1234 | if (usage_bytes == 0) | ||
| 1235 | usage_bytes = 20 * 1024; | ||
| 1236 | /* allocate some scratch memory */ | ||
| 1237 | ctx->scratch = kzalloc(usage_bytes, GFP_KERNEL); | ||
| 1238 | if (!ctx->scratch) | ||
| 1239 | return -ENOMEM; | ||
| 1240 | return 0; | ||
| 1241 | } | ||
diff --git a/drivers/gpu/drm/radeon/atom.h b/drivers/gpu/drm/radeon/atom.h index e6eb38f2bcae..6671848e5ea1 100644 --- a/drivers/gpu/drm/radeon/atom.h +++ b/drivers/gpu/drm/radeon/atom.h | |||
| @@ -132,6 +132,7 @@ struct atom_context { | |||
| 132 | uint8_t shift; | 132 | uint8_t shift; |
| 133 | int cs_equal, cs_above; | 133 | int cs_equal, cs_above; |
| 134 | int io_mode; | 134 | int io_mode; |
| 135 | uint32_t *scratch; | ||
| 135 | }; | 136 | }; |
| 136 | 137 | ||
| 137 | extern int atom_debug; | 138 | extern int atom_debug; |
| @@ -142,6 +143,7 @@ int atom_asic_init(struct atom_context *); | |||
| 142 | void atom_destroy(struct atom_context *); | 143 | void atom_destroy(struct atom_context *); |
| 143 | void atom_parse_data_header(struct atom_context *ctx, int index, uint16_t *size, uint8_t *frev, uint8_t *crev, uint16_t *data_start); | 144 | void atom_parse_data_header(struct atom_context *ctx, int index, uint16_t *size, uint8_t *frev, uint8_t *crev, uint16_t *data_start); |
| 144 | void atom_parse_cmd_header(struct atom_context *ctx, int index, uint8_t *frev, uint8_t *crev); | 145 | void atom_parse_cmd_header(struct atom_context *ctx, int index, uint8_t *frev, uint8_t *crev); |
| 146 | int atom_allocate_fb_scratch(struct atom_context *ctx); | ||
| 145 | #include "atom-types.h" | 147 | #include "atom-types.h" |
| 146 | #include "atombios.h" | 148 | #include "atombios.h" |
| 147 | #include "ObjectID.h" | 149 | #include "ObjectID.h" |
diff --git a/drivers/gpu/drm/radeon/atombios.h b/drivers/gpu/drm/radeon/atombios.h index 6643afc36cea..5f48515c77a7 100644 --- a/drivers/gpu/drm/radeon/atombios.h +++ b/drivers/gpu/drm/radeon/atombios.h | |||
| @@ -2680,7 +2680,7 @@ typedef struct _ATOM_I2C_RECORD { | |||
| 2680 | typedef struct _ATOM_HPD_INT_RECORD { | 2680 | typedef struct _ATOM_HPD_INT_RECORD { |
| 2681 | ATOM_COMMON_RECORD_HEADER sheader; | 2681 | ATOM_COMMON_RECORD_HEADER sheader; |
| 2682 | UCHAR ucHPDIntGPIOID; /* Corresponding block in GPIO_PIN_INFO table gives the pin info */ | 2682 | UCHAR ucHPDIntGPIOID; /* Corresponding block in GPIO_PIN_INFO table gives the pin info */ |
| 2683 | UCHAR ucPluggged_PinState; | 2683 | UCHAR ucPlugged_PinState; |
| 2684 | } ATOM_HPD_INT_RECORD; | 2684 | } ATOM_HPD_INT_RECORD; |
| 2685 | 2685 | ||
| 2686 | typedef struct _ATOM_OUTPUT_PROTECTION_RECORD { | 2686 | typedef struct _ATOM_OUTPUT_PROTECTION_RECORD { |
diff --git a/drivers/gpu/drm/radeon/atombios_crtc.c b/drivers/gpu/drm/radeon/atombios_crtc.c index c15287a590ff..260fcf59f00c 100644 --- a/drivers/gpu/drm/radeon/atombios_crtc.c +++ b/drivers/gpu/drm/radeon/atombios_crtc.c | |||
| @@ -241,6 +241,7 @@ void atombios_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 241 | { | 241 | { |
| 242 | struct drm_device *dev = crtc->dev; | 242 | struct drm_device *dev = crtc->dev; |
| 243 | struct radeon_device *rdev = dev->dev_private; | 243 | struct radeon_device *rdev = dev->dev_private; |
| 244 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc); | ||
| 244 | 245 | ||
| 245 | switch (mode) { | 246 | switch (mode) { |
| 246 | case DRM_MODE_DPMS_ON: | 247 | case DRM_MODE_DPMS_ON: |
| @@ -248,20 +249,19 @@ void atombios_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 248 | if (ASIC_IS_DCE3(rdev)) | 249 | if (ASIC_IS_DCE3(rdev)) |
| 249 | atombios_enable_crtc_memreq(crtc, 1); | 250 | atombios_enable_crtc_memreq(crtc, 1); |
| 250 | atombios_blank_crtc(crtc, 0); | 251 | atombios_blank_crtc(crtc, 0); |
| 252 | drm_vblank_post_modeset(dev, radeon_crtc->crtc_id); | ||
| 253 | radeon_crtc_load_lut(crtc); | ||
| 251 | break; | 254 | break; |
| 252 | case DRM_MODE_DPMS_STANDBY: | 255 | case DRM_MODE_DPMS_STANDBY: |
| 253 | case DRM_MODE_DPMS_SUSPEND: | 256 | case DRM_MODE_DPMS_SUSPEND: |
| 254 | case DRM_MODE_DPMS_OFF: | 257 | case DRM_MODE_DPMS_OFF: |
| 258 | drm_vblank_pre_modeset(dev, radeon_crtc->crtc_id); | ||
| 255 | atombios_blank_crtc(crtc, 1); | 259 | atombios_blank_crtc(crtc, 1); |
| 256 | if (ASIC_IS_DCE3(rdev)) | 260 | if (ASIC_IS_DCE3(rdev)) |
| 257 | atombios_enable_crtc_memreq(crtc, 0); | 261 | atombios_enable_crtc_memreq(crtc, 0); |
| 258 | atombios_enable_crtc(crtc, 0); | 262 | atombios_enable_crtc(crtc, 0); |
| 259 | break; | 263 | break; |
| 260 | } | 264 | } |
| 261 | |||
| 262 | if (mode != DRM_MODE_DPMS_OFF) { | ||
| 263 | radeon_crtc_load_lut(crtc); | ||
| 264 | } | ||
| 265 | } | 265 | } |
| 266 | 266 | ||
| 267 | static void | 267 | static void |
| @@ -457,9 +457,8 @@ void atombios_crtc_set_pll(struct drm_crtc *crtc, struct drm_display_mode *mode) | |||
| 457 | if (encoder->encoder_type != | 457 | if (encoder->encoder_type != |
| 458 | DRM_MODE_ENCODER_DAC) | 458 | DRM_MODE_ENCODER_DAC) |
| 459 | pll_flags |= RADEON_PLL_NO_ODD_POST_DIV; | 459 | pll_flags |= RADEON_PLL_NO_ODD_POST_DIV; |
| 460 | if (!ASIC_IS_AVIVO(rdev) | 460 | if (encoder->encoder_type == |
| 461 | && (encoder->encoder_type == | 461 | DRM_MODE_ENCODER_LVDS) |
| 462 | DRM_MODE_ENCODER_LVDS)) | ||
| 463 | pll_flags |= RADEON_PLL_USE_REF_DIV; | 462 | pll_flags |= RADEON_PLL_USE_REF_DIV; |
| 464 | } | 463 | } |
| 465 | radeon_encoder = to_radeon_encoder(encoder); | 464 | radeon_encoder = to_radeon_encoder(encoder); |
| @@ -500,8 +499,18 @@ void atombios_crtc_set_pll(struct drm_crtc *crtc, struct drm_display_mode *mode) | |||
| 500 | else | 499 | else |
| 501 | pll = &rdev->clock.p2pll; | 500 | pll = &rdev->clock.p2pll; |
| 502 | 501 | ||
| 503 | radeon_compute_pll(pll, adjusted_clock, &pll_clock, &fb_div, &frac_fb_div, | 502 | if (ASIC_IS_AVIVO(rdev)) { |
| 504 | &ref_div, &post_div, pll_flags); | 503 | if (radeon_new_pll) |
| 504 | radeon_compute_pll_avivo(pll, adjusted_clock, &pll_clock, | ||
| 505 | &fb_div, &frac_fb_div, | ||
| 506 | &ref_div, &post_div, pll_flags); | ||
| 507 | else | ||
| 508 | radeon_compute_pll(pll, adjusted_clock, &pll_clock, | ||
| 509 | &fb_div, &frac_fb_div, | ||
| 510 | &ref_div, &post_div, pll_flags); | ||
| 511 | } else | ||
| 512 | radeon_compute_pll(pll, adjusted_clock, &pll_clock, &fb_div, &frac_fb_div, | ||
| 513 | &ref_div, &post_div, pll_flags); | ||
| 505 | 514 | ||
| 506 | index = GetIndexIntoMasterTable(COMMAND, SetPixelClock); | 515 | index = GetIndexIntoMasterTable(COMMAND, SetPixelClock); |
| 507 | atom_parse_cmd_header(rdev->mode_info.atom_context, index, &frev, | 516 | atom_parse_cmd_header(rdev->mode_info.atom_context, index, &frev, |
| @@ -574,21 +583,32 @@ int atombios_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 574 | struct radeon_device *rdev = dev->dev_private; | 583 | struct radeon_device *rdev = dev->dev_private; |
| 575 | struct radeon_framebuffer *radeon_fb; | 584 | struct radeon_framebuffer *radeon_fb; |
| 576 | struct drm_gem_object *obj; | 585 | struct drm_gem_object *obj; |
| 577 | struct drm_radeon_gem_object *obj_priv; | 586 | struct radeon_bo *rbo; |
| 578 | uint64_t fb_location; | 587 | uint64_t fb_location; |
| 579 | uint32_t fb_format, fb_pitch_pixels, tiling_flags; | 588 | uint32_t fb_format, fb_pitch_pixels, tiling_flags; |
| 589 | int r; | ||
| 580 | 590 | ||
| 581 | if (!crtc->fb) | 591 | /* no fb bound */ |
| 582 | return -EINVAL; | 592 | if (!crtc->fb) { |
| 593 | DRM_DEBUG("No FB bound\n"); | ||
| 594 | return 0; | ||
| 595 | } | ||
| 583 | 596 | ||
| 584 | radeon_fb = to_radeon_framebuffer(crtc->fb); | 597 | radeon_fb = to_radeon_framebuffer(crtc->fb); |
| 585 | 598 | ||
| 599 | /* Pin framebuffer & get tilling informations */ | ||
| 586 | obj = radeon_fb->obj; | 600 | obj = radeon_fb->obj; |
| 587 | obj_priv = obj->driver_private; | 601 | rbo = obj->driver_private; |
| 588 | 602 | r = radeon_bo_reserve(rbo, false); | |
| 589 | if (radeon_gem_object_pin(obj, RADEON_GEM_DOMAIN_VRAM, &fb_location)) { | 603 | if (unlikely(r != 0)) |
| 604 | return r; | ||
| 605 | r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_location); | ||
| 606 | if (unlikely(r != 0)) { | ||
| 607 | radeon_bo_unreserve(rbo); | ||
| 590 | return -EINVAL; | 608 | return -EINVAL; |
| 591 | } | 609 | } |
| 610 | radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); | ||
| 611 | radeon_bo_unreserve(rbo); | ||
| 592 | 612 | ||
| 593 | switch (crtc->fb->bits_per_pixel) { | 613 | switch (crtc->fb->bits_per_pixel) { |
| 594 | case 8: | 614 | case 8: |
| @@ -618,8 +638,6 @@ int atombios_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 618 | return -EINVAL; | 638 | return -EINVAL; |
| 619 | } | 639 | } |
| 620 | 640 | ||
| 621 | radeon_object_get_tiling_flags(obj->driver_private, | ||
| 622 | &tiling_flags, NULL); | ||
| 623 | if (tiling_flags & RADEON_TILING_MACRO) | 641 | if (tiling_flags & RADEON_TILING_MACRO) |
| 624 | fb_format |= AVIVO_D1GRPH_MACRO_ADDRESS_MODE; | 642 | fb_format |= AVIVO_D1GRPH_MACRO_ADDRESS_MODE; |
| 625 | 643 | ||
| @@ -674,7 +692,12 @@ int atombios_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 674 | 692 | ||
| 675 | if (old_fb && old_fb != crtc->fb) { | 693 | if (old_fb && old_fb != crtc->fb) { |
| 676 | radeon_fb = to_radeon_framebuffer(old_fb); | 694 | radeon_fb = to_radeon_framebuffer(old_fb); |
| 677 | radeon_gem_object_unpin(radeon_fb->obj); | 695 | rbo = radeon_fb->obj->driver_private; |
| 696 | r = radeon_bo_reserve(rbo, false); | ||
| 697 | if (unlikely(r != 0)) | ||
| 698 | return r; | ||
| 699 | radeon_bo_unpin(rbo); | ||
| 700 | radeon_bo_unreserve(rbo); | ||
| 678 | } | 701 | } |
| 679 | 702 | ||
| 680 | /* Bytes per pixel may have changed */ | 703 | /* Bytes per pixel may have changed */ |
diff --git a/drivers/gpu/drm/radeon/atombios_dp.c b/drivers/gpu/drm/radeon/atombios_dp.c new file mode 100644 index 000000000000..0d63c4436e7c --- /dev/null +++ b/drivers/gpu/drm/radeon/atombios_dp.c | |||
| @@ -0,0 +1,790 @@ | |||
| 1 | /* | ||
| 2 | * Copyright 2007-8 Advanced Micro Devices, Inc. | ||
| 3 | * Copyright 2008 Red Hat Inc. | ||
| 4 | * | ||
| 5 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 6 | * copy of this software and associated documentation files (the "Software"), | ||
| 7 | * to deal in the Software without restriction, including without limitation | ||
| 8 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | ||
| 9 | * and/or sell copies of the Software, and to permit persons to whom the | ||
| 10 | * Software is furnished to do so, subject to the following conditions: | ||
| 11 | * | ||
| 12 | * The above copyright notice and this permission notice shall be included in | ||
| 13 | * all copies or substantial portions of the Software. | ||
| 14 | * | ||
| 15 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 16 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 17 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
| 18 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR | ||
| 19 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | ||
| 20 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | ||
| 21 | * OTHER DEALINGS IN THE SOFTWARE. | ||
| 22 | * | ||
| 23 | * Authors: Dave Airlie | ||
| 24 | * Alex Deucher | ||
| 25 | */ | ||
| 26 | #include "drmP.h" | ||
| 27 | #include "radeon_drm.h" | ||
| 28 | #include "radeon.h" | ||
| 29 | |||
| 30 | #include "atom.h" | ||
| 31 | #include "atom-bits.h" | ||
| 32 | #include "drm_dp_helper.h" | ||
| 33 | |||
| 34 | /* move these to drm_dp_helper.c/h */ | ||
| 35 | #define DP_LINK_CONFIGURATION_SIZE 9 | ||
| 36 | #define DP_LINK_STATUS_SIZE 6 | ||
| 37 | #define DP_DPCD_SIZE 8 | ||
| 38 | |||
| 39 | static char *voltage_names[] = { | ||
| 40 | "0.4V", "0.6V", "0.8V", "1.2V" | ||
| 41 | }; | ||
| 42 | static char *pre_emph_names[] = { | ||
| 43 | "0dB", "3.5dB", "6dB", "9.5dB" | ||
| 44 | }; | ||
| 45 | |||
| 46 | static const int dp_clocks[] = { | ||
| 47 | 54000, /* 1 lane, 1.62 Ghz */ | ||
| 48 | 90000, /* 1 lane, 2.70 Ghz */ | ||
| 49 | 108000, /* 2 lane, 1.62 Ghz */ | ||
| 50 | 180000, /* 2 lane, 2.70 Ghz */ | ||
| 51 | 216000, /* 4 lane, 1.62 Ghz */ | ||
| 52 | 360000, /* 4 lane, 2.70 Ghz */ | ||
| 53 | }; | ||
| 54 | |||
| 55 | static const int num_dp_clocks = sizeof(dp_clocks) / sizeof(int); | ||
| 56 | |||
| 57 | /* common helper functions */ | ||
| 58 | static int dp_lanes_for_mode_clock(u8 dpcd[DP_DPCD_SIZE], int mode_clock) | ||
| 59 | { | ||
| 60 | int i; | ||
| 61 | u8 max_link_bw; | ||
| 62 | u8 max_lane_count; | ||
| 63 | |||
| 64 | if (!dpcd) | ||
| 65 | return 0; | ||
| 66 | |||
| 67 | max_link_bw = dpcd[DP_MAX_LINK_RATE]; | ||
| 68 | max_lane_count = dpcd[DP_MAX_LANE_COUNT] & DP_MAX_LANE_COUNT_MASK; | ||
| 69 | |||
| 70 | switch (max_link_bw) { | ||
| 71 | case DP_LINK_BW_1_62: | ||
| 72 | default: | ||
| 73 | for (i = 0; i < num_dp_clocks; i++) { | ||
| 74 | if (i % 2) | ||
| 75 | continue; | ||
| 76 | switch (max_lane_count) { | ||
| 77 | case 1: | ||
| 78 | if (i > 1) | ||
| 79 | return 0; | ||
| 80 | break; | ||
| 81 | case 2: | ||
| 82 | if (i > 3) | ||
| 83 | return 0; | ||
| 84 | break; | ||
| 85 | case 4: | ||
| 86 | default: | ||
| 87 | break; | ||
| 88 | } | ||
| 89 | if (dp_clocks[i] > mode_clock) { | ||
| 90 | if (i < 2) | ||
| 91 | return 1; | ||
| 92 | else if (i < 4) | ||
| 93 | return 2; | ||
| 94 | else | ||
| 95 | return 4; | ||
| 96 | } | ||
| 97 | } | ||
| 98 | break; | ||
| 99 | case DP_LINK_BW_2_7: | ||
| 100 | for (i = 0; i < num_dp_clocks; i++) { | ||
| 101 | switch (max_lane_count) { | ||
| 102 | case 1: | ||
| 103 | if (i > 1) | ||
| 104 | return 0; | ||
| 105 | break; | ||
| 106 | case 2: | ||
| 107 | if (i > 3) | ||
| 108 | return 0; | ||
| 109 | break; | ||
| 110 | case 4: | ||
| 111 | default: | ||
| 112 | break; | ||
| 113 | } | ||
| 114 | if (dp_clocks[i] > mode_clock) { | ||
| 115 | if (i < 2) | ||
| 116 | return 1; | ||
| 117 | else if (i < 4) | ||
| 118 | return 2; | ||
| 119 | else | ||
| 120 | return 4; | ||
| 121 | } | ||
| 122 | } | ||
| 123 | break; | ||
| 124 | } | ||
| 125 | |||
| 126 | return 0; | ||
| 127 | } | ||
| 128 | |||
| 129 | static int dp_link_clock_for_mode_clock(u8 dpcd[DP_DPCD_SIZE], int mode_clock) | ||
| 130 | { | ||
| 131 | int i; | ||
| 132 | u8 max_link_bw; | ||
| 133 | u8 max_lane_count; | ||
| 134 | |||
| 135 | if (!dpcd) | ||
| 136 | return 0; | ||
| 137 | |||
| 138 | max_link_bw = dpcd[DP_MAX_LINK_RATE]; | ||
| 139 | max_lane_count = dpcd[DP_MAX_LANE_COUNT] & DP_MAX_LANE_COUNT_MASK; | ||
| 140 | |||
| 141 | switch (max_link_bw) { | ||
| 142 | case DP_LINK_BW_1_62: | ||
| 143 | default: | ||
| 144 | for (i = 0; i < num_dp_clocks; i++) { | ||
| 145 | if (i % 2) | ||
| 146 | continue; | ||
| 147 | switch (max_lane_count) { | ||
| 148 | case 1: | ||
| 149 | if (i > 1) | ||
| 150 | return 0; | ||
| 151 | break; | ||
| 152 | case 2: | ||
| 153 | if (i > 3) | ||
| 154 | return 0; | ||
| 155 | break; | ||
| 156 | case 4: | ||
| 157 | default: | ||
| 158 | break; | ||
| 159 | } | ||
| 160 | if (dp_clocks[i] > mode_clock) | ||
| 161 | return 162000; | ||
| 162 | } | ||
| 163 | break; | ||
| 164 | case DP_LINK_BW_2_7: | ||
| 165 | for (i = 0; i < num_dp_clocks; i++) { | ||
| 166 | switch (max_lane_count) { | ||
| 167 | case 1: | ||
| 168 | if (i > 1) | ||
| 169 | return 0; | ||
| 170 | break; | ||
| 171 | case 2: | ||
| 172 | if (i > 3) | ||
| 173 | return 0; | ||
| 174 | break; | ||
| 175 | case 4: | ||
| 176 | default: | ||
| 177 | break; | ||
| 178 | } | ||
| 179 | if (dp_clocks[i] > mode_clock) | ||
| 180 | return (i % 2) ? 270000 : 162000; | ||
| 181 | } | ||
| 182 | } | ||
| 183 | |||
| 184 | return 0; | ||
| 185 | } | ||
| 186 | |||
| 187 | int dp_mode_valid(u8 dpcd[DP_DPCD_SIZE], int mode_clock) | ||
| 188 | { | ||
| 189 | int lanes = dp_lanes_for_mode_clock(dpcd, mode_clock); | ||
| 190 | int bw = dp_lanes_for_mode_clock(dpcd, mode_clock); | ||
| 191 | |||
| 192 | if ((lanes == 0) || (bw == 0)) | ||
| 193 | return MODE_CLOCK_HIGH; | ||
| 194 | |||
| 195 | return MODE_OK; | ||
| 196 | } | ||
| 197 | |||
| 198 | static u8 dp_link_status(u8 link_status[DP_LINK_STATUS_SIZE], int r) | ||
| 199 | { | ||
| 200 | return link_status[r - DP_LANE0_1_STATUS]; | ||
| 201 | } | ||
| 202 | |||
| 203 | static u8 dp_get_lane_status(u8 link_status[DP_LINK_STATUS_SIZE], | ||
| 204 | int lane) | ||
| 205 | { | ||
| 206 | int i = DP_LANE0_1_STATUS + (lane >> 1); | ||
| 207 | int s = (lane & 1) * 4; | ||
| 208 | u8 l = dp_link_status(link_status, i); | ||
| 209 | return (l >> s) & 0xf; | ||
| 210 | } | ||
| 211 | |||
| 212 | static bool dp_clock_recovery_ok(u8 link_status[DP_LINK_STATUS_SIZE], | ||
| 213 | int lane_count) | ||
| 214 | { | ||
| 215 | int lane; | ||
| 216 | u8 lane_status; | ||
| 217 | |||
| 218 | for (lane = 0; lane < lane_count; lane++) { | ||
| 219 | lane_status = dp_get_lane_status(link_status, lane); | ||
| 220 | if ((lane_status & DP_LANE_CR_DONE) == 0) | ||
| 221 | return false; | ||
| 222 | } | ||
| 223 | return true; | ||
| 224 | } | ||
| 225 | |||
| 226 | static bool dp_channel_eq_ok(u8 link_status[DP_LINK_STATUS_SIZE], | ||
| 227 | int lane_count) | ||
| 228 | { | ||
| 229 | u8 lane_align; | ||
| 230 | u8 lane_status; | ||
| 231 | int lane; | ||
| 232 | |||
| 233 | lane_align = dp_link_status(link_status, | ||
| 234 | DP_LANE_ALIGN_STATUS_UPDATED); | ||
| 235 | if ((lane_align & DP_INTERLANE_ALIGN_DONE) == 0) | ||
| 236 | return false; | ||
| 237 | for (lane = 0; lane < lane_count; lane++) { | ||
| 238 | lane_status = dp_get_lane_status(link_status, lane); | ||
| 239 | if ((lane_status & DP_CHANNEL_EQ_BITS) != DP_CHANNEL_EQ_BITS) | ||
| 240 | return false; | ||
| 241 | } | ||
| 242 | return true; | ||
| 243 | } | ||
| 244 | |||
| 245 | static u8 dp_get_adjust_request_voltage(uint8_t link_status[DP_LINK_STATUS_SIZE], | ||
| 246 | int lane) | ||
| 247 | |||
| 248 | { | ||
| 249 | int i = DP_ADJUST_REQUEST_LANE0_1 + (lane >> 1); | ||
| 250 | int s = ((lane & 1) ? | ||
| 251 | DP_ADJUST_VOLTAGE_SWING_LANE1_SHIFT : | ||
| 252 | DP_ADJUST_VOLTAGE_SWING_LANE0_SHIFT); | ||
| 253 | u8 l = dp_link_status(link_status, i); | ||
| 254 | |||
| 255 | return ((l >> s) & 0x3) << DP_TRAIN_VOLTAGE_SWING_SHIFT; | ||
| 256 | } | ||
| 257 | |||
| 258 | static u8 dp_get_adjust_request_pre_emphasis(uint8_t link_status[DP_LINK_STATUS_SIZE], | ||
| 259 | int lane) | ||
| 260 | { | ||
| 261 | int i = DP_ADJUST_REQUEST_LANE0_1 + (lane >> 1); | ||
| 262 | int s = ((lane & 1) ? | ||
| 263 | DP_ADJUST_PRE_EMPHASIS_LANE1_SHIFT : | ||
| 264 | DP_ADJUST_PRE_EMPHASIS_LANE0_SHIFT); | ||
| 265 | u8 l = dp_link_status(link_status, i); | ||
| 266 | |||
| 267 | return ((l >> s) & 0x3) << DP_TRAIN_PRE_EMPHASIS_SHIFT; | ||
| 268 | } | ||
| 269 | |||
| 270 | /* XXX fix me -- chip specific */ | ||
| 271 | #define DP_VOLTAGE_MAX DP_TRAIN_VOLTAGE_SWING_1200 | ||
| 272 | static u8 dp_pre_emphasis_max(u8 voltage_swing) | ||
| 273 | { | ||
| 274 | switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) { | ||
| 275 | case DP_TRAIN_VOLTAGE_SWING_400: | ||
| 276 | return DP_TRAIN_PRE_EMPHASIS_6; | ||
| 277 | case DP_TRAIN_VOLTAGE_SWING_600: | ||
| 278 | return DP_TRAIN_PRE_EMPHASIS_6; | ||
| 279 | case DP_TRAIN_VOLTAGE_SWING_800: | ||
| 280 | return DP_TRAIN_PRE_EMPHASIS_3_5; | ||
| 281 | case DP_TRAIN_VOLTAGE_SWING_1200: | ||
| 282 | default: | ||
| 283 | return DP_TRAIN_PRE_EMPHASIS_0; | ||
| 284 | } | ||
| 285 | } | ||
| 286 | |||
| 287 | static void dp_get_adjust_train(u8 link_status[DP_LINK_STATUS_SIZE], | ||
| 288 | int lane_count, | ||
| 289 | u8 train_set[4]) | ||
| 290 | { | ||
| 291 | u8 v = 0; | ||
| 292 | u8 p = 0; | ||
| 293 | int lane; | ||
| 294 | |||
| 295 | for (lane = 0; lane < lane_count; lane++) { | ||
| 296 | u8 this_v = dp_get_adjust_request_voltage(link_status, lane); | ||
| 297 | u8 this_p = dp_get_adjust_request_pre_emphasis(link_status, lane); | ||
| 298 | |||
| 299 | DRM_DEBUG("requested signal parameters: lane %d voltage %s pre_emph %s\n", | ||
| 300 | lane, | ||
| 301 | voltage_names[this_v >> DP_TRAIN_VOLTAGE_SWING_SHIFT], | ||
| 302 | pre_emph_names[this_p >> DP_TRAIN_PRE_EMPHASIS_SHIFT]); | ||
| 303 | |||
| 304 | if (this_v > v) | ||
| 305 | v = this_v; | ||
| 306 | if (this_p > p) | ||
| 307 | p = this_p; | ||
| 308 | } | ||
| 309 | |||
| 310 | if (v >= DP_VOLTAGE_MAX) | ||
| 311 | v = DP_VOLTAGE_MAX | DP_TRAIN_MAX_SWING_REACHED; | ||
| 312 | |||
| 313 | if (p >= dp_pre_emphasis_max(v)) | ||
| 314 | p = dp_pre_emphasis_max(v) | DP_TRAIN_MAX_PRE_EMPHASIS_REACHED; | ||
| 315 | |||
| 316 | DRM_DEBUG("using signal parameters: voltage %s pre_emph %s\n", | ||
| 317 | voltage_names[(v & DP_TRAIN_VOLTAGE_SWING_MASK) >> DP_TRAIN_VOLTAGE_SWING_SHIFT], | ||
| 318 | pre_emph_names[(p & DP_TRAIN_PRE_EMPHASIS_MASK) >> DP_TRAIN_PRE_EMPHASIS_SHIFT]); | ||
| 319 | |||
| 320 | for (lane = 0; lane < 4; lane++) | ||
| 321 | train_set[lane] = v | p; | ||
| 322 | } | ||
| 323 | |||
| 324 | |||
| 325 | /* radeon aux chan functions */ | ||
| 326 | bool radeon_process_aux_ch(struct radeon_i2c_chan *chan, u8 *req_bytes, | ||
| 327 | int num_bytes, u8 *read_byte, | ||
| 328 | u8 read_buf_len, u8 delay) | ||
| 329 | { | ||
| 330 | struct drm_device *dev = chan->dev; | ||
| 331 | struct radeon_device *rdev = dev->dev_private; | ||
| 332 | PROCESS_AUX_CHANNEL_TRANSACTION_PS_ALLOCATION args; | ||
| 333 | int index = GetIndexIntoMasterTable(COMMAND, ProcessAuxChannelTransaction); | ||
| 334 | unsigned char *base; | ||
| 335 | |||
| 336 | memset(&args, 0, sizeof(args)); | ||
| 337 | |||
| 338 | base = (unsigned char *)rdev->mode_info.atom_context->scratch; | ||
| 339 | |||
| 340 | memcpy(base, req_bytes, num_bytes); | ||
| 341 | |||
| 342 | args.lpAuxRequest = 0; | ||
| 343 | args.lpDataOut = 16; | ||
| 344 | args.ucDataOutLen = 0; | ||
| 345 | args.ucChannelID = chan->rec.i2c_id; | ||
| 346 | args.ucDelay = delay / 10; | ||
| 347 | |||
| 348 | atom_execute_table(rdev->mode_info.atom_context, index, (uint32_t *)&args); | ||
| 349 | |||
| 350 | if (args.ucReplyStatus) { | ||
| 351 | DRM_DEBUG("failed to get auxch %02x%02x %02x %02x 0x%02x %02x\n", | ||
| 352 | req_bytes[1], req_bytes[0], req_bytes[2], req_bytes[3], | ||
| 353 | chan->rec.i2c_id, args.ucReplyStatus); | ||
| 354 | return false; | ||
| 355 | } | ||
| 356 | |||
| 357 | if (args.ucDataOutLen && read_byte && read_buf_len) { | ||
| 358 | if (read_buf_len < args.ucDataOutLen) { | ||
| 359 | DRM_ERROR("Buffer to small for return answer %d %d\n", | ||
| 360 | read_buf_len, args.ucDataOutLen); | ||
| 361 | return false; | ||
| 362 | } | ||
| 363 | { | ||
| 364 | int len = min(read_buf_len, args.ucDataOutLen); | ||
| 365 | memcpy(read_byte, base + 16, len); | ||
| 366 | } | ||
| 367 | } | ||
| 368 | return true; | ||
| 369 | } | ||
| 370 | |||
| 371 | bool radeon_dp_aux_native_write(struct radeon_connector *radeon_connector, uint16_t address, | ||
| 372 | uint8_t send_bytes, uint8_t *send) | ||
| 373 | { | ||
| 374 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 375 | u8 msg[20]; | ||
| 376 | u8 msg_len, dp_msg_len; | ||
| 377 | bool ret; | ||
| 378 | |||
| 379 | dp_msg_len = 4; | ||
| 380 | msg[0] = address; | ||
| 381 | msg[1] = address >> 8; | ||
| 382 | msg[2] = AUX_NATIVE_WRITE << 4; | ||
| 383 | dp_msg_len += send_bytes; | ||
| 384 | msg[3] = (dp_msg_len << 4) | (send_bytes - 1); | ||
| 385 | |||
| 386 | if (send_bytes > 16) | ||
| 387 | return false; | ||
| 388 | |||
| 389 | memcpy(&msg[4], send, send_bytes); | ||
| 390 | msg_len = 4 + send_bytes; | ||
| 391 | ret = radeon_process_aux_ch(dig_connector->dp_i2c_bus, msg, msg_len, NULL, 0, 0); | ||
| 392 | return ret; | ||
| 393 | } | ||
| 394 | |||
| 395 | bool radeon_dp_aux_native_read(struct radeon_connector *radeon_connector, uint16_t address, | ||
| 396 | uint8_t delay, uint8_t expected_bytes, | ||
| 397 | uint8_t *read_p) | ||
| 398 | { | ||
| 399 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 400 | u8 msg[20]; | ||
| 401 | u8 msg_len, dp_msg_len; | ||
| 402 | bool ret = false; | ||
| 403 | msg_len = 4; | ||
| 404 | dp_msg_len = 4; | ||
| 405 | msg[0] = address; | ||
| 406 | msg[1] = address >> 8; | ||
| 407 | msg[2] = AUX_NATIVE_READ << 4; | ||
| 408 | msg[3] = (dp_msg_len) << 4; | ||
| 409 | msg[3] |= expected_bytes - 1; | ||
| 410 | |||
| 411 | ret = radeon_process_aux_ch(dig_connector->dp_i2c_bus, msg, msg_len, read_p, expected_bytes, delay); | ||
| 412 | return ret; | ||
| 413 | } | ||
| 414 | |||
| 415 | /* radeon dp functions */ | ||
| 416 | static u8 radeon_dp_encoder_service(struct radeon_device *rdev, int action, int dp_clock, | ||
| 417 | uint8_t ucconfig, uint8_t lane_num) | ||
| 418 | { | ||
| 419 | DP_ENCODER_SERVICE_PARAMETERS args; | ||
| 420 | int index = GetIndexIntoMasterTable(COMMAND, DPEncoderService); | ||
| 421 | |||
| 422 | memset(&args, 0, sizeof(args)); | ||
| 423 | args.ucLinkClock = dp_clock / 10; | ||
| 424 | args.ucConfig = ucconfig; | ||
| 425 | args.ucAction = action; | ||
| 426 | args.ucLaneNum = lane_num; | ||
| 427 | args.ucStatus = 0; | ||
| 428 | |||
| 429 | atom_execute_table(rdev->mode_info.atom_context, index, (uint32_t *)&args); | ||
| 430 | return args.ucStatus; | ||
| 431 | } | ||
| 432 | |||
| 433 | u8 radeon_dp_getsinktype(struct radeon_connector *radeon_connector) | ||
| 434 | { | ||
| 435 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 436 | struct drm_device *dev = radeon_connector->base.dev; | ||
| 437 | struct radeon_device *rdev = dev->dev_private; | ||
| 438 | |||
| 439 | return radeon_dp_encoder_service(rdev, ATOM_DP_ACTION_GET_SINK_TYPE, 0, | ||
| 440 | dig_connector->dp_i2c_bus->rec.i2c_id, 0); | ||
| 441 | } | ||
| 442 | |||
| 443 | bool radeon_dp_getdpcd(struct radeon_connector *radeon_connector) | ||
| 444 | { | ||
| 445 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 446 | u8 msg[25]; | ||
| 447 | int ret; | ||
| 448 | |||
| 449 | ret = radeon_dp_aux_native_read(radeon_connector, DP_DPCD_REV, 0, 8, msg); | ||
| 450 | if (ret) { | ||
| 451 | memcpy(dig_connector->dpcd, msg, 8); | ||
| 452 | { | ||
| 453 | int i; | ||
| 454 | DRM_DEBUG("DPCD: "); | ||
| 455 | for (i = 0; i < 8; i++) | ||
| 456 | DRM_DEBUG("%02x ", msg[i]); | ||
| 457 | DRM_DEBUG("\n"); | ||
| 458 | } | ||
| 459 | return true; | ||
| 460 | } | ||
| 461 | dig_connector->dpcd[0] = 0; | ||
| 462 | return false; | ||
| 463 | } | ||
| 464 | |||
| 465 | void radeon_dp_set_link_config(struct drm_connector *connector, | ||
| 466 | struct drm_display_mode *mode) | ||
| 467 | { | ||
| 468 | struct radeon_connector *radeon_connector; | ||
| 469 | struct radeon_connector_atom_dig *dig_connector; | ||
| 470 | |||
| 471 | if (connector->connector_type != DRM_MODE_CONNECTOR_DisplayPort) | ||
| 472 | return; | ||
| 473 | |||
| 474 | radeon_connector = to_radeon_connector(connector); | ||
| 475 | if (!radeon_connector->con_priv) | ||
| 476 | return; | ||
| 477 | dig_connector = radeon_connector->con_priv; | ||
| 478 | |||
| 479 | dig_connector->dp_clock = | ||
| 480 | dp_link_clock_for_mode_clock(dig_connector->dpcd, mode->clock); | ||
| 481 | dig_connector->dp_lane_count = | ||
| 482 | dp_lanes_for_mode_clock(dig_connector->dpcd, mode->clock); | ||
| 483 | } | ||
| 484 | |||
| 485 | int radeon_dp_mode_valid_helper(struct radeon_connector *radeon_connector, | ||
| 486 | struct drm_display_mode *mode) | ||
| 487 | { | ||
| 488 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 489 | |||
| 490 | return dp_mode_valid(dig_connector->dpcd, mode->clock); | ||
| 491 | } | ||
| 492 | |||
| 493 | static bool atom_dp_get_link_status(struct radeon_connector *radeon_connector, | ||
| 494 | u8 link_status[DP_LINK_STATUS_SIZE]) | ||
| 495 | { | ||
| 496 | int ret; | ||
| 497 | ret = radeon_dp_aux_native_read(radeon_connector, DP_LANE0_1_STATUS, 100, | ||
| 498 | DP_LINK_STATUS_SIZE, link_status); | ||
| 499 | if (!ret) { | ||
| 500 | DRM_ERROR("displayport link status failed\n"); | ||
| 501 | return false; | ||
| 502 | } | ||
| 503 | |||
| 504 | DRM_DEBUG("link status %02x %02x %02x %02x %02x %02x\n", | ||
| 505 | link_status[0], link_status[1], link_status[2], | ||
| 506 | link_status[3], link_status[4], link_status[5]); | ||
| 507 | return true; | ||
| 508 | } | ||
| 509 | |||
| 510 | bool radeon_dp_needs_link_train(struct radeon_connector *radeon_connector) | ||
| 511 | { | ||
| 512 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 513 | u8 link_status[DP_LINK_STATUS_SIZE]; | ||
| 514 | |||
| 515 | if (!atom_dp_get_link_status(radeon_connector, link_status)) | ||
| 516 | return false; | ||
| 517 | if (dp_channel_eq_ok(link_status, dig_connector->dp_lane_count)) | ||
| 518 | return false; | ||
| 519 | return true; | ||
| 520 | } | ||
| 521 | |||
| 522 | static void dp_set_power(struct radeon_connector *radeon_connector, u8 power_state) | ||
| 523 | { | ||
| 524 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 525 | |||
| 526 | if (dig_connector->dpcd[0] >= 0x11) { | ||
| 527 | radeon_dp_aux_native_write(radeon_connector, DP_SET_POWER, 1, | ||
| 528 | &power_state); | ||
| 529 | } | ||
| 530 | } | ||
| 531 | |||
| 532 | static void dp_set_downspread(struct radeon_connector *radeon_connector, u8 downspread) | ||
| 533 | { | ||
| 534 | radeon_dp_aux_native_write(radeon_connector, DP_DOWNSPREAD_CTRL, 1, | ||
| 535 | &downspread); | ||
| 536 | } | ||
| 537 | |||
| 538 | static void dp_set_link_bw_lanes(struct radeon_connector *radeon_connector, | ||
| 539 | u8 link_configuration[DP_LINK_CONFIGURATION_SIZE]) | ||
| 540 | { | ||
| 541 | radeon_dp_aux_native_write(radeon_connector, DP_LINK_BW_SET, 2, | ||
| 542 | link_configuration); | ||
| 543 | } | ||
| 544 | |||
| 545 | static void dp_update_dpvs_emph(struct radeon_connector *radeon_connector, | ||
| 546 | struct drm_encoder *encoder, | ||
| 547 | u8 train_set[4]) | ||
| 548 | { | ||
| 549 | struct radeon_connector_atom_dig *dig_connector = radeon_connector->con_priv; | ||
| 550 | int i; | ||
| 551 | |||
| 552 | for (i = 0; i < dig_connector->dp_lane_count; i++) | ||
| 553 | atombios_dig_transmitter_setup(encoder, | ||
| 554 | ATOM_TRANSMITTER_ACTION_SETUP_VSEMPH, | ||
| 555 | i, train_set[i]); | ||
| 556 | |||
| 557 | radeon_dp_aux_native_write(radeon_connector, DP_TRAINING_LANE0_SET, | ||
| 558 | dig_connector->dp_lane_count, train_set); | ||
| 559 | } | ||
| 560 | |||
| 561 | static void dp_set_training(struct radeon_connector *radeon_connector, | ||
| 562 | u8 training) | ||
| 563 | { | ||
| 564 | radeon_dp_aux_native_write(radeon_connector, DP_TRAINING_PATTERN_SET, | ||
| 565 | 1, &training); | ||
| 566 | } | ||
| 567 | |||
| 568 | void dp_link_train(struct drm_encoder *encoder, | ||
| 569 | struct drm_connector *connector) | ||
| 570 | { | ||
| 571 | struct drm_device *dev = encoder->dev; | ||
| 572 | struct radeon_device *rdev = dev->dev_private; | ||
| 573 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 574 | struct radeon_encoder_atom_dig *dig; | ||
| 575 | struct radeon_connector *radeon_connector; | ||
| 576 | struct radeon_connector_atom_dig *dig_connector; | ||
| 577 | int enc_id = 0; | ||
| 578 | bool clock_recovery, channel_eq; | ||
| 579 | u8 link_status[DP_LINK_STATUS_SIZE]; | ||
| 580 | u8 link_configuration[DP_LINK_CONFIGURATION_SIZE]; | ||
| 581 | u8 tries, voltage; | ||
| 582 | u8 train_set[4]; | ||
| 583 | int i; | ||
| 584 | |||
| 585 | if (connector->connector_type != DRM_MODE_CONNECTOR_DisplayPort) | ||
| 586 | return; | ||
| 587 | |||
| 588 | if (!radeon_encoder->enc_priv) | ||
| 589 | return; | ||
| 590 | dig = radeon_encoder->enc_priv; | ||
| 591 | |||
| 592 | radeon_connector = to_radeon_connector(connector); | ||
| 593 | if (!radeon_connector->con_priv) | ||
| 594 | return; | ||
| 595 | dig_connector = radeon_connector->con_priv; | ||
| 596 | |||
| 597 | if (ASIC_IS_DCE32(rdev)) { | ||
| 598 | if (dig->dig_block) | ||
| 599 | enc_id |= ATOM_DP_CONFIG_DIG2_ENCODER; | ||
| 600 | else | ||
| 601 | enc_id |= ATOM_DP_CONFIG_DIG1_ENCODER; | ||
| 602 | if (dig_connector->linkb) | ||
| 603 | enc_id |= ATOM_DP_CONFIG_LINK_B; | ||
| 604 | else | ||
| 605 | enc_id |= ATOM_DP_CONFIG_LINK_A; | ||
| 606 | } else { | ||
| 607 | if (dig_connector->linkb) | ||
| 608 | enc_id |= ATOM_DP_CONFIG_DIG2_ENCODER | ATOM_DP_CONFIG_LINK_B; | ||
| 609 | else | ||
| 610 | enc_id |= ATOM_DP_CONFIG_DIG1_ENCODER | ATOM_DP_CONFIG_LINK_A; | ||
| 611 | } | ||
| 612 | |||
| 613 | memset(link_configuration, 0, DP_LINK_CONFIGURATION_SIZE); | ||
| 614 | if (dig_connector->dp_clock == 270000) | ||
| 615 | link_configuration[0] = DP_LINK_BW_2_7; | ||
| 616 | else | ||
| 617 | link_configuration[0] = DP_LINK_BW_1_62; | ||
| 618 | link_configuration[1] = dig_connector->dp_lane_count; | ||
| 619 | if (dig_connector->dpcd[0] >= 0x11) | ||
| 620 | link_configuration[1] |= DP_LANE_COUNT_ENHANCED_FRAME_EN; | ||
| 621 | |||
| 622 | /* power up the sink */ | ||
| 623 | dp_set_power(radeon_connector, DP_SET_POWER_D0); | ||
| 624 | /* disable the training pattern on the sink */ | ||
| 625 | dp_set_training(radeon_connector, DP_TRAINING_PATTERN_DISABLE); | ||
| 626 | /* set link bw and lanes on the sink */ | ||
| 627 | dp_set_link_bw_lanes(radeon_connector, link_configuration); | ||
| 628 | /* disable downspread on the sink */ | ||
| 629 | dp_set_downspread(radeon_connector, 0); | ||
| 630 | /* start training on the source */ | ||
| 631 | radeon_dp_encoder_service(rdev, ATOM_DP_ACTION_TRAINING_START, | ||
| 632 | dig_connector->dp_clock, enc_id, 0); | ||
| 633 | /* set training pattern 1 on the source */ | ||
| 634 | radeon_dp_encoder_service(rdev, ATOM_DP_ACTION_TRAINING_PATTERN_SEL, | ||
| 635 | dig_connector->dp_clock, enc_id, 0); | ||
| 636 | |||
| 637 | /* set initial vs/emph */ | ||
| 638 | memset(train_set, 0, 4); | ||
| 639 | udelay(400); | ||
| 640 | /* set training pattern 1 on the sink */ | ||
| 641 | dp_set_training(radeon_connector, DP_TRAINING_PATTERN_1); | ||
| 642 | |||
| 643 | dp_update_dpvs_emph(radeon_connector, encoder, train_set); | ||
| 644 | |||
| 645 | /* clock recovery loop */ | ||
| 646 | clock_recovery = false; | ||
| 647 | tries = 0; | ||
| 648 | voltage = 0xff; | ||
| 649 | for (;;) { | ||
| 650 | udelay(100); | ||
| 651 | if (!atom_dp_get_link_status(radeon_connector, link_status)) | ||
| 652 | break; | ||
| 653 | |||
| 654 | if (dp_clock_recovery_ok(link_status, dig_connector->dp_lane_count)) { | ||
| 655 | clock_recovery = true; | ||
| 656 | break; | ||
| 657 | } | ||
| 658 | |||
| 659 | for (i = 0; i < dig_connector->dp_lane_count; i++) { | ||
| 660 | if ((train_set[i] & DP_TRAIN_MAX_SWING_REACHED) == 0) | ||
| 661 | break; | ||
| 662 | } | ||
| 663 | if (i == dig_connector->dp_lane_count) { | ||
| 664 | DRM_ERROR("clock recovery reached max voltage\n"); | ||
| 665 | break; | ||
| 666 | } | ||
| 667 | |||
| 668 | if ((train_set[0] & DP_TRAIN_VOLTAGE_SWING_MASK) == voltage) { | ||
| 669 | ++tries; | ||
| 670 | if (tries == 5) { | ||
| 671 | DRM_ERROR("clock recovery tried 5 times\n"); | ||
| 672 | break; | ||
| 673 | } | ||
| 674 | } else | ||
| 675 | tries = 0; | ||
| 676 | |||
| 677 | voltage = train_set[0] & DP_TRAIN_VOLTAGE_SWING_MASK; | ||
| 678 | |||
| 679 | /* Compute new train_set as requested by sink */ | ||
| 680 | dp_get_adjust_train(link_status, dig_connector->dp_lane_count, train_set); | ||
| 681 | dp_update_dpvs_emph(radeon_connector, encoder, train_set); | ||
| 682 | } | ||
| 683 | if (!clock_recovery) | ||
| 684 | DRM_ERROR("clock recovery failed\n"); | ||
| 685 | else | ||
| 686 | DRM_DEBUG("clock recovery at voltage %d pre-emphasis %d\n", | ||
| 687 | train_set[0] & DP_TRAIN_VOLTAGE_SWING_MASK, | ||
| 688 | (train_set[0] & DP_TRAIN_PRE_EMPHASIS_MASK) >> | ||
| 689 | DP_TRAIN_PRE_EMPHASIS_SHIFT); | ||
| 690 | |||
| 691 | |||
| 692 | /* set training pattern 2 on the sink */ | ||
| 693 | dp_set_training(radeon_connector, DP_TRAINING_PATTERN_2); | ||
| 694 | /* set training pattern 2 on the source */ | ||
| 695 | radeon_dp_encoder_service(rdev, ATOM_DP_ACTION_TRAINING_PATTERN_SEL, | ||
| 696 | dig_connector->dp_clock, enc_id, 1); | ||
| 697 | |||
| 698 | /* channel equalization loop */ | ||
| 699 | tries = 0; | ||
| 700 | channel_eq = false; | ||
| 701 | for (;;) { | ||
| 702 | udelay(400); | ||
| 703 | if (!atom_dp_get_link_status(radeon_connector, link_status)) | ||
| 704 | break; | ||
| 705 | |||
| 706 | if (dp_channel_eq_ok(link_status, dig_connector->dp_lane_count)) { | ||
| 707 | channel_eq = true; | ||
| 708 | break; | ||
| 709 | } | ||
| 710 | |||
| 711 | /* Try 5 times */ | ||
| 712 | if (tries > 5) { | ||
| 713 | DRM_ERROR("channel eq failed: 5 tries\n"); | ||
| 714 | break; | ||
| 715 | } | ||
| 716 | |||
| 717 | /* Compute new train_set as requested by sink */ | ||
| 718 | dp_get_adjust_train(link_status, dig_connector->dp_lane_count, train_set); | ||
| 719 | dp_update_dpvs_emph(radeon_connector, encoder, train_set); | ||
| 720 | |||
| 721 | tries++; | ||
| 722 | } | ||
| 723 | |||
| 724 | if (!channel_eq) | ||
| 725 | DRM_ERROR("channel eq failed\n"); | ||
| 726 | else | ||
| 727 | DRM_DEBUG("channel eq at voltage %d pre-emphasis %d\n", | ||
| 728 | train_set[0] & DP_TRAIN_VOLTAGE_SWING_MASK, | ||
| 729 | (train_set[0] & DP_TRAIN_PRE_EMPHASIS_MASK) | ||
| 730 | >> DP_TRAIN_PRE_EMPHASIS_SHIFT); | ||
| 731 | |||
| 732 | /* disable the training pattern on the sink */ | ||
| 733 | dp_set_training(radeon_connector, DP_TRAINING_PATTERN_DISABLE); | ||
| 734 | |||
| 735 | radeon_dp_encoder_service(rdev, ATOM_DP_ACTION_TRAINING_COMPLETE, | ||
| 736 | dig_connector->dp_clock, enc_id, 0); | ||
| 737 | } | ||
| 738 | |||
| 739 | int radeon_dp_i2c_aux_ch(struct i2c_adapter *adapter, int mode, | ||
| 740 | uint8_t write_byte, uint8_t *read_byte) | ||
| 741 | { | ||
| 742 | struct i2c_algo_dp_aux_data *algo_data = adapter->algo_data; | ||
| 743 | struct radeon_i2c_chan *auxch = (struct radeon_i2c_chan *)adapter; | ||
| 744 | int ret = 0; | ||
| 745 | uint16_t address = algo_data->address; | ||
| 746 | uint8_t msg[5]; | ||
| 747 | uint8_t reply[2]; | ||
| 748 | int msg_len, dp_msg_len; | ||
| 749 | int reply_bytes; | ||
| 750 | |||
| 751 | /* Set up the command byte */ | ||
| 752 | if (mode & MODE_I2C_READ) | ||
| 753 | msg[2] = AUX_I2C_READ << 4; | ||
| 754 | else | ||
| 755 | msg[2] = AUX_I2C_WRITE << 4; | ||
| 756 | |||
| 757 | if (!(mode & MODE_I2C_STOP)) | ||
| 758 | msg[2] |= AUX_I2C_MOT << 4; | ||
| 759 | |||
| 760 | msg[0] = address; | ||
| 761 | msg[1] = address >> 8; | ||
| 762 | |||
| 763 | reply_bytes = 1; | ||
| 764 | |||
| 765 | msg_len = 4; | ||
| 766 | dp_msg_len = 3; | ||
| 767 | switch (mode) { | ||
| 768 | case MODE_I2C_WRITE: | ||
| 769 | msg[4] = write_byte; | ||
| 770 | msg_len++; | ||
| 771 | dp_msg_len += 2; | ||
| 772 | break; | ||
| 773 | case MODE_I2C_READ: | ||
| 774 | dp_msg_len += 1; | ||
| 775 | break; | ||
| 776 | default: | ||
| 777 | break; | ||
| 778 | } | ||
| 779 | |||
| 780 | msg[3] = (dp_msg_len) << 4; | ||
| 781 | ret = radeon_process_aux_ch(auxch, msg, msg_len, reply, reply_bytes, 0); | ||
| 782 | |||
| 783 | if (ret) { | ||
| 784 | if (read_byte) | ||
| 785 | *read_byte = reply[0]; | ||
| 786 | return reply_bytes; | ||
| 787 | } | ||
| 788 | return -EREMOTEIO; | ||
| 789 | } | ||
| 790 | |||
diff --git a/drivers/gpu/drm/radeon/r100.c b/drivers/gpu/drm/radeon/r100.c index c9e93eabcf16..824cc6480a06 100644 --- a/drivers/gpu/drm/radeon/r100.c +++ b/drivers/gpu/drm/radeon/r100.c | |||
| @@ -65,6 +65,95 @@ MODULE_FIRMWARE(FIRMWARE_R520); | |||
| 65 | * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280 | 65 | * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280 |
| 66 | */ | 66 | */ |
| 67 | 67 | ||
| 68 | /* hpd for digital panel detect/disconnect */ | ||
| 69 | bool r100_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd) | ||
| 70 | { | ||
| 71 | bool connected = false; | ||
| 72 | |||
| 73 | switch (hpd) { | ||
| 74 | case RADEON_HPD_1: | ||
| 75 | if (RREG32(RADEON_FP_GEN_CNTL) & RADEON_FP_DETECT_SENSE) | ||
| 76 | connected = true; | ||
| 77 | break; | ||
| 78 | case RADEON_HPD_2: | ||
| 79 | if (RREG32(RADEON_FP2_GEN_CNTL) & RADEON_FP2_DETECT_SENSE) | ||
| 80 | connected = true; | ||
| 81 | break; | ||
| 82 | default: | ||
| 83 | break; | ||
| 84 | } | ||
| 85 | return connected; | ||
| 86 | } | ||
| 87 | |||
| 88 | void r100_hpd_set_polarity(struct radeon_device *rdev, | ||
| 89 | enum radeon_hpd_id hpd) | ||
| 90 | { | ||
| 91 | u32 tmp; | ||
| 92 | bool connected = r100_hpd_sense(rdev, hpd); | ||
| 93 | |||
| 94 | switch (hpd) { | ||
| 95 | case RADEON_HPD_1: | ||
| 96 | tmp = RREG32(RADEON_FP_GEN_CNTL); | ||
| 97 | if (connected) | ||
| 98 | tmp &= ~RADEON_FP_DETECT_INT_POL; | ||
| 99 | else | ||
| 100 | tmp |= RADEON_FP_DETECT_INT_POL; | ||
| 101 | WREG32(RADEON_FP_GEN_CNTL, tmp); | ||
| 102 | break; | ||
| 103 | case RADEON_HPD_2: | ||
| 104 | tmp = RREG32(RADEON_FP2_GEN_CNTL); | ||
| 105 | if (connected) | ||
| 106 | tmp &= ~RADEON_FP2_DETECT_INT_POL; | ||
| 107 | else | ||
| 108 | tmp |= RADEON_FP2_DETECT_INT_POL; | ||
| 109 | WREG32(RADEON_FP2_GEN_CNTL, tmp); | ||
| 110 | break; | ||
| 111 | default: | ||
| 112 | break; | ||
| 113 | } | ||
| 114 | } | ||
| 115 | |||
| 116 | void r100_hpd_init(struct radeon_device *rdev) | ||
| 117 | { | ||
| 118 | struct drm_device *dev = rdev->ddev; | ||
| 119 | struct drm_connector *connector; | ||
| 120 | |||
| 121 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 122 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 123 | switch (radeon_connector->hpd.hpd) { | ||
| 124 | case RADEON_HPD_1: | ||
| 125 | rdev->irq.hpd[0] = true; | ||
| 126 | break; | ||
| 127 | case RADEON_HPD_2: | ||
| 128 | rdev->irq.hpd[1] = true; | ||
| 129 | break; | ||
| 130 | default: | ||
| 131 | break; | ||
| 132 | } | ||
| 133 | } | ||
| 134 | r100_irq_set(rdev); | ||
| 135 | } | ||
| 136 | |||
| 137 | void r100_hpd_fini(struct radeon_device *rdev) | ||
| 138 | { | ||
| 139 | struct drm_device *dev = rdev->ddev; | ||
| 140 | struct drm_connector *connector; | ||
| 141 | |||
| 142 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 143 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 144 | switch (radeon_connector->hpd.hpd) { | ||
| 145 | case RADEON_HPD_1: | ||
| 146 | rdev->irq.hpd[0] = false; | ||
| 147 | break; | ||
| 148 | case RADEON_HPD_2: | ||
| 149 | rdev->irq.hpd[1] = false; | ||
| 150 | break; | ||
| 151 | default: | ||
| 152 | break; | ||
| 153 | } | ||
| 154 | } | ||
| 155 | } | ||
| 156 | |||
| 68 | /* | 157 | /* |
| 69 | * PCI GART | 158 | * PCI GART |
| 70 | */ | 159 | */ |
| @@ -94,6 +183,15 @@ int r100_pci_gart_init(struct radeon_device *rdev) | |||
| 94 | return radeon_gart_table_ram_alloc(rdev); | 183 | return radeon_gart_table_ram_alloc(rdev); |
| 95 | } | 184 | } |
| 96 | 185 | ||
| 186 | /* required on r1xx, r2xx, r300, r(v)350, r420/r481, rs400/rs480 */ | ||
| 187 | void r100_enable_bm(struct radeon_device *rdev) | ||
| 188 | { | ||
| 189 | uint32_t tmp; | ||
| 190 | /* Enable bus mastering */ | ||
| 191 | tmp = RREG32(RADEON_BUS_CNTL) & ~RADEON_BUS_MASTER_DIS; | ||
| 192 | WREG32(RADEON_BUS_CNTL, tmp); | ||
| 193 | } | ||
| 194 | |||
| 97 | int r100_pci_gart_enable(struct radeon_device *rdev) | 195 | int r100_pci_gart_enable(struct radeon_device *rdev) |
| 98 | { | 196 | { |
| 99 | uint32_t tmp; | 197 | uint32_t tmp; |
| @@ -105,9 +203,6 @@ int r100_pci_gart_enable(struct radeon_device *rdev) | |||
| 105 | WREG32(RADEON_AIC_LO_ADDR, rdev->mc.gtt_location); | 203 | WREG32(RADEON_AIC_LO_ADDR, rdev->mc.gtt_location); |
| 106 | tmp = rdev->mc.gtt_location + rdev->mc.gtt_size - 1; | 204 | tmp = rdev->mc.gtt_location + rdev->mc.gtt_size - 1; |
| 107 | WREG32(RADEON_AIC_HI_ADDR, tmp); | 205 | WREG32(RADEON_AIC_HI_ADDR, tmp); |
| 108 | /* Enable bus mastering */ | ||
| 109 | tmp = RREG32(RADEON_BUS_CNTL) & ~RADEON_BUS_MASTER_DIS; | ||
| 110 | WREG32(RADEON_BUS_CNTL, tmp); | ||
| 111 | /* set PCI GART page-table base address */ | 206 | /* set PCI GART page-table base address */ |
| 112 | WREG32(RADEON_AIC_PT_BASE, rdev->gart.table_addr); | 207 | WREG32(RADEON_AIC_PT_BASE, rdev->gart.table_addr); |
| 113 | tmp = RREG32(RADEON_AIC_CNTL) | RADEON_PCIGART_TRANSLATE_EN; | 208 | tmp = RREG32(RADEON_AIC_CNTL) | RADEON_PCIGART_TRANSLATE_EN; |
| @@ -157,6 +252,12 @@ int r100_irq_set(struct radeon_device *rdev) | |||
| 157 | if (rdev->irq.crtc_vblank_int[1]) { | 252 | if (rdev->irq.crtc_vblank_int[1]) { |
| 158 | tmp |= RADEON_CRTC2_VBLANK_MASK; | 253 | tmp |= RADEON_CRTC2_VBLANK_MASK; |
| 159 | } | 254 | } |
| 255 | if (rdev->irq.hpd[0]) { | ||
| 256 | tmp |= RADEON_FP_DETECT_MASK; | ||
| 257 | } | ||
| 258 | if (rdev->irq.hpd[1]) { | ||
| 259 | tmp |= RADEON_FP2_DETECT_MASK; | ||
| 260 | } | ||
| 160 | WREG32(RADEON_GEN_INT_CNTL, tmp); | 261 | WREG32(RADEON_GEN_INT_CNTL, tmp); |
| 161 | return 0; | 262 | return 0; |
| 162 | } | 263 | } |
| @@ -175,8 +276,9 @@ void r100_irq_disable(struct radeon_device *rdev) | |||
| 175 | static inline uint32_t r100_irq_ack(struct radeon_device *rdev) | 276 | static inline uint32_t r100_irq_ack(struct radeon_device *rdev) |
| 176 | { | 277 | { |
| 177 | uint32_t irqs = RREG32(RADEON_GEN_INT_STATUS); | 278 | uint32_t irqs = RREG32(RADEON_GEN_INT_STATUS); |
| 178 | uint32_t irq_mask = RADEON_SW_INT_TEST | RADEON_CRTC_VBLANK_STAT | | 279 | uint32_t irq_mask = RADEON_SW_INT_TEST | |
| 179 | RADEON_CRTC2_VBLANK_STAT; | 280 | RADEON_CRTC_VBLANK_STAT | RADEON_CRTC2_VBLANK_STAT | |
| 281 | RADEON_FP_DETECT_STAT | RADEON_FP2_DETECT_STAT; | ||
| 180 | 282 | ||
| 181 | if (irqs) { | 283 | if (irqs) { |
| 182 | WREG32(RADEON_GEN_INT_STATUS, irqs); | 284 | WREG32(RADEON_GEN_INT_STATUS, irqs); |
| @@ -187,6 +289,7 @@ static inline uint32_t r100_irq_ack(struct radeon_device *rdev) | |||
| 187 | int r100_irq_process(struct radeon_device *rdev) | 289 | int r100_irq_process(struct radeon_device *rdev) |
| 188 | { | 290 | { |
| 189 | uint32_t status, msi_rearm; | 291 | uint32_t status, msi_rearm; |
| 292 | bool queue_hotplug = false; | ||
| 190 | 293 | ||
| 191 | status = r100_irq_ack(rdev); | 294 | status = r100_irq_ack(rdev); |
| 192 | if (!status) { | 295 | if (!status) { |
| @@ -207,8 +310,18 @@ int r100_irq_process(struct radeon_device *rdev) | |||
| 207 | if (status & RADEON_CRTC2_VBLANK_STAT) { | 310 | if (status & RADEON_CRTC2_VBLANK_STAT) { |
| 208 | drm_handle_vblank(rdev->ddev, 1); | 311 | drm_handle_vblank(rdev->ddev, 1); |
| 209 | } | 312 | } |
| 313 | if (status & RADEON_FP_DETECT_STAT) { | ||
| 314 | queue_hotplug = true; | ||
| 315 | DRM_DEBUG("HPD1\n"); | ||
| 316 | } | ||
| 317 | if (status & RADEON_FP2_DETECT_STAT) { | ||
| 318 | queue_hotplug = true; | ||
| 319 | DRM_DEBUG("HPD2\n"); | ||
| 320 | } | ||
| 210 | status = r100_irq_ack(rdev); | 321 | status = r100_irq_ack(rdev); |
| 211 | } | 322 | } |
| 323 | if (queue_hotplug) | ||
| 324 | queue_work(rdev->wq, &rdev->hotplug_work); | ||
| 212 | if (rdev->msi_enabled) { | 325 | if (rdev->msi_enabled) { |
| 213 | switch (rdev->family) { | 326 | switch (rdev->family) { |
| 214 | case CHIP_RS400: | 327 | case CHIP_RS400: |
| @@ -255,24 +368,27 @@ int r100_wb_init(struct radeon_device *rdev) | |||
| 255 | int r; | 368 | int r; |
| 256 | 369 | ||
| 257 | if (rdev->wb.wb_obj == NULL) { | 370 | if (rdev->wb.wb_obj == NULL) { |
| 258 | r = radeon_object_create(rdev, NULL, RADEON_GPU_PAGE_SIZE, | 371 | r = radeon_bo_create(rdev, NULL, RADEON_GPU_PAGE_SIZE, true, |
| 259 | true, | 372 | RADEON_GEM_DOMAIN_GTT, |
| 260 | RADEON_GEM_DOMAIN_GTT, | 373 | &rdev->wb.wb_obj); |
| 261 | false, &rdev->wb.wb_obj); | ||
| 262 | if (r) { | 374 | if (r) { |
| 263 | DRM_ERROR("radeon: failed to create WB buffer (%d).\n", r); | 375 | dev_err(rdev->dev, "(%d) create WB buffer failed\n", r); |
| 264 | return r; | 376 | return r; |
| 265 | } | 377 | } |
| 266 | r = radeon_object_pin(rdev->wb.wb_obj, | 378 | r = radeon_bo_reserve(rdev->wb.wb_obj, false); |
| 267 | RADEON_GEM_DOMAIN_GTT, | 379 | if (unlikely(r != 0)) |
| 268 | &rdev->wb.gpu_addr); | 380 | return r; |
| 381 | r = radeon_bo_pin(rdev->wb.wb_obj, RADEON_GEM_DOMAIN_GTT, | ||
| 382 | &rdev->wb.gpu_addr); | ||
| 269 | if (r) { | 383 | if (r) { |
| 270 | DRM_ERROR("radeon: failed to pin WB buffer (%d).\n", r); | 384 | dev_err(rdev->dev, "(%d) pin WB buffer failed\n", r); |
| 385 | radeon_bo_unreserve(rdev->wb.wb_obj); | ||
| 271 | return r; | 386 | return r; |
| 272 | } | 387 | } |
| 273 | r = radeon_object_kmap(rdev->wb.wb_obj, (void **)&rdev->wb.wb); | 388 | r = radeon_bo_kmap(rdev->wb.wb_obj, (void **)&rdev->wb.wb); |
| 389 | radeon_bo_unreserve(rdev->wb.wb_obj); | ||
| 274 | if (r) { | 390 | if (r) { |
| 275 | DRM_ERROR("radeon: failed to map WB buffer (%d).\n", r); | 391 | dev_err(rdev->dev, "(%d) map WB buffer failed\n", r); |
| 276 | return r; | 392 | return r; |
| 277 | } | 393 | } |
| 278 | } | 394 | } |
| @@ -290,11 +406,19 @@ void r100_wb_disable(struct radeon_device *rdev) | |||
| 290 | 406 | ||
| 291 | void r100_wb_fini(struct radeon_device *rdev) | 407 | void r100_wb_fini(struct radeon_device *rdev) |
| 292 | { | 408 | { |
| 409 | int r; | ||
| 410 | |||
| 293 | r100_wb_disable(rdev); | 411 | r100_wb_disable(rdev); |
| 294 | if (rdev->wb.wb_obj) { | 412 | if (rdev->wb.wb_obj) { |
| 295 | radeon_object_kunmap(rdev->wb.wb_obj); | 413 | r = radeon_bo_reserve(rdev->wb.wb_obj, false); |
| 296 | radeon_object_unpin(rdev->wb.wb_obj); | 414 | if (unlikely(r != 0)) { |
| 297 | radeon_object_unref(&rdev->wb.wb_obj); | 415 | dev_err(rdev->dev, "(%d) can't finish WB\n", r); |
| 416 | return; | ||
| 417 | } | ||
| 418 | radeon_bo_kunmap(rdev->wb.wb_obj); | ||
| 419 | radeon_bo_unpin(rdev->wb.wb_obj); | ||
| 420 | radeon_bo_unreserve(rdev->wb.wb_obj); | ||
| 421 | radeon_bo_unref(&rdev->wb.wb_obj); | ||
| 298 | rdev->wb.wb = NULL; | 422 | rdev->wb.wb = NULL; |
| 299 | rdev->wb.wb_obj = NULL; | 423 | rdev->wb.wb_obj = NULL; |
| 300 | } | 424 | } |
| @@ -1288,17 +1412,17 @@ static int r100_packet0_check(struct radeon_cs_parser *p, | |||
| 1288 | 1412 | ||
| 1289 | int r100_cs_track_check_pkt3_indx_buffer(struct radeon_cs_parser *p, | 1413 | int r100_cs_track_check_pkt3_indx_buffer(struct radeon_cs_parser *p, |
| 1290 | struct radeon_cs_packet *pkt, | 1414 | struct radeon_cs_packet *pkt, |
| 1291 | struct radeon_object *robj) | 1415 | struct radeon_bo *robj) |
| 1292 | { | 1416 | { |
| 1293 | unsigned idx; | 1417 | unsigned idx; |
| 1294 | u32 value; | 1418 | u32 value; |
| 1295 | idx = pkt->idx + 1; | 1419 | idx = pkt->idx + 1; |
| 1296 | value = radeon_get_ib_value(p, idx + 2); | 1420 | value = radeon_get_ib_value(p, idx + 2); |
| 1297 | if ((value + 1) > radeon_object_size(robj)) { | 1421 | if ((value + 1) > radeon_bo_size(robj)) { |
| 1298 | DRM_ERROR("[drm] Buffer too small for PACKET3 INDX_BUFFER " | 1422 | DRM_ERROR("[drm] Buffer too small for PACKET3 INDX_BUFFER " |
| 1299 | "(need %u have %lu) !\n", | 1423 | "(need %u have %lu) !\n", |
| 1300 | value + 1, | 1424 | value + 1, |
| 1301 | radeon_object_size(robj)); | 1425 | radeon_bo_size(robj)); |
| 1302 | return -EINVAL; | 1426 | return -EINVAL; |
| 1303 | } | 1427 | } |
| 1304 | return 0; | 1428 | return 0; |
| @@ -1583,6 +1707,14 @@ void r100_gpu_init(struct radeon_device *rdev) | |||
| 1583 | r100_hdp_reset(rdev); | 1707 | r100_hdp_reset(rdev); |
| 1584 | } | 1708 | } |
| 1585 | 1709 | ||
| 1710 | void r100_hdp_flush(struct radeon_device *rdev) | ||
| 1711 | { | ||
| 1712 | u32 tmp; | ||
| 1713 | tmp = RREG32(RADEON_HOST_PATH_CNTL); | ||
| 1714 | tmp |= RADEON_HDP_READ_BUFFER_INVALIDATE; | ||
| 1715 | WREG32(RADEON_HOST_PATH_CNTL, tmp); | ||
| 1716 | } | ||
| 1717 | |||
| 1586 | void r100_hdp_reset(struct radeon_device *rdev) | 1718 | void r100_hdp_reset(struct radeon_device *rdev) |
| 1587 | { | 1719 | { |
| 1588 | uint32_t tmp; | 1720 | uint32_t tmp; |
| @@ -1650,6 +1782,17 @@ int r100_gpu_reset(struct radeon_device *rdev) | |||
| 1650 | return 0; | 1782 | return 0; |
| 1651 | } | 1783 | } |
| 1652 | 1784 | ||
| 1785 | void r100_set_common_regs(struct radeon_device *rdev) | ||
| 1786 | { | ||
| 1787 | /* set these so they don't interfere with anything */ | ||
| 1788 | WREG32(RADEON_OV0_SCALE_CNTL, 0); | ||
| 1789 | WREG32(RADEON_SUBPIC_CNTL, 0); | ||
| 1790 | WREG32(RADEON_VIPH_CONTROL, 0); | ||
| 1791 | WREG32(RADEON_I2C_CNTL_1, 0); | ||
| 1792 | WREG32(RADEON_DVI_I2C_CNTL_1, 0); | ||
| 1793 | WREG32(RADEON_CAP0_TRIG_CNTL, 0); | ||
| 1794 | WREG32(RADEON_CAP1_TRIG_CNTL, 0); | ||
| 1795 | } | ||
| 1653 | 1796 | ||
| 1654 | /* | 1797 | /* |
| 1655 | * VRAM info | 1798 | * VRAM info |
| @@ -2594,7 +2737,7 @@ static int r100_cs_track_cube(struct radeon_device *rdev, | |||
| 2594 | struct r100_cs_track *track, unsigned idx) | 2737 | struct r100_cs_track *track, unsigned idx) |
| 2595 | { | 2738 | { |
| 2596 | unsigned face, w, h; | 2739 | unsigned face, w, h; |
| 2597 | struct radeon_object *cube_robj; | 2740 | struct radeon_bo *cube_robj; |
| 2598 | unsigned long size; | 2741 | unsigned long size; |
| 2599 | 2742 | ||
| 2600 | for (face = 0; face < 5; face++) { | 2743 | for (face = 0; face < 5; face++) { |
| @@ -2607,9 +2750,9 @@ static int r100_cs_track_cube(struct radeon_device *rdev, | |||
| 2607 | 2750 | ||
| 2608 | size += track->textures[idx].cube_info[face].offset; | 2751 | size += track->textures[idx].cube_info[face].offset; |
| 2609 | 2752 | ||
| 2610 | if (size > radeon_object_size(cube_robj)) { | 2753 | if (size > radeon_bo_size(cube_robj)) { |
| 2611 | DRM_ERROR("Cube texture offset greater than object size %lu %lu\n", | 2754 | DRM_ERROR("Cube texture offset greater than object size %lu %lu\n", |
| 2612 | size, radeon_object_size(cube_robj)); | 2755 | size, radeon_bo_size(cube_robj)); |
| 2613 | r100_cs_track_texture_print(&track->textures[idx]); | 2756 | r100_cs_track_texture_print(&track->textures[idx]); |
| 2614 | return -1; | 2757 | return -1; |
| 2615 | } | 2758 | } |
| @@ -2620,7 +2763,7 @@ static int r100_cs_track_cube(struct radeon_device *rdev, | |||
| 2620 | static int r100_cs_track_texture_check(struct radeon_device *rdev, | 2763 | static int r100_cs_track_texture_check(struct radeon_device *rdev, |
| 2621 | struct r100_cs_track *track) | 2764 | struct r100_cs_track *track) |
| 2622 | { | 2765 | { |
| 2623 | struct radeon_object *robj; | 2766 | struct radeon_bo *robj; |
| 2624 | unsigned long size; | 2767 | unsigned long size; |
| 2625 | unsigned u, i, w, h; | 2768 | unsigned u, i, w, h; |
| 2626 | int ret; | 2769 | int ret; |
| @@ -2676,9 +2819,9 @@ static int r100_cs_track_texture_check(struct radeon_device *rdev, | |||
| 2676 | "%u\n", track->textures[u].tex_coord_type, u); | 2819 | "%u\n", track->textures[u].tex_coord_type, u); |
| 2677 | return -EINVAL; | 2820 | return -EINVAL; |
| 2678 | } | 2821 | } |
| 2679 | if (size > radeon_object_size(robj)) { | 2822 | if (size > radeon_bo_size(robj)) { |
| 2680 | DRM_ERROR("Texture of unit %u needs %lu bytes but is " | 2823 | DRM_ERROR("Texture of unit %u needs %lu bytes but is " |
| 2681 | "%lu\n", u, size, radeon_object_size(robj)); | 2824 | "%lu\n", u, size, radeon_bo_size(robj)); |
| 2682 | r100_cs_track_texture_print(&track->textures[u]); | 2825 | r100_cs_track_texture_print(&track->textures[u]); |
| 2683 | return -EINVAL; | 2826 | return -EINVAL; |
| 2684 | } | 2827 | } |
| @@ -2700,10 +2843,10 @@ int r100_cs_track_check(struct radeon_device *rdev, struct r100_cs_track *track) | |||
| 2700 | } | 2843 | } |
| 2701 | size = track->cb[i].pitch * track->cb[i].cpp * track->maxy; | 2844 | size = track->cb[i].pitch * track->cb[i].cpp * track->maxy; |
| 2702 | size += track->cb[i].offset; | 2845 | size += track->cb[i].offset; |
| 2703 | if (size > radeon_object_size(track->cb[i].robj)) { | 2846 | if (size > radeon_bo_size(track->cb[i].robj)) { |
| 2704 | DRM_ERROR("[drm] Buffer too small for color buffer %d " | 2847 | DRM_ERROR("[drm] Buffer too small for color buffer %d " |
| 2705 | "(need %lu have %lu) !\n", i, size, | 2848 | "(need %lu have %lu) !\n", i, size, |
| 2706 | radeon_object_size(track->cb[i].robj)); | 2849 | radeon_bo_size(track->cb[i].robj)); |
| 2707 | DRM_ERROR("[drm] color buffer %d (%u %u %u %u)\n", | 2850 | DRM_ERROR("[drm] color buffer %d (%u %u %u %u)\n", |
| 2708 | i, track->cb[i].pitch, track->cb[i].cpp, | 2851 | i, track->cb[i].pitch, track->cb[i].cpp, |
| 2709 | track->cb[i].offset, track->maxy); | 2852 | track->cb[i].offset, track->maxy); |
| @@ -2717,10 +2860,10 @@ int r100_cs_track_check(struct radeon_device *rdev, struct r100_cs_track *track) | |||
| 2717 | } | 2860 | } |
| 2718 | size = track->zb.pitch * track->zb.cpp * track->maxy; | 2861 | size = track->zb.pitch * track->zb.cpp * track->maxy; |
| 2719 | size += track->zb.offset; | 2862 | size += track->zb.offset; |
| 2720 | if (size > radeon_object_size(track->zb.robj)) { | 2863 | if (size > radeon_bo_size(track->zb.robj)) { |
| 2721 | DRM_ERROR("[drm] Buffer too small for z buffer " | 2864 | DRM_ERROR("[drm] Buffer too small for z buffer " |
| 2722 | "(need %lu have %lu) !\n", size, | 2865 | "(need %lu have %lu) !\n", size, |
| 2723 | radeon_object_size(track->zb.robj)); | 2866 | radeon_bo_size(track->zb.robj)); |
| 2724 | DRM_ERROR("[drm] zbuffer (%u %u %u %u)\n", | 2867 | DRM_ERROR("[drm] zbuffer (%u %u %u %u)\n", |
| 2725 | track->zb.pitch, track->zb.cpp, | 2868 | track->zb.pitch, track->zb.cpp, |
| 2726 | track->zb.offset, track->maxy); | 2869 | track->zb.offset, track->maxy); |
| @@ -2738,11 +2881,12 @@ int r100_cs_track_check(struct radeon_device *rdev, struct r100_cs_track *track) | |||
| 2738 | "bound\n", prim_walk, i); | 2881 | "bound\n", prim_walk, i); |
| 2739 | return -EINVAL; | 2882 | return -EINVAL; |
| 2740 | } | 2883 | } |
| 2741 | if (size > radeon_object_size(track->arrays[i].robj)) { | 2884 | if (size > radeon_bo_size(track->arrays[i].robj)) { |
| 2742 | DRM_ERROR("(PW %u) Vertex array %u need %lu dwords " | 2885 | dev_err(rdev->dev, "(PW %u) Vertex array %u " |
| 2743 | "have %lu dwords\n", prim_walk, i, | 2886 | "need %lu dwords have %lu dwords\n", |
| 2744 | size >> 2, | 2887 | prim_walk, i, size >> 2, |
| 2745 | radeon_object_size(track->arrays[i].robj) >> 2); | 2888 | radeon_bo_size(track->arrays[i].robj) |
| 2889 | >> 2); | ||
| 2746 | DRM_ERROR("Max indices %u\n", track->max_indx); | 2890 | DRM_ERROR("Max indices %u\n", track->max_indx); |
| 2747 | return -EINVAL; | 2891 | return -EINVAL; |
| 2748 | } | 2892 | } |
| @@ -2756,10 +2900,12 @@ int r100_cs_track_check(struct radeon_device *rdev, struct r100_cs_track *track) | |||
| 2756 | "bound\n", prim_walk, i); | 2900 | "bound\n", prim_walk, i); |
| 2757 | return -EINVAL; | 2901 | return -EINVAL; |
| 2758 | } | 2902 | } |
| 2759 | if (size > radeon_object_size(track->arrays[i].robj)) { | 2903 | if (size > radeon_bo_size(track->arrays[i].robj)) { |
| 2760 | DRM_ERROR("(PW %u) Vertex array %u need %lu dwords " | 2904 | dev_err(rdev->dev, "(PW %u) Vertex array %u " |
| 2761 | "have %lu dwords\n", prim_walk, i, size >> 2, | 2905 | "need %lu dwords have %lu dwords\n", |
| 2762 | radeon_object_size(track->arrays[i].robj) >> 2); | 2906 | prim_walk, i, size >> 2, |
| 2907 | radeon_bo_size(track->arrays[i].robj) | ||
| 2908 | >> 2); | ||
| 2763 | return -EINVAL; | 2909 | return -EINVAL; |
| 2764 | } | 2910 | } |
| 2765 | } | 2911 | } |
| @@ -3101,6 +3247,9 @@ static int r100_startup(struct radeon_device *rdev) | |||
| 3101 | { | 3247 | { |
| 3102 | int r; | 3248 | int r; |
| 3103 | 3249 | ||
| 3250 | /* set common regs */ | ||
| 3251 | r100_set_common_regs(rdev); | ||
| 3252 | /* program mc */ | ||
| 3104 | r100_mc_program(rdev); | 3253 | r100_mc_program(rdev); |
| 3105 | /* Resume clock */ | 3254 | /* Resume clock */ |
| 3106 | r100_clock_startup(rdev); | 3255 | r100_clock_startup(rdev); |
| @@ -3108,13 +3257,13 @@ static int r100_startup(struct radeon_device *rdev) | |||
| 3108 | r100_gpu_init(rdev); | 3257 | r100_gpu_init(rdev); |
| 3109 | /* Initialize GART (initialize after TTM so we can allocate | 3258 | /* Initialize GART (initialize after TTM so we can allocate |
| 3110 | * memory through TTM but finalize after TTM) */ | 3259 | * memory through TTM but finalize after TTM) */ |
| 3260 | r100_enable_bm(rdev); | ||
| 3111 | if (rdev->flags & RADEON_IS_PCI) { | 3261 | if (rdev->flags & RADEON_IS_PCI) { |
| 3112 | r = r100_pci_gart_enable(rdev); | 3262 | r = r100_pci_gart_enable(rdev); |
| 3113 | if (r) | 3263 | if (r) |
| 3114 | return r; | 3264 | return r; |
| 3115 | } | 3265 | } |
| 3116 | /* Enable IRQ */ | 3266 | /* Enable IRQ */ |
| 3117 | rdev->irq.sw_int = true; | ||
| 3118 | r100_irq_set(rdev); | 3267 | r100_irq_set(rdev); |
| 3119 | /* 1M ring buffer */ | 3268 | /* 1M ring buffer */ |
| 3120 | r = r100_cp_init(rdev, 1024 * 1024); | 3269 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -3150,6 +3299,8 @@ int r100_resume(struct radeon_device *rdev) | |||
| 3150 | radeon_combios_asic_init(rdev->ddev); | 3299 | radeon_combios_asic_init(rdev->ddev); |
| 3151 | /* Resume clock after posting */ | 3300 | /* Resume clock after posting */ |
| 3152 | r100_clock_startup(rdev); | 3301 | r100_clock_startup(rdev); |
| 3302 | /* Initialize surface registers */ | ||
| 3303 | radeon_surface_init(rdev); | ||
| 3153 | return r100_startup(rdev); | 3304 | return r100_startup(rdev); |
| 3154 | } | 3305 | } |
| 3155 | 3306 | ||
| @@ -3174,7 +3325,7 @@ void r100_fini(struct radeon_device *rdev) | |||
| 3174 | r100_pci_gart_fini(rdev); | 3325 | r100_pci_gart_fini(rdev); |
| 3175 | radeon_irq_kms_fini(rdev); | 3326 | radeon_irq_kms_fini(rdev); |
| 3176 | radeon_fence_driver_fini(rdev); | 3327 | radeon_fence_driver_fini(rdev); |
| 3177 | radeon_object_fini(rdev); | 3328 | radeon_bo_fini(rdev); |
| 3178 | radeon_atombios_fini(rdev); | 3329 | radeon_atombios_fini(rdev); |
| 3179 | kfree(rdev->bios); | 3330 | kfree(rdev->bios); |
| 3180 | rdev->bios = NULL; | 3331 | rdev->bios = NULL; |
| @@ -3242,10 +3393,8 @@ int r100_init(struct radeon_device *rdev) | |||
| 3242 | RREG32(R_0007C0_CP_STAT)); | 3393 | RREG32(R_0007C0_CP_STAT)); |
| 3243 | } | 3394 | } |
| 3244 | /* check if cards are posted or not */ | 3395 | /* check if cards are posted or not */ |
| 3245 | if (!radeon_card_posted(rdev) && rdev->bios) { | 3396 | if (radeon_boot_test_post_card(rdev) == false) |
| 3246 | DRM_INFO("GPU not posted. posting now...\n"); | 3397 | return -EINVAL; |
| 3247 | radeon_combios_asic_init(rdev->ddev); | ||
| 3248 | } | ||
| 3249 | /* Set asic errata */ | 3398 | /* Set asic errata */ |
| 3250 | r100_errata(rdev); | 3399 | r100_errata(rdev); |
| 3251 | /* Initialize clocks */ | 3400 | /* Initialize clocks */ |
| @@ -3264,7 +3413,7 @@ int r100_init(struct radeon_device *rdev) | |||
| 3264 | if (r) | 3413 | if (r) |
| 3265 | return r; | 3414 | return r; |
| 3266 | /* Memory manager */ | 3415 | /* Memory manager */ |
| 3267 | r = radeon_object_init(rdev); | 3416 | r = radeon_bo_init(rdev); |
| 3268 | if (r) | 3417 | if (r) |
| 3269 | return r; | 3418 | return r; |
| 3270 | if (rdev->flags & RADEON_IS_PCI) { | 3419 | if (rdev->flags & RADEON_IS_PCI) { |
diff --git a/drivers/gpu/drm/radeon/r100_track.h b/drivers/gpu/drm/radeon/r100_track.h index 0daf0d76a891..ca50903dd2bb 100644 --- a/drivers/gpu/drm/radeon/r100_track.h +++ b/drivers/gpu/drm/radeon/r100_track.h | |||
| @@ -10,26 +10,26 @@ | |||
| 10 | * CS functions | 10 | * CS functions |
| 11 | */ | 11 | */ |
| 12 | struct r100_cs_track_cb { | 12 | struct r100_cs_track_cb { |
| 13 | struct radeon_object *robj; | 13 | struct radeon_bo *robj; |
| 14 | unsigned pitch; | 14 | unsigned pitch; |
| 15 | unsigned cpp; | 15 | unsigned cpp; |
| 16 | unsigned offset; | 16 | unsigned offset; |
| 17 | }; | 17 | }; |
| 18 | 18 | ||
| 19 | struct r100_cs_track_array { | 19 | struct r100_cs_track_array { |
| 20 | struct radeon_object *robj; | 20 | struct radeon_bo *robj; |
| 21 | unsigned esize; | 21 | unsigned esize; |
| 22 | }; | 22 | }; |
| 23 | 23 | ||
| 24 | struct r100_cs_cube_info { | 24 | struct r100_cs_cube_info { |
| 25 | struct radeon_object *robj; | 25 | struct radeon_bo *robj; |
| 26 | unsigned offset; | 26 | unsigned offset; |
| 27 | unsigned width; | 27 | unsigned width; |
| 28 | unsigned height; | 28 | unsigned height; |
| 29 | }; | 29 | }; |
| 30 | 30 | ||
| 31 | struct r100_cs_track_texture { | 31 | struct r100_cs_track_texture { |
| 32 | struct radeon_object *robj; | 32 | struct radeon_bo *robj; |
| 33 | struct r100_cs_cube_info cube_info[5]; /* info for 5 non-primary faces */ | 33 | struct r100_cs_cube_info cube_info[5]; /* info for 5 non-primary faces */ |
| 34 | unsigned pitch; | 34 | unsigned pitch; |
| 35 | unsigned width; | 35 | unsigned width; |
diff --git a/drivers/gpu/drm/radeon/r300.c b/drivers/gpu/drm/radeon/r300.c index 2f43ee8e4048..83378c39d0e3 100644 --- a/drivers/gpu/drm/radeon/r300.c +++ b/drivers/gpu/drm/radeon/r300.c | |||
| @@ -137,14 +137,19 @@ int rv370_pcie_gart_enable(struct radeon_device *rdev) | |||
| 137 | 137 | ||
| 138 | void rv370_pcie_gart_disable(struct radeon_device *rdev) | 138 | void rv370_pcie_gart_disable(struct radeon_device *rdev) |
| 139 | { | 139 | { |
| 140 | uint32_t tmp; | 140 | u32 tmp; |
| 141 | int r; | ||
| 141 | 142 | ||
| 142 | tmp = RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL); | 143 | tmp = RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL); |
| 143 | tmp |= RADEON_PCIE_TX_GART_UNMAPPED_ACCESS_DISCARD; | 144 | tmp |= RADEON_PCIE_TX_GART_UNMAPPED_ACCESS_DISCARD; |
| 144 | WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp & ~RADEON_PCIE_TX_GART_EN); | 145 | WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp & ~RADEON_PCIE_TX_GART_EN); |
| 145 | if (rdev->gart.table.vram.robj) { | 146 | if (rdev->gart.table.vram.robj) { |
| 146 | radeon_object_kunmap(rdev->gart.table.vram.robj); | 147 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 147 | radeon_object_unpin(rdev->gart.table.vram.robj); | 148 | if (likely(r == 0)) { |
| 149 | radeon_bo_kunmap(rdev->gart.table.vram.robj); | ||
| 150 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 151 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 152 | } | ||
| 148 | } | 153 | } |
| 149 | } | 154 | } |
| 150 | 155 | ||
| @@ -1181,6 +1186,9 @@ static int r300_startup(struct radeon_device *rdev) | |||
| 1181 | { | 1186 | { |
| 1182 | int r; | 1187 | int r; |
| 1183 | 1188 | ||
| 1189 | /* set common regs */ | ||
| 1190 | r100_set_common_regs(rdev); | ||
| 1191 | /* program mc */ | ||
| 1184 | r300_mc_program(rdev); | 1192 | r300_mc_program(rdev); |
| 1185 | /* Resume clock */ | 1193 | /* Resume clock */ |
| 1186 | r300_clock_startup(rdev); | 1194 | r300_clock_startup(rdev); |
| @@ -1193,13 +1201,18 @@ static int r300_startup(struct radeon_device *rdev) | |||
| 1193 | if (r) | 1201 | if (r) |
| 1194 | return r; | 1202 | return r; |
| 1195 | } | 1203 | } |
| 1204 | |||
| 1205 | if (rdev->family == CHIP_R300 || | ||
| 1206 | rdev->family == CHIP_R350 || | ||
| 1207 | rdev->family == CHIP_RV350) | ||
| 1208 | r100_enable_bm(rdev); | ||
| 1209 | |||
| 1196 | if (rdev->flags & RADEON_IS_PCI) { | 1210 | if (rdev->flags & RADEON_IS_PCI) { |
| 1197 | r = r100_pci_gart_enable(rdev); | 1211 | r = r100_pci_gart_enable(rdev); |
| 1198 | if (r) | 1212 | if (r) |
| 1199 | return r; | 1213 | return r; |
| 1200 | } | 1214 | } |
| 1201 | /* Enable IRQ */ | 1215 | /* Enable IRQ */ |
| 1202 | rdev->irq.sw_int = true; | ||
| 1203 | r100_irq_set(rdev); | 1216 | r100_irq_set(rdev); |
| 1204 | /* 1M ring buffer */ | 1217 | /* 1M ring buffer */ |
| 1205 | r = r100_cp_init(rdev, 1024 * 1024); | 1218 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -1237,6 +1250,8 @@ int r300_resume(struct radeon_device *rdev) | |||
| 1237 | radeon_combios_asic_init(rdev->ddev); | 1250 | radeon_combios_asic_init(rdev->ddev); |
| 1238 | /* Resume clock after posting */ | 1251 | /* Resume clock after posting */ |
| 1239 | r300_clock_startup(rdev); | 1252 | r300_clock_startup(rdev); |
| 1253 | /* Initialize surface registers */ | ||
| 1254 | radeon_surface_init(rdev); | ||
| 1240 | return r300_startup(rdev); | 1255 | return r300_startup(rdev); |
| 1241 | } | 1256 | } |
| 1242 | 1257 | ||
| @@ -1265,7 +1280,7 @@ void r300_fini(struct radeon_device *rdev) | |||
| 1265 | r100_pci_gart_fini(rdev); | 1280 | r100_pci_gart_fini(rdev); |
| 1266 | radeon_irq_kms_fini(rdev); | 1281 | radeon_irq_kms_fini(rdev); |
| 1267 | radeon_fence_driver_fini(rdev); | 1282 | radeon_fence_driver_fini(rdev); |
| 1268 | radeon_object_fini(rdev); | 1283 | radeon_bo_fini(rdev); |
| 1269 | radeon_atombios_fini(rdev); | 1284 | radeon_atombios_fini(rdev); |
| 1270 | kfree(rdev->bios); | 1285 | kfree(rdev->bios); |
| 1271 | rdev->bios = NULL; | 1286 | rdev->bios = NULL; |
| @@ -1303,10 +1318,8 @@ int r300_init(struct radeon_device *rdev) | |||
| 1303 | RREG32(R_0007C0_CP_STAT)); | 1318 | RREG32(R_0007C0_CP_STAT)); |
| 1304 | } | 1319 | } |
| 1305 | /* check if cards are posted or not */ | 1320 | /* check if cards are posted or not */ |
| 1306 | if (!radeon_card_posted(rdev) && rdev->bios) { | 1321 | if (radeon_boot_test_post_card(rdev) == false) |
| 1307 | DRM_INFO("GPU not posted. posting now...\n"); | 1322 | return -EINVAL; |
| 1308 | radeon_combios_asic_init(rdev->ddev); | ||
| 1309 | } | ||
| 1310 | /* Set asic errata */ | 1323 | /* Set asic errata */ |
| 1311 | r300_errata(rdev); | 1324 | r300_errata(rdev); |
| 1312 | /* Initialize clocks */ | 1325 | /* Initialize clocks */ |
| @@ -1325,7 +1338,7 @@ int r300_init(struct radeon_device *rdev) | |||
| 1325 | if (r) | 1338 | if (r) |
| 1326 | return r; | 1339 | return r; |
| 1327 | /* Memory manager */ | 1340 | /* Memory manager */ |
| 1328 | r = radeon_object_init(rdev); | 1341 | r = radeon_bo_init(rdev); |
| 1329 | if (r) | 1342 | if (r) |
| 1330 | return r; | 1343 | return r; |
| 1331 | if (rdev->flags & RADEON_IS_PCIE) { | 1344 | if (rdev->flags & RADEON_IS_PCIE) { |
diff --git a/drivers/gpu/drm/radeon/r420.c b/drivers/gpu/drm/radeon/r420.c index 1cefdbcc0850..c05a7270cf0c 100644 --- a/drivers/gpu/drm/radeon/r420.c +++ b/drivers/gpu/drm/radeon/r420.c | |||
| @@ -169,6 +169,9 @@ static int r420_startup(struct radeon_device *rdev) | |||
| 169 | { | 169 | { |
| 170 | int r; | 170 | int r; |
| 171 | 171 | ||
| 172 | /* set common regs */ | ||
| 173 | r100_set_common_regs(rdev); | ||
| 174 | /* program mc */ | ||
| 172 | r300_mc_program(rdev); | 175 | r300_mc_program(rdev); |
| 173 | /* Resume clock */ | 176 | /* Resume clock */ |
| 174 | r420_clock_resume(rdev); | 177 | r420_clock_resume(rdev); |
| @@ -186,7 +189,6 @@ static int r420_startup(struct radeon_device *rdev) | |||
| 186 | } | 189 | } |
| 187 | r420_pipes_init(rdev); | 190 | r420_pipes_init(rdev); |
| 188 | /* Enable IRQ */ | 191 | /* Enable IRQ */ |
| 189 | rdev->irq.sw_int = true; | ||
| 190 | r100_irq_set(rdev); | 192 | r100_irq_set(rdev); |
| 191 | /* 1M ring buffer */ | 193 | /* 1M ring buffer */ |
| 192 | r = r100_cp_init(rdev, 1024 * 1024); | 194 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -229,7 +231,8 @@ int r420_resume(struct radeon_device *rdev) | |||
| 229 | } | 231 | } |
| 230 | /* Resume clock after posting */ | 232 | /* Resume clock after posting */ |
| 231 | r420_clock_resume(rdev); | 233 | r420_clock_resume(rdev); |
| 232 | 234 | /* Initialize surface registers */ | |
| 235 | radeon_surface_init(rdev); | ||
| 233 | return r420_startup(rdev); | 236 | return r420_startup(rdev); |
| 234 | } | 237 | } |
| 235 | 238 | ||
| @@ -258,7 +261,7 @@ void r420_fini(struct radeon_device *rdev) | |||
| 258 | radeon_agp_fini(rdev); | 261 | radeon_agp_fini(rdev); |
| 259 | radeon_irq_kms_fini(rdev); | 262 | radeon_irq_kms_fini(rdev); |
| 260 | radeon_fence_driver_fini(rdev); | 263 | radeon_fence_driver_fini(rdev); |
| 261 | radeon_object_fini(rdev); | 264 | radeon_bo_fini(rdev); |
| 262 | if (rdev->is_atom_bios) { | 265 | if (rdev->is_atom_bios) { |
| 263 | radeon_atombios_fini(rdev); | 266 | radeon_atombios_fini(rdev); |
| 264 | } else { | 267 | } else { |
| @@ -301,14 +304,9 @@ int r420_init(struct radeon_device *rdev) | |||
| 301 | RREG32(R_0007C0_CP_STAT)); | 304 | RREG32(R_0007C0_CP_STAT)); |
| 302 | } | 305 | } |
| 303 | /* check if cards are posted or not */ | 306 | /* check if cards are posted or not */ |
| 304 | if (!radeon_card_posted(rdev) && rdev->bios) { | 307 | if (radeon_boot_test_post_card(rdev) == false) |
| 305 | DRM_INFO("GPU not posted. posting now...\n"); | 308 | return -EINVAL; |
| 306 | if (rdev->is_atom_bios) { | 309 | |
| 307 | atom_asic_init(rdev->mode_info.atom_context); | ||
| 308 | } else { | ||
| 309 | radeon_combios_asic_init(rdev->ddev); | ||
| 310 | } | ||
| 311 | } | ||
| 312 | /* Initialize clocks */ | 310 | /* Initialize clocks */ |
| 313 | radeon_get_clock_info(rdev->ddev); | 311 | radeon_get_clock_info(rdev->ddev); |
| 314 | /* Initialize power management */ | 312 | /* Initialize power management */ |
| @@ -331,10 +329,13 @@ int r420_init(struct radeon_device *rdev) | |||
| 331 | return r; | 329 | return r; |
| 332 | } | 330 | } |
| 333 | /* Memory manager */ | 331 | /* Memory manager */ |
| 334 | r = radeon_object_init(rdev); | 332 | r = radeon_bo_init(rdev); |
| 335 | if (r) { | 333 | if (r) { |
| 336 | return r; | 334 | return r; |
| 337 | } | 335 | } |
| 336 | if (rdev->family == CHIP_R420) | ||
| 337 | r100_enable_bm(rdev); | ||
| 338 | |||
| 338 | if (rdev->flags & RADEON_IS_PCIE) { | 339 | if (rdev->flags & RADEON_IS_PCIE) { |
| 339 | r = rv370_pcie_gart_init(rdev); | 340 | r = rv370_pcie_gart_init(rdev); |
| 340 | if (r) | 341 | if (r) |
diff --git a/drivers/gpu/drm/radeon/r500_reg.h b/drivers/gpu/drm/radeon/r500_reg.h index 7baa73955563..74ad89bdf2b5 100644 --- a/drivers/gpu/drm/radeon/r500_reg.h +++ b/drivers/gpu/drm/radeon/r500_reg.h | |||
| @@ -716,6 +716,8 @@ | |||
| 716 | 716 | ||
| 717 | #define AVIVO_DVOA_BIT_DEPTH_CONTROL 0x7988 | 717 | #define AVIVO_DVOA_BIT_DEPTH_CONTROL 0x7988 |
| 718 | 718 | ||
| 719 | #define AVIVO_DC_GPIO_HPD_A 0x7e94 | ||
| 720 | |||
| 719 | #define AVIVO_GPIO_0 0x7e30 | 721 | #define AVIVO_GPIO_0 0x7e30 |
| 720 | #define AVIVO_GPIO_1 0x7e40 | 722 | #define AVIVO_GPIO_1 0x7e40 |
| 721 | #define AVIVO_GPIO_2 0x7e50 | 723 | #define AVIVO_GPIO_2 0x7e50 |
diff --git a/drivers/gpu/drm/radeon/r520.c b/drivers/gpu/drm/radeon/r520.c index f7435185c0a6..0f3843b6dac7 100644 --- a/drivers/gpu/drm/radeon/r520.c +++ b/drivers/gpu/drm/radeon/r520.c | |||
| @@ -185,7 +185,6 @@ static int r520_startup(struct radeon_device *rdev) | |||
| 185 | return r; | 185 | return r; |
| 186 | } | 186 | } |
| 187 | /* Enable IRQ */ | 187 | /* Enable IRQ */ |
| 188 | rdev->irq.sw_int = true; | ||
| 189 | rs600_irq_set(rdev); | 188 | rs600_irq_set(rdev); |
| 190 | /* 1M ring buffer */ | 189 | /* 1M ring buffer */ |
| 191 | r = r100_cp_init(rdev, 1024 * 1024); | 190 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -221,6 +220,8 @@ int r520_resume(struct radeon_device *rdev) | |||
| 221 | atom_asic_init(rdev->mode_info.atom_context); | 220 | atom_asic_init(rdev->mode_info.atom_context); |
| 222 | /* Resume clock after posting */ | 221 | /* Resume clock after posting */ |
| 223 | rv515_clock_startup(rdev); | 222 | rv515_clock_startup(rdev); |
| 223 | /* Initialize surface registers */ | ||
| 224 | radeon_surface_init(rdev); | ||
| 224 | return r520_startup(rdev); | 225 | return r520_startup(rdev); |
| 225 | } | 226 | } |
| 226 | 227 | ||
| @@ -254,6 +255,9 @@ int r520_init(struct radeon_device *rdev) | |||
| 254 | RREG32(R_0007C0_CP_STAT)); | 255 | RREG32(R_0007C0_CP_STAT)); |
| 255 | } | 256 | } |
| 256 | /* check if cards are posted or not */ | 257 | /* check if cards are posted or not */ |
| 258 | if (radeon_boot_test_post_card(rdev) == false) | ||
| 259 | return -EINVAL; | ||
| 260 | |||
| 257 | if (!radeon_card_posted(rdev) && rdev->bios) { | 261 | if (!radeon_card_posted(rdev) && rdev->bios) { |
| 258 | DRM_INFO("GPU not posted. posting now...\n"); | 262 | DRM_INFO("GPU not posted. posting now...\n"); |
| 259 | atom_asic_init(rdev->mode_info.atom_context); | 263 | atom_asic_init(rdev->mode_info.atom_context); |
| @@ -277,7 +281,7 @@ int r520_init(struct radeon_device *rdev) | |||
| 277 | if (r) | 281 | if (r) |
| 278 | return r; | 282 | return r; |
| 279 | /* Memory manager */ | 283 | /* Memory manager */ |
| 280 | r = radeon_object_init(rdev); | 284 | r = radeon_bo_init(rdev); |
| 281 | if (r) | 285 | if (r) |
| 282 | return r; | 286 | return r; |
| 283 | r = rv370_pcie_gart_init(rdev); | 287 | r = rv370_pcie_gart_init(rdev); |
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index 6740ed24358f..36656bd110bf 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c | |||
| @@ -38,8 +38,10 @@ | |||
| 38 | 38 | ||
| 39 | #define PFP_UCODE_SIZE 576 | 39 | #define PFP_UCODE_SIZE 576 |
| 40 | #define PM4_UCODE_SIZE 1792 | 40 | #define PM4_UCODE_SIZE 1792 |
| 41 | #define RLC_UCODE_SIZE 768 | ||
| 41 | #define R700_PFP_UCODE_SIZE 848 | 42 | #define R700_PFP_UCODE_SIZE 848 |
| 42 | #define R700_PM4_UCODE_SIZE 1360 | 43 | #define R700_PM4_UCODE_SIZE 1360 |
| 44 | #define R700_RLC_UCODE_SIZE 1024 | ||
| 43 | 45 | ||
| 44 | /* Firmware Names */ | 46 | /* Firmware Names */ |
| 45 | MODULE_FIRMWARE("radeon/R600_pfp.bin"); | 47 | MODULE_FIRMWARE("radeon/R600_pfp.bin"); |
| @@ -62,6 +64,8 @@ MODULE_FIRMWARE("radeon/RV730_pfp.bin"); | |||
| 62 | MODULE_FIRMWARE("radeon/RV730_me.bin"); | 64 | MODULE_FIRMWARE("radeon/RV730_me.bin"); |
| 63 | MODULE_FIRMWARE("radeon/RV710_pfp.bin"); | 65 | MODULE_FIRMWARE("radeon/RV710_pfp.bin"); |
| 64 | MODULE_FIRMWARE("radeon/RV710_me.bin"); | 66 | MODULE_FIRMWARE("radeon/RV710_me.bin"); |
| 67 | MODULE_FIRMWARE("radeon/R600_rlc.bin"); | ||
| 68 | MODULE_FIRMWARE("radeon/R700_rlc.bin"); | ||
| 65 | 69 | ||
| 66 | int r600_debugfs_mc_info_init(struct radeon_device *rdev); | 70 | int r600_debugfs_mc_info_init(struct radeon_device *rdev); |
| 67 | 71 | ||
| @@ -70,6 +74,281 @@ int r600_mc_wait_for_idle(struct radeon_device *rdev); | |||
| 70 | void r600_gpu_init(struct radeon_device *rdev); | 74 | void r600_gpu_init(struct radeon_device *rdev); |
| 71 | void r600_fini(struct radeon_device *rdev); | 75 | void r600_fini(struct radeon_device *rdev); |
| 72 | 76 | ||
| 77 | /* hpd for digital panel detect/disconnect */ | ||
| 78 | bool r600_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd) | ||
| 79 | { | ||
| 80 | bool connected = false; | ||
| 81 | |||
| 82 | if (ASIC_IS_DCE3(rdev)) { | ||
| 83 | switch (hpd) { | ||
| 84 | case RADEON_HPD_1: | ||
| 85 | if (RREG32(DC_HPD1_INT_STATUS) & DC_HPDx_SENSE) | ||
| 86 | connected = true; | ||
| 87 | break; | ||
| 88 | case RADEON_HPD_2: | ||
| 89 | if (RREG32(DC_HPD2_INT_STATUS) & DC_HPDx_SENSE) | ||
| 90 | connected = true; | ||
| 91 | break; | ||
| 92 | case RADEON_HPD_3: | ||
| 93 | if (RREG32(DC_HPD3_INT_STATUS) & DC_HPDx_SENSE) | ||
| 94 | connected = true; | ||
| 95 | break; | ||
| 96 | case RADEON_HPD_4: | ||
| 97 | if (RREG32(DC_HPD4_INT_STATUS) & DC_HPDx_SENSE) | ||
| 98 | connected = true; | ||
| 99 | break; | ||
| 100 | /* DCE 3.2 */ | ||
| 101 | case RADEON_HPD_5: | ||
| 102 | if (RREG32(DC_HPD5_INT_STATUS) & DC_HPDx_SENSE) | ||
| 103 | connected = true; | ||
| 104 | break; | ||
| 105 | case RADEON_HPD_6: | ||
| 106 | if (RREG32(DC_HPD6_INT_STATUS) & DC_HPDx_SENSE) | ||
| 107 | connected = true; | ||
| 108 | break; | ||
| 109 | default: | ||
| 110 | break; | ||
| 111 | } | ||
| 112 | } else { | ||
| 113 | switch (hpd) { | ||
| 114 | case RADEON_HPD_1: | ||
| 115 | if (RREG32(DC_HOT_PLUG_DETECT1_INT_STATUS) & DC_HOT_PLUG_DETECTx_SENSE) | ||
| 116 | connected = true; | ||
| 117 | break; | ||
| 118 | case RADEON_HPD_2: | ||
| 119 | if (RREG32(DC_HOT_PLUG_DETECT2_INT_STATUS) & DC_HOT_PLUG_DETECTx_SENSE) | ||
| 120 | connected = true; | ||
| 121 | break; | ||
| 122 | case RADEON_HPD_3: | ||
| 123 | if (RREG32(DC_HOT_PLUG_DETECT3_INT_STATUS) & DC_HOT_PLUG_DETECTx_SENSE) | ||
| 124 | connected = true; | ||
| 125 | break; | ||
| 126 | default: | ||
| 127 | break; | ||
| 128 | } | ||
| 129 | } | ||
| 130 | return connected; | ||
| 131 | } | ||
| 132 | |||
| 133 | void r600_hpd_set_polarity(struct radeon_device *rdev, | ||
| 134 | enum radeon_hpd_id hpd) | ||
| 135 | { | ||
| 136 | u32 tmp; | ||
| 137 | bool connected = r600_hpd_sense(rdev, hpd); | ||
| 138 | |||
| 139 | if (ASIC_IS_DCE3(rdev)) { | ||
| 140 | switch (hpd) { | ||
| 141 | case RADEON_HPD_1: | ||
| 142 | tmp = RREG32(DC_HPD1_INT_CONTROL); | ||
| 143 | if (connected) | ||
| 144 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 145 | else | ||
| 146 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 147 | WREG32(DC_HPD1_INT_CONTROL, tmp); | ||
| 148 | break; | ||
| 149 | case RADEON_HPD_2: | ||
| 150 | tmp = RREG32(DC_HPD2_INT_CONTROL); | ||
| 151 | if (connected) | ||
| 152 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 153 | else | ||
| 154 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 155 | WREG32(DC_HPD2_INT_CONTROL, tmp); | ||
| 156 | break; | ||
| 157 | case RADEON_HPD_3: | ||
| 158 | tmp = RREG32(DC_HPD3_INT_CONTROL); | ||
| 159 | if (connected) | ||
| 160 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 161 | else | ||
| 162 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 163 | WREG32(DC_HPD3_INT_CONTROL, tmp); | ||
| 164 | break; | ||
| 165 | case RADEON_HPD_4: | ||
| 166 | tmp = RREG32(DC_HPD4_INT_CONTROL); | ||
| 167 | if (connected) | ||
| 168 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 169 | else | ||
| 170 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 171 | WREG32(DC_HPD4_INT_CONTROL, tmp); | ||
| 172 | break; | ||
| 173 | case RADEON_HPD_5: | ||
| 174 | tmp = RREG32(DC_HPD5_INT_CONTROL); | ||
| 175 | if (connected) | ||
| 176 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 177 | else | ||
| 178 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 179 | WREG32(DC_HPD5_INT_CONTROL, tmp); | ||
| 180 | break; | ||
| 181 | /* DCE 3.2 */ | ||
| 182 | case RADEON_HPD_6: | ||
| 183 | tmp = RREG32(DC_HPD6_INT_CONTROL); | ||
| 184 | if (connected) | ||
| 185 | tmp &= ~DC_HPDx_INT_POLARITY; | ||
| 186 | else | ||
| 187 | tmp |= DC_HPDx_INT_POLARITY; | ||
| 188 | WREG32(DC_HPD6_INT_CONTROL, tmp); | ||
| 189 | break; | ||
| 190 | default: | ||
| 191 | break; | ||
| 192 | } | ||
| 193 | } else { | ||
| 194 | switch (hpd) { | ||
| 195 | case RADEON_HPD_1: | ||
| 196 | tmp = RREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL); | ||
| 197 | if (connected) | ||
| 198 | tmp &= ~DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 199 | else | ||
| 200 | tmp |= DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 201 | WREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL, tmp); | ||
| 202 | break; | ||
| 203 | case RADEON_HPD_2: | ||
| 204 | tmp = RREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL); | ||
| 205 | if (connected) | ||
| 206 | tmp &= ~DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 207 | else | ||
| 208 | tmp |= DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 209 | WREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL, tmp); | ||
| 210 | break; | ||
| 211 | case RADEON_HPD_3: | ||
| 212 | tmp = RREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL); | ||
| 213 | if (connected) | ||
| 214 | tmp &= ~DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 215 | else | ||
| 216 | tmp |= DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 217 | WREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL, tmp); | ||
| 218 | break; | ||
| 219 | default: | ||
| 220 | break; | ||
| 221 | } | ||
| 222 | } | ||
| 223 | } | ||
| 224 | |||
| 225 | void r600_hpd_init(struct radeon_device *rdev) | ||
| 226 | { | ||
| 227 | struct drm_device *dev = rdev->ddev; | ||
| 228 | struct drm_connector *connector; | ||
| 229 | |||
| 230 | if (ASIC_IS_DCE3(rdev)) { | ||
| 231 | u32 tmp = DC_HPDx_CONNECTION_TIMER(0x9c4) | DC_HPDx_RX_INT_TIMER(0xfa); | ||
| 232 | if (ASIC_IS_DCE32(rdev)) | ||
| 233 | tmp |= DC_HPDx_EN; | ||
| 234 | |||
| 235 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 236 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 237 | switch (radeon_connector->hpd.hpd) { | ||
| 238 | case RADEON_HPD_1: | ||
| 239 | WREG32(DC_HPD1_CONTROL, tmp); | ||
| 240 | rdev->irq.hpd[0] = true; | ||
| 241 | break; | ||
| 242 | case RADEON_HPD_2: | ||
| 243 | WREG32(DC_HPD2_CONTROL, tmp); | ||
| 244 | rdev->irq.hpd[1] = true; | ||
| 245 | break; | ||
| 246 | case RADEON_HPD_3: | ||
| 247 | WREG32(DC_HPD3_CONTROL, tmp); | ||
| 248 | rdev->irq.hpd[2] = true; | ||
| 249 | break; | ||
| 250 | case RADEON_HPD_4: | ||
| 251 | WREG32(DC_HPD4_CONTROL, tmp); | ||
| 252 | rdev->irq.hpd[3] = true; | ||
| 253 | break; | ||
| 254 | /* DCE 3.2 */ | ||
| 255 | case RADEON_HPD_5: | ||
| 256 | WREG32(DC_HPD5_CONTROL, tmp); | ||
| 257 | rdev->irq.hpd[4] = true; | ||
| 258 | break; | ||
| 259 | case RADEON_HPD_6: | ||
| 260 | WREG32(DC_HPD6_CONTROL, tmp); | ||
| 261 | rdev->irq.hpd[5] = true; | ||
| 262 | break; | ||
| 263 | default: | ||
| 264 | break; | ||
| 265 | } | ||
| 266 | } | ||
| 267 | } else { | ||
| 268 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 269 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 270 | switch (radeon_connector->hpd.hpd) { | ||
| 271 | case RADEON_HPD_1: | ||
| 272 | WREG32(DC_HOT_PLUG_DETECT1_CONTROL, DC_HOT_PLUG_DETECTx_EN); | ||
| 273 | rdev->irq.hpd[0] = true; | ||
| 274 | break; | ||
| 275 | case RADEON_HPD_2: | ||
| 276 | WREG32(DC_HOT_PLUG_DETECT2_CONTROL, DC_HOT_PLUG_DETECTx_EN); | ||
| 277 | rdev->irq.hpd[1] = true; | ||
| 278 | break; | ||
| 279 | case RADEON_HPD_3: | ||
| 280 | WREG32(DC_HOT_PLUG_DETECT3_CONTROL, DC_HOT_PLUG_DETECTx_EN); | ||
| 281 | rdev->irq.hpd[2] = true; | ||
| 282 | break; | ||
| 283 | default: | ||
| 284 | break; | ||
| 285 | } | ||
| 286 | } | ||
| 287 | } | ||
| 288 | r600_irq_set(rdev); | ||
| 289 | } | ||
| 290 | |||
| 291 | void r600_hpd_fini(struct radeon_device *rdev) | ||
| 292 | { | ||
| 293 | struct drm_device *dev = rdev->ddev; | ||
| 294 | struct drm_connector *connector; | ||
| 295 | |||
| 296 | if (ASIC_IS_DCE3(rdev)) { | ||
| 297 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 298 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 299 | switch (radeon_connector->hpd.hpd) { | ||
| 300 | case RADEON_HPD_1: | ||
| 301 | WREG32(DC_HPD1_CONTROL, 0); | ||
| 302 | rdev->irq.hpd[0] = false; | ||
| 303 | break; | ||
| 304 | case RADEON_HPD_2: | ||
| 305 | WREG32(DC_HPD2_CONTROL, 0); | ||
| 306 | rdev->irq.hpd[1] = false; | ||
| 307 | break; | ||
| 308 | case RADEON_HPD_3: | ||
| 309 | WREG32(DC_HPD3_CONTROL, 0); | ||
| 310 | rdev->irq.hpd[2] = false; | ||
| 311 | break; | ||
| 312 | case RADEON_HPD_4: | ||
| 313 | WREG32(DC_HPD4_CONTROL, 0); | ||
| 314 | rdev->irq.hpd[3] = false; | ||
| 315 | break; | ||
| 316 | /* DCE 3.2 */ | ||
| 317 | case RADEON_HPD_5: | ||
| 318 | WREG32(DC_HPD5_CONTROL, 0); | ||
| 319 | rdev->irq.hpd[4] = false; | ||
| 320 | break; | ||
| 321 | case RADEON_HPD_6: | ||
| 322 | WREG32(DC_HPD6_CONTROL, 0); | ||
| 323 | rdev->irq.hpd[5] = false; | ||
| 324 | break; | ||
| 325 | default: | ||
| 326 | break; | ||
| 327 | } | ||
| 328 | } | ||
| 329 | } else { | ||
| 330 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 331 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 332 | switch (radeon_connector->hpd.hpd) { | ||
| 333 | case RADEON_HPD_1: | ||
| 334 | WREG32(DC_HOT_PLUG_DETECT1_CONTROL, 0); | ||
| 335 | rdev->irq.hpd[0] = false; | ||
| 336 | break; | ||
| 337 | case RADEON_HPD_2: | ||
| 338 | WREG32(DC_HOT_PLUG_DETECT2_CONTROL, 0); | ||
| 339 | rdev->irq.hpd[1] = false; | ||
| 340 | break; | ||
| 341 | case RADEON_HPD_3: | ||
| 342 | WREG32(DC_HOT_PLUG_DETECT3_CONTROL, 0); | ||
| 343 | rdev->irq.hpd[2] = false; | ||
| 344 | break; | ||
| 345 | default: | ||
| 346 | break; | ||
| 347 | } | ||
| 348 | } | ||
| 349 | } | ||
| 350 | } | ||
| 351 | |||
| 73 | /* | 352 | /* |
| 74 | * R600 PCIE GART | 353 | * R600 PCIE GART |
| 75 | */ | 354 | */ |
| @@ -180,7 +459,7 @@ int r600_pcie_gart_enable(struct radeon_device *rdev) | |||
| 180 | void r600_pcie_gart_disable(struct radeon_device *rdev) | 459 | void r600_pcie_gart_disable(struct radeon_device *rdev) |
| 181 | { | 460 | { |
| 182 | u32 tmp; | 461 | u32 tmp; |
| 183 | int i; | 462 | int i, r; |
| 184 | 463 | ||
| 185 | /* Disable all tables */ | 464 | /* Disable all tables */ |
| 186 | for (i = 0; i < 7; i++) | 465 | for (i = 0; i < 7; i++) |
| @@ -208,8 +487,12 @@ void r600_pcie_gart_disable(struct radeon_device *rdev) | |||
| 208 | WREG32(MC_VM_L1_TLB_MCB_RD_HDP_CNTL, tmp); | 487 | WREG32(MC_VM_L1_TLB_MCB_RD_HDP_CNTL, tmp); |
| 209 | WREG32(MC_VM_L1_TLB_MCB_WR_HDP_CNTL, tmp); | 488 | WREG32(MC_VM_L1_TLB_MCB_WR_HDP_CNTL, tmp); |
| 210 | if (rdev->gart.table.vram.robj) { | 489 | if (rdev->gart.table.vram.robj) { |
| 211 | radeon_object_kunmap(rdev->gart.table.vram.robj); | 490 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 212 | radeon_object_unpin(rdev->gart.table.vram.robj); | 491 | if (likely(r == 0)) { |
| 492 | radeon_bo_kunmap(rdev->gart.table.vram.robj); | ||
| 493 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 494 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 495 | } | ||
| 213 | } | 496 | } |
| 214 | } | 497 | } |
| 215 | 498 | ||
| @@ -1101,6 +1384,10 @@ void r600_pciep_wreg(struct radeon_device *rdev, u32 reg, u32 v) | |||
| 1101 | (void)RREG32(PCIE_PORT_DATA); | 1384 | (void)RREG32(PCIE_PORT_DATA); |
| 1102 | } | 1385 | } |
| 1103 | 1386 | ||
| 1387 | void r600_hdp_flush(struct radeon_device *rdev) | ||
| 1388 | { | ||
| 1389 | WREG32(R_005480_HDP_MEM_COHERENCY_FLUSH_CNTL, 0x1); | ||
| 1390 | } | ||
| 1104 | 1391 | ||
| 1105 | /* | 1392 | /* |
| 1106 | * CP & Ring | 1393 | * CP & Ring |
| @@ -1110,11 +1397,12 @@ void r600_cp_stop(struct radeon_device *rdev) | |||
| 1110 | WREG32(R_0086D8_CP_ME_CNTL, S_0086D8_CP_ME_HALT(1)); | 1397 | WREG32(R_0086D8_CP_ME_CNTL, S_0086D8_CP_ME_HALT(1)); |
| 1111 | } | 1398 | } |
| 1112 | 1399 | ||
| 1113 | int r600_cp_init_microcode(struct radeon_device *rdev) | 1400 | int r600_init_microcode(struct radeon_device *rdev) |
| 1114 | { | 1401 | { |
| 1115 | struct platform_device *pdev; | 1402 | struct platform_device *pdev; |
| 1116 | const char *chip_name; | 1403 | const char *chip_name; |
| 1117 | size_t pfp_req_size, me_req_size; | 1404 | const char *rlc_chip_name; |
| 1405 | size_t pfp_req_size, me_req_size, rlc_req_size; | ||
| 1118 | char fw_name[30]; | 1406 | char fw_name[30]; |
| 1119 | int err; | 1407 | int err; |
| 1120 | 1408 | ||
| @@ -1128,30 +1416,62 @@ int r600_cp_init_microcode(struct radeon_device *rdev) | |||
| 1128 | } | 1416 | } |
| 1129 | 1417 | ||
| 1130 | switch (rdev->family) { | 1418 | switch (rdev->family) { |
| 1131 | case CHIP_R600: chip_name = "R600"; break; | 1419 | case CHIP_R600: |
| 1132 | case CHIP_RV610: chip_name = "RV610"; break; | 1420 | chip_name = "R600"; |
| 1133 | case CHIP_RV630: chip_name = "RV630"; break; | 1421 | rlc_chip_name = "R600"; |
| 1134 | case CHIP_RV620: chip_name = "RV620"; break; | 1422 | break; |
| 1135 | case CHIP_RV635: chip_name = "RV635"; break; | 1423 | case CHIP_RV610: |
| 1136 | case CHIP_RV670: chip_name = "RV670"; break; | 1424 | chip_name = "RV610"; |
| 1425 | rlc_chip_name = "R600"; | ||
| 1426 | break; | ||
| 1427 | case CHIP_RV630: | ||
| 1428 | chip_name = "RV630"; | ||
| 1429 | rlc_chip_name = "R600"; | ||
| 1430 | break; | ||
| 1431 | case CHIP_RV620: | ||
| 1432 | chip_name = "RV620"; | ||
| 1433 | rlc_chip_name = "R600"; | ||
| 1434 | break; | ||
| 1435 | case CHIP_RV635: | ||
| 1436 | chip_name = "RV635"; | ||
| 1437 | rlc_chip_name = "R600"; | ||
| 1438 | break; | ||
| 1439 | case CHIP_RV670: | ||
| 1440 | chip_name = "RV670"; | ||
| 1441 | rlc_chip_name = "R600"; | ||
| 1442 | break; | ||
| 1137 | case CHIP_RS780: | 1443 | case CHIP_RS780: |
| 1138 | case CHIP_RS880: chip_name = "RS780"; break; | 1444 | case CHIP_RS880: |
| 1139 | case CHIP_RV770: chip_name = "RV770"; break; | 1445 | chip_name = "RS780"; |
| 1446 | rlc_chip_name = "R600"; | ||
| 1447 | break; | ||
| 1448 | case CHIP_RV770: | ||
| 1449 | chip_name = "RV770"; | ||
| 1450 | rlc_chip_name = "R700"; | ||
| 1451 | break; | ||
| 1140 | case CHIP_RV730: | 1452 | case CHIP_RV730: |
| 1141 | case CHIP_RV740: chip_name = "RV730"; break; | 1453 | case CHIP_RV740: |
| 1142 | case CHIP_RV710: chip_name = "RV710"; break; | 1454 | chip_name = "RV730"; |
| 1455 | rlc_chip_name = "R700"; | ||
| 1456 | break; | ||
| 1457 | case CHIP_RV710: | ||
| 1458 | chip_name = "RV710"; | ||
| 1459 | rlc_chip_name = "R700"; | ||
| 1460 | break; | ||
| 1143 | default: BUG(); | 1461 | default: BUG(); |
| 1144 | } | 1462 | } |
| 1145 | 1463 | ||
| 1146 | if (rdev->family >= CHIP_RV770) { | 1464 | if (rdev->family >= CHIP_RV770) { |
| 1147 | pfp_req_size = R700_PFP_UCODE_SIZE * 4; | 1465 | pfp_req_size = R700_PFP_UCODE_SIZE * 4; |
| 1148 | me_req_size = R700_PM4_UCODE_SIZE * 4; | 1466 | me_req_size = R700_PM4_UCODE_SIZE * 4; |
| 1467 | rlc_req_size = R700_RLC_UCODE_SIZE * 4; | ||
| 1149 | } else { | 1468 | } else { |
| 1150 | pfp_req_size = PFP_UCODE_SIZE * 4; | 1469 | pfp_req_size = PFP_UCODE_SIZE * 4; |
| 1151 | me_req_size = PM4_UCODE_SIZE * 12; | 1470 | me_req_size = PM4_UCODE_SIZE * 12; |
| 1471 | rlc_req_size = RLC_UCODE_SIZE * 4; | ||
| 1152 | } | 1472 | } |
| 1153 | 1473 | ||
| 1154 | DRM_INFO("Loading %s CP Microcode\n", chip_name); | 1474 | DRM_INFO("Loading %s Microcode\n", chip_name); |
| 1155 | 1475 | ||
| 1156 | snprintf(fw_name, sizeof(fw_name), "radeon/%s_pfp.bin", chip_name); | 1476 | snprintf(fw_name, sizeof(fw_name), "radeon/%s_pfp.bin", chip_name); |
| 1157 | err = request_firmware(&rdev->pfp_fw, fw_name, &pdev->dev); | 1477 | err = request_firmware(&rdev->pfp_fw, fw_name, &pdev->dev); |
| @@ -1175,6 +1495,18 @@ int r600_cp_init_microcode(struct radeon_device *rdev) | |||
| 1175 | rdev->me_fw->size, fw_name); | 1495 | rdev->me_fw->size, fw_name); |
| 1176 | err = -EINVAL; | 1496 | err = -EINVAL; |
| 1177 | } | 1497 | } |
| 1498 | |||
| 1499 | snprintf(fw_name, sizeof(fw_name), "radeon/%s_rlc.bin", rlc_chip_name); | ||
| 1500 | err = request_firmware(&rdev->rlc_fw, fw_name, &pdev->dev); | ||
| 1501 | if (err) | ||
| 1502 | goto out; | ||
| 1503 | if (rdev->rlc_fw->size != rlc_req_size) { | ||
| 1504 | printk(KERN_ERR | ||
| 1505 | "r600_rlc: Bogus length %zu in firmware \"%s\"\n", | ||
| 1506 | rdev->rlc_fw->size, fw_name); | ||
| 1507 | err = -EINVAL; | ||
| 1508 | } | ||
| 1509 | |||
| 1178 | out: | 1510 | out: |
| 1179 | platform_device_unregister(pdev); | 1511 | platform_device_unregister(pdev); |
| 1180 | 1512 | ||
| @@ -1187,6 +1519,8 @@ out: | |||
| 1187 | rdev->pfp_fw = NULL; | 1519 | rdev->pfp_fw = NULL; |
| 1188 | release_firmware(rdev->me_fw); | 1520 | release_firmware(rdev->me_fw); |
| 1189 | rdev->me_fw = NULL; | 1521 | rdev->me_fw = NULL; |
| 1522 | release_firmware(rdev->rlc_fw); | ||
| 1523 | rdev->rlc_fw = NULL; | ||
| 1190 | } | 1524 | } |
| 1191 | return err; | 1525 | return err; |
| 1192 | } | 1526 | } |
| @@ -1381,10 +1715,16 @@ int r600_ring_test(struct radeon_device *rdev) | |||
| 1381 | 1715 | ||
| 1382 | void r600_wb_disable(struct radeon_device *rdev) | 1716 | void r600_wb_disable(struct radeon_device *rdev) |
| 1383 | { | 1717 | { |
| 1718 | int r; | ||
| 1719 | |||
| 1384 | WREG32(SCRATCH_UMSK, 0); | 1720 | WREG32(SCRATCH_UMSK, 0); |
| 1385 | if (rdev->wb.wb_obj) { | 1721 | if (rdev->wb.wb_obj) { |
| 1386 | radeon_object_kunmap(rdev->wb.wb_obj); | 1722 | r = radeon_bo_reserve(rdev->wb.wb_obj, false); |
| 1387 | radeon_object_unpin(rdev->wb.wb_obj); | 1723 | if (unlikely(r != 0)) |
| 1724 | return; | ||
| 1725 | radeon_bo_kunmap(rdev->wb.wb_obj); | ||
| 1726 | radeon_bo_unpin(rdev->wb.wb_obj); | ||
| 1727 | radeon_bo_unreserve(rdev->wb.wb_obj); | ||
| 1388 | } | 1728 | } |
| 1389 | } | 1729 | } |
| 1390 | 1730 | ||
| @@ -1392,7 +1732,7 @@ void r600_wb_fini(struct radeon_device *rdev) | |||
| 1392 | { | 1732 | { |
| 1393 | r600_wb_disable(rdev); | 1733 | r600_wb_disable(rdev); |
| 1394 | if (rdev->wb.wb_obj) { | 1734 | if (rdev->wb.wb_obj) { |
| 1395 | radeon_object_unref(&rdev->wb.wb_obj); | 1735 | radeon_bo_unref(&rdev->wb.wb_obj); |
| 1396 | rdev->wb.wb = NULL; | 1736 | rdev->wb.wb = NULL; |
| 1397 | rdev->wb.wb_obj = NULL; | 1737 | rdev->wb.wb_obj = NULL; |
| 1398 | } | 1738 | } |
| @@ -1403,22 +1743,29 @@ int r600_wb_enable(struct radeon_device *rdev) | |||
| 1403 | int r; | 1743 | int r; |
| 1404 | 1744 | ||
| 1405 | if (rdev->wb.wb_obj == NULL) { | 1745 | if (rdev->wb.wb_obj == NULL) { |
| 1406 | r = radeon_object_create(rdev, NULL, RADEON_GPU_PAGE_SIZE, true, | 1746 | r = radeon_bo_create(rdev, NULL, RADEON_GPU_PAGE_SIZE, true, |
| 1407 | RADEON_GEM_DOMAIN_GTT, false, &rdev->wb.wb_obj); | 1747 | RADEON_GEM_DOMAIN_GTT, &rdev->wb.wb_obj); |
| 1408 | if (r) { | 1748 | if (r) { |
| 1409 | dev_warn(rdev->dev, "failed to create WB buffer (%d).\n", r); | 1749 | dev_warn(rdev->dev, "(%d) create WB bo failed\n", r); |
| 1410 | return r; | 1750 | return r; |
| 1411 | } | 1751 | } |
| 1412 | r = radeon_object_pin(rdev->wb.wb_obj, RADEON_GEM_DOMAIN_GTT, | 1752 | r = radeon_bo_reserve(rdev->wb.wb_obj, false); |
| 1753 | if (unlikely(r != 0)) { | ||
| 1754 | r600_wb_fini(rdev); | ||
| 1755 | return r; | ||
| 1756 | } | ||
| 1757 | r = radeon_bo_pin(rdev->wb.wb_obj, RADEON_GEM_DOMAIN_GTT, | ||
| 1413 | &rdev->wb.gpu_addr); | 1758 | &rdev->wb.gpu_addr); |
| 1414 | if (r) { | 1759 | if (r) { |
| 1415 | dev_warn(rdev->dev, "failed to pin WB buffer (%d).\n", r); | 1760 | radeon_bo_unreserve(rdev->wb.wb_obj); |
| 1761 | dev_warn(rdev->dev, "(%d) pin WB bo failed\n", r); | ||
| 1416 | r600_wb_fini(rdev); | 1762 | r600_wb_fini(rdev); |
| 1417 | return r; | 1763 | return r; |
| 1418 | } | 1764 | } |
| 1419 | r = radeon_object_kmap(rdev->wb.wb_obj, (void **)&rdev->wb.wb); | 1765 | r = radeon_bo_kmap(rdev->wb.wb_obj, (void **)&rdev->wb.wb); |
| 1766 | radeon_bo_unreserve(rdev->wb.wb_obj); | ||
| 1420 | if (r) { | 1767 | if (r) { |
| 1421 | dev_warn(rdev->dev, "failed to map WB buffer (%d).\n", r); | 1768 | dev_warn(rdev->dev, "(%d) map WB bo failed\n", r); |
| 1422 | r600_wb_fini(rdev); | 1769 | r600_wb_fini(rdev); |
| 1423 | return r; | 1770 | return r; |
| 1424 | } | 1771 | } |
| @@ -1433,10 +1780,14 @@ int r600_wb_enable(struct radeon_device *rdev) | |||
| 1433 | void r600_fence_ring_emit(struct radeon_device *rdev, | 1780 | void r600_fence_ring_emit(struct radeon_device *rdev, |
| 1434 | struct radeon_fence *fence) | 1781 | struct radeon_fence *fence) |
| 1435 | { | 1782 | { |
| 1783 | /* Also consider EVENT_WRITE_EOP. it handles the interrupts + timestamps + events */ | ||
| 1436 | /* Emit fence sequence & fire IRQ */ | 1784 | /* Emit fence sequence & fire IRQ */ |
| 1437 | radeon_ring_write(rdev, PACKET3(PACKET3_SET_CONFIG_REG, 1)); | 1785 | radeon_ring_write(rdev, PACKET3(PACKET3_SET_CONFIG_REG, 1)); |
| 1438 | radeon_ring_write(rdev, ((rdev->fence_drv.scratch_reg - PACKET3_SET_CONFIG_REG_OFFSET) >> 2)); | 1786 | radeon_ring_write(rdev, ((rdev->fence_drv.scratch_reg - PACKET3_SET_CONFIG_REG_OFFSET) >> 2)); |
| 1439 | radeon_ring_write(rdev, fence->seq); | 1787 | radeon_ring_write(rdev, fence->seq); |
| 1788 | /* CP_INTERRUPT packet 3 no longer exists, use packet 0 */ | ||
| 1789 | radeon_ring_write(rdev, PACKET0(CP_INT_STATUS, 0)); | ||
| 1790 | radeon_ring_write(rdev, RB_INT_STAT); | ||
| 1440 | } | 1791 | } |
| 1441 | 1792 | ||
| 1442 | int r600_copy_dma(struct radeon_device *rdev, | 1793 | int r600_copy_dma(struct radeon_device *rdev, |
| @@ -1459,18 +1810,6 @@ int r600_copy_blit(struct radeon_device *rdev, | |||
| 1459 | return 0; | 1810 | return 0; |
| 1460 | } | 1811 | } |
| 1461 | 1812 | ||
| 1462 | int r600_irq_process(struct radeon_device *rdev) | ||
| 1463 | { | ||
| 1464 | /* FIXME: implement */ | ||
| 1465 | return 0; | ||
| 1466 | } | ||
| 1467 | |||
| 1468 | int r600_irq_set(struct radeon_device *rdev) | ||
| 1469 | { | ||
| 1470 | /* FIXME: implement */ | ||
| 1471 | return 0; | ||
| 1472 | } | ||
| 1473 | |||
| 1474 | int r600_set_surface_reg(struct radeon_device *rdev, int reg, | 1813 | int r600_set_surface_reg(struct radeon_device *rdev, int reg, |
| 1475 | uint32_t tiling_flags, uint32_t pitch, | 1814 | uint32_t tiling_flags, uint32_t pitch, |
| 1476 | uint32_t offset, uint32_t obj_size) | 1815 | uint32_t offset, uint32_t obj_size) |
| @@ -1506,6 +1845,14 @@ int r600_startup(struct radeon_device *rdev) | |||
| 1506 | { | 1845 | { |
| 1507 | int r; | 1846 | int r; |
| 1508 | 1847 | ||
| 1848 | if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) { | ||
| 1849 | r = r600_init_microcode(rdev); | ||
| 1850 | if (r) { | ||
| 1851 | DRM_ERROR("Failed to load firmware!\n"); | ||
| 1852 | return r; | ||
| 1853 | } | ||
| 1854 | } | ||
| 1855 | |||
| 1509 | r600_mc_program(rdev); | 1856 | r600_mc_program(rdev); |
| 1510 | if (rdev->flags & RADEON_IS_AGP) { | 1857 | if (rdev->flags & RADEON_IS_AGP) { |
| 1511 | r600_agp_enable(rdev); | 1858 | r600_agp_enable(rdev); |
| @@ -1516,13 +1863,26 @@ int r600_startup(struct radeon_device *rdev) | |||
| 1516 | } | 1863 | } |
| 1517 | r600_gpu_init(rdev); | 1864 | r600_gpu_init(rdev); |
| 1518 | 1865 | ||
| 1519 | r = radeon_object_pin(rdev->r600_blit.shader_obj, RADEON_GEM_DOMAIN_VRAM, | 1866 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); |
| 1520 | &rdev->r600_blit.shader_gpu_addr); | 1867 | if (unlikely(r != 0)) |
| 1868 | return r; | ||
| 1869 | r = radeon_bo_pin(rdev->r600_blit.shader_obj, RADEON_GEM_DOMAIN_VRAM, | ||
| 1870 | &rdev->r600_blit.shader_gpu_addr); | ||
| 1871 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); | ||
| 1521 | if (r) { | 1872 | if (r) { |
| 1522 | DRM_ERROR("failed to pin blit object %d\n", r); | 1873 | dev_err(rdev->dev, "(%d) pin blit object failed\n", r); |
| 1523 | return r; | 1874 | return r; |
| 1524 | } | 1875 | } |
| 1525 | 1876 | ||
| 1877 | /* Enable IRQ */ | ||
| 1878 | r = r600_irq_init(rdev); | ||
| 1879 | if (r) { | ||
| 1880 | DRM_ERROR("radeon: IH init failed (%d).\n", r); | ||
| 1881 | radeon_irq_kms_fini(rdev); | ||
| 1882 | return r; | ||
| 1883 | } | ||
| 1884 | r600_irq_set(rdev); | ||
| 1885 | |||
| 1526 | r = radeon_ring_init(rdev, rdev->cp.ring_size); | 1886 | r = radeon_ring_init(rdev, rdev->cp.ring_size); |
| 1527 | if (r) | 1887 | if (r) |
| 1528 | return r; | 1888 | return r; |
| @@ -1583,13 +1943,19 @@ int r600_resume(struct radeon_device *rdev) | |||
| 1583 | 1943 | ||
| 1584 | int r600_suspend(struct radeon_device *rdev) | 1944 | int r600_suspend(struct radeon_device *rdev) |
| 1585 | { | 1945 | { |
| 1946 | int r; | ||
| 1947 | |||
| 1586 | /* FIXME: we should wait for ring to be empty */ | 1948 | /* FIXME: we should wait for ring to be empty */ |
| 1587 | r600_cp_stop(rdev); | 1949 | r600_cp_stop(rdev); |
| 1588 | rdev->cp.ready = false; | 1950 | rdev->cp.ready = false; |
| 1589 | r600_wb_disable(rdev); | 1951 | r600_wb_disable(rdev); |
| 1590 | r600_pcie_gart_disable(rdev); | 1952 | r600_pcie_gart_disable(rdev); |
| 1591 | /* unpin shaders bo */ | 1953 | /* unpin shaders bo */ |
| 1592 | radeon_object_unpin(rdev->r600_blit.shader_obj); | 1954 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); |
| 1955 | if (unlikely(r != 0)) | ||
| 1956 | return r; | ||
| 1957 | radeon_bo_unpin(rdev->r600_blit.shader_obj); | ||
| 1958 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); | ||
| 1593 | return 0; | 1959 | return 0; |
| 1594 | } | 1960 | } |
| 1595 | 1961 | ||
| @@ -1627,7 +1993,11 @@ int r600_init(struct radeon_device *rdev) | |||
| 1627 | if (r) | 1993 | if (r) |
| 1628 | return r; | 1994 | return r; |
| 1629 | /* Post card if necessary */ | 1995 | /* Post card if necessary */ |
| 1630 | if (!r600_card_posted(rdev) && rdev->bios) { | 1996 | if (!r600_card_posted(rdev)) { |
| 1997 | if (!rdev->bios) { | ||
| 1998 | dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n"); | ||
| 1999 | return -EINVAL; | ||
| 2000 | } | ||
| 1631 | DRM_INFO("GPU not posted. posting now...\n"); | 2001 | DRM_INFO("GPU not posted. posting now...\n"); |
| 1632 | atom_asic_init(rdev->mode_info.atom_context); | 2002 | atom_asic_init(rdev->mode_info.atom_context); |
| 1633 | } | 2003 | } |
| @@ -1650,31 +2020,31 @@ int r600_init(struct radeon_device *rdev) | |||
| 1650 | if (r) | 2020 | if (r) |
| 1651 | return r; | 2021 | return r; |
| 1652 | /* Memory manager */ | 2022 | /* Memory manager */ |
| 1653 | r = radeon_object_init(rdev); | 2023 | r = radeon_bo_init(rdev); |
| 1654 | if (r) | 2024 | if (r) |
| 1655 | return r; | 2025 | return r; |
| 2026 | |||
| 2027 | r = radeon_irq_kms_init(rdev); | ||
| 2028 | if (r) | ||
| 2029 | return r; | ||
| 2030 | |||
| 1656 | rdev->cp.ring_obj = NULL; | 2031 | rdev->cp.ring_obj = NULL; |
| 1657 | r600_ring_init(rdev, 1024 * 1024); | 2032 | r600_ring_init(rdev, 1024 * 1024); |
| 1658 | 2033 | ||
| 1659 | if (!rdev->me_fw || !rdev->pfp_fw) { | 2034 | rdev->ih.ring_obj = NULL; |
| 1660 | r = r600_cp_init_microcode(rdev); | 2035 | r600_ih_ring_init(rdev, 64 * 1024); |
| 1661 | if (r) { | ||
| 1662 | DRM_ERROR("Failed to load firmware!\n"); | ||
| 1663 | return r; | ||
| 1664 | } | ||
| 1665 | } | ||
| 1666 | 2036 | ||
| 1667 | r = r600_pcie_gart_init(rdev); | 2037 | r = r600_pcie_gart_init(rdev); |
| 1668 | if (r) | 2038 | if (r) |
| 1669 | return r; | 2039 | return r; |
| 1670 | 2040 | ||
| 1671 | rdev->accel_working = true; | ||
| 1672 | r = r600_blit_init(rdev); | 2041 | r = r600_blit_init(rdev); |
| 1673 | if (r) { | 2042 | if (r) { |
| 1674 | DRM_ERROR("radeon: failled blitter (%d).\n", r); | 2043 | DRM_ERROR("radeon: failed blitter (%d).\n", r); |
| 1675 | return r; | 2044 | return r; |
| 1676 | } | 2045 | } |
| 1677 | 2046 | ||
| 2047 | rdev->accel_working = true; | ||
| 1678 | r = r600_startup(rdev); | 2048 | r = r600_startup(rdev); |
| 1679 | if (r) { | 2049 | if (r) { |
| 1680 | r600_suspend(rdev); | 2050 | r600_suspend(rdev); |
| @@ -1686,12 +2056,12 @@ int r600_init(struct radeon_device *rdev) | |||
| 1686 | if (rdev->accel_working) { | 2056 | if (rdev->accel_working) { |
| 1687 | r = radeon_ib_pool_init(rdev); | 2057 | r = radeon_ib_pool_init(rdev); |
| 1688 | if (r) { | 2058 | if (r) { |
| 1689 | DRM_ERROR("radeon: failled initializing IB pool (%d).\n", r); | 2059 | DRM_ERROR("radeon: failed initializing IB pool (%d).\n", r); |
| 1690 | rdev->accel_working = false; | 2060 | rdev->accel_working = false; |
| 1691 | } | 2061 | } |
| 1692 | r = r600_ib_test(rdev); | 2062 | r = r600_ib_test(rdev); |
| 1693 | if (r) { | 2063 | if (r) { |
| 1694 | DRM_ERROR("radeon: failled testing IB (%d).\n", r); | 2064 | DRM_ERROR("radeon: failed testing IB (%d).\n", r); |
| 1695 | rdev->accel_working = false; | 2065 | rdev->accel_working = false; |
| 1696 | } | 2066 | } |
| 1697 | } | 2067 | } |
| @@ -1704,6 +2074,8 @@ void r600_fini(struct radeon_device *rdev) | |||
| 1704 | r600_suspend(rdev); | 2074 | r600_suspend(rdev); |
| 1705 | 2075 | ||
| 1706 | r600_blit_fini(rdev); | 2076 | r600_blit_fini(rdev); |
| 2077 | r600_irq_fini(rdev); | ||
| 2078 | radeon_irq_kms_fini(rdev); | ||
| 1707 | radeon_ring_fini(rdev); | 2079 | radeon_ring_fini(rdev); |
| 1708 | r600_wb_fini(rdev); | 2080 | r600_wb_fini(rdev); |
| 1709 | r600_pcie_gart_fini(rdev); | 2081 | r600_pcie_gart_fini(rdev); |
| @@ -1712,7 +2084,7 @@ void r600_fini(struct radeon_device *rdev) | |||
| 1712 | radeon_clocks_fini(rdev); | 2084 | radeon_clocks_fini(rdev); |
| 1713 | if (rdev->flags & RADEON_IS_AGP) | 2085 | if (rdev->flags & RADEON_IS_AGP) |
| 1714 | radeon_agp_fini(rdev); | 2086 | radeon_agp_fini(rdev); |
| 1715 | radeon_object_fini(rdev); | 2087 | radeon_bo_fini(rdev); |
| 1716 | radeon_atombios_fini(rdev); | 2088 | radeon_atombios_fini(rdev); |
| 1717 | kfree(rdev->bios); | 2089 | kfree(rdev->bios); |
| 1718 | rdev->bios = NULL; | 2090 | rdev->bios = NULL; |
| @@ -1798,8 +2170,657 @@ int r600_ib_test(struct radeon_device *rdev) | |||
| 1798 | return r; | 2170 | return r; |
| 1799 | } | 2171 | } |
| 1800 | 2172 | ||
| 2173 | /* | ||
| 2174 | * Interrupts | ||
| 2175 | * | ||
| 2176 | * Interrupts use a ring buffer on r6xx/r7xx hardware. It works pretty | ||
| 2177 | * the same as the CP ring buffer, but in reverse. Rather than the CPU | ||
| 2178 | * writing to the ring and the GPU consuming, the GPU writes to the ring | ||
| 2179 | * and host consumes. As the host irq handler processes interrupts, it | ||
| 2180 | * increments the rptr. When the rptr catches up with the wptr, all the | ||
| 2181 | * current interrupts have been processed. | ||
| 2182 | */ | ||
| 2183 | |||
| 2184 | void r600_ih_ring_init(struct radeon_device *rdev, unsigned ring_size) | ||
| 2185 | { | ||
| 2186 | u32 rb_bufsz; | ||
| 2187 | |||
| 2188 | /* Align ring size */ | ||
| 2189 | rb_bufsz = drm_order(ring_size / 4); | ||
| 2190 | ring_size = (1 << rb_bufsz) * 4; | ||
| 2191 | rdev->ih.ring_size = ring_size; | ||
| 2192 | rdev->ih.align_mask = 4 - 1; | ||
| 2193 | } | ||
| 2194 | |||
| 2195 | static int r600_ih_ring_alloc(struct radeon_device *rdev, unsigned ring_size) | ||
| 2196 | { | ||
| 2197 | int r; | ||
| 2198 | |||
| 2199 | rdev->ih.ring_size = ring_size; | ||
| 2200 | /* Allocate ring buffer */ | ||
| 2201 | if (rdev->ih.ring_obj == NULL) { | ||
| 2202 | r = radeon_bo_create(rdev, NULL, rdev->ih.ring_size, | ||
| 2203 | true, | ||
| 2204 | RADEON_GEM_DOMAIN_GTT, | ||
| 2205 | &rdev->ih.ring_obj); | ||
| 2206 | if (r) { | ||
| 2207 | DRM_ERROR("radeon: failed to create ih ring buffer (%d).\n", r); | ||
| 2208 | return r; | ||
| 2209 | } | ||
| 2210 | r = radeon_bo_reserve(rdev->ih.ring_obj, false); | ||
| 2211 | if (unlikely(r != 0)) | ||
| 2212 | return r; | ||
| 2213 | r = radeon_bo_pin(rdev->ih.ring_obj, | ||
| 2214 | RADEON_GEM_DOMAIN_GTT, | ||
| 2215 | &rdev->ih.gpu_addr); | ||
| 2216 | if (r) { | ||
| 2217 | radeon_bo_unreserve(rdev->ih.ring_obj); | ||
| 2218 | DRM_ERROR("radeon: failed to pin ih ring buffer (%d).\n", r); | ||
| 2219 | return r; | ||
| 2220 | } | ||
| 2221 | r = radeon_bo_kmap(rdev->ih.ring_obj, | ||
| 2222 | (void **)&rdev->ih.ring); | ||
| 2223 | radeon_bo_unreserve(rdev->ih.ring_obj); | ||
| 2224 | if (r) { | ||
| 2225 | DRM_ERROR("radeon: failed to map ih ring buffer (%d).\n", r); | ||
| 2226 | return r; | ||
| 2227 | } | ||
| 2228 | } | ||
| 2229 | rdev->ih.ptr_mask = (rdev->cp.ring_size / 4) - 1; | ||
| 2230 | rdev->ih.rptr = 0; | ||
| 2231 | |||
| 2232 | return 0; | ||
| 2233 | } | ||
| 2234 | |||
| 2235 | static void r600_ih_ring_fini(struct radeon_device *rdev) | ||
| 2236 | { | ||
| 2237 | int r; | ||
| 2238 | if (rdev->ih.ring_obj) { | ||
| 2239 | r = radeon_bo_reserve(rdev->ih.ring_obj, false); | ||
| 2240 | if (likely(r == 0)) { | ||
| 2241 | radeon_bo_kunmap(rdev->ih.ring_obj); | ||
| 2242 | radeon_bo_unpin(rdev->ih.ring_obj); | ||
| 2243 | radeon_bo_unreserve(rdev->ih.ring_obj); | ||
| 2244 | } | ||
| 2245 | radeon_bo_unref(&rdev->ih.ring_obj); | ||
| 2246 | rdev->ih.ring = NULL; | ||
| 2247 | rdev->ih.ring_obj = NULL; | ||
| 2248 | } | ||
| 2249 | } | ||
| 2250 | |||
| 2251 | static void r600_rlc_stop(struct radeon_device *rdev) | ||
| 2252 | { | ||
| 2253 | |||
| 2254 | if (rdev->family >= CHIP_RV770) { | ||
| 2255 | /* r7xx asics need to soft reset RLC before halting */ | ||
| 2256 | WREG32(SRBM_SOFT_RESET, SOFT_RESET_RLC); | ||
| 2257 | RREG32(SRBM_SOFT_RESET); | ||
| 2258 | udelay(15000); | ||
| 2259 | WREG32(SRBM_SOFT_RESET, 0); | ||
| 2260 | RREG32(SRBM_SOFT_RESET); | ||
| 2261 | } | ||
| 2262 | |||
| 2263 | WREG32(RLC_CNTL, 0); | ||
| 2264 | } | ||
| 2265 | |||
| 2266 | static void r600_rlc_start(struct radeon_device *rdev) | ||
| 2267 | { | ||
| 2268 | WREG32(RLC_CNTL, RLC_ENABLE); | ||
| 2269 | } | ||
| 2270 | |||
| 2271 | static int r600_rlc_init(struct radeon_device *rdev) | ||
| 2272 | { | ||
| 2273 | u32 i; | ||
| 2274 | const __be32 *fw_data; | ||
| 2275 | |||
| 2276 | if (!rdev->rlc_fw) | ||
| 2277 | return -EINVAL; | ||
| 2278 | |||
| 2279 | r600_rlc_stop(rdev); | ||
| 2280 | |||
| 2281 | WREG32(RLC_HB_BASE, 0); | ||
| 2282 | WREG32(RLC_HB_CNTL, 0); | ||
| 2283 | WREG32(RLC_HB_RPTR, 0); | ||
| 2284 | WREG32(RLC_HB_WPTR, 0); | ||
| 2285 | WREG32(RLC_HB_WPTR_LSB_ADDR, 0); | ||
| 2286 | WREG32(RLC_HB_WPTR_MSB_ADDR, 0); | ||
| 2287 | WREG32(RLC_MC_CNTL, 0); | ||
| 2288 | WREG32(RLC_UCODE_CNTL, 0); | ||
| 2289 | |||
| 2290 | fw_data = (const __be32 *)rdev->rlc_fw->data; | ||
| 2291 | if (rdev->family >= CHIP_RV770) { | ||
| 2292 | for (i = 0; i < R700_RLC_UCODE_SIZE; i++) { | ||
| 2293 | WREG32(RLC_UCODE_ADDR, i); | ||
| 2294 | WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++)); | ||
| 2295 | } | ||
| 2296 | } else { | ||
| 2297 | for (i = 0; i < RLC_UCODE_SIZE; i++) { | ||
| 2298 | WREG32(RLC_UCODE_ADDR, i); | ||
| 2299 | WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++)); | ||
| 2300 | } | ||
| 2301 | } | ||
| 2302 | WREG32(RLC_UCODE_ADDR, 0); | ||
| 2303 | |||
| 2304 | r600_rlc_start(rdev); | ||
| 2305 | |||
| 2306 | return 0; | ||
| 2307 | } | ||
| 2308 | |||
| 2309 | static void r600_enable_interrupts(struct radeon_device *rdev) | ||
| 2310 | { | ||
| 2311 | u32 ih_cntl = RREG32(IH_CNTL); | ||
| 2312 | u32 ih_rb_cntl = RREG32(IH_RB_CNTL); | ||
| 2313 | |||
| 2314 | ih_cntl |= ENABLE_INTR; | ||
| 2315 | ih_rb_cntl |= IH_RB_ENABLE; | ||
| 2316 | WREG32(IH_CNTL, ih_cntl); | ||
| 2317 | WREG32(IH_RB_CNTL, ih_rb_cntl); | ||
| 2318 | rdev->ih.enabled = true; | ||
| 2319 | } | ||
| 2320 | |||
| 2321 | static void r600_disable_interrupts(struct radeon_device *rdev) | ||
| 2322 | { | ||
| 2323 | u32 ih_rb_cntl = RREG32(IH_RB_CNTL); | ||
| 2324 | u32 ih_cntl = RREG32(IH_CNTL); | ||
| 2325 | |||
| 2326 | ih_rb_cntl &= ~IH_RB_ENABLE; | ||
| 2327 | ih_cntl &= ~ENABLE_INTR; | ||
| 2328 | WREG32(IH_RB_CNTL, ih_rb_cntl); | ||
| 2329 | WREG32(IH_CNTL, ih_cntl); | ||
| 2330 | /* set rptr, wptr to 0 */ | ||
| 2331 | WREG32(IH_RB_RPTR, 0); | ||
| 2332 | WREG32(IH_RB_WPTR, 0); | ||
| 2333 | rdev->ih.enabled = false; | ||
| 2334 | rdev->ih.wptr = 0; | ||
| 2335 | rdev->ih.rptr = 0; | ||
| 2336 | } | ||
| 2337 | |||
| 2338 | static void r600_disable_interrupt_state(struct radeon_device *rdev) | ||
| 2339 | { | ||
| 2340 | u32 tmp; | ||
| 2341 | |||
| 2342 | WREG32(CP_INT_CNTL, 0); | ||
| 2343 | WREG32(GRBM_INT_CNTL, 0); | ||
| 2344 | WREG32(DxMODE_INT_MASK, 0); | ||
| 2345 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2346 | WREG32(DCE3_DACA_AUTODETECT_INT_CONTROL, 0); | ||
| 2347 | WREG32(DCE3_DACB_AUTODETECT_INT_CONTROL, 0); | ||
| 2348 | tmp = RREG32(DC_HPD1_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2349 | WREG32(DC_HPD1_INT_CONTROL, tmp); | ||
| 2350 | tmp = RREG32(DC_HPD2_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2351 | WREG32(DC_HPD2_INT_CONTROL, tmp); | ||
| 2352 | tmp = RREG32(DC_HPD3_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2353 | WREG32(DC_HPD3_INT_CONTROL, tmp); | ||
| 2354 | tmp = RREG32(DC_HPD4_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2355 | WREG32(DC_HPD4_INT_CONTROL, tmp); | ||
| 2356 | if (ASIC_IS_DCE32(rdev)) { | ||
| 2357 | tmp = RREG32(DC_HPD5_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2358 | WREG32(DC_HPD5_INT_CONTROL, 0); | ||
| 2359 | tmp = RREG32(DC_HPD6_INT_CONTROL) & DC_HPDx_INT_POLARITY; | ||
| 2360 | WREG32(DC_HPD6_INT_CONTROL, 0); | ||
| 2361 | } | ||
| 2362 | } else { | ||
| 2363 | WREG32(DACA_AUTODETECT_INT_CONTROL, 0); | ||
| 2364 | WREG32(DACB_AUTODETECT_INT_CONTROL, 0); | ||
| 2365 | tmp = RREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL) & DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 2366 | WREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL, 0); | ||
| 2367 | tmp = RREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL) & DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 2368 | WREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL, 0); | ||
| 2369 | tmp = RREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL) & DC_HOT_PLUG_DETECTx_INT_POLARITY; | ||
| 2370 | WREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL, 0); | ||
| 2371 | } | ||
| 2372 | } | ||
| 2373 | |||
| 2374 | int r600_irq_init(struct radeon_device *rdev) | ||
| 2375 | { | ||
| 2376 | int ret = 0; | ||
| 2377 | int rb_bufsz; | ||
| 2378 | u32 interrupt_cntl, ih_cntl, ih_rb_cntl; | ||
| 2379 | |||
| 2380 | /* allocate ring */ | ||
| 2381 | ret = r600_ih_ring_alloc(rdev, rdev->ih.ring_size); | ||
| 2382 | if (ret) | ||
| 2383 | return ret; | ||
| 2384 | |||
| 2385 | /* disable irqs */ | ||
| 2386 | r600_disable_interrupts(rdev); | ||
| 2387 | |||
| 2388 | /* init rlc */ | ||
| 2389 | ret = r600_rlc_init(rdev); | ||
| 2390 | if (ret) { | ||
| 2391 | r600_ih_ring_fini(rdev); | ||
| 2392 | return ret; | ||
| 2393 | } | ||
| 2394 | |||
| 2395 | /* setup interrupt control */ | ||
| 2396 | /* set dummy read address to ring address */ | ||
| 2397 | WREG32(INTERRUPT_CNTL2, rdev->ih.gpu_addr >> 8); | ||
| 2398 | interrupt_cntl = RREG32(INTERRUPT_CNTL); | ||
| 2399 | /* IH_DUMMY_RD_OVERRIDE=0 - dummy read disabled with msi, enabled without msi | ||
| 2400 | * IH_DUMMY_RD_OVERRIDE=1 - dummy read controlled by IH_DUMMY_RD_EN | ||
| 2401 | */ | ||
| 2402 | interrupt_cntl &= ~IH_DUMMY_RD_OVERRIDE; | ||
| 2403 | /* IH_REQ_NONSNOOP_EN=1 if ring is in non-cacheable memory, e.g., vram */ | ||
| 2404 | interrupt_cntl &= ~IH_REQ_NONSNOOP_EN; | ||
| 2405 | WREG32(INTERRUPT_CNTL, interrupt_cntl); | ||
| 2406 | |||
| 2407 | WREG32(IH_RB_BASE, rdev->ih.gpu_addr >> 8); | ||
| 2408 | rb_bufsz = drm_order(rdev->ih.ring_size / 4); | ||
| 2409 | |||
| 2410 | ih_rb_cntl = (IH_WPTR_OVERFLOW_ENABLE | | ||
| 2411 | IH_WPTR_OVERFLOW_CLEAR | | ||
| 2412 | (rb_bufsz << 1)); | ||
| 2413 | /* WPTR writeback, not yet */ | ||
| 2414 | /*ih_rb_cntl |= IH_WPTR_WRITEBACK_ENABLE;*/ | ||
| 2415 | WREG32(IH_RB_WPTR_ADDR_LO, 0); | ||
| 2416 | WREG32(IH_RB_WPTR_ADDR_HI, 0); | ||
| 2417 | |||
| 2418 | WREG32(IH_RB_CNTL, ih_rb_cntl); | ||
| 2419 | |||
| 2420 | /* set rptr, wptr to 0 */ | ||
| 2421 | WREG32(IH_RB_RPTR, 0); | ||
| 2422 | WREG32(IH_RB_WPTR, 0); | ||
| 2423 | |||
| 2424 | /* Default settings for IH_CNTL (disabled at first) */ | ||
| 2425 | ih_cntl = MC_WRREQ_CREDIT(0x10) | MC_WR_CLEAN_CNT(0x10); | ||
| 2426 | /* RPTR_REARM only works if msi's are enabled */ | ||
| 2427 | if (rdev->msi_enabled) | ||
| 2428 | ih_cntl |= RPTR_REARM; | ||
| 2429 | |||
| 2430 | #ifdef __BIG_ENDIAN | ||
| 2431 | ih_cntl |= IH_MC_SWAP(IH_MC_SWAP_32BIT); | ||
| 2432 | #endif | ||
| 2433 | WREG32(IH_CNTL, ih_cntl); | ||
| 2434 | |||
| 2435 | /* force the active interrupt state to all disabled */ | ||
| 2436 | r600_disable_interrupt_state(rdev); | ||
| 2437 | |||
| 2438 | /* enable irqs */ | ||
| 2439 | r600_enable_interrupts(rdev); | ||
| 2440 | |||
| 2441 | return ret; | ||
| 2442 | } | ||
| 2443 | |||
| 2444 | void r600_irq_fini(struct radeon_device *rdev) | ||
| 2445 | { | ||
| 2446 | r600_disable_interrupts(rdev); | ||
| 2447 | r600_rlc_stop(rdev); | ||
| 2448 | r600_ih_ring_fini(rdev); | ||
| 2449 | } | ||
| 2450 | |||
| 2451 | int r600_irq_set(struct radeon_device *rdev) | ||
| 2452 | { | ||
| 2453 | u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE; | ||
| 2454 | u32 mode_int = 0; | ||
| 2455 | u32 hpd1, hpd2, hpd3, hpd4 = 0, hpd5 = 0, hpd6 = 0; | ||
| 2456 | |||
| 2457 | /* don't enable anything if the ih is disabled */ | ||
| 2458 | if (!rdev->ih.enabled) | ||
| 2459 | return 0; | ||
| 2460 | |||
| 2461 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2462 | hpd1 = RREG32(DC_HPD1_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2463 | hpd2 = RREG32(DC_HPD2_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2464 | hpd3 = RREG32(DC_HPD3_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2465 | hpd4 = RREG32(DC_HPD4_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2466 | if (ASIC_IS_DCE32(rdev)) { | ||
| 2467 | hpd5 = RREG32(DC_HPD5_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2468 | hpd6 = RREG32(DC_HPD6_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2469 | } | ||
| 2470 | } else { | ||
| 2471 | hpd1 = RREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2472 | hpd2 = RREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2473 | hpd3 = RREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL) & ~DC_HPDx_INT_EN; | ||
| 2474 | } | ||
| 2475 | |||
| 2476 | if (rdev->irq.sw_int) { | ||
| 2477 | DRM_DEBUG("r600_irq_set: sw int\n"); | ||
| 2478 | cp_int_cntl |= RB_INT_ENABLE; | ||
| 2479 | } | ||
| 2480 | if (rdev->irq.crtc_vblank_int[0]) { | ||
| 2481 | DRM_DEBUG("r600_irq_set: vblank 0\n"); | ||
| 2482 | mode_int |= D1MODE_VBLANK_INT_MASK; | ||
| 2483 | } | ||
| 2484 | if (rdev->irq.crtc_vblank_int[1]) { | ||
| 2485 | DRM_DEBUG("r600_irq_set: vblank 1\n"); | ||
| 2486 | mode_int |= D2MODE_VBLANK_INT_MASK; | ||
| 2487 | } | ||
| 2488 | if (rdev->irq.hpd[0]) { | ||
| 2489 | DRM_DEBUG("r600_irq_set: hpd 1\n"); | ||
| 2490 | hpd1 |= DC_HPDx_INT_EN; | ||
| 2491 | } | ||
| 2492 | if (rdev->irq.hpd[1]) { | ||
| 2493 | DRM_DEBUG("r600_irq_set: hpd 2\n"); | ||
| 2494 | hpd2 |= DC_HPDx_INT_EN; | ||
| 2495 | } | ||
| 2496 | if (rdev->irq.hpd[2]) { | ||
| 2497 | DRM_DEBUG("r600_irq_set: hpd 3\n"); | ||
| 2498 | hpd3 |= DC_HPDx_INT_EN; | ||
| 2499 | } | ||
| 2500 | if (rdev->irq.hpd[3]) { | ||
| 2501 | DRM_DEBUG("r600_irq_set: hpd 4\n"); | ||
| 2502 | hpd4 |= DC_HPDx_INT_EN; | ||
| 2503 | } | ||
| 2504 | if (rdev->irq.hpd[4]) { | ||
| 2505 | DRM_DEBUG("r600_irq_set: hpd 5\n"); | ||
| 2506 | hpd5 |= DC_HPDx_INT_EN; | ||
| 2507 | } | ||
| 2508 | if (rdev->irq.hpd[5]) { | ||
| 2509 | DRM_DEBUG("r600_irq_set: hpd 6\n"); | ||
| 2510 | hpd6 |= DC_HPDx_INT_EN; | ||
| 2511 | } | ||
| 2512 | |||
| 2513 | WREG32(CP_INT_CNTL, cp_int_cntl); | ||
| 2514 | WREG32(DxMODE_INT_MASK, mode_int); | ||
| 2515 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2516 | WREG32(DC_HPD1_INT_CONTROL, hpd1); | ||
| 2517 | WREG32(DC_HPD2_INT_CONTROL, hpd2); | ||
| 2518 | WREG32(DC_HPD3_INT_CONTROL, hpd3); | ||
| 2519 | WREG32(DC_HPD4_INT_CONTROL, hpd4); | ||
| 2520 | if (ASIC_IS_DCE32(rdev)) { | ||
| 2521 | WREG32(DC_HPD5_INT_CONTROL, hpd5); | ||
| 2522 | WREG32(DC_HPD6_INT_CONTROL, hpd6); | ||
| 2523 | } | ||
| 2524 | } else { | ||
| 2525 | WREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL, hpd1); | ||
| 2526 | WREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL, hpd2); | ||
| 2527 | WREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL, hpd3); | ||
| 2528 | } | ||
| 2529 | |||
| 2530 | return 0; | ||
| 2531 | } | ||
| 2532 | |||
| 2533 | static inline void r600_irq_ack(struct radeon_device *rdev, | ||
| 2534 | u32 *disp_int, | ||
| 2535 | u32 *disp_int_cont, | ||
| 2536 | u32 *disp_int_cont2) | ||
| 2537 | { | ||
| 2538 | u32 tmp; | ||
| 2539 | |||
| 2540 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2541 | *disp_int = RREG32(DCE3_DISP_INTERRUPT_STATUS); | ||
| 2542 | *disp_int_cont = RREG32(DCE3_DISP_INTERRUPT_STATUS_CONTINUE); | ||
| 2543 | *disp_int_cont2 = RREG32(DCE3_DISP_INTERRUPT_STATUS_CONTINUE2); | ||
| 2544 | } else { | ||
| 2545 | *disp_int = RREG32(DISP_INTERRUPT_STATUS); | ||
| 2546 | *disp_int_cont = RREG32(DISP_INTERRUPT_STATUS_CONTINUE); | ||
| 2547 | *disp_int_cont2 = 0; | ||
| 2548 | } | ||
| 2549 | |||
| 2550 | if (*disp_int & LB_D1_VBLANK_INTERRUPT) | ||
| 2551 | WREG32(D1MODE_VBLANK_STATUS, DxMODE_VBLANK_ACK); | ||
| 2552 | if (*disp_int & LB_D1_VLINE_INTERRUPT) | ||
| 2553 | WREG32(D1MODE_VLINE_STATUS, DxMODE_VLINE_ACK); | ||
| 2554 | if (*disp_int & LB_D2_VBLANK_INTERRUPT) | ||
| 2555 | WREG32(D2MODE_VBLANK_STATUS, DxMODE_VBLANK_ACK); | ||
| 2556 | if (*disp_int & LB_D2_VLINE_INTERRUPT) | ||
| 2557 | WREG32(D2MODE_VLINE_STATUS, DxMODE_VLINE_ACK); | ||
| 2558 | if (*disp_int & DC_HPD1_INTERRUPT) { | ||
| 2559 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2560 | tmp = RREG32(DC_HPD1_INT_CONTROL); | ||
| 2561 | tmp |= DC_HPDx_INT_ACK; | ||
| 2562 | WREG32(DC_HPD1_INT_CONTROL, tmp); | ||
| 2563 | } else { | ||
| 2564 | tmp = RREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL); | ||
| 2565 | tmp |= DC_HPDx_INT_ACK; | ||
| 2566 | WREG32(DC_HOT_PLUG_DETECT1_INT_CONTROL, tmp); | ||
| 2567 | } | ||
| 2568 | } | ||
| 2569 | if (*disp_int & DC_HPD2_INTERRUPT) { | ||
| 2570 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2571 | tmp = RREG32(DC_HPD2_INT_CONTROL); | ||
| 2572 | tmp |= DC_HPDx_INT_ACK; | ||
| 2573 | WREG32(DC_HPD2_INT_CONTROL, tmp); | ||
| 2574 | } else { | ||
| 2575 | tmp = RREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL); | ||
| 2576 | tmp |= DC_HPDx_INT_ACK; | ||
| 2577 | WREG32(DC_HOT_PLUG_DETECT2_INT_CONTROL, tmp); | ||
| 2578 | } | ||
| 2579 | } | ||
| 2580 | if (*disp_int_cont & DC_HPD3_INTERRUPT) { | ||
| 2581 | if (ASIC_IS_DCE3(rdev)) { | ||
| 2582 | tmp = RREG32(DC_HPD3_INT_CONTROL); | ||
| 2583 | tmp |= DC_HPDx_INT_ACK; | ||
| 2584 | WREG32(DC_HPD3_INT_CONTROL, tmp); | ||
| 2585 | } else { | ||
| 2586 | tmp = RREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL); | ||
| 2587 | tmp |= DC_HPDx_INT_ACK; | ||
| 2588 | WREG32(DC_HOT_PLUG_DETECT3_INT_CONTROL, tmp); | ||
| 2589 | } | ||
| 2590 | } | ||
| 2591 | if (*disp_int_cont & DC_HPD4_INTERRUPT) { | ||
| 2592 | tmp = RREG32(DC_HPD4_INT_CONTROL); | ||
| 2593 | tmp |= DC_HPDx_INT_ACK; | ||
| 2594 | WREG32(DC_HPD4_INT_CONTROL, tmp); | ||
| 2595 | } | ||
| 2596 | if (ASIC_IS_DCE32(rdev)) { | ||
| 2597 | if (*disp_int_cont2 & DC_HPD5_INTERRUPT) { | ||
| 2598 | tmp = RREG32(DC_HPD5_INT_CONTROL); | ||
| 2599 | tmp |= DC_HPDx_INT_ACK; | ||
| 2600 | WREG32(DC_HPD5_INT_CONTROL, tmp); | ||
| 2601 | } | ||
| 2602 | if (*disp_int_cont2 & DC_HPD6_INTERRUPT) { | ||
| 2603 | tmp = RREG32(DC_HPD5_INT_CONTROL); | ||
| 2604 | tmp |= DC_HPDx_INT_ACK; | ||
| 2605 | WREG32(DC_HPD6_INT_CONTROL, tmp); | ||
| 2606 | } | ||
| 2607 | } | ||
| 2608 | } | ||
| 2609 | |||
| 2610 | void r600_irq_disable(struct radeon_device *rdev) | ||
| 2611 | { | ||
| 2612 | u32 disp_int, disp_int_cont, disp_int_cont2; | ||
| 2613 | |||
| 2614 | r600_disable_interrupts(rdev); | ||
| 2615 | /* Wait and acknowledge irq */ | ||
| 2616 | mdelay(1); | ||
| 2617 | r600_irq_ack(rdev, &disp_int, &disp_int_cont, &disp_int_cont2); | ||
| 2618 | r600_disable_interrupt_state(rdev); | ||
| 2619 | } | ||
| 2620 | |||
| 2621 | static inline u32 r600_get_ih_wptr(struct radeon_device *rdev) | ||
| 2622 | { | ||
| 2623 | u32 wptr, tmp; | ||
| 1801 | 2624 | ||
| 2625 | /* XXX use writeback */ | ||
| 2626 | wptr = RREG32(IH_RB_WPTR); | ||
| 1802 | 2627 | ||
| 2628 | if (wptr & RB_OVERFLOW) { | ||
| 2629 | WARN_ON(1); | ||
| 2630 | /* XXX deal with overflow */ | ||
| 2631 | DRM_ERROR("IH RB overflow\n"); | ||
| 2632 | tmp = RREG32(IH_RB_CNTL); | ||
| 2633 | tmp |= IH_WPTR_OVERFLOW_CLEAR; | ||
| 2634 | WREG32(IH_RB_CNTL, tmp); | ||
| 2635 | } | ||
| 2636 | wptr = wptr & WPTR_OFFSET_MASK; | ||
| 2637 | |||
| 2638 | return wptr; | ||
| 2639 | } | ||
| 2640 | |||
| 2641 | /* r600 IV Ring | ||
| 2642 | * Each IV ring entry is 128 bits: | ||
| 2643 | * [7:0] - interrupt source id | ||
| 2644 | * [31:8] - reserved | ||
| 2645 | * [59:32] - interrupt source data | ||
| 2646 | * [127:60] - reserved | ||
| 2647 | * | ||
| 2648 | * The basic interrupt vector entries | ||
| 2649 | * are decoded as follows: | ||
| 2650 | * src_id src_data description | ||
| 2651 | * 1 0 D1 Vblank | ||
| 2652 | * 1 1 D1 Vline | ||
| 2653 | * 5 0 D2 Vblank | ||
| 2654 | * 5 1 D2 Vline | ||
| 2655 | * 19 0 FP Hot plug detection A | ||
| 2656 | * 19 1 FP Hot plug detection B | ||
| 2657 | * 19 2 DAC A auto-detection | ||
| 2658 | * 19 3 DAC B auto-detection | ||
| 2659 | * 176 - CP_INT RB | ||
| 2660 | * 177 - CP_INT IB1 | ||
| 2661 | * 178 - CP_INT IB2 | ||
| 2662 | * 181 - EOP Interrupt | ||
| 2663 | * 233 - GUI Idle | ||
| 2664 | * | ||
| 2665 | * Note, these are based on r600 and may need to be | ||
| 2666 | * adjusted or added to on newer asics | ||
| 2667 | */ | ||
| 2668 | |||
| 2669 | int r600_irq_process(struct radeon_device *rdev) | ||
| 2670 | { | ||
| 2671 | u32 wptr = r600_get_ih_wptr(rdev); | ||
| 2672 | u32 rptr = rdev->ih.rptr; | ||
| 2673 | u32 src_id, src_data; | ||
| 2674 | u32 last_entry = rdev->ih.ring_size - 16; | ||
| 2675 | u32 ring_index, disp_int, disp_int_cont, disp_int_cont2; | ||
| 2676 | unsigned long flags; | ||
| 2677 | bool queue_hotplug = false; | ||
| 2678 | |||
| 2679 | DRM_DEBUG("r600_irq_process start: rptr %d, wptr %d\n", rptr, wptr); | ||
| 2680 | |||
| 2681 | spin_lock_irqsave(&rdev->ih.lock, flags); | ||
| 2682 | |||
| 2683 | if (rptr == wptr) { | ||
| 2684 | spin_unlock_irqrestore(&rdev->ih.lock, flags); | ||
| 2685 | return IRQ_NONE; | ||
| 2686 | } | ||
| 2687 | if (rdev->shutdown) { | ||
| 2688 | spin_unlock_irqrestore(&rdev->ih.lock, flags); | ||
| 2689 | return IRQ_NONE; | ||
| 2690 | } | ||
| 2691 | |||
| 2692 | restart_ih: | ||
| 2693 | /* display interrupts */ | ||
| 2694 | r600_irq_ack(rdev, &disp_int, &disp_int_cont, &disp_int_cont2); | ||
| 2695 | |||
| 2696 | rdev->ih.wptr = wptr; | ||
| 2697 | while (rptr != wptr) { | ||
| 2698 | /* wptr/rptr are in bytes! */ | ||
| 2699 | ring_index = rptr / 4; | ||
| 2700 | src_id = rdev->ih.ring[ring_index] & 0xff; | ||
| 2701 | src_data = rdev->ih.ring[ring_index + 1] & 0xfffffff; | ||
| 2702 | |||
| 2703 | switch (src_id) { | ||
| 2704 | case 1: /* D1 vblank/vline */ | ||
| 2705 | switch (src_data) { | ||
| 2706 | case 0: /* D1 vblank */ | ||
| 2707 | if (disp_int & LB_D1_VBLANK_INTERRUPT) { | ||
| 2708 | drm_handle_vblank(rdev->ddev, 0); | ||
| 2709 | disp_int &= ~LB_D1_VBLANK_INTERRUPT; | ||
| 2710 | DRM_DEBUG("IH: D1 vblank\n"); | ||
| 2711 | } | ||
| 2712 | break; | ||
| 2713 | case 1: /* D1 vline */ | ||
| 2714 | if (disp_int & LB_D1_VLINE_INTERRUPT) { | ||
| 2715 | disp_int &= ~LB_D1_VLINE_INTERRUPT; | ||
| 2716 | DRM_DEBUG("IH: D1 vline\n"); | ||
| 2717 | } | ||
| 2718 | break; | ||
| 2719 | default: | ||
| 2720 | DRM_ERROR("Unhandled interrupt: %d %d\n", src_id, src_data); | ||
| 2721 | break; | ||
| 2722 | } | ||
| 2723 | break; | ||
| 2724 | case 5: /* D2 vblank/vline */ | ||
| 2725 | switch (src_data) { | ||
| 2726 | case 0: /* D2 vblank */ | ||
| 2727 | if (disp_int & LB_D2_VBLANK_INTERRUPT) { | ||
| 2728 | drm_handle_vblank(rdev->ddev, 1); | ||
| 2729 | disp_int &= ~LB_D2_VBLANK_INTERRUPT; | ||
| 2730 | DRM_DEBUG("IH: D2 vblank\n"); | ||
| 2731 | } | ||
| 2732 | break; | ||
| 2733 | case 1: /* D1 vline */ | ||
| 2734 | if (disp_int & LB_D2_VLINE_INTERRUPT) { | ||
| 2735 | disp_int &= ~LB_D2_VLINE_INTERRUPT; | ||
| 2736 | DRM_DEBUG("IH: D2 vline\n"); | ||
| 2737 | } | ||
| 2738 | break; | ||
| 2739 | default: | ||
| 2740 | DRM_ERROR("Unhandled interrupt: %d %d\n", src_id, src_data); | ||
| 2741 | break; | ||
| 2742 | } | ||
| 2743 | break; | ||
| 2744 | case 19: /* HPD/DAC hotplug */ | ||
| 2745 | switch (src_data) { | ||
| 2746 | case 0: | ||
| 2747 | if (disp_int & DC_HPD1_INTERRUPT) { | ||
| 2748 | disp_int &= ~DC_HPD1_INTERRUPT; | ||
| 2749 | queue_hotplug = true; | ||
| 2750 | DRM_DEBUG("IH: HPD1\n"); | ||
| 2751 | } | ||
| 2752 | break; | ||
| 2753 | case 1: | ||
| 2754 | if (disp_int & DC_HPD2_INTERRUPT) { | ||
| 2755 | disp_int &= ~DC_HPD2_INTERRUPT; | ||
| 2756 | queue_hotplug = true; | ||
| 2757 | DRM_DEBUG("IH: HPD2\n"); | ||
| 2758 | } | ||
| 2759 | break; | ||
| 2760 | case 4: | ||
| 2761 | if (disp_int_cont & DC_HPD3_INTERRUPT) { | ||
| 2762 | disp_int_cont &= ~DC_HPD3_INTERRUPT; | ||
| 2763 | queue_hotplug = true; | ||
| 2764 | DRM_DEBUG("IH: HPD3\n"); | ||
| 2765 | } | ||
| 2766 | break; | ||
| 2767 | case 5: | ||
| 2768 | if (disp_int_cont & DC_HPD4_INTERRUPT) { | ||
| 2769 | disp_int_cont &= ~DC_HPD4_INTERRUPT; | ||
| 2770 | queue_hotplug = true; | ||
| 2771 | DRM_DEBUG("IH: HPD4\n"); | ||
| 2772 | } | ||
| 2773 | break; | ||
| 2774 | case 10: | ||
| 2775 | if (disp_int_cont2 & DC_HPD5_INTERRUPT) { | ||
| 2776 | disp_int_cont &= ~DC_HPD5_INTERRUPT; | ||
| 2777 | queue_hotplug = true; | ||
| 2778 | DRM_DEBUG("IH: HPD5\n"); | ||
| 2779 | } | ||
| 2780 | break; | ||
| 2781 | case 12: | ||
| 2782 | if (disp_int_cont2 & DC_HPD6_INTERRUPT) { | ||
| 2783 | disp_int_cont &= ~DC_HPD6_INTERRUPT; | ||
| 2784 | queue_hotplug = true; | ||
| 2785 | DRM_DEBUG("IH: HPD6\n"); | ||
| 2786 | } | ||
| 2787 | break; | ||
| 2788 | default: | ||
| 2789 | DRM_ERROR("Unhandled interrupt: %d %d\n", src_id, src_data); | ||
| 2790 | break; | ||
| 2791 | } | ||
| 2792 | break; | ||
| 2793 | case 176: /* CP_INT in ring buffer */ | ||
| 2794 | case 177: /* CP_INT in IB1 */ | ||
| 2795 | case 178: /* CP_INT in IB2 */ | ||
| 2796 | DRM_DEBUG("IH: CP int: 0x%08x\n", src_data); | ||
| 2797 | radeon_fence_process(rdev); | ||
| 2798 | break; | ||
| 2799 | case 181: /* CP EOP event */ | ||
| 2800 | DRM_DEBUG("IH: CP EOP\n"); | ||
| 2801 | break; | ||
| 2802 | default: | ||
| 2803 | DRM_ERROR("Unhandled interrupt: %d %d\n", src_id, src_data); | ||
| 2804 | break; | ||
| 2805 | } | ||
| 2806 | |||
| 2807 | /* wptr/rptr are in bytes! */ | ||
| 2808 | if (rptr == last_entry) | ||
| 2809 | rptr = 0; | ||
| 2810 | else | ||
| 2811 | rptr += 16; | ||
| 2812 | } | ||
| 2813 | /* make sure wptr hasn't changed while processing */ | ||
| 2814 | wptr = r600_get_ih_wptr(rdev); | ||
| 2815 | if (wptr != rdev->ih.wptr) | ||
| 2816 | goto restart_ih; | ||
| 2817 | if (queue_hotplug) | ||
| 2818 | queue_work(rdev->wq, &rdev->hotplug_work); | ||
| 2819 | rdev->ih.rptr = rptr; | ||
| 2820 | WREG32(IH_RB_RPTR, rdev->ih.rptr); | ||
| 2821 | spin_unlock_irqrestore(&rdev->ih.lock, flags); | ||
| 2822 | return IRQ_HANDLED; | ||
| 2823 | } | ||
| 1803 | 2824 | ||
| 1804 | /* | 2825 | /* |
| 1805 | * Debugfs info | 2826 | * Debugfs info |
| @@ -1811,21 +2832,21 @@ static int r600_debugfs_cp_ring_info(struct seq_file *m, void *data) | |||
| 1811 | struct drm_info_node *node = (struct drm_info_node *) m->private; | 2832 | struct drm_info_node *node = (struct drm_info_node *) m->private; |
| 1812 | struct drm_device *dev = node->minor->dev; | 2833 | struct drm_device *dev = node->minor->dev; |
| 1813 | struct radeon_device *rdev = dev->dev_private; | 2834 | struct radeon_device *rdev = dev->dev_private; |
| 1814 | uint32_t rdp, wdp; | ||
| 1815 | unsigned count, i, j; | 2835 | unsigned count, i, j; |
| 1816 | 2836 | ||
| 1817 | radeon_ring_free_size(rdev); | 2837 | radeon_ring_free_size(rdev); |
| 1818 | rdp = RREG32(CP_RB_RPTR); | 2838 | count = (rdev->cp.ring_size / 4) - rdev->cp.ring_free_dw; |
| 1819 | wdp = RREG32(CP_RB_WPTR); | ||
| 1820 | count = (rdp + rdev->cp.ring_size - wdp) & rdev->cp.ptr_mask; | ||
| 1821 | seq_printf(m, "CP_STAT 0x%08x\n", RREG32(CP_STAT)); | 2839 | seq_printf(m, "CP_STAT 0x%08x\n", RREG32(CP_STAT)); |
| 1822 | seq_printf(m, "CP_RB_WPTR 0x%08x\n", wdp); | 2840 | seq_printf(m, "CP_RB_WPTR 0x%08x\n", RREG32(CP_RB_WPTR)); |
| 1823 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", rdp); | 2841 | seq_printf(m, "CP_RB_RPTR 0x%08x\n", RREG32(CP_RB_RPTR)); |
| 2842 | seq_printf(m, "driver's copy of the CP_RB_WPTR 0x%08x\n", rdev->cp.wptr); | ||
| 2843 | seq_printf(m, "driver's copy of the CP_RB_RPTR 0x%08x\n", rdev->cp.rptr); | ||
| 1824 | seq_printf(m, "%u free dwords in ring\n", rdev->cp.ring_free_dw); | 2844 | seq_printf(m, "%u free dwords in ring\n", rdev->cp.ring_free_dw); |
| 1825 | seq_printf(m, "%u dwords in ring\n", count); | 2845 | seq_printf(m, "%u dwords in ring\n", count); |
| 2846 | i = rdev->cp.rptr; | ||
| 1826 | for (j = 0; j <= count; j++) { | 2847 | for (j = 0; j <= count; j++) { |
| 1827 | i = (rdp + j) & rdev->cp.ptr_mask; | ||
| 1828 | seq_printf(m, "r[%04d]=0x%08x\n", i, rdev->cp.ring[i]); | 2848 | seq_printf(m, "r[%04d]=0x%08x\n", i, rdev->cp.ring[i]); |
| 2849 | i = (i + 1) & rdev->cp.ptr_mask; | ||
| 1829 | } | 2850 | } |
| 1830 | return 0; | 2851 | return 0; |
| 1831 | } | 2852 | } |
diff --git a/drivers/gpu/drm/radeon/r600_blit_kms.c b/drivers/gpu/drm/radeon/r600_blit_kms.c index dbf716e1fbf3..9aecafb51b66 100644 --- a/drivers/gpu/drm/radeon/r600_blit_kms.c +++ b/drivers/gpu/drm/radeon/r600_blit_kms.c | |||
| @@ -473,9 +473,8 @@ int r600_blit_init(struct radeon_device *rdev) | |||
| 473 | obj_size += r6xx_ps_size * 4; | 473 | obj_size += r6xx_ps_size * 4; |
| 474 | obj_size = ALIGN(obj_size, 256); | 474 | obj_size = ALIGN(obj_size, 256); |
| 475 | 475 | ||
| 476 | r = radeon_object_create(rdev, NULL, obj_size, | 476 | r = radeon_bo_create(rdev, NULL, obj_size, true, RADEON_GEM_DOMAIN_VRAM, |
| 477 | true, RADEON_GEM_DOMAIN_VRAM, | 477 | &rdev->r600_blit.shader_obj); |
| 478 | false, &rdev->r600_blit.shader_obj); | ||
| 479 | if (r) { | 478 | if (r) { |
| 480 | DRM_ERROR("r600 failed to allocate shader\n"); | 479 | DRM_ERROR("r600 failed to allocate shader\n"); |
| 481 | return r; | 480 | return r; |
| @@ -485,12 +484,14 @@ int r600_blit_init(struct radeon_device *rdev) | |||
| 485 | obj_size, | 484 | obj_size, |
| 486 | rdev->r600_blit.vs_offset, rdev->r600_blit.ps_offset); | 485 | rdev->r600_blit.vs_offset, rdev->r600_blit.ps_offset); |
| 487 | 486 | ||
| 488 | r = radeon_object_kmap(rdev->r600_blit.shader_obj, &ptr); | 487 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); |
| 488 | if (unlikely(r != 0)) | ||
| 489 | return r; | ||
| 490 | r = radeon_bo_kmap(rdev->r600_blit.shader_obj, &ptr); | ||
| 489 | if (r) { | 491 | if (r) { |
| 490 | DRM_ERROR("failed to map blit object %d\n", r); | 492 | DRM_ERROR("failed to map blit object %d\n", r); |
| 491 | return r; | 493 | return r; |
| 492 | } | 494 | } |
| 493 | |||
| 494 | if (rdev->family >= CHIP_RV770) | 495 | if (rdev->family >= CHIP_RV770) |
| 495 | memcpy_toio(ptr + rdev->r600_blit.state_offset, | 496 | memcpy_toio(ptr + rdev->r600_blit.state_offset, |
| 496 | r7xx_default_state, rdev->r600_blit.state_len * 4); | 497 | r7xx_default_state, rdev->r600_blit.state_len * 4); |
| @@ -500,19 +501,26 @@ int r600_blit_init(struct radeon_device *rdev) | |||
| 500 | if (num_packet2s) | 501 | if (num_packet2s) |
| 501 | memcpy_toio(ptr + rdev->r600_blit.state_offset + (rdev->r600_blit.state_len * 4), | 502 | memcpy_toio(ptr + rdev->r600_blit.state_offset + (rdev->r600_blit.state_len * 4), |
| 502 | packet2s, num_packet2s * 4); | 503 | packet2s, num_packet2s * 4); |
| 503 | |||
| 504 | |||
| 505 | memcpy(ptr + rdev->r600_blit.vs_offset, r6xx_vs, r6xx_vs_size * 4); | 504 | memcpy(ptr + rdev->r600_blit.vs_offset, r6xx_vs, r6xx_vs_size * 4); |
| 506 | memcpy(ptr + rdev->r600_blit.ps_offset, r6xx_ps, r6xx_ps_size * 4); | 505 | memcpy(ptr + rdev->r600_blit.ps_offset, r6xx_ps, r6xx_ps_size * 4); |
| 507 | 506 | radeon_bo_kunmap(rdev->r600_blit.shader_obj); | |
| 508 | radeon_object_kunmap(rdev->r600_blit.shader_obj); | 507 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); |
| 509 | return 0; | 508 | return 0; |
| 510 | } | 509 | } |
| 511 | 510 | ||
| 512 | void r600_blit_fini(struct radeon_device *rdev) | 511 | void r600_blit_fini(struct radeon_device *rdev) |
| 513 | { | 512 | { |
| 514 | radeon_object_unpin(rdev->r600_blit.shader_obj); | 513 | int r; |
| 515 | radeon_object_unref(&rdev->r600_blit.shader_obj); | 514 | |
| 515 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); | ||
| 516 | if (unlikely(r != 0)) { | ||
| 517 | dev_err(rdev->dev, "(%d) can't finish r600 blit\n", r); | ||
| 518 | goto out_unref; | ||
| 519 | } | ||
| 520 | radeon_bo_unpin(rdev->r600_blit.shader_obj); | ||
| 521 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); | ||
| 522 | out_unref: | ||
| 523 | radeon_bo_unref(&rdev->r600_blit.shader_obj); | ||
| 516 | } | 524 | } |
| 517 | 525 | ||
| 518 | int r600_vb_ib_get(struct radeon_device *rdev) | 526 | int r600_vb_ib_get(struct radeon_device *rdev) |
| @@ -569,9 +577,9 @@ int r600_blit_prepare_copy(struct radeon_device *rdev, int size_bytes) | |||
| 569 | ring_size = num_loops * dwords_per_loop; | 577 | ring_size = num_loops * dwords_per_loop; |
| 570 | /* set default + shaders */ | 578 | /* set default + shaders */ |
| 571 | ring_size += 40; /* shaders + def state */ | 579 | ring_size += 40; /* shaders + def state */ |
| 572 | ring_size += 3; /* fence emit for VB IB */ | 580 | ring_size += 5; /* fence emit for VB IB */ |
| 573 | ring_size += 5; /* done copy */ | 581 | ring_size += 5; /* done copy */ |
| 574 | ring_size += 3; /* fence emit for done copy */ | 582 | ring_size += 5; /* fence emit for done copy */ |
| 575 | r = radeon_ring_lock(rdev, ring_size); | 583 | r = radeon_ring_lock(rdev, ring_size); |
| 576 | WARN_ON(r); | 584 | WARN_ON(r); |
| 577 | 585 | ||
diff --git a/drivers/gpu/drm/radeon/r600d.h b/drivers/gpu/drm/radeon/r600d.h index 27ab428b149b..05894edadab4 100644 --- a/drivers/gpu/drm/radeon/r600d.h +++ b/drivers/gpu/drm/radeon/r600d.h | |||
| @@ -456,7 +456,215 @@ | |||
| 456 | #define WAIT_2D_IDLECLEAN_bit (1 << 16) | 456 | #define WAIT_2D_IDLECLEAN_bit (1 << 16) |
| 457 | #define WAIT_3D_IDLECLEAN_bit (1 << 17) | 457 | #define WAIT_3D_IDLECLEAN_bit (1 << 17) |
| 458 | 458 | ||
| 459 | 459 | #define IH_RB_CNTL 0x3e00 | |
| 460 | # define IH_RB_ENABLE (1 << 0) | ||
| 461 | # define IH_IB_SIZE(x) ((x) << 1) /* log2 */ | ||
| 462 | # define IH_RB_FULL_DRAIN_ENABLE (1 << 6) | ||
| 463 | # define IH_WPTR_WRITEBACK_ENABLE (1 << 8) | ||
| 464 | # define IH_WPTR_WRITEBACK_TIMER(x) ((x) << 9) /* log2 */ | ||
| 465 | # define IH_WPTR_OVERFLOW_ENABLE (1 << 16) | ||
| 466 | # define IH_WPTR_OVERFLOW_CLEAR (1 << 31) | ||
| 467 | #define IH_RB_BASE 0x3e04 | ||
| 468 | #define IH_RB_RPTR 0x3e08 | ||
| 469 | #define IH_RB_WPTR 0x3e0c | ||
| 470 | # define RB_OVERFLOW (1 << 0) | ||
| 471 | # define WPTR_OFFSET_MASK 0x3fffc | ||
| 472 | #define IH_RB_WPTR_ADDR_HI 0x3e10 | ||
| 473 | #define IH_RB_WPTR_ADDR_LO 0x3e14 | ||
| 474 | #define IH_CNTL 0x3e18 | ||
| 475 | # define ENABLE_INTR (1 << 0) | ||
| 476 | # define IH_MC_SWAP(x) ((x) << 2) | ||
| 477 | # define IH_MC_SWAP_NONE 0 | ||
| 478 | # define IH_MC_SWAP_16BIT 1 | ||
| 479 | # define IH_MC_SWAP_32BIT 2 | ||
| 480 | # define IH_MC_SWAP_64BIT 3 | ||
| 481 | # define RPTR_REARM (1 << 4) | ||
| 482 | # define MC_WRREQ_CREDIT(x) ((x) << 15) | ||
| 483 | # define MC_WR_CLEAN_CNT(x) ((x) << 20) | ||
| 484 | |||
| 485 | #define RLC_CNTL 0x3f00 | ||
| 486 | # define RLC_ENABLE (1 << 0) | ||
| 487 | #define RLC_HB_BASE 0x3f10 | ||
| 488 | #define RLC_HB_CNTL 0x3f0c | ||
| 489 | #define RLC_HB_RPTR 0x3f20 | ||
| 490 | #define RLC_HB_WPTR 0x3f1c | ||
| 491 | #define RLC_HB_WPTR_LSB_ADDR 0x3f14 | ||
| 492 | #define RLC_HB_WPTR_MSB_ADDR 0x3f18 | ||
| 493 | #define RLC_MC_CNTL 0x3f44 | ||
| 494 | #define RLC_UCODE_CNTL 0x3f48 | ||
| 495 | #define RLC_UCODE_ADDR 0x3f2c | ||
| 496 | #define RLC_UCODE_DATA 0x3f30 | ||
| 497 | |||
| 498 | #define SRBM_SOFT_RESET 0xe60 | ||
| 499 | # define SOFT_RESET_RLC (1 << 13) | ||
| 500 | |||
| 501 | #define CP_INT_CNTL 0xc124 | ||
| 502 | # define CNTX_BUSY_INT_ENABLE (1 << 19) | ||
| 503 | # define CNTX_EMPTY_INT_ENABLE (1 << 20) | ||
| 504 | # define SCRATCH_INT_ENABLE (1 << 25) | ||
| 505 | # define TIME_STAMP_INT_ENABLE (1 << 26) | ||
| 506 | # define IB2_INT_ENABLE (1 << 29) | ||
| 507 | # define IB1_INT_ENABLE (1 << 30) | ||
| 508 | # define RB_INT_ENABLE (1 << 31) | ||
| 509 | #define CP_INT_STATUS 0xc128 | ||
| 510 | # define SCRATCH_INT_STAT (1 << 25) | ||
| 511 | # define TIME_STAMP_INT_STAT (1 << 26) | ||
| 512 | # define IB2_INT_STAT (1 << 29) | ||
| 513 | # define IB1_INT_STAT (1 << 30) | ||
| 514 | # define RB_INT_STAT (1 << 31) | ||
| 515 | |||
| 516 | #define GRBM_INT_CNTL 0x8060 | ||
| 517 | # define RDERR_INT_ENABLE (1 << 0) | ||
| 518 | # define WAIT_COUNT_TIMEOUT_INT_ENABLE (1 << 1) | ||
| 519 | # define GUI_IDLE_INT_ENABLE (1 << 19) | ||
| 520 | |||
| 521 | #define INTERRUPT_CNTL 0x5468 | ||
| 522 | # define IH_DUMMY_RD_OVERRIDE (1 << 0) | ||
| 523 | # define IH_DUMMY_RD_EN (1 << 1) | ||
| 524 | # define IH_REQ_NONSNOOP_EN (1 << 3) | ||
| 525 | # define GEN_IH_INT_EN (1 << 8) | ||
| 526 | #define INTERRUPT_CNTL2 0x546c | ||
| 527 | |||
| 528 | #define D1MODE_VBLANK_STATUS 0x6534 | ||
| 529 | #define D2MODE_VBLANK_STATUS 0x6d34 | ||
| 530 | # define DxMODE_VBLANK_OCCURRED (1 << 0) | ||
| 531 | # define DxMODE_VBLANK_ACK (1 << 4) | ||
| 532 | # define DxMODE_VBLANK_STAT (1 << 12) | ||
| 533 | # define DxMODE_VBLANK_INTERRUPT (1 << 16) | ||
| 534 | # define DxMODE_VBLANK_INTERRUPT_TYPE (1 << 17) | ||
| 535 | #define D1MODE_VLINE_STATUS 0x653c | ||
| 536 | #define D2MODE_VLINE_STATUS 0x6d3c | ||
| 537 | # define DxMODE_VLINE_OCCURRED (1 << 0) | ||
| 538 | # define DxMODE_VLINE_ACK (1 << 4) | ||
| 539 | # define DxMODE_VLINE_STAT (1 << 12) | ||
| 540 | # define DxMODE_VLINE_INTERRUPT (1 << 16) | ||
| 541 | # define DxMODE_VLINE_INTERRUPT_TYPE (1 << 17) | ||
| 542 | #define DxMODE_INT_MASK 0x6540 | ||
| 543 | # define D1MODE_VBLANK_INT_MASK (1 << 0) | ||
| 544 | # define D1MODE_VLINE_INT_MASK (1 << 4) | ||
| 545 | # define D2MODE_VBLANK_INT_MASK (1 << 8) | ||
| 546 | # define D2MODE_VLINE_INT_MASK (1 << 12) | ||
| 547 | #define DCE3_DISP_INTERRUPT_STATUS 0x7ddc | ||
| 548 | # define DC_HPD1_INTERRUPT (1 << 18) | ||
| 549 | # define DC_HPD2_INTERRUPT (1 << 19) | ||
| 550 | #define DISP_INTERRUPT_STATUS 0x7edc | ||
| 551 | # define LB_D1_VLINE_INTERRUPT (1 << 2) | ||
| 552 | # define LB_D2_VLINE_INTERRUPT (1 << 3) | ||
| 553 | # define LB_D1_VBLANK_INTERRUPT (1 << 4) | ||
| 554 | # define LB_D2_VBLANK_INTERRUPT (1 << 5) | ||
| 555 | # define DACA_AUTODETECT_INTERRUPT (1 << 16) | ||
| 556 | # define DACB_AUTODETECT_INTERRUPT (1 << 17) | ||
| 557 | # define DC_HOT_PLUG_DETECT1_INTERRUPT (1 << 18) | ||
| 558 | # define DC_HOT_PLUG_DETECT2_INTERRUPT (1 << 19) | ||
| 559 | # define DC_I2C_SW_DONE_INTERRUPT (1 << 20) | ||
| 560 | # define DC_I2C_HW_DONE_INTERRUPT (1 << 21) | ||
| 561 | #define DISP_INTERRUPT_STATUS_CONTINUE 0x7ee8 | ||
| 562 | #define DCE3_DISP_INTERRUPT_STATUS_CONTINUE 0x7de8 | ||
| 563 | # define DC_HPD4_INTERRUPT (1 << 14) | ||
| 564 | # define DC_HPD4_RX_INTERRUPT (1 << 15) | ||
| 565 | # define DC_HPD3_INTERRUPT (1 << 28) | ||
| 566 | # define DC_HPD1_RX_INTERRUPT (1 << 29) | ||
| 567 | # define DC_HPD2_RX_INTERRUPT (1 << 30) | ||
| 568 | #define DCE3_DISP_INTERRUPT_STATUS_CONTINUE2 0x7dec | ||
| 569 | # define DC_HPD3_RX_INTERRUPT (1 << 0) | ||
| 570 | # define DIGA_DP_VID_STREAM_DISABLE_INTERRUPT (1 << 1) | ||
| 571 | # define DIGA_DP_STEER_FIFO_OVERFLOW_INTERRUPT (1 << 2) | ||
| 572 | # define DIGB_DP_VID_STREAM_DISABLE_INTERRUPT (1 << 3) | ||
| 573 | # define DIGB_DP_STEER_FIFO_OVERFLOW_INTERRUPT (1 << 4) | ||
| 574 | # define AUX1_SW_DONE_INTERRUPT (1 << 5) | ||
| 575 | # define AUX1_LS_DONE_INTERRUPT (1 << 6) | ||
| 576 | # define AUX2_SW_DONE_INTERRUPT (1 << 7) | ||
| 577 | # define AUX2_LS_DONE_INTERRUPT (1 << 8) | ||
| 578 | # define AUX3_SW_DONE_INTERRUPT (1 << 9) | ||
| 579 | # define AUX3_LS_DONE_INTERRUPT (1 << 10) | ||
| 580 | # define AUX4_SW_DONE_INTERRUPT (1 << 11) | ||
| 581 | # define AUX4_LS_DONE_INTERRUPT (1 << 12) | ||
| 582 | # define DIGA_DP_FAST_TRAINING_COMPLETE_INTERRUPT (1 << 13) | ||
| 583 | # define DIGB_DP_FAST_TRAINING_COMPLETE_INTERRUPT (1 << 14) | ||
| 584 | /* DCE 3.2 */ | ||
| 585 | # define AUX5_SW_DONE_INTERRUPT (1 << 15) | ||
| 586 | # define AUX5_LS_DONE_INTERRUPT (1 << 16) | ||
| 587 | # define AUX6_SW_DONE_INTERRUPT (1 << 17) | ||
| 588 | # define AUX6_LS_DONE_INTERRUPT (1 << 18) | ||
| 589 | # define DC_HPD5_INTERRUPT (1 << 19) | ||
| 590 | # define DC_HPD5_RX_INTERRUPT (1 << 20) | ||
| 591 | # define DC_HPD6_INTERRUPT (1 << 21) | ||
| 592 | # define DC_HPD6_RX_INTERRUPT (1 << 22) | ||
| 593 | |||
| 594 | #define DACA_AUTO_DETECT_CONTROL 0x7828 | ||
| 595 | #define DACB_AUTO_DETECT_CONTROL 0x7a28 | ||
| 596 | #define DCE3_DACA_AUTO_DETECT_CONTROL 0x7028 | ||
| 597 | #define DCE3_DACB_AUTO_DETECT_CONTROL 0x7128 | ||
| 598 | # define DACx_AUTODETECT_MODE(x) ((x) << 0) | ||
| 599 | # define DACx_AUTODETECT_MODE_NONE 0 | ||
| 600 | # define DACx_AUTODETECT_MODE_CONNECT 1 | ||
| 601 | # define DACx_AUTODETECT_MODE_DISCONNECT 2 | ||
| 602 | # define DACx_AUTODETECT_FRAME_TIME_COUNTER(x) ((x) << 8) | ||
| 603 | /* bit 18 = R/C, 17 = G/Y, 16 = B/Comp */ | ||
| 604 | # define DACx_AUTODETECT_CHECK_MASK(x) ((x) << 16) | ||
| 605 | |||
| 606 | #define DCE3_DACA_AUTODETECT_INT_CONTROL 0x7038 | ||
| 607 | #define DCE3_DACB_AUTODETECT_INT_CONTROL 0x7138 | ||
| 608 | #define DACA_AUTODETECT_INT_CONTROL 0x7838 | ||
| 609 | #define DACB_AUTODETECT_INT_CONTROL 0x7a38 | ||
| 610 | # define DACx_AUTODETECT_ACK (1 << 0) | ||
| 611 | # define DACx_AUTODETECT_INT_ENABLE (1 << 16) | ||
| 612 | |||
| 613 | #define DC_HOT_PLUG_DETECT1_CONTROL 0x7d00 | ||
| 614 | #define DC_HOT_PLUG_DETECT2_CONTROL 0x7d10 | ||
| 615 | #define DC_HOT_PLUG_DETECT3_CONTROL 0x7d24 | ||
| 616 | # define DC_HOT_PLUG_DETECTx_EN (1 << 0) | ||
| 617 | |||
| 618 | #define DC_HOT_PLUG_DETECT1_INT_STATUS 0x7d04 | ||
| 619 | #define DC_HOT_PLUG_DETECT2_INT_STATUS 0x7d14 | ||
| 620 | #define DC_HOT_PLUG_DETECT3_INT_STATUS 0x7d28 | ||
| 621 | # define DC_HOT_PLUG_DETECTx_INT_STATUS (1 << 0) | ||
| 622 | # define DC_HOT_PLUG_DETECTx_SENSE (1 << 1) | ||
| 623 | |||
| 624 | /* DCE 3.0 */ | ||
| 625 | #define DC_HPD1_INT_STATUS 0x7d00 | ||
| 626 | #define DC_HPD2_INT_STATUS 0x7d0c | ||
| 627 | #define DC_HPD3_INT_STATUS 0x7d18 | ||
| 628 | #define DC_HPD4_INT_STATUS 0x7d24 | ||
| 629 | /* DCE 3.2 */ | ||
| 630 | #define DC_HPD5_INT_STATUS 0x7dc0 | ||
| 631 | #define DC_HPD6_INT_STATUS 0x7df4 | ||
| 632 | # define DC_HPDx_INT_STATUS (1 << 0) | ||
| 633 | # define DC_HPDx_SENSE (1 << 1) | ||
| 634 | # define DC_HPDx_RX_INT_STATUS (1 << 8) | ||
| 635 | |||
| 636 | #define DC_HOT_PLUG_DETECT1_INT_CONTROL 0x7d08 | ||
| 637 | #define DC_HOT_PLUG_DETECT2_INT_CONTROL 0x7d18 | ||
| 638 | #define DC_HOT_PLUG_DETECT3_INT_CONTROL 0x7d2c | ||
| 639 | # define DC_HOT_PLUG_DETECTx_INT_ACK (1 << 0) | ||
| 640 | # define DC_HOT_PLUG_DETECTx_INT_POLARITY (1 << 8) | ||
| 641 | # define DC_HOT_PLUG_DETECTx_INT_EN (1 << 16) | ||
| 642 | /* DCE 3.0 */ | ||
| 643 | #define DC_HPD1_INT_CONTROL 0x7d04 | ||
| 644 | #define DC_HPD2_INT_CONTROL 0x7d10 | ||
| 645 | #define DC_HPD3_INT_CONTROL 0x7d1c | ||
| 646 | #define DC_HPD4_INT_CONTROL 0x7d28 | ||
| 647 | /* DCE 3.2 */ | ||
| 648 | #define DC_HPD5_INT_CONTROL 0x7dc4 | ||
| 649 | #define DC_HPD6_INT_CONTROL 0x7df8 | ||
| 650 | # define DC_HPDx_INT_ACK (1 << 0) | ||
| 651 | # define DC_HPDx_INT_POLARITY (1 << 8) | ||
| 652 | # define DC_HPDx_INT_EN (1 << 16) | ||
| 653 | # define DC_HPDx_RX_INT_ACK (1 << 20) | ||
| 654 | # define DC_HPDx_RX_INT_EN (1 << 24) | ||
| 655 | |||
| 656 | /* DCE 3.0 */ | ||
| 657 | #define DC_HPD1_CONTROL 0x7d08 | ||
| 658 | #define DC_HPD2_CONTROL 0x7d14 | ||
| 659 | #define DC_HPD3_CONTROL 0x7d20 | ||
| 660 | #define DC_HPD4_CONTROL 0x7d2c | ||
| 661 | /* DCE 3.2 */ | ||
| 662 | #define DC_HPD5_CONTROL 0x7dc8 | ||
| 663 | #define DC_HPD6_CONTROL 0x7dfc | ||
| 664 | # define DC_HPDx_CONNECTION_TIMER(x) ((x) << 0) | ||
| 665 | # define DC_HPDx_RX_INT_TIMER(x) ((x) << 16) | ||
| 666 | /* DCE 3.2 */ | ||
| 667 | # define DC_HPDx_EN (1 << 28) | ||
| 460 | 668 | ||
| 461 | /* | 669 | /* |
| 462 | * PM4 | 670 | * PM4 |
| @@ -500,7 +708,6 @@ | |||
| 500 | #define PACKET3_WAIT_REG_MEM 0x3C | 708 | #define PACKET3_WAIT_REG_MEM 0x3C |
| 501 | #define PACKET3_MEM_WRITE 0x3D | 709 | #define PACKET3_MEM_WRITE 0x3D |
| 502 | #define PACKET3_INDIRECT_BUFFER 0x32 | 710 | #define PACKET3_INDIRECT_BUFFER 0x32 |
| 503 | #define PACKET3_CP_INTERRUPT 0x40 | ||
| 504 | #define PACKET3_SURFACE_SYNC 0x43 | 711 | #define PACKET3_SURFACE_SYNC 0x43 |
| 505 | # define PACKET3_CB0_DEST_BASE_ENA (1 << 6) | 712 | # define PACKET3_CB0_DEST_BASE_ENA (1 << 6) |
| 506 | # define PACKET3_TC_ACTION_ENA (1 << 23) | 713 | # define PACKET3_TC_ACTION_ENA (1 << 23) |
| @@ -674,4 +881,5 @@ | |||
| 674 | #define S_000E60_SOFT_RESET_TSC(x) (((x) & 1) << 16) | 881 | #define S_000E60_SOFT_RESET_TSC(x) (((x) & 1) << 16) |
| 675 | #define S_000E60_SOFT_RESET_VMC(x) (((x) & 1) << 17) | 882 | #define S_000E60_SOFT_RESET_VMC(x) (((x) & 1) << 17) |
| 676 | 883 | ||
| 884 | #define R_005480_HDP_MEM_COHERENCY_FLUSH_CNTL 0x5480 | ||
| 677 | #endif | 885 | #endif |
diff --git a/drivers/gpu/drm/radeon/radeon.h b/drivers/gpu/drm/radeon/radeon.h index 224506a2f7b1..c938bb54123c 100644 --- a/drivers/gpu/drm/radeon/radeon.h +++ b/drivers/gpu/drm/radeon/radeon.h | |||
| @@ -28,8 +28,6 @@ | |||
| 28 | #ifndef __RADEON_H__ | 28 | #ifndef __RADEON_H__ |
| 29 | #define __RADEON_H__ | 29 | #define __RADEON_H__ |
| 30 | 30 | ||
| 31 | #include "radeon_object.h" | ||
| 32 | |||
| 33 | /* TODO: Here are things that needs to be done : | 31 | /* TODO: Here are things that needs to be done : |
| 34 | * - surface allocator & initializer : (bit like scratch reg) should | 32 | * - surface allocator & initializer : (bit like scratch reg) should |
| 35 | * initialize HDP_ stuff on RS600, R600, R700 hw, well anythings | 33 | * initialize HDP_ stuff on RS600, R600, R700 hw, well anythings |
| @@ -67,6 +65,11 @@ | |||
| 67 | #include <linux/list.h> | 65 | #include <linux/list.h> |
| 68 | #include <linux/kref.h> | 66 | #include <linux/kref.h> |
| 69 | 67 | ||
| 68 | #include <ttm/ttm_bo_api.h> | ||
| 69 | #include <ttm/ttm_bo_driver.h> | ||
| 70 | #include <ttm/ttm_placement.h> | ||
| 71 | #include <ttm/ttm_module.h> | ||
| 72 | |||
| 70 | #include "radeon_family.h" | 73 | #include "radeon_family.h" |
| 71 | #include "radeon_mode.h" | 74 | #include "radeon_mode.h" |
| 72 | #include "radeon_reg.h" | 75 | #include "radeon_reg.h" |
| @@ -85,6 +88,7 @@ extern int radeon_benchmarking; | |||
| 85 | extern int radeon_testing; | 88 | extern int radeon_testing; |
| 86 | extern int radeon_connector_table; | 89 | extern int radeon_connector_table; |
| 87 | extern int radeon_tv; | 90 | extern int radeon_tv; |
| 91 | extern int radeon_new_pll; | ||
| 88 | 92 | ||
| 89 | /* | 93 | /* |
| 90 | * Copy from radeon_drv.h so we don't have to include both and have conflicting | 94 | * Copy from radeon_drv.h so we don't have to include both and have conflicting |
| @@ -186,76 +190,62 @@ void radeon_fence_unref(struct radeon_fence **fence); | |||
| 186 | * Tiling registers | 190 | * Tiling registers |
| 187 | */ | 191 | */ |
| 188 | struct radeon_surface_reg { | 192 | struct radeon_surface_reg { |
| 189 | struct radeon_object *robj; | 193 | struct radeon_bo *bo; |
| 190 | }; | 194 | }; |
| 191 | 195 | ||
| 192 | #define RADEON_GEM_MAX_SURFACES 8 | 196 | #define RADEON_GEM_MAX_SURFACES 8 |
| 193 | 197 | ||
| 194 | /* | 198 | /* |
| 195 | * Radeon buffer. | 199 | * TTM. |
| 196 | */ | 200 | */ |
| 197 | struct radeon_object; | 201 | struct radeon_mman { |
| 202 | struct ttm_bo_global_ref bo_global_ref; | ||
| 203 | struct ttm_global_reference mem_global_ref; | ||
| 204 | bool mem_global_referenced; | ||
| 205 | struct ttm_bo_device bdev; | ||
| 206 | }; | ||
| 207 | |||
| 208 | struct radeon_bo { | ||
| 209 | /* Protected by gem.mutex */ | ||
| 210 | struct list_head list; | ||
| 211 | /* Protected by tbo.reserved */ | ||
| 212 | u32 placements[3]; | ||
| 213 | struct ttm_placement placement; | ||
| 214 | struct ttm_buffer_object tbo; | ||
| 215 | struct ttm_bo_kmap_obj kmap; | ||
| 216 | unsigned pin_count; | ||
| 217 | void *kptr; | ||
| 218 | u32 tiling_flags; | ||
| 219 | u32 pitch; | ||
| 220 | int surface_reg; | ||
| 221 | /* Constant after initialization */ | ||
| 222 | struct radeon_device *rdev; | ||
| 223 | struct drm_gem_object *gobj; | ||
| 224 | }; | ||
| 198 | 225 | ||
| 199 | struct radeon_object_list { | 226 | struct radeon_bo_list { |
| 200 | struct list_head list; | 227 | struct list_head list; |
| 201 | struct radeon_object *robj; | 228 | struct radeon_bo *bo; |
| 202 | uint64_t gpu_offset; | 229 | uint64_t gpu_offset; |
| 203 | unsigned rdomain; | 230 | unsigned rdomain; |
| 204 | unsigned wdomain; | 231 | unsigned wdomain; |
| 205 | uint32_t tiling_flags; | 232 | u32 tiling_flags; |
| 206 | }; | 233 | }; |
| 207 | 234 | ||
| 208 | int radeon_object_init(struct radeon_device *rdev); | ||
| 209 | void radeon_object_fini(struct radeon_device *rdev); | ||
| 210 | int radeon_object_create(struct radeon_device *rdev, | ||
| 211 | struct drm_gem_object *gobj, | ||
| 212 | unsigned long size, | ||
| 213 | bool kernel, | ||
| 214 | uint32_t domain, | ||
| 215 | bool interruptible, | ||
| 216 | struct radeon_object **robj_ptr); | ||
| 217 | int radeon_object_kmap(struct radeon_object *robj, void **ptr); | ||
| 218 | void radeon_object_kunmap(struct radeon_object *robj); | ||
| 219 | void radeon_object_unref(struct radeon_object **robj); | ||
| 220 | int radeon_object_pin(struct radeon_object *robj, uint32_t domain, | ||
| 221 | uint64_t *gpu_addr); | ||
| 222 | void radeon_object_unpin(struct radeon_object *robj); | ||
| 223 | int radeon_object_wait(struct radeon_object *robj); | ||
| 224 | int radeon_object_busy_domain(struct radeon_object *robj, uint32_t *cur_placement); | ||
| 225 | int radeon_object_evict_vram(struct radeon_device *rdev); | ||
| 226 | int radeon_object_mmap(struct radeon_object *robj, uint64_t *offset); | ||
| 227 | void radeon_object_force_delete(struct radeon_device *rdev); | ||
| 228 | void radeon_object_list_add_object(struct radeon_object_list *lobj, | ||
| 229 | struct list_head *head); | ||
| 230 | int radeon_object_list_validate(struct list_head *head, void *fence); | ||
| 231 | void radeon_object_list_unvalidate(struct list_head *head); | ||
| 232 | void radeon_object_list_clean(struct list_head *head); | ||
| 233 | int radeon_object_fbdev_mmap(struct radeon_object *robj, | ||
| 234 | struct vm_area_struct *vma); | ||
| 235 | unsigned long radeon_object_size(struct radeon_object *robj); | ||
| 236 | void radeon_object_clear_surface_reg(struct radeon_object *robj); | ||
| 237 | int radeon_object_check_tiling(struct radeon_object *robj, bool has_moved, | ||
| 238 | bool force_drop); | ||
| 239 | void radeon_object_set_tiling_flags(struct radeon_object *robj, | ||
| 240 | uint32_t tiling_flags, uint32_t pitch); | ||
| 241 | void radeon_object_get_tiling_flags(struct radeon_object *robj, uint32_t *tiling_flags, uint32_t *pitch); | ||
| 242 | void radeon_bo_move_notify(struct ttm_buffer_object *bo, | ||
| 243 | struct ttm_mem_reg *mem); | ||
| 244 | void radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo); | ||
| 245 | /* | 235 | /* |
| 246 | * GEM objects. | 236 | * GEM objects. |
| 247 | */ | 237 | */ |
| 248 | struct radeon_gem { | 238 | struct radeon_gem { |
| 239 | struct mutex mutex; | ||
| 249 | struct list_head objects; | 240 | struct list_head objects; |
| 250 | }; | 241 | }; |
| 251 | 242 | ||
| 252 | int radeon_gem_init(struct radeon_device *rdev); | 243 | int radeon_gem_init(struct radeon_device *rdev); |
| 253 | void radeon_gem_fini(struct radeon_device *rdev); | 244 | void radeon_gem_fini(struct radeon_device *rdev); |
| 254 | int radeon_gem_object_create(struct radeon_device *rdev, int size, | 245 | int radeon_gem_object_create(struct radeon_device *rdev, int size, |
| 255 | int alignment, int initial_domain, | 246 | int alignment, int initial_domain, |
| 256 | bool discardable, bool kernel, | 247 | bool discardable, bool kernel, |
| 257 | bool interruptible, | 248 | struct drm_gem_object **obj); |
| 258 | struct drm_gem_object **obj); | ||
| 259 | int radeon_gem_object_pin(struct drm_gem_object *obj, uint32_t pin_domain, | 249 | int radeon_gem_object_pin(struct drm_gem_object *obj, uint32_t pin_domain, |
| 260 | uint64_t *gpu_addr); | 250 | uint64_t *gpu_addr); |
| 261 | void radeon_gem_object_unpin(struct drm_gem_object *obj); | 251 | void radeon_gem_object_unpin(struct drm_gem_object *obj); |
| @@ -271,7 +261,7 @@ struct radeon_gart_table_ram { | |||
| 271 | }; | 261 | }; |
| 272 | 262 | ||
| 273 | struct radeon_gart_table_vram { | 263 | struct radeon_gart_table_vram { |
| 274 | struct radeon_object *robj; | 264 | struct radeon_bo *robj; |
| 275 | volatile uint32_t *ptr; | 265 | volatile uint32_t *ptr; |
| 276 | }; | 266 | }; |
| 277 | 267 | ||
| @@ -352,11 +342,16 @@ struct radeon_irq { | |||
| 352 | bool sw_int; | 342 | bool sw_int; |
| 353 | /* FIXME: use a define max crtc rather than hardcode it */ | 343 | /* FIXME: use a define max crtc rather than hardcode it */ |
| 354 | bool crtc_vblank_int[2]; | 344 | bool crtc_vblank_int[2]; |
| 345 | /* FIXME: use defines for max hpd/dacs */ | ||
| 346 | bool hpd[6]; | ||
| 347 | spinlock_t sw_lock; | ||
| 348 | int sw_refcount; | ||
| 355 | }; | 349 | }; |
| 356 | 350 | ||
| 357 | int radeon_irq_kms_init(struct radeon_device *rdev); | 351 | int radeon_irq_kms_init(struct radeon_device *rdev); |
| 358 | void radeon_irq_kms_fini(struct radeon_device *rdev); | 352 | void radeon_irq_kms_fini(struct radeon_device *rdev); |
| 359 | 353 | void radeon_irq_kms_sw_irq_get(struct radeon_device *rdev); | |
| 354 | void radeon_irq_kms_sw_irq_put(struct radeon_device *rdev); | ||
| 360 | 355 | ||
| 361 | /* | 356 | /* |
| 362 | * CP & ring. | 357 | * CP & ring. |
| @@ -376,7 +371,7 @@ struct radeon_ib { | |||
| 376 | */ | 371 | */ |
| 377 | struct radeon_ib_pool { | 372 | struct radeon_ib_pool { |
| 378 | struct mutex mutex; | 373 | struct mutex mutex; |
| 379 | struct radeon_object *robj; | 374 | struct radeon_bo *robj; |
| 380 | struct list_head scheduled_ibs; | 375 | struct list_head scheduled_ibs; |
| 381 | struct radeon_ib ibs[RADEON_IB_POOL_SIZE]; | 376 | struct radeon_ib ibs[RADEON_IB_POOL_SIZE]; |
| 382 | bool ready; | 377 | bool ready; |
| @@ -384,7 +379,7 @@ struct radeon_ib_pool { | |||
| 384 | }; | 379 | }; |
| 385 | 380 | ||
| 386 | struct radeon_cp { | 381 | struct radeon_cp { |
| 387 | struct radeon_object *ring_obj; | 382 | struct radeon_bo *ring_obj; |
| 388 | volatile uint32_t *ring; | 383 | volatile uint32_t *ring; |
| 389 | unsigned rptr; | 384 | unsigned rptr; |
| 390 | unsigned wptr; | 385 | unsigned wptr; |
| @@ -399,8 +394,25 @@ struct radeon_cp { | |||
| 399 | bool ready; | 394 | bool ready; |
| 400 | }; | 395 | }; |
| 401 | 396 | ||
| 397 | /* | ||
| 398 | * R6xx+ IH ring | ||
| 399 | */ | ||
| 400 | struct r600_ih { | ||
| 401 | struct radeon_bo *ring_obj; | ||
| 402 | volatile uint32_t *ring; | ||
| 403 | unsigned rptr; | ||
| 404 | unsigned wptr; | ||
| 405 | unsigned wptr_old; | ||
| 406 | unsigned ring_size; | ||
| 407 | uint64_t gpu_addr; | ||
| 408 | uint32_t align_mask; | ||
| 409 | uint32_t ptr_mask; | ||
| 410 | spinlock_t lock; | ||
| 411 | bool enabled; | ||
| 412 | }; | ||
| 413 | |||
| 402 | struct r600_blit { | 414 | struct r600_blit { |
| 403 | struct radeon_object *shader_obj; | 415 | struct radeon_bo *shader_obj; |
| 404 | u64 shader_gpu_addr; | 416 | u64 shader_gpu_addr; |
| 405 | u32 vs_offset, ps_offset; | 417 | u32 vs_offset, ps_offset; |
| 406 | u32 state_offset; | 418 | u32 state_offset; |
| @@ -430,8 +442,8 @@ void radeon_ring_fini(struct radeon_device *rdev); | |||
| 430 | */ | 442 | */ |
| 431 | struct radeon_cs_reloc { | 443 | struct radeon_cs_reloc { |
| 432 | struct drm_gem_object *gobj; | 444 | struct drm_gem_object *gobj; |
| 433 | struct radeon_object *robj; | 445 | struct radeon_bo *robj; |
| 434 | struct radeon_object_list lobj; | 446 | struct radeon_bo_list lobj; |
| 435 | uint32_t handle; | 447 | uint32_t handle; |
| 436 | uint32_t flags; | 448 | uint32_t flags; |
| 437 | }; | 449 | }; |
| @@ -527,7 +539,7 @@ void radeon_agp_fini(struct radeon_device *rdev); | |||
| 527 | * Writeback | 539 | * Writeback |
| 528 | */ | 540 | */ |
| 529 | struct radeon_wb { | 541 | struct radeon_wb { |
| 530 | struct radeon_object *wb_obj; | 542 | struct radeon_bo *wb_obj; |
| 531 | volatile uint32_t *wb; | 543 | volatile uint32_t *wb; |
| 532 | uint64_t gpu_addr; | 544 | uint64_t gpu_addr; |
| 533 | }; | 545 | }; |
| @@ -639,6 +651,11 @@ struct radeon_asic { | |||
| 639 | uint32_t offset, uint32_t obj_size); | 651 | uint32_t offset, uint32_t obj_size); |
| 640 | int (*clear_surface_reg)(struct radeon_device *rdev, int reg); | 652 | int (*clear_surface_reg)(struct radeon_device *rdev, int reg); |
| 641 | void (*bandwidth_update)(struct radeon_device *rdev); | 653 | void (*bandwidth_update)(struct radeon_device *rdev); |
| 654 | void (*hdp_flush)(struct radeon_device *rdev); | ||
| 655 | void (*hpd_init)(struct radeon_device *rdev); | ||
| 656 | void (*hpd_fini)(struct radeon_device *rdev); | ||
| 657 | bool (*hpd_sense)(struct radeon_device *rdev, enum radeon_hpd_id hpd); | ||
| 658 | void (*hpd_set_polarity)(struct radeon_device *rdev, enum radeon_hpd_id hpd); | ||
| 642 | }; | 659 | }; |
| 643 | 660 | ||
| 644 | /* | 661 | /* |
| @@ -751,9 +768,9 @@ struct radeon_device { | |||
| 751 | uint8_t *bios; | 768 | uint8_t *bios; |
| 752 | bool is_atom_bios; | 769 | bool is_atom_bios; |
| 753 | uint16_t bios_header_start; | 770 | uint16_t bios_header_start; |
| 754 | struct radeon_object *stollen_vga_memory; | 771 | struct radeon_bo *stollen_vga_memory; |
| 755 | struct fb_info *fbdev_info; | 772 | struct fb_info *fbdev_info; |
| 756 | struct radeon_object *fbdev_robj; | 773 | struct radeon_bo *fbdev_rbo; |
| 757 | struct radeon_framebuffer *fbdev_rfb; | 774 | struct radeon_framebuffer *fbdev_rfb; |
| 758 | /* Register mmio */ | 775 | /* Register mmio */ |
| 759 | resource_size_t rmmio_base; | 776 | resource_size_t rmmio_base; |
| @@ -791,8 +808,12 @@ struct radeon_device { | |||
| 791 | struct radeon_surface_reg surface_regs[RADEON_GEM_MAX_SURFACES]; | 808 | struct radeon_surface_reg surface_regs[RADEON_GEM_MAX_SURFACES]; |
| 792 | const struct firmware *me_fw; /* all family ME firmware */ | 809 | const struct firmware *me_fw; /* all family ME firmware */ |
| 793 | const struct firmware *pfp_fw; /* r6/700 PFP firmware */ | 810 | const struct firmware *pfp_fw; /* r6/700 PFP firmware */ |
| 811 | const struct firmware *rlc_fw; /* r6/700 RLC firmware */ | ||
| 794 | struct r600_blit r600_blit; | 812 | struct r600_blit r600_blit; |
| 795 | int msi_enabled; /* msi enabled */ | 813 | int msi_enabled; /* msi enabled */ |
| 814 | struct r600_ih ih; /* r6/700 interrupt ring */ | ||
| 815 | struct workqueue_struct *wq; | ||
| 816 | struct work_struct hotplug_work; | ||
| 796 | }; | 817 | }; |
| 797 | 818 | ||
| 798 | int radeon_device_init(struct radeon_device *rdev, | 819 | int radeon_device_init(struct radeon_device *rdev, |
| @@ -829,6 +850,10 @@ static inline void r100_mm_wreg(struct radeon_device *rdev, uint32_t reg, uint32 | |||
| 829 | } | 850 | } |
| 830 | } | 851 | } |
| 831 | 852 | ||
| 853 | /* | ||
| 854 | * Cast helper | ||
| 855 | */ | ||
| 856 | #define to_radeon_fence(p) ((struct radeon_fence *)(p)) | ||
| 832 | 857 | ||
| 833 | /* | 858 | /* |
| 834 | * Registers read & write functions. | 859 | * Registers read & write functions. |
| @@ -965,18 +990,24 @@ static inline void radeon_ring_write(struct radeon_device *rdev, uint32_t v) | |||
| 965 | #define radeon_get_engine_clock(rdev) (rdev)->asic->get_engine_clock((rdev)) | 990 | #define radeon_get_engine_clock(rdev) (rdev)->asic->get_engine_clock((rdev)) |
| 966 | #define radeon_set_engine_clock(rdev, e) (rdev)->asic->set_engine_clock((rdev), (e)) | 991 | #define radeon_set_engine_clock(rdev, e) (rdev)->asic->set_engine_clock((rdev), (e)) |
| 967 | #define radeon_get_memory_clock(rdev) (rdev)->asic->get_memory_clock((rdev)) | 992 | #define radeon_get_memory_clock(rdev) (rdev)->asic->get_memory_clock((rdev)) |
| 968 | #define radeon_set_memory_clock(rdev, e) (rdev)->asic->set_engine_clock((rdev), (e)) | 993 | #define radeon_set_memory_clock(rdev, e) (rdev)->asic->set_memory_clock((rdev), (e)) |
| 969 | #define radeon_set_pcie_lanes(rdev, l) (rdev)->asic->set_pcie_lanes((rdev), (l)) | 994 | #define radeon_set_pcie_lanes(rdev, l) (rdev)->asic->set_pcie_lanes((rdev), (l)) |
| 970 | #define radeon_set_clock_gating(rdev, e) (rdev)->asic->set_clock_gating((rdev), (e)) | 995 | #define radeon_set_clock_gating(rdev, e) (rdev)->asic->set_clock_gating((rdev), (e)) |
| 971 | #define radeon_set_surface_reg(rdev, r, f, p, o, s) ((rdev)->asic->set_surface_reg((rdev), (r), (f), (p), (o), (s))) | 996 | #define radeon_set_surface_reg(rdev, r, f, p, o, s) ((rdev)->asic->set_surface_reg((rdev), (r), (f), (p), (o), (s))) |
| 972 | #define radeon_clear_surface_reg(rdev, r) ((rdev)->asic->clear_surface_reg((rdev), (r))) | 997 | #define radeon_clear_surface_reg(rdev, r) ((rdev)->asic->clear_surface_reg((rdev), (r))) |
| 973 | #define radeon_bandwidth_update(rdev) (rdev)->asic->bandwidth_update((rdev)) | 998 | #define radeon_bandwidth_update(rdev) (rdev)->asic->bandwidth_update((rdev)) |
| 999 | #define radeon_hdp_flush(rdev) (rdev)->asic->hdp_flush((rdev)) | ||
| 1000 | #define radeon_hpd_init(rdev) (rdev)->asic->hpd_init((rdev)) | ||
| 1001 | #define radeon_hpd_fini(rdev) (rdev)->asic->hpd_fini((rdev)) | ||
| 1002 | #define radeon_hpd_sense(rdev, hpd) (rdev)->asic->hpd_sense((rdev), (hpd)) | ||
| 1003 | #define radeon_hpd_set_polarity(rdev, hpd) (rdev)->asic->hpd_set_polarity((rdev), (hpd)) | ||
| 974 | 1004 | ||
| 975 | /* Common functions */ | 1005 | /* Common functions */ |
| 976 | extern int radeon_gart_table_vram_pin(struct radeon_device *rdev); | 1006 | extern int radeon_gart_table_vram_pin(struct radeon_device *rdev); |
| 977 | extern int radeon_modeset_init(struct radeon_device *rdev); | 1007 | extern int radeon_modeset_init(struct radeon_device *rdev); |
| 978 | extern void radeon_modeset_fini(struct radeon_device *rdev); | 1008 | extern void radeon_modeset_fini(struct radeon_device *rdev); |
| 979 | extern bool radeon_card_posted(struct radeon_device *rdev); | 1009 | extern bool radeon_card_posted(struct radeon_device *rdev); |
| 1010 | extern bool radeon_boot_test_post_card(struct radeon_device *rdev); | ||
| 980 | extern int radeon_clocks_init(struct radeon_device *rdev); | 1011 | extern int radeon_clocks_init(struct radeon_device *rdev); |
| 981 | extern void radeon_clocks_fini(struct radeon_device *rdev); | 1012 | extern void radeon_clocks_fini(struct radeon_device *rdev); |
| 982 | extern void radeon_scratch_init(struct radeon_device *rdev); | 1013 | extern void radeon_scratch_init(struct radeon_device *rdev); |
| @@ -984,6 +1015,7 @@ extern void radeon_surface_init(struct radeon_device *rdev); | |||
| 984 | extern int radeon_cs_parser_init(struct radeon_cs_parser *p, void *data); | 1015 | extern int radeon_cs_parser_init(struct radeon_cs_parser *p, void *data); |
| 985 | extern void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable); | 1016 | extern void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable); |
| 986 | extern void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable); | 1017 | extern void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable); |
| 1018 | extern void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain); | ||
| 987 | 1019 | ||
| 988 | /* r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280 */ | 1020 | /* r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280 */ |
| 989 | struct r100_mc_save { | 1021 | struct r100_mc_save { |
| @@ -1021,7 +1053,7 @@ extern int r100_cp_reset(struct radeon_device *rdev); | |||
| 1021 | extern void r100_vga_render_disable(struct radeon_device *rdev); | 1053 | extern void r100_vga_render_disable(struct radeon_device *rdev); |
| 1022 | extern int r100_cs_track_check_pkt3_indx_buffer(struct radeon_cs_parser *p, | 1054 | extern int r100_cs_track_check_pkt3_indx_buffer(struct radeon_cs_parser *p, |
| 1023 | struct radeon_cs_packet *pkt, | 1055 | struct radeon_cs_packet *pkt, |
| 1024 | struct radeon_object *robj); | 1056 | struct radeon_bo *robj); |
| 1025 | extern int r100_cs_parse_packet0(struct radeon_cs_parser *p, | 1057 | extern int r100_cs_parse_packet0(struct radeon_cs_parser *p, |
| 1026 | struct radeon_cs_packet *pkt, | 1058 | struct radeon_cs_packet *pkt, |
| 1027 | const unsigned *auth, unsigned n, | 1059 | const unsigned *auth, unsigned n, |
| @@ -1029,6 +1061,8 @@ extern int r100_cs_parse_packet0(struct radeon_cs_parser *p, | |||
| 1029 | extern int r100_cs_packet_parse(struct radeon_cs_parser *p, | 1061 | extern int r100_cs_packet_parse(struct radeon_cs_parser *p, |
| 1030 | struct radeon_cs_packet *pkt, | 1062 | struct radeon_cs_packet *pkt, |
| 1031 | unsigned idx); | 1063 | unsigned idx); |
| 1064 | extern void r100_enable_bm(struct radeon_device *rdev); | ||
| 1065 | extern void r100_set_common_regs(struct radeon_device *rdev); | ||
| 1032 | 1066 | ||
| 1033 | /* rv200,rv250,rv280 */ | 1067 | /* rv200,rv250,rv280 */ |
| 1034 | extern void r200_set_safe_registers(struct radeon_device *rdev); | 1068 | extern void r200_set_safe_registers(struct radeon_device *rdev); |
| @@ -1104,7 +1138,14 @@ extern void r600_wb_disable(struct radeon_device *rdev); | |||
| 1104 | extern void r600_scratch_init(struct radeon_device *rdev); | 1138 | extern void r600_scratch_init(struct radeon_device *rdev); |
| 1105 | extern int r600_blit_init(struct radeon_device *rdev); | 1139 | extern int r600_blit_init(struct radeon_device *rdev); |
| 1106 | extern void r600_blit_fini(struct radeon_device *rdev); | 1140 | extern void r600_blit_fini(struct radeon_device *rdev); |
| 1107 | extern int r600_cp_init_microcode(struct radeon_device *rdev); | 1141 | extern int r600_init_microcode(struct radeon_device *rdev); |
| 1108 | extern int r600_gpu_reset(struct radeon_device *rdev); | 1142 | extern int r600_gpu_reset(struct radeon_device *rdev); |
| 1143 | /* r600 irq */ | ||
| 1144 | extern int r600_irq_init(struct radeon_device *rdev); | ||
| 1145 | extern void r600_irq_fini(struct radeon_device *rdev); | ||
| 1146 | extern void r600_ih_ring_init(struct radeon_device *rdev, unsigned ring_size); | ||
| 1147 | extern int r600_irq_set(struct radeon_device *rdev); | ||
| 1148 | |||
| 1149 | #include "radeon_object.h" | ||
| 1109 | 1150 | ||
| 1110 | #endif | 1151 | #endif |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.h b/drivers/gpu/drm/radeon/radeon_asic.h index c18fbee387d7..636116bedcb4 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.h +++ b/drivers/gpu/drm/radeon/radeon_asic.h | |||
| @@ -76,6 +76,12 @@ int r100_clear_surface_reg(struct radeon_device *rdev, int reg); | |||
| 76 | void r100_bandwidth_update(struct radeon_device *rdev); | 76 | void r100_bandwidth_update(struct radeon_device *rdev); |
| 77 | void r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib); | 77 | void r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib); |
| 78 | int r100_ring_test(struct radeon_device *rdev); | 78 | int r100_ring_test(struct radeon_device *rdev); |
| 79 | void r100_hdp_flush(struct radeon_device *rdev); | ||
| 80 | void r100_hpd_init(struct radeon_device *rdev); | ||
| 81 | void r100_hpd_fini(struct radeon_device *rdev); | ||
| 82 | bool r100_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd); | ||
| 83 | void r100_hpd_set_polarity(struct radeon_device *rdev, | ||
| 84 | enum radeon_hpd_id hpd); | ||
| 79 | 85 | ||
| 80 | static struct radeon_asic r100_asic = { | 86 | static struct radeon_asic r100_asic = { |
| 81 | .init = &r100_init, | 87 | .init = &r100_init, |
| @@ -107,6 +113,11 @@ static struct radeon_asic r100_asic = { | |||
| 107 | .set_surface_reg = r100_set_surface_reg, | 113 | .set_surface_reg = r100_set_surface_reg, |
| 108 | .clear_surface_reg = r100_clear_surface_reg, | 114 | .clear_surface_reg = r100_clear_surface_reg, |
| 109 | .bandwidth_update = &r100_bandwidth_update, | 115 | .bandwidth_update = &r100_bandwidth_update, |
| 116 | .hdp_flush = &r100_hdp_flush, | ||
| 117 | .hpd_init = &r100_hpd_init, | ||
| 118 | .hpd_fini = &r100_hpd_fini, | ||
| 119 | .hpd_sense = &r100_hpd_sense, | ||
| 120 | .hpd_set_polarity = &r100_hpd_set_polarity, | ||
| 110 | }; | 121 | }; |
| 111 | 122 | ||
| 112 | 123 | ||
| @@ -162,6 +173,11 @@ static struct radeon_asic r300_asic = { | |||
| 162 | .set_surface_reg = r100_set_surface_reg, | 173 | .set_surface_reg = r100_set_surface_reg, |
| 163 | .clear_surface_reg = r100_clear_surface_reg, | 174 | .clear_surface_reg = r100_clear_surface_reg, |
| 164 | .bandwidth_update = &r100_bandwidth_update, | 175 | .bandwidth_update = &r100_bandwidth_update, |
| 176 | .hdp_flush = &r100_hdp_flush, | ||
| 177 | .hpd_init = &r100_hpd_init, | ||
| 178 | .hpd_fini = &r100_hpd_fini, | ||
| 179 | .hpd_sense = &r100_hpd_sense, | ||
| 180 | .hpd_set_polarity = &r100_hpd_set_polarity, | ||
| 165 | }; | 181 | }; |
| 166 | 182 | ||
| 167 | /* | 183 | /* |
| @@ -201,6 +217,11 @@ static struct radeon_asic r420_asic = { | |||
| 201 | .set_surface_reg = r100_set_surface_reg, | 217 | .set_surface_reg = r100_set_surface_reg, |
| 202 | .clear_surface_reg = r100_clear_surface_reg, | 218 | .clear_surface_reg = r100_clear_surface_reg, |
| 203 | .bandwidth_update = &r100_bandwidth_update, | 219 | .bandwidth_update = &r100_bandwidth_update, |
| 220 | .hdp_flush = &r100_hdp_flush, | ||
| 221 | .hpd_init = &r100_hpd_init, | ||
| 222 | .hpd_fini = &r100_hpd_fini, | ||
| 223 | .hpd_sense = &r100_hpd_sense, | ||
| 224 | .hpd_set_polarity = &r100_hpd_set_polarity, | ||
| 204 | }; | 225 | }; |
| 205 | 226 | ||
| 206 | 227 | ||
| @@ -245,6 +266,11 @@ static struct radeon_asic rs400_asic = { | |||
| 245 | .set_surface_reg = r100_set_surface_reg, | 266 | .set_surface_reg = r100_set_surface_reg, |
| 246 | .clear_surface_reg = r100_clear_surface_reg, | 267 | .clear_surface_reg = r100_clear_surface_reg, |
| 247 | .bandwidth_update = &r100_bandwidth_update, | 268 | .bandwidth_update = &r100_bandwidth_update, |
| 269 | .hdp_flush = &r100_hdp_flush, | ||
| 270 | .hpd_init = &r100_hpd_init, | ||
| 271 | .hpd_fini = &r100_hpd_fini, | ||
| 272 | .hpd_sense = &r100_hpd_sense, | ||
| 273 | .hpd_set_polarity = &r100_hpd_set_polarity, | ||
| 248 | }; | 274 | }; |
| 249 | 275 | ||
| 250 | 276 | ||
| @@ -263,6 +289,12 @@ int rs600_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr); | |||
| 263 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg); | 289 | uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg); |
| 264 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v); | 290 | void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v); |
| 265 | void rs600_bandwidth_update(struct radeon_device *rdev); | 291 | void rs600_bandwidth_update(struct radeon_device *rdev); |
| 292 | void rs600_hpd_init(struct radeon_device *rdev); | ||
| 293 | void rs600_hpd_fini(struct radeon_device *rdev); | ||
| 294 | bool rs600_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd); | ||
| 295 | void rs600_hpd_set_polarity(struct radeon_device *rdev, | ||
| 296 | enum radeon_hpd_id hpd); | ||
| 297 | |||
| 266 | static struct radeon_asic rs600_asic = { | 298 | static struct radeon_asic rs600_asic = { |
| 267 | .init = &rs600_init, | 299 | .init = &rs600_init, |
| 268 | .fini = &rs600_fini, | 300 | .fini = &rs600_fini, |
| @@ -291,6 +323,11 @@ static struct radeon_asic rs600_asic = { | |||
| 291 | .set_pcie_lanes = NULL, | 323 | .set_pcie_lanes = NULL, |
| 292 | .set_clock_gating = &radeon_atom_set_clock_gating, | 324 | .set_clock_gating = &radeon_atom_set_clock_gating, |
| 293 | .bandwidth_update = &rs600_bandwidth_update, | 325 | .bandwidth_update = &rs600_bandwidth_update, |
| 326 | .hdp_flush = &r100_hdp_flush, | ||
| 327 | .hpd_init = &rs600_hpd_init, | ||
| 328 | .hpd_fini = &rs600_hpd_fini, | ||
| 329 | .hpd_sense = &rs600_hpd_sense, | ||
| 330 | .hpd_set_polarity = &rs600_hpd_set_polarity, | ||
| 294 | }; | 331 | }; |
| 295 | 332 | ||
| 296 | 333 | ||
| @@ -334,6 +371,11 @@ static struct radeon_asic rs690_asic = { | |||
| 334 | .set_surface_reg = r100_set_surface_reg, | 371 | .set_surface_reg = r100_set_surface_reg, |
| 335 | .clear_surface_reg = r100_clear_surface_reg, | 372 | .clear_surface_reg = r100_clear_surface_reg, |
| 336 | .bandwidth_update = &rs690_bandwidth_update, | 373 | .bandwidth_update = &rs690_bandwidth_update, |
| 374 | .hdp_flush = &r100_hdp_flush, | ||
| 375 | .hpd_init = &rs600_hpd_init, | ||
| 376 | .hpd_fini = &rs600_hpd_fini, | ||
| 377 | .hpd_sense = &rs600_hpd_sense, | ||
| 378 | .hpd_set_polarity = &rs600_hpd_set_polarity, | ||
| 337 | }; | 379 | }; |
| 338 | 380 | ||
| 339 | 381 | ||
| @@ -381,6 +423,11 @@ static struct radeon_asic rv515_asic = { | |||
| 381 | .set_surface_reg = r100_set_surface_reg, | 423 | .set_surface_reg = r100_set_surface_reg, |
| 382 | .clear_surface_reg = r100_clear_surface_reg, | 424 | .clear_surface_reg = r100_clear_surface_reg, |
| 383 | .bandwidth_update = &rv515_bandwidth_update, | 425 | .bandwidth_update = &rv515_bandwidth_update, |
| 426 | .hdp_flush = &r100_hdp_flush, | ||
| 427 | .hpd_init = &rs600_hpd_init, | ||
| 428 | .hpd_fini = &rs600_hpd_fini, | ||
| 429 | .hpd_sense = &rs600_hpd_sense, | ||
| 430 | .hpd_set_polarity = &rs600_hpd_set_polarity, | ||
| 384 | }; | 431 | }; |
| 385 | 432 | ||
| 386 | 433 | ||
| @@ -419,6 +466,11 @@ static struct radeon_asic r520_asic = { | |||
| 419 | .set_surface_reg = r100_set_surface_reg, | 466 | .set_surface_reg = r100_set_surface_reg, |
| 420 | .clear_surface_reg = r100_clear_surface_reg, | 467 | .clear_surface_reg = r100_clear_surface_reg, |
| 421 | .bandwidth_update = &rv515_bandwidth_update, | 468 | .bandwidth_update = &rv515_bandwidth_update, |
| 469 | .hdp_flush = &r100_hdp_flush, | ||
| 470 | .hpd_init = &rs600_hpd_init, | ||
| 471 | .hpd_fini = &rs600_hpd_fini, | ||
| 472 | .hpd_sense = &rs600_hpd_sense, | ||
| 473 | .hpd_set_polarity = &rs600_hpd_set_polarity, | ||
| 422 | }; | 474 | }; |
| 423 | 475 | ||
| 424 | /* | 476 | /* |
| @@ -455,6 +507,12 @@ int r600_ring_test(struct radeon_device *rdev); | |||
| 455 | int r600_copy_blit(struct radeon_device *rdev, | 507 | int r600_copy_blit(struct radeon_device *rdev, |
| 456 | uint64_t src_offset, uint64_t dst_offset, | 508 | uint64_t src_offset, uint64_t dst_offset, |
| 457 | unsigned num_pages, struct radeon_fence *fence); | 509 | unsigned num_pages, struct radeon_fence *fence); |
| 510 | void r600_hdp_flush(struct radeon_device *rdev); | ||
| 511 | void r600_hpd_init(struct radeon_device *rdev); | ||
| 512 | void r600_hpd_fini(struct radeon_device *rdev); | ||
| 513 | bool r600_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd); | ||
| 514 | void r600_hpd_set_polarity(struct radeon_device *rdev, | ||
| 515 | enum radeon_hpd_id hpd); | ||
| 458 | 516 | ||
| 459 | static struct radeon_asic r600_asic = { | 517 | static struct radeon_asic r600_asic = { |
| 460 | .init = &r600_init, | 518 | .init = &r600_init, |
| @@ -470,6 +528,7 @@ static struct radeon_asic r600_asic = { | |||
| 470 | .ring_ib_execute = &r600_ring_ib_execute, | 528 | .ring_ib_execute = &r600_ring_ib_execute, |
| 471 | .irq_set = &r600_irq_set, | 529 | .irq_set = &r600_irq_set, |
| 472 | .irq_process = &r600_irq_process, | 530 | .irq_process = &r600_irq_process, |
| 531 | .get_vblank_counter = &rs600_get_vblank_counter, | ||
| 473 | .fence_ring_emit = &r600_fence_ring_emit, | 532 | .fence_ring_emit = &r600_fence_ring_emit, |
| 474 | .cs_parse = &r600_cs_parse, | 533 | .cs_parse = &r600_cs_parse, |
| 475 | .copy_blit = &r600_copy_blit, | 534 | .copy_blit = &r600_copy_blit, |
| @@ -484,6 +543,11 @@ static struct radeon_asic r600_asic = { | |||
| 484 | .set_surface_reg = r600_set_surface_reg, | 543 | .set_surface_reg = r600_set_surface_reg, |
| 485 | .clear_surface_reg = r600_clear_surface_reg, | 544 | .clear_surface_reg = r600_clear_surface_reg, |
| 486 | .bandwidth_update = &rv515_bandwidth_update, | 545 | .bandwidth_update = &rv515_bandwidth_update, |
| 546 | .hdp_flush = &r600_hdp_flush, | ||
| 547 | .hpd_init = &r600_hpd_init, | ||
| 548 | .hpd_fini = &r600_hpd_fini, | ||
| 549 | .hpd_sense = &r600_hpd_sense, | ||
| 550 | .hpd_set_polarity = &r600_hpd_set_polarity, | ||
| 487 | }; | 551 | }; |
| 488 | 552 | ||
| 489 | /* | 553 | /* |
| @@ -509,6 +573,7 @@ static struct radeon_asic rv770_asic = { | |||
| 509 | .ring_ib_execute = &r600_ring_ib_execute, | 573 | .ring_ib_execute = &r600_ring_ib_execute, |
| 510 | .irq_set = &r600_irq_set, | 574 | .irq_set = &r600_irq_set, |
| 511 | .irq_process = &r600_irq_process, | 575 | .irq_process = &r600_irq_process, |
| 576 | .get_vblank_counter = &rs600_get_vblank_counter, | ||
| 512 | .fence_ring_emit = &r600_fence_ring_emit, | 577 | .fence_ring_emit = &r600_fence_ring_emit, |
| 513 | .cs_parse = &r600_cs_parse, | 578 | .cs_parse = &r600_cs_parse, |
| 514 | .copy_blit = &r600_copy_blit, | 579 | .copy_blit = &r600_copy_blit, |
| @@ -523,6 +588,11 @@ static struct radeon_asic rv770_asic = { | |||
| 523 | .set_surface_reg = r600_set_surface_reg, | 588 | .set_surface_reg = r600_set_surface_reg, |
| 524 | .clear_surface_reg = r600_clear_surface_reg, | 589 | .clear_surface_reg = r600_clear_surface_reg, |
| 525 | .bandwidth_update = &rv515_bandwidth_update, | 590 | .bandwidth_update = &rv515_bandwidth_update, |
| 591 | .hdp_flush = &r600_hdp_flush, | ||
| 592 | .hpd_init = &r600_hpd_init, | ||
| 593 | .hpd_fini = &r600_hpd_fini, | ||
| 594 | .hpd_sense = &r600_hpd_sense, | ||
| 595 | .hpd_set_polarity = &r600_hpd_set_polarity, | ||
| 526 | }; | 596 | }; |
| 527 | 597 | ||
| 528 | #endif | 598 | #endif |
diff --git a/drivers/gpu/drm/radeon/radeon_atombios.c b/drivers/gpu/drm/radeon/radeon_atombios.c index 2ed88a820935..12a0c760e7ff 100644 --- a/drivers/gpu/drm/radeon/radeon_atombios.c +++ b/drivers/gpu/drm/radeon/radeon_atombios.c | |||
| @@ -47,7 +47,8 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 47 | int connector_type, | 47 | int connector_type, |
| 48 | struct radeon_i2c_bus_rec *i2c_bus, | 48 | struct radeon_i2c_bus_rec *i2c_bus, |
| 49 | bool linkb, uint32_t igp_lane_info, | 49 | bool linkb, uint32_t igp_lane_info, |
| 50 | uint16_t connector_object_id); | 50 | uint16_t connector_object_id, |
| 51 | struct radeon_hpd *hpd); | ||
| 51 | 52 | ||
| 52 | /* from radeon_legacy_encoder.c */ | 53 | /* from radeon_legacy_encoder.c */ |
| 53 | extern void | 54 | extern void |
| @@ -60,16 +61,16 @@ union atom_supported_devices { | |||
| 60 | struct _ATOM_SUPPORTED_DEVICES_INFO_2d1 info_2d1; | 61 | struct _ATOM_SUPPORTED_DEVICES_INFO_2d1 info_2d1; |
| 61 | }; | 62 | }; |
| 62 | 63 | ||
| 63 | static inline struct radeon_i2c_bus_rec radeon_lookup_gpio(struct drm_device | 64 | static inline struct radeon_i2c_bus_rec radeon_lookup_i2c_gpio(struct radeon_device *rdev, |
| 64 | *dev, uint8_t id) | 65 | uint8_t id) |
| 65 | { | 66 | { |
| 66 | struct radeon_device *rdev = dev->dev_private; | ||
| 67 | struct atom_context *ctx = rdev->mode_info.atom_context; | 67 | struct atom_context *ctx = rdev->mode_info.atom_context; |
| 68 | ATOM_GPIO_I2C_ASSIGMENT gpio; | 68 | ATOM_GPIO_I2C_ASSIGMENT *gpio; |
| 69 | struct radeon_i2c_bus_rec i2c; | 69 | struct radeon_i2c_bus_rec i2c; |
| 70 | int index = GetIndexIntoMasterTable(DATA, GPIO_I2C_Info); | 70 | int index = GetIndexIntoMasterTable(DATA, GPIO_I2C_Info); |
| 71 | struct _ATOM_GPIO_I2C_INFO *i2c_info; | 71 | struct _ATOM_GPIO_I2C_INFO *i2c_info; |
| 72 | uint16_t data_offset; | 72 | uint16_t data_offset; |
| 73 | int i; | ||
| 73 | 74 | ||
| 74 | memset(&i2c, 0, sizeof(struct radeon_i2c_bus_rec)); | 75 | memset(&i2c, 0, sizeof(struct radeon_i2c_bus_rec)); |
| 75 | i2c.valid = false; | 76 | i2c.valid = false; |
| @@ -78,34 +79,121 @@ static inline struct radeon_i2c_bus_rec radeon_lookup_gpio(struct drm_device | |||
| 78 | 79 | ||
| 79 | i2c_info = (struct _ATOM_GPIO_I2C_INFO *)(ctx->bios + data_offset); | 80 | i2c_info = (struct _ATOM_GPIO_I2C_INFO *)(ctx->bios + data_offset); |
| 80 | 81 | ||
| 81 | gpio = i2c_info->asGPIO_Info[id]; | 82 | |
| 82 | 83 | for (i = 0; i < ATOM_MAX_SUPPORTED_DEVICE; i++) { | |
| 83 | i2c.mask_clk_reg = le16_to_cpu(gpio.usClkMaskRegisterIndex) * 4; | 84 | gpio = &i2c_info->asGPIO_Info[i]; |
| 84 | i2c.mask_data_reg = le16_to_cpu(gpio.usDataMaskRegisterIndex) * 4; | 85 | |
| 85 | i2c.put_clk_reg = le16_to_cpu(gpio.usClkEnRegisterIndex) * 4; | 86 | if (gpio->sucI2cId.ucAccess == id) { |
| 86 | i2c.put_data_reg = le16_to_cpu(gpio.usDataEnRegisterIndex) * 4; | 87 | i2c.mask_clk_reg = le16_to_cpu(gpio->usClkMaskRegisterIndex) * 4; |
| 87 | i2c.get_clk_reg = le16_to_cpu(gpio.usClkY_RegisterIndex) * 4; | 88 | i2c.mask_data_reg = le16_to_cpu(gpio->usDataMaskRegisterIndex) * 4; |
| 88 | i2c.get_data_reg = le16_to_cpu(gpio.usDataY_RegisterIndex) * 4; | 89 | i2c.en_clk_reg = le16_to_cpu(gpio->usClkEnRegisterIndex) * 4; |
| 89 | i2c.a_clk_reg = le16_to_cpu(gpio.usClkA_RegisterIndex) * 4; | 90 | i2c.en_data_reg = le16_to_cpu(gpio->usDataEnRegisterIndex) * 4; |
| 90 | i2c.a_data_reg = le16_to_cpu(gpio.usDataA_RegisterIndex) * 4; | 91 | i2c.y_clk_reg = le16_to_cpu(gpio->usClkY_RegisterIndex) * 4; |
| 91 | i2c.mask_clk_mask = (1 << gpio.ucClkMaskShift); | 92 | i2c.y_data_reg = le16_to_cpu(gpio->usDataY_RegisterIndex) * 4; |
| 92 | i2c.mask_data_mask = (1 << gpio.ucDataMaskShift); | 93 | i2c.a_clk_reg = le16_to_cpu(gpio->usClkA_RegisterIndex) * 4; |
| 93 | i2c.put_clk_mask = (1 << gpio.ucClkEnShift); | 94 | i2c.a_data_reg = le16_to_cpu(gpio->usDataA_RegisterIndex) * 4; |
| 94 | i2c.put_data_mask = (1 << gpio.ucDataEnShift); | 95 | i2c.mask_clk_mask = (1 << gpio->ucClkMaskShift); |
| 95 | i2c.get_clk_mask = (1 << gpio.ucClkY_Shift); | 96 | i2c.mask_data_mask = (1 << gpio->ucDataMaskShift); |
| 96 | i2c.get_data_mask = (1 << gpio.ucDataY_Shift); | 97 | i2c.en_clk_mask = (1 << gpio->ucClkEnShift); |
| 97 | i2c.a_clk_mask = (1 << gpio.ucClkA_Shift); | 98 | i2c.en_data_mask = (1 << gpio->ucDataEnShift); |
| 98 | i2c.a_data_mask = (1 << gpio.ucDataA_Shift); | 99 | i2c.y_clk_mask = (1 << gpio->ucClkY_Shift); |
| 99 | i2c.valid = true; | 100 | i2c.y_data_mask = (1 << gpio->ucDataY_Shift); |
| 101 | i2c.a_clk_mask = (1 << gpio->ucClkA_Shift); | ||
| 102 | i2c.a_data_mask = (1 << gpio->ucDataA_Shift); | ||
| 103 | |||
| 104 | if (gpio->sucI2cId.sbfAccess.bfHW_Capable) | ||
| 105 | i2c.hw_capable = true; | ||
| 106 | else | ||
| 107 | i2c.hw_capable = false; | ||
| 108 | |||
| 109 | if (gpio->sucI2cId.ucAccess == 0xa0) | ||
| 110 | i2c.mm_i2c = true; | ||
| 111 | else | ||
| 112 | i2c.mm_i2c = false; | ||
| 113 | |||
| 114 | i2c.i2c_id = gpio->sucI2cId.ucAccess; | ||
| 115 | |||
| 116 | i2c.valid = true; | ||
| 117 | } | ||
| 118 | } | ||
| 100 | 119 | ||
| 101 | return i2c; | 120 | return i2c; |
| 102 | } | 121 | } |
| 103 | 122 | ||
| 123 | static inline struct radeon_gpio_rec radeon_lookup_gpio(struct radeon_device *rdev, | ||
| 124 | u8 id) | ||
| 125 | { | ||
| 126 | struct atom_context *ctx = rdev->mode_info.atom_context; | ||
| 127 | struct radeon_gpio_rec gpio; | ||
| 128 | int index = GetIndexIntoMasterTable(DATA, GPIO_Pin_LUT); | ||
| 129 | struct _ATOM_GPIO_PIN_LUT *gpio_info; | ||
| 130 | ATOM_GPIO_PIN_ASSIGNMENT *pin; | ||
| 131 | u16 data_offset, size; | ||
| 132 | int i, num_indices; | ||
| 133 | |||
| 134 | memset(&gpio, 0, sizeof(struct radeon_gpio_rec)); | ||
| 135 | gpio.valid = false; | ||
| 136 | |||
| 137 | atom_parse_data_header(ctx, index, &size, NULL, NULL, &data_offset); | ||
| 138 | |||
| 139 | gpio_info = (struct _ATOM_GPIO_PIN_LUT *)(ctx->bios + data_offset); | ||
| 140 | |||
| 141 | num_indices = (size - sizeof(ATOM_COMMON_TABLE_HEADER)) / sizeof(ATOM_GPIO_PIN_ASSIGNMENT); | ||
| 142 | |||
| 143 | for (i = 0; i < num_indices; i++) { | ||
| 144 | pin = &gpio_info->asGPIO_Pin[i]; | ||
| 145 | if (id == pin->ucGPIO_ID) { | ||
| 146 | gpio.id = pin->ucGPIO_ID; | ||
| 147 | gpio.reg = pin->usGpioPin_AIndex * 4; | ||
| 148 | gpio.mask = (1 << pin->ucGpioPinBitShift); | ||
| 149 | gpio.valid = true; | ||
| 150 | break; | ||
| 151 | } | ||
| 152 | } | ||
| 153 | |||
| 154 | return gpio; | ||
| 155 | } | ||
| 156 | |||
| 157 | static struct radeon_hpd radeon_atom_get_hpd_info_from_gpio(struct radeon_device *rdev, | ||
| 158 | struct radeon_gpio_rec *gpio) | ||
| 159 | { | ||
| 160 | struct radeon_hpd hpd; | ||
| 161 | hpd.gpio = *gpio; | ||
| 162 | if (gpio->reg == AVIVO_DC_GPIO_HPD_A) { | ||
| 163 | switch(gpio->mask) { | ||
| 164 | case (1 << 0): | ||
| 165 | hpd.hpd = RADEON_HPD_1; | ||
| 166 | break; | ||
| 167 | case (1 << 8): | ||
| 168 | hpd.hpd = RADEON_HPD_2; | ||
| 169 | break; | ||
| 170 | case (1 << 16): | ||
| 171 | hpd.hpd = RADEON_HPD_3; | ||
| 172 | break; | ||
| 173 | case (1 << 24): | ||
| 174 | hpd.hpd = RADEON_HPD_4; | ||
| 175 | break; | ||
| 176 | case (1 << 26): | ||
| 177 | hpd.hpd = RADEON_HPD_5; | ||
| 178 | break; | ||
| 179 | case (1 << 28): | ||
| 180 | hpd.hpd = RADEON_HPD_6; | ||
| 181 | break; | ||
| 182 | default: | ||
| 183 | hpd.hpd = RADEON_HPD_NONE; | ||
| 184 | break; | ||
| 185 | } | ||
| 186 | } else | ||
| 187 | hpd.hpd = RADEON_HPD_NONE; | ||
| 188 | return hpd; | ||
| 189 | } | ||
| 190 | |||
| 104 | static bool radeon_atom_apply_quirks(struct drm_device *dev, | 191 | static bool radeon_atom_apply_quirks(struct drm_device *dev, |
| 105 | uint32_t supported_device, | 192 | uint32_t supported_device, |
| 106 | int *connector_type, | 193 | int *connector_type, |
| 107 | struct radeon_i2c_bus_rec *i2c_bus, | 194 | struct radeon_i2c_bus_rec *i2c_bus, |
| 108 | uint16_t *line_mux) | 195 | uint16_t *line_mux, |
| 196 | struct radeon_hpd *hpd) | ||
| 109 | { | 197 | { |
| 110 | 198 | ||
| 111 | /* Asus M2A-VM HDMI board lists the DVI port as HDMI */ | 199 | /* Asus M2A-VM HDMI board lists the DVI port as HDMI */ |
| @@ -135,6 +223,23 @@ static bool radeon_atom_apply_quirks(struct drm_device *dev, | |||
| 135 | } | 223 | } |
| 136 | } | 224 | } |
| 137 | 225 | ||
| 226 | /* HIS X1300 is DVI+VGA, not DVI+DVI */ | ||
| 227 | if ((dev->pdev->device == 0x7146) && | ||
| 228 | (dev->pdev->subsystem_vendor == 0x17af) && | ||
| 229 | (dev->pdev->subsystem_device == 0x2058)) { | ||
| 230 | if (supported_device == ATOM_DEVICE_DFP1_SUPPORT) | ||
| 231 | return false; | ||
| 232 | } | ||
| 233 | |||
| 234 | /* Gigabyte X1300 is DVI+VGA, not DVI+DVI */ | ||
| 235 | if ((dev->pdev->device == 0x7142) && | ||
| 236 | (dev->pdev->subsystem_vendor == 0x1458) && | ||
| 237 | (dev->pdev->subsystem_device == 0x2134)) { | ||
| 238 | if (supported_device == ATOM_DEVICE_DFP1_SUPPORT) | ||
| 239 | return false; | ||
| 240 | } | ||
| 241 | |||
| 242 | |||
| 138 | /* Funky macbooks */ | 243 | /* Funky macbooks */ |
| 139 | if ((dev->pdev->device == 0x71C5) && | 244 | if ((dev->pdev->device == 0x71C5) && |
| 140 | (dev->pdev->subsystem_vendor == 0x106b) && | 245 | (dev->pdev->subsystem_vendor == 0x106b) && |
| @@ -172,6 +277,15 @@ static bool radeon_atom_apply_quirks(struct drm_device *dev, | |||
| 172 | } | 277 | } |
| 173 | } | 278 | } |
| 174 | 279 | ||
| 280 | /* Acer laptop reports DVI-D as DVI-I */ | ||
| 281 | if ((dev->pdev->device == 0x95c4) && | ||
| 282 | (dev->pdev->subsystem_vendor == 0x1025) && | ||
| 283 | (dev->pdev->subsystem_device == 0x013c)) { | ||
| 284 | if ((*connector_type == DRM_MODE_CONNECTOR_DVII) && | ||
| 285 | (supported_device == ATOM_DEVICE_DFP1_SUPPORT)) | ||
| 286 | *connector_type = DRM_MODE_CONNECTOR_DVID; | ||
| 287 | } | ||
| 288 | |||
| 175 | return true; | 289 | return true; |
| 176 | } | 290 | } |
| 177 | 291 | ||
| @@ -240,16 +354,18 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 240 | struct radeon_mode_info *mode_info = &rdev->mode_info; | 354 | struct radeon_mode_info *mode_info = &rdev->mode_info; |
| 241 | struct atom_context *ctx = mode_info->atom_context; | 355 | struct atom_context *ctx = mode_info->atom_context; |
| 242 | int index = GetIndexIntoMasterTable(DATA, Object_Header); | 356 | int index = GetIndexIntoMasterTable(DATA, Object_Header); |
| 243 | uint16_t size, data_offset; | 357 | u16 size, data_offset; |
| 244 | uint8_t frev, crev, line_mux = 0; | 358 | u8 frev, crev; |
| 245 | ATOM_CONNECTOR_OBJECT_TABLE *con_obj; | 359 | ATOM_CONNECTOR_OBJECT_TABLE *con_obj; |
| 246 | ATOM_DISPLAY_OBJECT_PATH_TABLE *path_obj; | 360 | ATOM_DISPLAY_OBJECT_PATH_TABLE *path_obj; |
| 247 | ATOM_OBJECT_HEADER *obj_header; | 361 | ATOM_OBJECT_HEADER *obj_header; |
| 248 | int i, j, path_size, device_support; | 362 | int i, j, path_size, device_support; |
| 249 | int connector_type; | 363 | int connector_type; |
| 250 | uint16_t igp_lane_info, conn_id, connector_object_id; | 364 | u16 igp_lane_info, conn_id, connector_object_id; |
| 251 | bool linkb; | 365 | bool linkb; |
| 252 | struct radeon_i2c_bus_rec ddc_bus; | 366 | struct radeon_i2c_bus_rec ddc_bus; |
| 367 | struct radeon_gpio_rec gpio; | ||
| 368 | struct radeon_hpd hpd; | ||
| 253 | 369 | ||
| 254 | atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); | 370 | atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); |
| 255 | 371 | ||
| @@ -276,7 +392,6 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 276 | path = (ATOM_DISPLAY_OBJECT_PATH *) addr; | 392 | path = (ATOM_DISPLAY_OBJECT_PATH *) addr; |
| 277 | path_size += le16_to_cpu(path->usSize); | 393 | path_size += le16_to_cpu(path->usSize); |
| 278 | linkb = false; | 394 | linkb = false; |
| 279 | |||
| 280 | if (device_support & le16_to_cpu(path->usDeviceTag)) { | 395 | if (device_support & le16_to_cpu(path->usDeviceTag)) { |
| 281 | uint8_t con_obj_id, con_obj_num, con_obj_type; | 396 | uint8_t con_obj_id, con_obj_num, con_obj_type; |
| 282 | 397 | ||
| @@ -377,10 +492,9 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 377 | } | 492 | } |
| 378 | } | 493 | } |
| 379 | 494 | ||
| 380 | /* look up gpio for ddc */ | 495 | /* look up gpio for ddc, hpd */ |
| 381 | if ((le16_to_cpu(path->usDeviceTag) & | 496 | if ((le16_to_cpu(path->usDeviceTag) & |
| 382 | (ATOM_DEVICE_TV_SUPPORT | ATOM_DEVICE_CV_SUPPORT)) | 497 | (ATOM_DEVICE_TV_SUPPORT | ATOM_DEVICE_CV_SUPPORT)) == 0) { |
| 383 | == 0) { | ||
| 384 | for (j = 0; j < con_obj->ucNumberOfObjects; j++) { | 498 | for (j = 0; j < con_obj->ucNumberOfObjects; j++) { |
| 385 | if (le16_to_cpu(path->usConnObjectId) == | 499 | if (le16_to_cpu(path->usConnObjectId) == |
| 386 | le16_to_cpu(con_obj->asObjects[j]. | 500 | le16_to_cpu(con_obj->asObjects[j]. |
| @@ -394,21 +508,34 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 394 | asObjects[j]. | 508 | asObjects[j]. |
| 395 | usRecordOffset)); | 509 | usRecordOffset)); |
| 396 | ATOM_I2C_RECORD *i2c_record; | 510 | ATOM_I2C_RECORD *i2c_record; |
| 511 | ATOM_HPD_INT_RECORD *hpd_record; | ||
| 512 | ATOM_I2C_ID_CONFIG_ACCESS *i2c_config; | ||
| 513 | hpd.hpd = RADEON_HPD_NONE; | ||
| 397 | 514 | ||
| 398 | while (record->ucRecordType > 0 | 515 | while (record->ucRecordType > 0 |
| 399 | && record-> | 516 | && record-> |
| 400 | ucRecordType <= | 517 | ucRecordType <= |
| 401 | ATOM_MAX_OBJECT_RECORD_NUMBER) { | 518 | ATOM_MAX_OBJECT_RECORD_NUMBER) { |
| 402 | switch (record-> | 519 | switch (record->ucRecordType) { |
| 403 | ucRecordType) { | ||
| 404 | case ATOM_I2C_RECORD_TYPE: | 520 | case ATOM_I2C_RECORD_TYPE: |
| 405 | i2c_record = | 521 | i2c_record = |
| 406 | (ATOM_I2C_RECORD | 522 | (ATOM_I2C_RECORD *) |
| 407 | *) record; | 523 | record; |
| 408 | line_mux = | 524 | i2c_config = |
| 409 | i2c_record-> | 525 | (ATOM_I2C_ID_CONFIG_ACCESS *) |
| 410 | sucI2cId. | 526 | &i2c_record->sucI2cId; |
| 411 | bfI2C_LineMux; | 527 | ddc_bus = radeon_lookup_i2c_gpio(rdev, |
| 528 | i2c_config-> | ||
| 529 | ucAccess); | ||
| 530 | break; | ||
| 531 | case ATOM_HPD_INT_RECORD_TYPE: | ||
| 532 | hpd_record = | ||
| 533 | (ATOM_HPD_INT_RECORD *) | ||
| 534 | record; | ||
| 535 | gpio = radeon_lookup_gpio(rdev, | ||
| 536 | hpd_record->ucHPDIntGPIOID); | ||
| 537 | hpd = radeon_atom_get_hpd_info_from_gpio(rdev, &gpio); | ||
| 538 | hpd.plugged_state = hpd_record->ucPlugged_PinState; | ||
| 412 | break; | 539 | break; |
| 413 | } | 540 | } |
| 414 | record = | 541 | record = |
| @@ -421,24 +548,16 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 421 | break; | 548 | break; |
| 422 | } | 549 | } |
| 423 | } | 550 | } |
| 424 | } else | 551 | } else { |
| 425 | line_mux = 0; | 552 | hpd.hpd = RADEON_HPD_NONE; |
| 426 | |||
| 427 | if ((le16_to_cpu(path->usDeviceTag) == | ||
| 428 | ATOM_DEVICE_TV1_SUPPORT) | ||
| 429 | || (le16_to_cpu(path->usDeviceTag) == | ||
| 430 | ATOM_DEVICE_TV2_SUPPORT) | ||
| 431 | || (le16_to_cpu(path->usDeviceTag) == | ||
| 432 | ATOM_DEVICE_CV_SUPPORT)) | ||
| 433 | ddc_bus.valid = false; | 553 | ddc_bus.valid = false; |
| 434 | else | 554 | } |
| 435 | ddc_bus = radeon_lookup_gpio(dev, line_mux); | ||
| 436 | 555 | ||
| 437 | conn_id = le16_to_cpu(path->usConnObjectId); | 556 | conn_id = le16_to_cpu(path->usConnObjectId); |
| 438 | 557 | ||
| 439 | if (!radeon_atom_apply_quirks | 558 | if (!radeon_atom_apply_quirks |
| 440 | (dev, le16_to_cpu(path->usDeviceTag), &connector_type, | 559 | (dev, le16_to_cpu(path->usDeviceTag), &connector_type, |
| 441 | &ddc_bus, &conn_id)) | 560 | &ddc_bus, &conn_id, &hpd)) |
| 442 | continue; | 561 | continue; |
| 443 | 562 | ||
| 444 | radeon_add_atom_connector(dev, | 563 | radeon_add_atom_connector(dev, |
| @@ -447,7 +566,8 @@ bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev) | |||
| 447 | usDeviceTag), | 566 | usDeviceTag), |
| 448 | connector_type, &ddc_bus, | 567 | connector_type, &ddc_bus, |
| 449 | linkb, igp_lane_info, | 568 | linkb, igp_lane_info, |
| 450 | connector_object_id); | 569 | connector_object_id, |
| 570 | &hpd); | ||
| 451 | 571 | ||
| 452 | } | 572 | } |
| 453 | } | 573 | } |
| @@ -502,6 +622,7 @@ struct bios_connector { | |||
| 502 | uint16_t devices; | 622 | uint16_t devices; |
| 503 | int connector_type; | 623 | int connector_type; |
| 504 | struct radeon_i2c_bus_rec ddc_bus; | 624 | struct radeon_i2c_bus_rec ddc_bus; |
| 625 | struct radeon_hpd hpd; | ||
| 505 | }; | 626 | }; |
| 506 | 627 | ||
| 507 | bool radeon_get_atom_connector_info_from_supported_devices_table(struct | 628 | bool radeon_get_atom_connector_info_from_supported_devices_table(struct |
| @@ -517,7 +638,7 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 517 | uint16_t device_support; | 638 | uint16_t device_support; |
| 518 | uint8_t dac; | 639 | uint8_t dac; |
| 519 | union atom_supported_devices *supported_devices; | 640 | union atom_supported_devices *supported_devices; |
| 520 | int i, j; | 641 | int i, j, max_device; |
| 521 | struct bios_connector bios_connectors[ATOM_MAX_SUPPORTED_DEVICE]; | 642 | struct bios_connector bios_connectors[ATOM_MAX_SUPPORTED_DEVICE]; |
| 522 | 643 | ||
| 523 | atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); | 644 | atom_parse_data_header(ctx, index, &size, &frev, &crev, &data_offset); |
| @@ -527,7 +648,12 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 527 | 648 | ||
| 528 | device_support = le16_to_cpu(supported_devices->info.usDeviceSupport); | 649 | device_support = le16_to_cpu(supported_devices->info.usDeviceSupport); |
| 529 | 650 | ||
| 530 | for (i = 0; i < ATOM_MAX_SUPPORTED_DEVICE; i++) { | 651 | if (frev > 1) |
| 652 | max_device = ATOM_MAX_SUPPORTED_DEVICE; | ||
| 653 | else | ||
| 654 | max_device = ATOM_MAX_SUPPORTED_DEVICE_INFO; | ||
| 655 | |||
| 656 | for (i = 0; i < max_device; i++) { | ||
| 531 | ATOM_CONNECTOR_INFO_I2C ci = | 657 | ATOM_CONNECTOR_INFO_I2C ci = |
| 532 | supported_devices->info.asConnInfo[i]; | 658 | supported_devices->info.asConnInfo[i]; |
| 533 | 659 | ||
| @@ -553,22 +679,8 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 553 | 679 | ||
| 554 | dac = ci.sucConnectorInfo.sbfAccess.bfAssociatedDAC; | 680 | dac = ci.sucConnectorInfo.sbfAccess.bfAssociatedDAC; |
| 555 | 681 | ||
| 556 | if ((rdev->family == CHIP_RS690) || | 682 | bios_connectors[i].line_mux = |
| 557 | (rdev->family == CHIP_RS740)) { | 683 | ci.sucI2cId.ucAccess; |
| 558 | if ((i == ATOM_DEVICE_DFP2_INDEX) | ||
| 559 | && (ci.sucI2cId.sbfAccess.bfI2C_LineMux == 2)) | ||
| 560 | bios_connectors[i].line_mux = | ||
| 561 | ci.sucI2cId.sbfAccess.bfI2C_LineMux + 1; | ||
| 562 | else if ((i == ATOM_DEVICE_DFP3_INDEX) | ||
| 563 | && (ci.sucI2cId.sbfAccess.bfI2C_LineMux == 1)) | ||
| 564 | bios_connectors[i].line_mux = | ||
| 565 | ci.sucI2cId.sbfAccess.bfI2C_LineMux + 1; | ||
| 566 | else | ||
| 567 | bios_connectors[i].line_mux = | ||
| 568 | ci.sucI2cId.sbfAccess.bfI2C_LineMux; | ||
| 569 | } else | ||
| 570 | bios_connectors[i].line_mux = | ||
| 571 | ci.sucI2cId.sbfAccess.bfI2C_LineMux; | ||
| 572 | 684 | ||
| 573 | /* give tv unique connector ids */ | 685 | /* give tv unique connector ids */ |
| 574 | if (i == ATOM_DEVICE_TV1_INDEX) { | 686 | if (i == ATOM_DEVICE_TV1_INDEX) { |
| @@ -582,8 +694,30 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 582 | bios_connectors[i].line_mux = 52; | 694 | bios_connectors[i].line_mux = 52; |
| 583 | } else | 695 | } else |
| 584 | bios_connectors[i].ddc_bus = | 696 | bios_connectors[i].ddc_bus = |
| 585 | radeon_lookup_gpio(dev, | 697 | radeon_lookup_i2c_gpio(rdev, |
| 586 | bios_connectors[i].line_mux); | 698 | bios_connectors[i].line_mux); |
| 699 | |||
| 700 | if ((crev > 1) && (frev > 1)) { | ||
| 701 | u8 isb = supported_devices->info_2d1.asIntSrcInfo[i].ucIntSrcBitmap; | ||
| 702 | switch (isb) { | ||
| 703 | case 0x4: | ||
| 704 | bios_connectors[i].hpd.hpd = RADEON_HPD_1; | ||
| 705 | break; | ||
| 706 | case 0xa: | ||
| 707 | bios_connectors[i].hpd.hpd = RADEON_HPD_2; | ||
| 708 | break; | ||
| 709 | default: | ||
| 710 | bios_connectors[i].hpd.hpd = RADEON_HPD_NONE; | ||
| 711 | break; | ||
| 712 | } | ||
| 713 | } else { | ||
| 714 | if (i == ATOM_DEVICE_DFP1_INDEX) | ||
| 715 | bios_connectors[i].hpd.hpd = RADEON_HPD_1; | ||
| 716 | else if (i == ATOM_DEVICE_DFP2_INDEX) | ||
| 717 | bios_connectors[i].hpd.hpd = RADEON_HPD_2; | ||
| 718 | else | ||
| 719 | bios_connectors[i].hpd.hpd = RADEON_HPD_NONE; | ||
| 720 | } | ||
| 587 | 721 | ||
| 588 | /* Always set the connector type to VGA for CRT1/CRT2. if they are | 722 | /* Always set the connector type to VGA for CRT1/CRT2. if they are |
| 589 | * shared with a DVI port, we'll pick up the DVI connector when we | 723 | * shared with a DVI port, we'll pick up the DVI connector when we |
| @@ -595,7 +729,8 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 595 | 729 | ||
| 596 | if (!radeon_atom_apply_quirks | 730 | if (!radeon_atom_apply_quirks |
| 597 | (dev, (1 << i), &bios_connectors[i].connector_type, | 731 | (dev, (1 << i), &bios_connectors[i].connector_type, |
| 598 | &bios_connectors[i].ddc_bus, &bios_connectors[i].line_mux)) | 732 | &bios_connectors[i].ddc_bus, &bios_connectors[i].line_mux, |
| 733 | &bios_connectors[i].hpd)) | ||
| 599 | continue; | 734 | continue; |
| 600 | 735 | ||
| 601 | bios_connectors[i].valid = true; | 736 | bios_connectors[i].valid = true; |
| @@ -617,9 +752,9 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 617 | } | 752 | } |
| 618 | 753 | ||
| 619 | /* combine shared connectors */ | 754 | /* combine shared connectors */ |
| 620 | for (i = 0; i < ATOM_MAX_SUPPORTED_DEVICE; i++) { | 755 | for (i = 0; i < max_device; i++) { |
| 621 | if (bios_connectors[i].valid) { | 756 | if (bios_connectors[i].valid) { |
| 622 | for (j = 0; j < ATOM_MAX_SUPPORTED_DEVICE; j++) { | 757 | for (j = 0; j < max_device; j++) { |
| 623 | if (bios_connectors[j].valid && (i != j)) { | 758 | if (bios_connectors[j].valid && (i != j)) { |
| 624 | if (bios_connectors[i].line_mux == | 759 | if (bios_connectors[i].line_mux == |
| 625 | bios_connectors[j].line_mux) { | 760 | bios_connectors[j].line_mux) { |
| @@ -643,6 +778,10 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 643 | bios_connectors[i]. | 778 | bios_connectors[i]. |
| 644 | connector_type = | 779 | connector_type = |
| 645 | DRM_MODE_CONNECTOR_DVII; | 780 | DRM_MODE_CONNECTOR_DVII; |
| 781 | if (bios_connectors[j].devices & | ||
| 782 | (ATOM_DEVICE_DFP_SUPPORT)) | ||
| 783 | bios_connectors[i].hpd = | ||
| 784 | bios_connectors[j].hpd; | ||
| 646 | bios_connectors[j]. | 785 | bios_connectors[j]. |
| 647 | valid = false; | 786 | valid = false; |
| 648 | } | 787 | } |
| @@ -653,7 +792,7 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 653 | } | 792 | } |
| 654 | 793 | ||
| 655 | /* add the connectors */ | 794 | /* add the connectors */ |
| 656 | for (i = 0; i < ATOM_MAX_SUPPORTED_DEVICE; i++) { | 795 | for (i = 0; i < max_device; i++) { |
| 657 | if (bios_connectors[i].valid) { | 796 | if (bios_connectors[i].valid) { |
| 658 | uint16_t connector_object_id = | 797 | uint16_t connector_object_id = |
| 659 | atombios_get_connector_object_id(dev, | 798 | atombios_get_connector_object_id(dev, |
| @@ -666,7 +805,8 @@ bool radeon_get_atom_connector_info_from_supported_devices_table(struct | |||
| 666 | connector_type, | 805 | connector_type, |
| 667 | &bios_connectors[i].ddc_bus, | 806 | &bios_connectors[i].ddc_bus, |
| 668 | false, 0, | 807 | false, 0, |
| 669 | connector_object_id); | 808 | connector_object_id, |
| 809 | &bios_connectors[i].hpd); | ||
| 670 | } | 810 | } |
| 671 | } | 811 | } |
| 672 | 812 | ||
| @@ -731,7 +871,8 @@ bool radeon_atom_get_clock_info(struct drm_device *dev) | |||
| 731 | * pre-DCE 3.0 r6xx hardware. This might need to be adjusted per | 871 | * pre-DCE 3.0 r6xx hardware. This might need to be adjusted per |
| 732 | * family. | 872 | * family. |
| 733 | */ | 873 | */ |
| 734 | p1pll->pll_out_min = 64800; | 874 | if (!radeon_new_pll) |
| 875 | p1pll->pll_out_min = 64800; | ||
| 735 | } | 876 | } |
| 736 | 877 | ||
| 737 | p1pll->pll_in_min = | 878 | p1pll->pll_in_min = |
| @@ -861,6 +1002,7 @@ static struct radeon_atom_ss *radeon_atombios_get_ss_info(struct | |||
| 861 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; | 1002 | struct _ATOM_SPREAD_SPECTRUM_INFO *ss_info; |
| 862 | uint8_t frev, crev; | 1003 | uint8_t frev, crev; |
| 863 | struct radeon_atom_ss *ss = NULL; | 1004 | struct radeon_atom_ss *ss = NULL; |
| 1005 | int i; | ||
| 864 | 1006 | ||
| 865 | if (id > ATOM_MAX_SS_ENTRY) | 1007 | if (id > ATOM_MAX_SS_ENTRY) |
| 866 | return NULL; | 1008 | return NULL; |
| @@ -878,12 +1020,17 @@ static struct radeon_atom_ss *radeon_atombios_get_ss_info(struct | |||
| 878 | if (!ss) | 1020 | if (!ss) |
| 879 | return NULL; | 1021 | return NULL; |
| 880 | 1022 | ||
| 881 | ss->percentage = le16_to_cpu(ss_info->asSS_Info[id].usSpreadSpectrumPercentage); | 1023 | for (i = 0; i < ATOM_MAX_SS_ENTRY; i++) { |
| 882 | ss->type = ss_info->asSS_Info[id].ucSpreadSpectrumType; | 1024 | if (ss_info->asSS_Info[i].ucSS_Id == id) { |
| 883 | ss->step = ss_info->asSS_Info[id].ucSS_Step; | 1025 | ss->percentage = |
| 884 | ss->delay = ss_info->asSS_Info[id].ucSS_Delay; | 1026 | le16_to_cpu(ss_info->asSS_Info[i].usSpreadSpectrumPercentage); |
| 885 | ss->range = ss_info->asSS_Info[id].ucSS_Range; | 1027 | ss->type = ss_info->asSS_Info[i].ucSpreadSpectrumType; |
| 886 | ss->refdiv = ss_info->asSS_Info[id].ucRecommendedRef_Div; | 1028 | ss->step = ss_info->asSS_Info[i].ucSS_Step; |
| 1029 | ss->delay = ss_info->asSS_Info[i].ucSS_Delay; | ||
| 1030 | ss->range = ss_info->asSS_Info[i].ucSS_Range; | ||
| 1031 | ss->refdiv = ss_info->asSS_Info[i].ucRecommendedRef_Div; | ||
| 1032 | } | ||
| 1033 | } | ||
| 887 | } | 1034 | } |
| 888 | return ss; | 1035 | return ss; |
| 889 | } | 1036 | } |
| @@ -901,7 +1048,7 @@ struct radeon_encoder_atom_dig *radeon_atombios_get_lvds_info(struct | |||
| 901 | struct radeon_device *rdev = dev->dev_private; | 1048 | struct radeon_device *rdev = dev->dev_private; |
| 902 | struct radeon_mode_info *mode_info = &rdev->mode_info; | 1049 | struct radeon_mode_info *mode_info = &rdev->mode_info; |
| 903 | int index = GetIndexIntoMasterTable(DATA, LVDS_Info); | 1050 | int index = GetIndexIntoMasterTable(DATA, LVDS_Info); |
| 904 | uint16_t data_offset; | 1051 | uint16_t data_offset, misc; |
| 905 | union lvds_info *lvds_info; | 1052 | union lvds_info *lvds_info; |
| 906 | uint8_t frev, crev; | 1053 | uint8_t frev, crev; |
| 907 | struct radeon_encoder_atom_dig *lvds = NULL; | 1054 | struct radeon_encoder_atom_dig *lvds = NULL; |
| @@ -940,6 +1087,19 @@ struct radeon_encoder_atom_dig *radeon_atombios_get_lvds_info(struct | |||
| 940 | lvds->panel_pwr_delay = | 1087 | lvds->panel_pwr_delay = |
| 941 | le16_to_cpu(lvds_info->info.usOffDelayInMs); | 1088 | le16_to_cpu(lvds_info->info.usOffDelayInMs); |
| 942 | lvds->lvds_misc = lvds_info->info.ucLVDS_Misc; | 1089 | lvds->lvds_misc = lvds_info->info.ucLVDS_Misc; |
| 1090 | |||
| 1091 | misc = le16_to_cpu(lvds_info->info.sLCDTiming.susModeMiscInfo.usAccess); | ||
| 1092 | if (misc & ATOM_VSYNC_POLARITY) | ||
| 1093 | lvds->native_mode.flags |= DRM_MODE_FLAG_NVSYNC; | ||
| 1094 | if (misc & ATOM_HSYNC_POLARITY) | ||
| 1095 | lvds->native_mode.flags |= DRM_MODE_FLAG_NHSYNC; | ||
| 1096 | if (misc & ATOM_COMPOSITESYNC) | ||
| 1097 | lvds->native_mode.flags |= DRM_MODE_FLAG_CSYNC; | ||
| 1098 | if (misc & ATOM_INTERLACE) | ||
| 1099 | lvds->native_mode.flags |= DRM_MODE_FLAG_INTERLACE; | ||
| 1100 | if (misc & ATOM_DOUBLE_CLOCK_MODE) | ||
| 1101 | lvds->native_mode.flags |= DRM_MODE_FLAG_DBLSCAN; | ||
| 1102 | |||
| 943 | /* set crtc values */ | 1103 | /* set crtc values */ |
| 944 | drm_mode_set_crtcinfo(&lvds->native_mode, CRTC_INTERLACE_HALVE_V); | 1104 | drm_mode_set_crtcinfo(&lvds->native_mode, CRTC_INTERLACE_HALVE_V); |
| 945 | 1105 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_benchmark.c b/drivers/gpu/drm/radeon/radeon_benchmark.c index 10bd50a7db87..4ddfd4b5bc51 100644 --- a/drivers/gpu/drm/radeon/radeon_benchmark.c +++ b/drivers/gpu/drm/radeon/radeon_benchmark.c | |||
| @@ -29,8 +29,8 @@ | |||
| 29 | void radeon_benchmark_move(struct radeon_device *rdev, unsigned bsize, | 29 | void radeon_benchmark_move(struct radeon_device *rdev, unsigned bsize, |
| 30 | unsigned sdomain, unsigned ddomain) | 30 | unsigned sdomain, unsigned ddomain) |
| 31 | { | 31 | { |
| 32 | struct radeon_object *dobj = NULL; | 32 | struct radeon_bo *dobj = NULL; |
| 33 | struct radeon_object *sobj = NULL; | 33 | struct radeon_bo *sobj = NULL; |
| 34 | struct radeon_fence *fence = NULL; | 34 | struct radeon_fence *fence = NULL; |
| 35 | uint64_t saddr, daddr; | 35 | uint64_t saddr, daddr; |
| 36 | unsigned long start_jiffies; | 36 | unsigned long start_jiffies; |
| @@ -41,19 +41,27 @@ void radeon_benchmark_move(struct radeon_device *rdev, unsigned bsize, | |||
| 41 | 41 | ||
| 42 | size = bsize; | 42 | size = bsize; |
| 43 | n = 1024; | 43 | n = 1024; |
| 44 | r = radeon_object_create(rdev, NULL, size, true, sdomain, false, &sobj); | 44 | r = radeon_bo_create(rdev, NULL, size, true, sdomain, &sobj); |
| 45 | if (r) { | 45 | if (r) { |
| 46 | goto out_cleanup; | 46 | goto out_cleanup; |
| 47 | } | 47 | } |
| 48 | r = radeon_object_pin(sobj, sdomain, &saddr); | 48 | r = radeon_bo_reserve(sobj, false); |
| 49 | if (unlikely(r != 0)) | ||
| 50 | goto out_cleanup; | ||
| 51 | r = radeon_bo_pin(sobj, sdomain, &saddr); | ||
| 52 | radeon_bo_unreserve(sobj); | ||
| 49 | if (r) { | 53 | if (r) { |
| 50 | goto out_cleanup; | 54 | goto out_cleanup; |
| 51 | } | 55 | } |
| 52 | r = radeon_object_create(rdev, NULL, size, true, ddomain, false, &dobj); | 56 | r = radeon_bo_create(rdev, NULL, size, true, ddomain, &dobj); |
| 53 | if (r) { | 57 | if (r) { |
| 54 | goto out_cleanup; | 58 | goto out_cleanup; |
| 55 | } | 59 | } |
| 56 | r = radeon_object_pin(dobj, ddomain, &daddr); | 60 | r = radeon_bo_reserve(dobj, false); |
| 61 | if (unlikely(r != 0)) | ||
| 62 | goto out_cleanup; | ||
| 63 | r = radeon_bo_pin(dobj, ddomain, &daddr); | ||
| 64 | radeon_bo_unreserve(dobj); | ||
| 57 | if (r) { | 65 | if (r) { |
| 58 | goto out_cleanup; | 66 | goto out_cleanup; |
| 59 | } | 67 | } |
| @@ -109,12 +117,20 @@ void radeon_benchmark_move(struct radeon_device *rdev, unsigned bsize, | |||
| 109 | } | 117 | } |
| 110 | out_cleanup: | 118 | out_cleanup: |
| 111 | if (sobj) { | 119 | if (sobj) { |
| 112 | radeon_object_unpin(sobj); | 120 | r = radeon_bo_reserve(sobj, false); |
| 113 | radeon_object_unref(&sobj); | 121 | if (likely(r == 0)) { |
| 122 | radeon_bo_unpin(sobj); | ||
| 123 | radeon_bo_unreserve(sobj); | ||
| 124 | } | ||
| 125 | radeon_bo_unref(&sobj); | ||
| 114 | } | 126 | } |
| 115 | if (dobj) { | 127 | if (dobj) { |
| 116 | radeon_object_unpin(dobj); | 128 | r = radeon_bo_reserve(dobj, false); |
| 117 | radeon_object_unref(&dobj); | 129 | if (likely(r == 0)) { |
| 130 | radeon_bo_unpin(dobj); | ||
| 131 | radeon_bo_unreserve(dobj); | ||
| 132 | } | ||
| 133 | radeon_bo_unref(&dobj); | ||
| 118 | } | 134 | } |
| 119 | if (fence) { | 135 | if (fence) { |
| 120 | radeon_fence_unref(&fence); | 136 | radeon_fence_unref(&fence); |
diff --git a/drivers/gpu/drm/radeon/radeon_clocks.c b/drivers/gpu/drm/radeon/radeon_clocks.c index a81354167621..b062109efbee 100644 --- a/drivers/gpu/drm/radeon/radeon_clocks.c +++ b/drivers/gpu/drm/radeon/radeon_clocks.c | |||
| @@ -44,6 +44,10 @@ uint32_t radeon_legacy_get_engine_clock(struct radeon_device *rdev) | |||
| 44 | 44 | ||
| 45 | ref_div = | 45 | ref_div = |
| 46 | RREG32_PLL(RADEON_M_SPLL_REF_FB_DIV) & RADEON_M_SPLL_REF_DIV_MASK; | 46 | RREG32_PLL(RADEON_M_SPLL_REF_FB_DIV) & RADEON_M_SPLL_REF_DIV_MASK; |
| 47 | |||
| 48 | if (ref_div == 0) | ||
| 49 | return 0; | ||
| 50 | |||
| 47 | sclk = fb_div / ref_div; | 51 | sclk = fb_div / ref_div; |
| 48 | 52 | ||
| 49 | post_div = RREG32_PLL(RADEON_SCLK_CNTL) & RADEON_SCLK_SRC_SEL_MASK; | 53 | post_div = RREG32_PLL(RADEON_SCLK_CNTL) & RADEON_SCLK_SRC_SEL_MASK; |
| @@ -70,6 +74,10 @@ static uint32_t radeon_legacy_get_memory_clock(struct radeon_device *rdev) | |||
| 70 | 74 | ||
| 71 | ref_div = | 75 | ref_div = |
| 72 | RREG32_PLL(RADEON_M_SPLL_REF_FB_DIV) & RADEON_M_SPLL_REF_DIV_MASK; | 76 | RREG32_PLL(RADEON_M_SPLL_REF_FB_DIV) & RADEON_M_SPLL_REF_DIV_MASK; |
| 77 | |||
| 78 | if (ref_div == 0) | ||
| 79 | return 0; | ||
| 80 | |||
| 73 | mclk = fb_div / ref_div; | 81 | mclk = fb_div / ref_div; |
| 74 | 82 | ||
| 75 | post_div = RREG32_PLL(RADEON_MCLK_CNTL) & 0x7; | 83 | post_div = RREG32_PLL(RADEON_MCLK_CNTL) & 0x7; |
| @@ -98,8 +106,19 @@ void radeon_get_clock_info(struct drm_device *dev) | |||
| 98 | ret = radeon_combios_get_clock_info(dev); | 106 | ret = radeon_combios_get_clock_info(dev); |
| 99 | 107 | ||
| 100 | if (ret) { | 108 | if (ret) { |
| 101 | if (p1pll->reference_div < 2) | 109 | if (p1pll->reference_div < 2) { |
| 102 | p1pll->reference_div = 12; | 110 | if (!ASIC_IS_AVIVO(rdev)) { |
| 111 | u32 tmp = RREG32_PLL(RADEON_PPLL_REF_DIV); | ||
| 112 | if (ASIC_IS_R300(rdev)) | ||
| 113 | p1pll->reference_div = | ||
| 114 | (tmp & R300_PPLL_REF_DIV_ACC_MASK) >> R300_PPLL_REF_DIV_ACC_SHIFT; | ||
| 115 | else | ||
| 116 | p1pll->reference_div = tmp & RADEON_PPLL_REF_DIV_MASK; | ||
| 117 | if (p1pll->reference_div < 2) | ||
| 118 | p1pll->reference_div = 12; | ||
| 119 | } else | ||
| 120 | p1pll->reference_div = 12; | ||
| 121 | } | ||
| 103 | if (p2pll->reference_div < 2) | 122 | if (p2pll->reference_div < 2) |
| 104 | p2pll->reference_div = 12; | 123 | p2pll->reference_div = 12; |
| 105 | if (rdev->family < CHIP_RS600) { | 124 | if (rdev->family < CHIP_RS600) { |
diff --git a/drivers/gpu/drm/radeon/radeon_combios.c b/drivers/gpu/drm/radeon/radeon_combios.c index 5253cbf6db1f..c5021a3445de 100644 --- a/drivers/gpu/drm/radeon/radeon_combios.c +++ b/drivers/gpu/drm/radeon/radeon_combios.c | |||
| @@ -50,7 +50,8 @@ radeon_add_legacy_connector(struct drm_device *dev, | |||
| 50 | uint32_t supported_device, | 50 | uint32_t supported_device, |
| 51 | int connector_type, | 51 | int connector_type, |
| 52 | struct radeon_i2c_bus_rec *i2c_bus, | 52 | struct radeon_i2c_bus_rec *i2c_bus, |
| 53 | uint16_t connector_object_id); | 53 | uint16_t connector_object_id, |
| 54 | struct radeon_hpd *hpd); | ||
| 54 | 55 | ||
| 55 | /* from radeon_legacy_encoder.c */ | 56 | /* from radeon_legacy_encoder.c */ |
| 56 | extern void | 57 | extern void |
| @@ -442,38 +443,70 @@ static uint16_t combios_get_table_offset(struct drm_device *dev, | |||
| 442 | 443 | ||
| 443 | } | 444 | } |
| 444 | 445 | ||
| 445 | struct radeon_i2c_bus_rec combios_setup_i2c_bus(int ddc_line) | 446 | static struct radeon_i2c_bus_rec combios_setup_i2c_bus(struct radeon_device *rdev, |
| 447 | int ddc_line) | ||
| 446 | { | 448 | { |
| 447 | struct radeon_i2c_bus_rec i2c; | 449 | struct radeon_i2c_bus_rec i2c; |
| 448 | 450 | ||
| 449 | i2c.mask_clk_mask = RADEON_GPIO_EN_1; | 451 | if (ddc_line == RADEON_GPIOPAD_MASK) { |
| 450 | i2c.mask_data_mask = RADEON_GPIO_EN_0; | 452 | i2c.mask_clk_reg = RADEON_GPIOPAD_MASK; |
| 451 | i2c.a_clk_mask = RADEON_GPIO_A_1; | 453 | i2c.mask_data_reg = RADEON_GPIOPAD_MASK; |
| 452 | i2c.a_data_mask = RADEON_GPIO_A_0; | 454 | i2c.a_clk_reg = RADEON_GPIOPAD_A; |
| 453 | i2c.put_clk_mask = RADEON_GPIO_EN_1; | 455 | i2c.a_data_reg = RADEON_GPIOPAD_A; |
| 454 | i2c.put_data_mask = RADEON_GPIO_EN_0; | 456 | i2c.en_clk_reg = RADEON_GPIOPAD_EN; |
| 455 | i2c.get_clk_mask = RADEON_GPIO_Y_1; | 457 | i2c.en_data_reg = RADEON_GPIOPAD_EN; |
| 456 | i2c.get_data_mask = RADEON_GPIO_Y_0; | 458 | i2c.y_clk_reg = RADEON_GPIOPAD_Y; |
| 457 | if ((ddc_line == RADEON_LCD_GPIO_MASK) || | 459 | i2c.y_data_reg = RADEON_GPIOPAD_Y; |
| 458 | (ddc_line == RADEON_MDGPIO_EN_REG)) { | 460 | } else if (ddc_line == RADEON_MDGPIO_MASK) { |
| 459 | i2c.mask_clk_reg = ddc_line; | 461 | i2c.mask_clk_reg = RADEON_MDGPIO_MASK; |
| 460 | i2c.mask_data_reg = ddc_line; | 462 | i2c.mask_data_reg = RADEON_MDGPIO_MASK; |
| 461 | i2c.a_clk_reg = ddc_line; | 463 | i2c.a_clk_reg = RADEON_MDGPIO_A; |
| 462 | i2c.a_data_reg = ddc_line; | 464 | i2c.a_data_reg = RADEON_MDGPIO_A; |
| 463 | i2c.put_clk_reg = ddc_line; | 465 | i2c.en_clk_reg = RADEON_MDGPIO_EN; |
| 464 | i2c.put_data_reg = ddc_line; | 466 | i2c.en_data_reg = RADEON_MDGPIO_EN; |
| 465 | i2c.get_clk_reg = ddc_line + 4; | 467 | i2c.y_clk_reg = RADEON_MDGPIO_Y; |
| 466 | i2c.get_data_reg = ddc_line + 4; | 468 | i2c.y_data_reg = RADEON_MDGPIO_Y; |
| 467 | } else { | 469 | } else { |
| 470 | i2c.mask_clk_mask = RADEON_GPIO_EN_1; | ||
| 471 | i2c.mask_data_mask = RADEON_GPIO_EN_0; | ||
| 472 | i2c.a_clk_mask = RADEON_GPIO_A_1; | ||
| 473 | i2c.a_data_mask = RADEON_GPIO_A_0; | ||
| 474 | i2c.en_clk_mask = RADEON_GPIO_EN_1; | ||
| 475 | i2c.en_data_mask = RADEON_GPIO_EN_0; | ||
| 476 | i2c.y_clk_mask = RADEON_GPIO_Y_1; | ||
| 477 | i2c.y_data_mask = RADEON_GPIO_Y_0; | ||
| 478 | |||
| 468 | i2c.mask_clk_reg = ddc_line; | 479 | i2c.mask_clk_reg = ddc_line; |
| 469 | i2c.mask_data_reg = ddc_line; | 480 | i2c.mask_data_reg = ddc_line; |
| 470 | i2c.a_clk_reg = ddc_line; | 481 | i2c.a_clk_reg = ddc_line; |
| 471 | i2c.a_data_reg = ddc_line; | 482 | i2c.a_data_reg = ddc_line; |
| 472 | i2c.put_clk_reg = ddc_line; | 483 | i2c.en_clk_reg = ddc_line; |
| 473 | i2c.put_data_reg = ddc_line; | 484 | i2c.en_data_reg = ddc_line; |
| 474 | i2c.get_clk_reg = ddc_line; | 485 | i2c.y_clk_reg = ddc_line; |
| 475 | i2c.get_data_reg = ddc_line; | 486 | i2c.y_data_reg = ddc_line; |
| 487 | } | ||
| 488 | |||
| 489 | if (rdev->family < CHIP_R200) | ||
| 490 | i2c.hw_capable = false; | ||
| 491 | else { | ||
| 492 | switch (ddc_line) { | ||
| 493 | case RADEON_GPIO_VGA_DDC: | ||
| 494 | case RADEON_GPIO_DVI_DDC: | ||
| 495 | i2c.hw_capable = true; | ||
| 496 | break; | ||
| 497 | case RADEON_GPIO_MONID: | ||
| 498 | /* hw i2c on RADEON_GPIO_MONID doesn't seem to work | ||
| 499 | * reliably on some pre-r4xx hardware; not sure why. | ||
| 500 | */ | ||
| 501 | i2c.hw_capable = false; | ||
| 502 | break; | ||
| 503 | default: | ||
| 504 | i2c.hw_capable = false; | ||
| 505 | break; | ||
| 506 | } | ||
| 476 | } | 507 | } |
| 508 | i2c.mm_i2c = false; | ||
| 509 | i2c.i2c_id = 0; | ||
| 477 | 510 | ||
| 478 | if (ddc_line) | 511 | if (ddc_line) |
| 479 | i2c.valid = true; | 512 | i2c.valid = true; |
| @@ -495,7 +528,7 @@ bool radeon_combios_get_clock_info(struct drm_device *dev) | |||
| 495 | uint16_t sclk, mclk; | 528 | uint16_t sclk, mclk; |
| 496 | 529 | ||
| 497 | if (rdev->bios == NULL) | 530 | if (rdev->bios == NULL) |
| 498 | return NULL; | 531 | return false; |
| 499 | 532 | ||
| 500 | pll_info = combios_get_table_offset(dev, COMBIOS_PLL_INFO_TABLE); | 533 | pll_info = combios_get_table_offset(dev, COMBIOS_PLL_INFO_TABLE); |
| 501 | if (pll_info) { | 534 | if (pll_info) { |
| @@ -993,8 +1026,8 @@ static const struct radeon_tmds_pll default_tmds_pll[CHIP_LAST][4] = { | |||
| 993 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_R420 */ | 1026 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_R420 */ |
| 994 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_R423 */ | 1027 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_R423 */ |
| 995 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_RV410 */ | 1028 | {{0xffffffff, 0xb01cb}, {0, 0}, {0, 0}, {0, 0}}, /* CHIP_RV410 */ |
| 996 | {{15000, 0xb0155}, {0xffffffff, 0xb01cb}, {0, 0}, {0, 0}}, /* CHIP_RS400 */ | 1029 | { {0, 0}, {0, 0}, {0, 0}, {0, 0} }, /* CHIP_RS400 */ |
| 997 | {{15000, 0xb0155}, {0xffffffff, 0xb01cb}, {0, 0}, {0, 0}}, /* CHIP_RS480 */ | 1030 | { {0, 0}, {0, 0}, {0, 0}, {0, 0} }, /* CHIP_RS480 */ |
| 998 | }; | 1031 | }; |
| 999 | 1032 | ||
| 1000 | bool radeon_legacy_get_tmds_info_from_table(struct radeon_encoder *encoder, | 1033 | bool radeon_legacy_get_tmds_info_from_table(struct radeon_encoder *encoder, |
| @@ -1028,7 +1061,6 @@ bool radeon_legacy_get_tmds_info_from_combios(struct radeon_encoder *encoder, | |||
| 1028 | tmds_info = combios_get_table_offset(dev, COMBIOS_DFP_INFO_TABLE); | 1061 | tmds_info = combios_get_table_offset(dev, COMBIOS_DFP_INFO_TABLE); |
| 1029 | 1062 | ||
| 1030 | if (tmds_info) { | 1063 | if (tmds_info) { |
| 1031 | |||
| 1032 | ver = RBIOS8(tmds_info); | 1064 | ver = RBIOS8(tmds_info); |
| 1033 | DRM_INFO("DFP table revision: %d\n", ver); | 1065 | DRM_INFO("DFP table revision: %d\n", ver); |
| 1034 | if (ver == 3) { | 1066 | if (ver == 3) { |
| @@ -1063,51 +1095,139 @@ bool radeon_legacy_get_tmds_info_from_combios(struct radeon_encoder *encoder, | |||
| 1063 | tmds->tmds_pll[i].value); | 1095 | tmds->tmds_pll[i].value); |
| 1064 | } | 1096 | } |
| 1065 | } | 1097 | } |
| 1066 | } else | 1098 | } else { |
| 1067 | DRM_INFO("No TMDS info found in BIOS\n"); | 1099 | DRM_INFO("No TMDS info found in BIOS\n"); |
| 1100 | return false; | ||
| 1101 | } | ||
| 1068 | return true; | 1102 | return true; |
| 1069 | } | 1103 | } |
| 1070 | 1104 | ||
| 1071 | struct radeon_encoder_int_tmds *radeon_combios_get_tmds_info(struct radeon_encoder *encoder) | 1105 | bool radeon_legacy_get_ext_tmds_info_from_table(struct radeon_encoder *encoder, |
| 1106 | struct radeon_encoder_ext_tmds *tmds) | ||
| 1072 | { | 1107 | { |
| 1073 | struct radeon_encoder_int_tmds *tmds = NULL; | 1108 | struct drm_device *dev = encoder->base.dev; |
| 1074 | bool ret; | 1109 | struct radeon_device *rdev = dev->dev_private; |
| 1110 | struct radeon_i2c_bus_rec i2c_bus; | ||
| 1075 | 1111 | ||
| 1076 | tmds = kzalloc(sizeof(struct radeon_encoder_int_tmds), GFP_KERNEL); | 1112 | /* default for macs */ |
| 1113 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); | ||
| 1114 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1077 | 1115 | ||
| 1078 | if (!tmds) | 1116 | /* XXX some macs have duallink chips */ |
| 1079 | return NULL; | 1117 | switch (rdev->mode_info.connector_table) { |
| 1080 | 1118 | case CT_POWERBOOK_EXTERNAL: | |
| 1081 | ret = radeon_legacy_get_tmds_info_from_combios(encoder, tmds); | 1119 | case CT_MINI_EXTERNAL: |
| 1082 | if (ret == false) | 1120 | default: |
| 1083 | radeon_legacy_get_tmds_info_from_table(encoder, tmds); | 1121 | tmds->dvo_chip = DVO_SIL164; |
| 1122 | tmds->slave_addr = 0x70 >> 1; /* 7 bit addressing */ | ||
| 1123 | break; | ||
| 1124 | } | ||
| 1084 | 1125 | ||
| 1085 | return tmds; | 1126 | return true; |
| 1086 | } | 1127 | } |
| 1087 | 1128 | ||
| 1088 | void radeon_combios_get_ext_tmds_info(struct radeon_encoder *encoder) | 1129 | bool radeon_legacy_get_ext_tmds_info_from_combios(struct radeon_encoder *encoder, |
| 1130 | struct radeon_encoder_ext_tmds *tmds) | ||
| 1089 | { | 1131 | { |
| 1090 | struct drm_device *dev = encoder->base.dev; | 1132 | struct drm_device *dev = encoder->base.dev; |
| 1091 | struct radeon_device *rdev = dev->dev_private; | 1133 | struct radeon_device *rdev = dev->dev_private; |
| 1092 | uint16_t ext_tmds_info; | 1134 | uint16_t offset; |
| 1093 | uint8_t ver; | 1135 | uint8_t ver, id, blocks, clk, data; |
| 1136 | int i; | ||
| 1137 | enum radeon_combios_ddc gpio; | ||
| 1138 | struct radeon_i2c_bus_rec i2c_bus; | ||
| 1094 | 1139 | ||
| 1095 | if (rdev->bios == NULL) | 1140 | if (rdev->bios == NULL) |
| 1096 | return; | 1141 | return false; |
| 1097 | 1142 | ||
| 1098 | ext_tmds_info = | 1143 | tmds->i2c_bus = NULL; |
| 1099 | combios_get_table_offset(dev, COMBIOS_EXT_TMDS_INFO_TABLE); | 1144 | if (rdev->flags & RADEON_IS_IGP) { |
| 1100 | if (ext_tmds_info) { | 1145 | offset = combios_get_table_offset(dev, COMBIOS_I2C_INFO_TABLE); |
| 1101 | ver = RBIOS8(ext_tmds_info); | 1146 | if (offset) { |
| 1102 | DRM_INFO("External TMDS Table revision: %d\n", ver); | 1147 | ver = RBIOS8(offset); |
| 1103 | // TODO | 1148 | DRM_INFO("GPIO Table revision: %d\n", ver); |
| 1149 | blocks = RBIOS8(offset + 2); | ||
| 1150 | for (i = 0; i < blocks; i++) { | ||
| 1151 | id = RBIOS8(offset + 3 + (i * 5) + 0); | ||
| 1152 | if (id == 136) { | ||
| 1153 | clk = RBIOS8(offset + 3 + (i * 5) + 3); | ||
| 1154 | data = RBIOS8(offset + 3 + (i * 5) + 4); | ||
| 1155 | i2c_bus.valid = true; | ||
| 1156 | i2c_bus.mask_clk_mask = (1 << clk); | ||
| 1157 | i2c_bus.mask_data_mask = (1 << data); | ||
| 1158 | i2c_bus.a_clk_mask = (1 << clk); | ||
| 1159 | i2c_bus.a_data_mask = (1 << data); | ||
| 1160 | i2c_bus.en_clk_mask = (1 << clk); | ||
| 1161 | i2c_bus.en_data_mask = (1 << data); | ||
| 1162 | i2c_bus.y_clk_mask = (1 << clk); | ||
| 1163 | i2c_bus.y_data_mask = (1 << data); | ||
| 1164 | i2c_bus.mask_clk_reg = RADEON_GPIOPAD_MASK; | ||
| 1165 | i2c_bus.mask_data_reg = RADEON_GPIOPAD_MASK; | ||
| 1166 | i2c_bus.a_clk_reg = RADEON_GPIOPAD_A; | ||
| 1167 | i2c_bus.a_data_reg = RADEON_GPIOPAD_A; | ||
| 1168 | i2c_bus.en_clk_reg = RADEON_GPIOPAD_EN; | ||
| 1169 | i2c_bus.en_data_reg = RADEON_GPIOPAD_EN; | ||
| 1170 | i2c_bus.y_clk_reg = RADEON_GPIOPAD_Y; | ||
| 1171 | i2c_bus.y_data_reg = RADEON_GPIOPAD_Y; | ||
| 1172 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1173 | tmds->dvo_chip = DVO_SIL164; | ||
| 1174 | tmds->slave_addr = 0x70 >> 1; /* 7 bit addressing */ | ||
| 1175 | break; | ||
| 1176 | } | ||
| 1177 | } | ||
| 1178 | } | ||
| 1179 | } else { | ||
| 1180 | offset = combios_get_table_offset(dev, COMBIOS_EXT_TMDS_INFO_TABLE); | ||
| 1181 | if (offset) { | ||
| 1182 | ver = RBIOS8(offset); | ||
| 1183 | DRM_INFO("External TMDS Table revision: %d\n", ver); | ||
| 1184 | tmds->slave_addr = RBIOS8(offset + 4 + 2); | ||
| 1185 | tmds->slave_addr >>= 1; /* 7 bit addressing */ | ||
| 1186 | gpio = RBIOS8(offset + 4 + 3); | ||
| 1187 | switch (gpio) { | ||
| 1188 | case DDC_MONID: | ||
| 1189 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); | ||
| 1190 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1191 | break; | ||
| 1192 | case DDC_DVI: | ||
| 1193 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); | ||
| 1194 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1195 | break; | ||
| 1196 | case DDC_VGA: | ||
| 1197 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); | ||
| 1198 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1199 | break; | ||
| 1200 | case DDC_CRT2: | ||
| 1201 | /* R3xx+ chips don't have GPIO_CRT2_DDC gpio pad */ | ||
| 1202 | if (rdev->family >= CHIP_R300) | ||
| 1203 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); | ||
| 1204 | else | ||
| 1205 | i2c_bus = combios_setup_i2c_bus(rdev, RADEON_GPIO_CRT2_DDC); | ||
| 1206 | tmds->i2c_bus = radeon_i2c_create(dev, &i2c_bus, "DVO"); | ||
| 1207 | break; | ||
| 1208 | case DDC_LCD: /* MM i2c */ | ||
| 1209 | DRM_ERROR("MM i2c requires hw i2c engine\n"); | ||
| 1210 | break; | ||
| 1211 | default: | ||
| 1212 | DRM_ERROR("Unsupported gpio %d\n", gpio); | ||
| 1213 | break; | ||
| 1214 | } | ||
| 1215 | } | ||
| 1104 | } | 1216 | } |
| 1217 | |||
| 1218 | if (!tmds->i2c_bus) { | ||
| 1219 | DRM_INFO("No valid Ext TMDS info found in BIOS\n"); | ||
| 1220 | return false; | ||
| 1221 | } | ||
| 1222 | |||
| 1223 | return true; | ||
| 1105 | } | 1224 | } |
| 1106 | 1225 | ||
| 1107 | bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | 1226 | bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) |
| 1108 | { | 1227 | { |
| 1109 | struct radeon_device *rdev = dev->dev_private; | 1228 | struct radeon_device *rdev = dev->dev_private; |
| 1110 | struct radeon_i2c_bus_rec ddc_i2c; | 1229 | struct radeon_i2c_bus_rec ddc_i2c; |
| 1230 | struct radeon_hpd hpd; | ||
| 1111 | 1231 | ||
| 1112 | rdev->mode_info.connector_table = radeon_connector_table; | 1232 | rdev->mode_info.connector_table = radeon_connector_table; |
| 1113 | if (rdev->mode_info.connector_table == CT_NONE) { | 1233 | if (rdev->mode_info.connector_table == CT_NONE) { |
| @@ -1168,7 +1288,8 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1168 | /* these are the most common settings */ | 1288 | /* these are the most common settings */ |
| 1169 | if (rdev->flags & RADEON_SINGLE_CRTC) { | 1289 | if (rdev->flags & RADEON_SINGLE_CRTC) { |
| 1170 | /* VGA - primary dac */ | 1290 | /* VGA - primary dac */ |
| 1171 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1291 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1292 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1172 | radeon_add_legacy_encoder(dev, | 1293 | radeon_add_legacy_encoder(dev, |
| 1173 | radeon_get_encoder_id(dev, | 1294 | radeon_get_encoder_id(dev, |
| 1174 | ATOM_DEVICE_CRT1_SUPPORT, | 1295 | ATOM_DEVICE_CRT1_SUPPORT, |
| @@ -1178,10 +1299,12 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1178 | ATOM_DEVICE_CRT1_SUPPORT, | 1299 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1179 | DRM_MODE_CONNECTOR_VGA, | 1300 | DRM_MODE_CONNECTOR_VGA, |
| 1180 | &ddc_i2c, | 1301 | &ddc_i2c, |
| 1181 | CONNECTOR_OBJECT_ID_VGA); | 1302 | CONNECTOR_OBJECT_ID_VGA, |
| 1303 | &hpd); | ||
| 1182 | } else if (rdev->flags & RADEON_IS_MOBILITY) { | 1304 | } else if (rdev->flags & RADEON_IS_MOBILITY) { |
| 1183 | /* LVDS */ | 1305 | /* LVDS */ |
| 1184 | ddc_i2c = combios_setup_i2c_bus(RADEON_LCD_GPIO_MASK); | 1306 | ddc_i2c = combios_setup_i2c_bus(rdev, 0); |
| 1307 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1185 | radeon_add_legacy_encoder(dev, | 1308 | radeon_add_legacy_encoder(dev, |
| 1186 | radeon_get_encoder_id(dev, | 1309 | radeon_get_encoder_id(dev, |
| 1187 | ATOM_DEVICE_LCD1_SUPPORT, | 1310 | ATOM_DEVICE_LCD1_SUPPORT, |
| @@ -1191,10 +1314,12 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1191 | ATOM_DEVICE_LCD1_SUPPORT, | 1314 | ATOM_DEVICE_LCD1_SUPPORT, |
| 1192 | DRM_MODE_CONNECTOR_LVDS, | 1315 | DRM_MODE_CONNECTOR_LVDS, |
| 1193 | &ddc_i2c, | 1316 | &ddc_i2c, |
| 1194 | CONNECTOR_OBJECT_ID_LVDS); | 1317 | CONNECTOR_OBJECT_ID_LVDS, |
| 1318 | &hpd); | ||
| 1195 | 1319 | ||
| 1196 | /* VGA - primary dac */ | 1320 | /* VGA - primary dac */ |
| 1197 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1321 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1322 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1198 | radeon_add_legacy_encoder(dev, | 1323 | radeon_add_legacy_encoder(dev, |
| 1199 | radeon_get_encoder_id(dev, | 1324 | radeon_get_encoder_id(dev, |
| 1200 | ATOM_DEVICE_CRT1_SUPPORT, | 1325 | ATOM_DEVICE_CRT1_SUPPORT, |
| @@ -1204,10 +1329,12 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1204 | ATOM_DEVICE_CRT1_SUPPORT, | 1329 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1205 | DRM_MODE_CONNECTOR_VGA, | 1330 | DRM_MODE_CONNECTOR_VGA, |
| 1206 | &ddc_i2c, | 1331 | &ddc_i2c, |
| 1207 | CONNECTOR_OBJECT_ID_VGA); | 1332 | CONNECTOR_OBJECT_ID_VGA, |
| 1333 | &hpd); | ||
| 1208 | } else { | 1334 | } else { |
| 1209 | /* DVI-I - tv dac, int tmds */ | 1335 | /* DVI-I - tv dac, int tmds */ |
| 1210 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1336 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1337 | hpd.hpd = RADEON_HPD_1; | ||
| 1211 | radeon_add_legacy_encoder(dev, | 1338 | radeon_add_legacy_encoder(dev, |
| 1212 | radeon_get_encoder_id(dev, | 1339 | radeon_get_encoder_id(dev, |
| 1213 | ATOM_DEVICE_DFP1_SUPPORT, | 1340 | ATOM_DEVICE_DFP1_SUPPORT, |
| @@ -1223,10 +1350,12 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1223 | ATOM_DEVICE_CRT2_SUPPORT, | 1350 | ATOM_DEVICE_CRT2_SUPPORT, |
| 1224 | DRM_MODE_CONNECTOR_DVII, | 1351 | DRM_MODE_CONNECTOR_DVII, |
| 1225 | &ddc_i2c, | 1352 | &ddc_i2c, |
| 1226 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I); | 1353 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I, |
| 1354 | &hpd); | ||
| 1227 | 1355 | ||
| 1228 | /* VGA - primary dac */ | 1356 | /* VGA - primary dac */ |
| 1229 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1357 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1358 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1230 | radeon_add_legacy_encoder(dev, | 1359 | radeon_add_legacy_encoder(dev, |
| 1231 | radeon_get_encoder_id(dev, | 1360 | radeon_get_encoder_id(dev, |
| 1232 | ATOM_DEVICE_CRT1_SUPPORT, | 1361 | ATOM_DEVICE_CRT1_SUPPORT, |
| @@ -1236,11 +1365,14 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1236 | ATOM_DEVICE_CRT1_SUPPORT, | 1365 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1237 | DRM_MODE_CONNECTOR_VGA, | 1366 | DRM_MODE_CONNECTOR_VGA, |
| 1238 | &ddc_i2c, | 1367 | &ddc_i2c, |
| 1239 | CONNECTOR_OBJECT_ID_VGA); | 1368 | CONNECTOR_OBJECT_ID_VGA, |
| 1369 | &hpd); | ||
| 1240 | } | 1370 | } |
| 1241 | 1371 | ||
| 1242 | if (rdev->family != CHIP_R100 && rdev->family != CHIP_R200) { | 1372 | if (rdev->family != CHIP_R100 && rdev->family != CHIP_R200) { |
| 1243 | /* TV - tv dac */ | 1373 | /* TV - tv dac */ |
| 1374 | ddc_i2c.valid = false; | ||
| 1375 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1244 | radeon_add_legacy_encoder(dev, | 1376 | radeon_add_legacy_encoder(dev, |
| 1245 | radeon_get_encoder_id(dev, | 1377 | radeon_get_encoder_id(dev, |
| 1246 | ATOM_DEVICE_TV1_SUPPORT, | 1378 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1250,14 +1382,16 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1250 | ATOM_DEVICE_TV1_SUPPORT, | 1382 | ATOM_DEVICE_TV1_SUPPORT, |
| 1251 | DRM_MODE_CONNECTOR_SVIDEO, | 1383 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1252 | &ddc_i2c, | 1384 | &ddc_i2c, |
| 1253 | CONNECTOR_OBJECT_ID_SVIDEO); | 1385 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1386 | &hpd); | ||
| 1254 | } | 1387 | } |
| 1255 | break; | 1388 | break; |
| 1256 | case CT_IBOOK: | 1389 | case CT_IBOOK: |
| 1257 | DRM_INFO("Connector Table: %d (ibook)\n", | 1390 | DRM_INFO("Connector Table: %d (ibook)\n", |
| 1258 | rdev->mode_info.connector_table); | 1391 | rdev->mode_info.connector_table); |
| 1259 | /* LVDS */ | 1392 | /* LVDS */ |
| 1260 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1393 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1394 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1261 | radeon_add_legacy_encoder(dev, | 1395 | radeon_add_legacy_encoder(dev, |
| 1262 | radeon_get_encoder_id(dev, | 1396 | radeon_get_encoder_id(dev, |
| 1263 | ATOM_DEVICE_LCD1_SUPPORT, | 1397 | ATOM_DEVICE_LCD1_SUPPORT, |
| @@ -1265,9 +1399,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1265 | ATOM_DEVICE_LCD1_SUPPORT); | 1399 | ATOM_DEVICE_LCD1_SUPPORT); |
| 1266 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, | 1400 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, |
| 1267 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, | 1401 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, |
| 1268 | CONNECTOR_OBJECT_ID_LVDS); | 1402 | CONNECTOR_OBJECT_ID_LVDS, |
| 1403 | &hpd); | ||
| 1269 | /* VGA - TV DAC */ | 1404 | /* VGA - TV DAC */ |
| 1270 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1405 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1406 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1271 | radeon_add_legacy_encoder(dev, | 1407 | radeon_add_legacy_encoder(dev, |
| 1272 | radeon_get_encoder_id(dev, | 1408 | radeon_get_encoder_id(dev, |
| 1273 | ATOM_DEVICE_CRT2_SUPPORT, | 1409 | ATOM_DEVICE_CRT2_SUPPORT, |
| @@ -1275,8 +1411,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1275 | ATOM_DEVICE_CRT2_SUPPORT); | 1411 | ATOM_DEVICE_CRT2_SUPPORT); |
| 1276 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, | 1412 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, |
| 1277 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, | 1413 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, |
| 1278 | CONNECTOR_OBJECT_ID_VGA); | 1414 | CONNECTOR_OBJECT_ID_VGA, |
| 1415 | &hpd); | ||
| 1279 | /* TV - TV DAC */ | 1416 | /* TV - TV DAC */ |
| 1417 | ddc_i2c.valid = false; | ||
| 1418 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1280 | radeon_add_legacy_encoder(dev, | 1419 | radeon_add_legacy_encoder(dev, |
| 1281 | radeon_get_encoder_id(dev, | 1420 | radeon_get_encoder_id(dev, |
| 1282 | ATOM_DEVICE_TV1_SUPPORT, | 1421 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1285,13 +1424,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1285 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1424 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1286 | DRM_MODE_CONNECTOR_SVIDEO, | 1425 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1287 | &ddc_i2c, | 1426 | &ddc_i2c, |
| 1288 | CONNECTOR_OBJECT_ID_SVIDEO); | 1427 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1428 | &hpd); | ||
| 1289 | break; | 1429 | break; |
| 1290 | case CT_POWERBOOK_EXTERNAL: | 1430 | case CT_POWERBOOK_EXTERNAL: |
| 1291 | DRM_INFO("Connector Table: %d (powerbook external tmds)\n", | 1431 | DRM_INFO("Connector Table: %d (powerbook external tmds)\n", |
| 1292 | rdev->mode_info.connector_table); | 1432 | rdev->mode_info.connector_table); |
| 1293 | /* LVDS */ | 1433 | /* LVDS */ |
| 1294 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1434 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1435 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1295 | radeon_add_legacy_encoder(dev, | 1436 | radeon_add_legacy_encoder(dev, |
| 1296 | radeon_get_encoder_id(dev, | 1437 | radeon_get_encoder_id(dev, |
| 1297 | ATOM_DEVICE_LCD1_SUPPORT, | 1438 | ATOM_DEVICE_LCD1_SUPPORT, |
| @@ -1299,9 +1440,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1299 | ATOM_DEVICE_LCD1_SUPPORT); | 1440 | ATOM_DEVICE_LCD1_SUPPORT); |
| 1300 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, | 1441 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, |
| 1301 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, | 1442 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, |
| 1302 | CONNECTOR_OBJECT_ID_LVDS); | 1443 | CONNECTOR_OBJECT_ID_LVDS, |
| 1444 | &hpd); | ||
| 1303 | /* DVI-I - primary dac, ext tmds */ | 1445 | /* DVI-I - primary dac, ext tmds */ |
| 1304 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1446 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1447 | hpd.hpd = RADEON_HPD_2; /* ??? */ | ||
| 1305 | radeon_add_legacy_encoder(dev, | 1448 | radeon_add_legacy_encoder(dev, |
| 1306 | radeon_get_encoder_id(dev, | 1449 | radeon_get_encoder_id(dev, |
| 1307 | ATOM_DEVICE_DFP2_SUPPORT, | 1450 | ATOM_DEVICE_DFP2_SUPPORT, |
| @@ -1317,8 +1460,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1317 | ATOM_DEVICE_DFP2_SUPPORT | | 1460 | ATOM_DEVICE_DFP2_SUPPORT | |
| 1318 | ATOM_DEVICE_CRT1_SUPPORT, | 1461 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1319 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, | 1462 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, |
| 1320 | CONNECTOR_OBJECT_ID_DUAL_LINK_DVI_I); | 1463 | CONNECTOR_OBJECT_ID_DUAL_LINK_DVI_I, |
| 1464 | &hpd); | ||
| 1321 | /* TV - TV DAC */ | 1465 | /* TV - TV DAC */ |
| 1466 | ddc_i2c.valid = false; | ||
| 1467 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1322 | radeon_add_legacy_encoder(dev, | 1468 | radeon_add_legacy_encoder(dev, |
| 1323 | radeon_get_encoder_id(dev, | 1469 | radeon_get_encoder_id(dev, |
| 1324 | ATOM_DEVICE_TV1_SUPPORT, | 1470 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1327,13 +1473,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1327 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1473 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1328 | DRM_MODE_CONNECTOR_SVIDEO, | 1474 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1329 | &ddc_i2c, | 1475 | &ddc_i2c, |
| 1330 | CONNECTOR_OBJECT_ID_SVIDEO); | 1476 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1477 | &hpd); | ||
| 1331 | break; | 1478 | break; |
| 1332 | case CT_POWERBOOK_INTERNAL: | 1479 | case CT_POWERBOOK_INTERNAL: |
| 1333 | DRM_INFO("Connector Table: %d (powerbook internal tmds)\n", | 1480 | DRM_INFO("Connector Table: %d (powerbook internal tmds)\n", |
| 1334 | rdev->mode_info.connector_table); | 1481 | rdev->mode_info.connector_table); |
| 1335 | /* LVDS */ | 1482 | /* LVDS */ |
| 1336 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1483 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1484 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1337 | radeon_add_legacy_encoder(dev, | 1485 | radeon_add_legacy_encoder(dev, |
| 1338 | radeon_get_encoder_id(dev, | 1486 | radeon_get_encoder_id(dev, |
| 1339 | ATOM_DEVICE_LCD1_SUPPORT, | 1487 | ATOM_DEVICE_LCD1_SUPPORT, |
| @@ -1341,9 +1489,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1341 | ATOM_DEVICE_LCD1_SUPPORT); | 1489 | ATOM_DEVICE_LCD1_SUPPORT); |
| 1342 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, | 1490 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, |
| 1343 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, | 1491 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, |
| 1344 | CONNECTOR_OBJECT_ID_LVDS); | 1492 | CONNECTOR_OBJECT_ID_LVDS, |
| 1493 | &hpd); | ||
| 1345 | /* DVI-I - primary dac, int tmds */ | 1494 | /* DVI-I - primary dac, int tmds */ |
| 1346 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1495 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1496 | hpd.hpd = RADEON_HPD_1; /* ??? */ | ||
| 1347 | radeon_add_legacy_encoder(dev, | 1497 | radeon_add_legacy_encoder(dev, |
| 1348 | radeon_get_encoder_id(dev, | 1498 | radeon_get_encoder_id(dev, |
| 1349 | ATOM_DEVICE_DFP1_SUPPORT, | 1499 | ATOM_DEVICE_DFP1_SUPPORT, |
| @@ -1358,8 +1508,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1358 | ATOM_DEVICE_DFP1_SUPPORT | | 1508 | ATOM_DEVICE_DFP1_SUPPORT | |
| 1359 | ATOM_DEVICE_CRT1_SUPPORT, | 1509 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1360 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, | 1510 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, |
| 1361 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I); | 1511 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I, |
| 1512 | &hpd); | ||
| 1362 | /* TV - TV DAC */ | 1513 | /* TV - TV DAC */ |
| 1514 | ddc_i2c.valid = false; | ||
| 1515 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1363 | radeon_add_legacy_encoder(dev, | 1516 | radeon_add_legacy_encoder(dev, |
| 1364 | radeon_get_encoder_id(dev, | 1517 | radeon_get_encoder_id(dev, |
| 1365 | ATOM_DEVICE_TV1_SUPPORT, | 1518 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1368,13 +1521,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1368 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1521 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1369 | DRM_MODE_CONNECTOR_SVIDEO, | 1522 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1370 | &ddc_i2c, | 1523 | &ddc_i2c, |
| 1371 | CONNECTOR_OBJECT_ID_SVIDEO); | 1524 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1525 | &hpd); | ||
| 1372 | break; | 1526 | break; |
| 1373 | case CT_POWERBOOK_VGA: | 1527 | case CT_POWERBOOK_VGA: |
| 1374 | DRM_INFO("Connector Table: %d (powerbook vga)\n", | 1528 | DRM_INFO("Connector Table: %d (powerbook vga)\n", |
| 1375 | rdev->mode_info.connector_table); | 1529 | rdev->mode_info.connector_table); |
| 1376 | /* LVDS */ | 1530 | /* LVDS */ |
| 1377 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1531 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1532 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1378 | radeon_add_legacy_encoder(dev, | 1533 | radeon_add_legacy_encoder(dev, |
| 1379 | radeon_get_encoder_id(dev, | 1534 | radeon_get_encoder_id(dev, |
| 1380 | ATOM_DEVICE_LCD1_SUPPORT, | 1535 | ATOM_DEVICE_LCD1_SUPPORT, |
| @@ -1382,9 +1537,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1382 | ATOM_DEVICE_LCD1_SUPPORT); | 1537 | ATOM_DEVICE_LCD1_SUPPORT); |
| 1383 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, | 1538 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_LCD1_SUPPORT, |
| 1384 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, | 1539 | DRM_MODE_CONNECTOR_LVDS, &ddc_i2c, |
| 1385 | CONNECTOR_OBJECT_ID_LVDS); | 1540 | CONNECTOR_OBJECT_ID_LVDS, |
| 1541 | &hpd); | ||
| 1386 | /* VGA - primary dac */ | 1542 | /* VGA - primary dac */ |
| 1387 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1543 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1544 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1388 | radeon_add_legacy_encoder(dev, | 1545 | radeon_add_legacy_encoder(dev, |
| 1389 | radeon_get_encoder_id(dev, | 1546 | radeon_get_encoder_id(dev, |
| 1390 | ATOM_DEVICE_CRT1_SUPPORT, | 1547 | ATOM_DEVICE_CRT1_SUPPORT, |
| @@ -1392,8 +1549,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1392 | ATOM_DEVICE_CRT1_SUPPORT); | 1549 | ATOM_DEVICE_CRT1_SUPPORT); |
| 1393 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT1_SUPPORT, | 1550 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT1_SUPPORT, |
| 1394 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, | 1551 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, |
| 1395 | CONNECTOR_OBJECT_ID_VGA); | 1552 | CONNECTOR_OBJECT_ID_VGA, |
| 1553 | &hpd); | ||
| 1396 | /* TV - TV DAC */ | 1554 | /* TV - TV DAC */ |
| 1555 | ddc_i2c.valid = false; | ||
| 1556 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1397 | radeon_add_legacy_encoder(dev, | 1557 | radeon_add_legacy_encoder(dev, |
| 1398 | radeon_get_encoder_id(dev, | 1558 | radeon_get_encoder_id(dev, |
| 1399 | ATOM_DEVICE_TV1_SUPPORT, | 1559 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1402,13 +1562,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1402 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1562 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1403 | DRM_MODE_CONNECTOR_SVIDEO, | 1563 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1404 | &ddc_i2c, | 1564 | &ddc_i2c, |
| 1405 | CONNECTOR_OBJECT_ID_SVIDEO); | 1565 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1566 | &hpd); | ||
| 1406 | break; | 1567 | break; |
| 1407 | case CT_MINI_EXTERNAL: | 1568 | case CT_MINI_EXTERNAL: |
| 1408 | DRM_INFO("Connector Table: %d (mini external tmds)\n", | 1569 | DRM_INFO("Connector Table: %d (mini external tmds)\n", |
| 1409 | rdev->mode_info.connector_table); | 1570 | rdev->mode_info.connector_table); |
| 1410 | /* DVI-I - tv dac, ext tmds */ | 1571 | /* DVI-I - tv dac, ext tmds */ |
| 1411 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_CRT2_DDC); | 1572 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_CRT2_DDC); |
| 1573 | hpd.hpd = RADEON_HPD_2; /* ??? */ | ||
| 1412 | radeon_add_legacy_encoder(dev, | 1574 | radeon_add_legacy_encoder(dev, |
| 1413 | radeon_get_encoder_id(dev, | 1575 | radeon_get_encoder_id(dev, |
| 1414 | ATOM_DEVICE_DFP2_SUPPORT, | 1576 | ATOM_DEVICE_DFP2_SUPPORT, |
| @@ -1424,8 +1586,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1424 | ATOM_DEVICE_DFP2_SUPPORT | | 1586 | ATOM_DEVICE_DFP2_SUPPORT | |
| 1425 | ATOM_DEVICE_CRT2_SUPPORT, | 1587 | ATOM_DEVICE_CRT2_SUPPORT, |
| 1426 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, | 1588 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, |
| 1427 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I); | 1589 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I, |
| 1590 | &hpd); | ||
| 1428 | /* TV - TV DAC */ | 1591 | /* TV - TV DAC */ |
| 1592 | ddc_i2c.valid = false; | ||
| 1593 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1429 | radeon_add_legacy_encoder(dev, | 1594 | radeon_add_legacy_encoder(dev, |
| 1430 | radeon_get_encoder_id(dev, | 1595 | radeon_get_encoder_id(dev, |
| 1431 | ATOM_DEVICE_TV1_SUPPORT, | 1596 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1434,13 +1599,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1434 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_TV1_SUPPORT, | 1599 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_TV1_SUPPORT, |
| 1435 | DRM_MODE_CONNECTOR_SVIDEO, | 1600 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1436 | &ddc_i2c, | 1601 | &ddc_i2c, |
| 1437 | CONNECTOR_OBJECT_ID_SVIDEO); | 1602 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1603 | &hpd); | ||
| 1438 | break; | 1604 | break; |
| 1439 | case CT_MINI_INTERNAL: | 1605 | case CT_MINI_INTERNAL: |
| 1440 | DRM_INFO("Connector Table: %d (mini internal tmds)\n", | 1606 | DRM_INFO("Connector Table: %d (mini internal tmds)\n", |
| 1441 | rdev->mode_info.connector_table); | 1607 | rdev->mode_info.connector_table); |
| 1442 | /* DVI-I - tv dac, int tmds */ | 1608 | /* DVI-I - tv dac, int tmds */ |
| 1443 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_CRT2_DDC); | 1609 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_CRT2_DDC); |
| 1610 | hpd.hpd = RADEON_HPD_1; /* ??? */ | ||
| 1444 | radeon_add_legacy_encoder(dev, | 1611 | radeon_add_legacy_encoder(dev, |
| 1445 | radeon_get_encoder_id(dev, | 1612 | radeon_get_encoder_id(dev, |
| 1446 | ATOM_DEVICE_DFP1_SUPPORT, | 1613 | ATOM_DEVICE_DFP1_SUPPORT, |
| @@ -1455,8 +1622,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1455 | ATOM_DEVICE_DFP1_SUPPORT | | 1622 | ATOM_DEVICE_DFP1_SUPPORT | |
| 1456 | ATOM_DEVICE_CRT2_SUPPORT, | 1623 | ATOM_DEVICE_CRT2_SUPPORT, |
| 1457 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, | 1624 | DRM_MODE_CONNECTOR_DVII, &ddc_i2c, |
| 1458 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I); | 1625 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I, |
| 1626 | &hpd); | ||
| 1459 | /* TV - TV DAC */ | 1627 | /* TV - TV DAC */ |
| 1628 | ddc_i2c.valid = false; | ||
| 1629 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1460 | radeon_add_legacy_encoder(dev, | 1630 | radeon_add_legacy_encoder(dev, |
| 1461 | radeon_get_encoder_id(dev, | 1631 | radeon_get_encoder_id(dev, |
| 1462 | ATOM_DEVICE_TV1_SUPPORT, | 1632 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1465,13 +1635,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1465 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_TV1_SUPPORT, | 1635 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_TV1_SUPPORT, |
| 1466 | DRM_MODE_CONNECTOR_SVIDEO, | 1636 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1467 | &ddc_i2c, | 1637 | &ddc_i2c, |
| 1468 | CONNECTOR_OBJECT_ID_SVIDEO); | 1638 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1639 | &hpd); | ||
| 1469 | break; | 1640 | break; |
| 1470 | case CT_IMAC_G5_ISIGHT: | 1641 | case CT_IMAC_G5_ISIGHT: |
| 1471 | DRM_INFO("Connector Table: %d (imac g5 isight)\n", | 1642 | DRM_INFO("Connector Table: %d (imac g5 isight)\n", |
| 1472 | rdev->mode_info.connector_table); | 1643 | rdev->mode_info.connector_table); |
| 1473 | /* DVI-D - int tmds */ | 1644 | /* DVI-D - int tmds */ |
| 1474 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_MONID); | 1645 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); |
| 1646 | hpd.hpd = RADEON_HPD_1; /* ??? */ | ||
| 1475 | radeon_add_legacy_encoder(dev, | 1647 | radeon_add_legacy_encoder(dev, |
| 1476 | radeon_get_encoder_id(dev, | 1648 | radeon_get_encoder_id(dev, |
| 1477 | ATOM_DEVICE_DFP1_SUPPORT, | 1649 | ATOM_DEVICE_DFP1_SUPPORT, |
| @@ -1479,9 +1651,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1479 | ATOM_DEVICE_DFP1_SUPPORT); | 1651 | ATOM_DEVICE_DFP1_SUPPORT); |
| 1480 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_DFP1_SUPPORT, | 1652 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_DFP1_SUPPORT, |
| 1481 | DRM_MODE_CONNECTOR_DVID, &ddc_i2c, | 1653 | DRM_MODE_CONNECTOR_DVID, &ddc_i2c, |
| 1482 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_D); | 1654 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_D, |
| 1655 | &hpd); | ||
| 1483 | /* VGA - tv dac */ | 1656 | /* VGA - tv dac */ |
| 1484 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1657 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1658 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1485 | radeon_add_legacy_encoder(dev, | 1659 | radeon_add_legacy_encoder(dev, |
| 1486 | radeon_get_encoder_id(dev, | 1660 | radeon_get_encoder_id(dev, |
| 1487 | ATOM_DEVICE_CRT2_SUPPORT, | 1661 | ATOM_DEVICE_CRT2_SUPPORT, |
| @@ -1489,8 +1663,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1489 | ATOM_DEVICE_CRT2_SUPPORT); | 1663 | ATOM_DEVICE_CRT2_SUPPORT); |
| 1490 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, | 1664 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, |
| 1491 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, | 1665 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, |
| 1492 | CONNECTOR_OBJECT_ID_VGA); | 1666 | CONNECTOR_OBJECT_ID_VGA, |
| 1667 | &hpd); | ||
| 1493 | /* TV - TV DAC */ | 1668 | /* TV - TV DAC */ |
| 1669 | ddc_i2c.valid = false; | ||
| 1670 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1494 | radeon_add_legacy_encoder(dev, | 1671 | radeon_add_legacy_encoder(dev, |
| 1495 | radeon_get_encoder_id(dev, | 1672 | radeon_get_encoder_id(dev, |
| 1496 | ATOM_DEVICE_TV1_SUPPORT, | 1673 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1499,13 +1676,15 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1499 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1676 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1500 | DRM_MODE_CONNECTOR_SVIDEO, | 1677 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1501 | &ddc_i2c, | 1678 | &ddc_i2c, |
| 1502 | CONNECTOR_OBJECT_ID_SVIDEO); | 1679 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1680 | &hpd); | ||
| 1503 | break; | 1681 | break; |
| 1504 | case CT_EMAC: | 1682 | case CT_EMAC: |
| 1505 | DRM_INFO("Connector Table: %d (emac)\n", | 1683 | DRM_INFO("Connector Table: %d (emac)\n", |
| 1506 | rdev->mode_info.connector_table); | 1684 | rdev->mode_info.connector_table); |
| 1507 | /* VGA - primary dac */ | 1685 | /* VGA - primary dac */ |
| 1508 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1686 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1687 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1509 | radeon_add_legacy_encoder(dev, | 1688 | radeon_add_legacy_encoder(dev, |
| 1510 | radeon_get_encoder_id(dev, | 1689 | radeon_get_encoder_id(dev, |
| 1511 | ATOM_DEVICE_CRT1_SUPPORT, | 1690 | ATOM_DEVICE_CRT1_SUPPORT, |
| @@ -1513,9 +1692,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1513 | ATOM_DEVICE_CRT1_SUPPORT); | 1692 | ATOM_DEVICE_CRT1_SUPPORT); |
| 1514 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_CRT1_SUPPORT, | 1693 | radeon_add_legacy_connector(dev, 0, ATOM_DEVICE_CRT1_SUPPORT, |
| 1515 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, | 1694 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, |
| 1516 | CONNECTOR_OBJECT_ID_VGA); | 1695 | CONNECTOR_OBJECT_ID_VGA, |
| 1696 | &hpd); | ||
| 1517 | /* VGA - tv dac */ | 1697 | /* VGA - tv dac */ |
| 1518 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_CRT2_DDC); | 1698 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_CRT2_DDC); |
| 1699 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1519 | radeon_add_legacy_encoder(dev, | 1700 | radeon_add_legacy_encoder(dev, |
| 1520 | radeon_get_encoder_id(dev, | 1701 | radeon_get_encoder_id(dev, |
| 1521 | ATOM_DEVICE_CRT2_SUPPORT, | 1702 | ATOM_DEVICE_CRT2_SUPPORT, |
| @@ -1523,8 +1704,11 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1523 | ATOM_DEVICE_CRT2_SUPPORT); | 1704 | ATOM_DEVICE_CRT2_SUPPORT); |
| 1524 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, | 1705 | radeon_add_legacy_connector(dev, 1, ATOM_DEVICE_CRT2_SUPPORT, |
| 1525 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, | 1706 | DRM_MODE_CONNECTOR_VGA, &ddc_i2c, |
| 1526 | CONNECTOR_OBJECT_ID_VGA); | 1707 | CONNECTOR_OBJECT_ID_VGA, |
| 1708 | &hpd); | ||
| 1527 | /* TV - TV DAC */ | 1709 | /* TV - TV DAC */ |
| 1710 | ddc_i2c.valid = false; | ||
| 1711 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1528 | radeon_add_legacy_encoder(dev, | 1712 | radeon_add_legacy_encoder(dev, |
| 1529 | radeon_get_encoder_id(dev, | 1713 | radeon_get_encoder_id(dev, |
| 1530 | ATOM_DEVICE_TV1_SUPPORT, | 1714 | ATOM_DEVICE_TV1_SUPPORT, |
| @@ -1533,7 +1717,8 @@ bool radeon_get_legacy_connector_info_from_table(struct drm_device *dev) | |||
| 1533 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, | 1717 | radeon_add_legacy_connector(dev, 2, ATOM_DEVICE_TV1_SUPPORT, |
| 1534 | DRM_MODE_CONNECTOR_SVIDEO, | 1718 | DRM_MODE_CONNECTOR_SVIDEO, |
| 1535 | &ddc_i2c, | 1719 | &ddc_i2c, |
| 1536 | CONNECTOR_OBJECT_ID_SVIDEO); | 1720 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 1721 | &hpd); | ||
| 1537 | break; | 1722 | break; |
| 1538 | default: | 1723 | default: |
| 1539 | DRM_INFO("Connector table: %d (invalid)\n", | 1724 | DRM_INFO("Connector table: %d (invalid)\n", |
| @@ -1550,7 +1735,8 @@ static bool radeon_apply_legacy_quirks(struct drm_device *dev, | |||
| 1550 | int bios_index, | 1735 | int bios_index, |
| 1551 | enum radeon_combios_connector | 1736 | enum radeon_combios_connector |
| 1552 | *legacy_connector, | 1737 | *legacy_connector, |
| 1553 | struct radeon_i2c_bus_rec *ddc_i2c) | 1738 | struct radeon_i2c_bus_rec *ddc_i2c, |
| 1739 | struct radeon_hpd *hpd) | ||
| 1554 | { | 1740 | { |
| 1555 | struct radeon_device *rdev = dev->dev_private; | 1741 | struct radeon_device *rdev = dev->dev_private; |
| 1556 | 1742 | ||
| @@ -1558,29 +1744,26 @@ static bool radeon_apply_legacy_quirks(struct drm_device *dev, | |||
| 1558 | if ((rdev->family == CHIP_RS400 || | 1744 | if ((rdev->family == CHIP_RS400 || |
| 1559 | rdev->family == CHIP_RS480) && | 1745 | rdev->family == CHIP_RS480) && |
| 1560 | ddc_i2c->mask_clk_reg == RADEON_GPIO_CRT2_DDC) | 1746 | ddc_i2c->mask_clk_reg == RADEON_GPIO_CRT2_DDC) |
| 1561 | *ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_MONID); | 1747 | *ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); |
| 1562 | else if ((rdev->family == CHIP_RS400 || | 1748 | else if ((rdev->family == CHIP_RS400 || |
| 1563 | rdev->family == CHIP_RS480) && | 1749 | rdev->family == CHIP_RS480) && |
| 1564 | ddc_i2c->mask_clk_reg == RADEON_GPIO_MONID) { | 1750 | ddc_i2c->mask_clk_reg == RADEON_GPIO_MONID) { |
| 1565 | ddc_i2c->valid = true; | 1751 | *ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIOPAD_MASK); |
| 1566 | ddc_i2c->mask_clk_mask = (0x20 << 8); | 1752 | ddc_i2c->mask_clk_mask = (0x20 << 8); |
| 1567 | ddc_i2c->mask_data_mask = 0x80; | 1753 | ddc_i2c->mask_data_mask = 0x80; |
| 1568 | ddc_i2c->a_clk_mask = (0x20 << 8); | 1754 | ddc_i2c->a_clk_mask = (0x20 << 8); |
| 1569 | ddc_i2c->a_data_mask = 0x80; | 1755 | ddc_i2c->a_data_mask = 0x80; |
| 1570 | ddc_i2c->put_clk_mask = (0x20 << 8); | 1756 | ddc_i2c->en_clk_mask = (0x20 << 8); |
| 1571 | ddc_i2c->put_data_mask = 0x80; | 1757 | ddc_i2c->en_data_mask = 0x80; |
| 1572 | ddc_i2c->get_clk_mask = (0x20 << 8); | 1758 | ddc_i2c->y_clk_mask = (0x20 << 8); |
| 1573 | ddc_i2c->get_data_mask = 0x80; | 1759 | ddc_i2c->y_data_mask = 0x80; |
| 1574 | ddc_i2c->mask_clk_reg = RADEON_GPIOPAD_MASK; | ||
| 1575 | ddc_i2c->mask_data_reg = RADEON_GPIOPAD_MASK; | ||
| 1576 | ddc_i2c->a_clk_reg = RADEON_GPIOPAD_A; | ||
| 1577 | ddc_i2c->a_data_reg = RADEON_GPIOPAD_A; | ||
| 1578 | ddc_i2c->put_clk_reg = RADEON_GPIOPAD_EN; | ||
| 1579 | ddc_i2c->put_data_reg = RADEON_GPIOPAD_EN; | ||
| 1580 | ddc_i2c->get_clk_reg = RADEON_LCD_GPIO_Y_REG; | ||
| 1581 | ddc_i2c->get_data_reg = RADEON_LCD_GPIO_Y_REG; | ||
| 1582 | } | 1760 | } |
| 1583 | 1761 | ||
| 1762 | /* R3xx+ chips don't have GPIO_CRT2_DDC gpio pad */ | ||
| 1763 | if ((rdev->family >= CHIP_R300) && | ||
| 1764 | ddc_i2c->mask_clk_reg == RADEON_GPIO_CRT2_DDC) | ||
| 1765 | *ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); | ||
| 1766 | |||
| 1584 | /* Certain IBM chipset RN50s have a BIOS reporting two VGAs, | 1767 | /* Certain IBM chipset RN50s have a BIOS reporting two VGAs, |
| 1585 | one with VGA DDC and one with CRT2 DDC. - kill the CRT2 DDC one */ | 1768 | one with VGA DDC and one with CRT2 DDC. - kill the CRT2 DDC one */ |
| 1586 | if (dev->pdev->device == 0x515e && | 1769 | if (dev->pdev->device == 0x515e && |
| @@ -1624,6 +1807,12 @@ static bool radeon_apply_legacy_tv_quirks(struct drm_device *dev) | |||
| 1624 | dev->pdev->subsystem_device == 0x280a) | 1807 | dev->pdev->subsystem_device == 0x280a) |
| 1625 | return false; | 1808 | return false; |
| 1626 | 1809 | ||
| 1810 | /* MSI S270 has non-existent TV port */ | ||
| 1811 | if (dev->pdev->device == 0x5955 && | ||
| 1812 | dev->pdev->subsystem_vendor == 0x1462 && | ||
| 1813 | dev->pdev->subsystem_device == 0x0131) | ||
| 1814 | return false; | ||
| 1815 | |||
| 1627 | return true; | 1816 | return true; |
| 1628 | } | 1817 | } |
| 1629 | 1818 | ||
| @@ -1671,6 +1860,7 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1671 | enum radeon_combios_connector connector; | 1860 | enum radeon_combios_connector connector; |
| 1672 | int i = 0; | 1861 | int i = 0; |
| 1673 | struct radeon_i2c_bus_rec ddc_i2c; | 1862 | struct radeon_i2c_bus_rec ddc_i2c; |
| 1863 | struct radeon_hpd hpd; | ||
| 1674 | 1864 | ||
| 1675 | if (rdev->bios == NULL) | 1865 | if (rdev->bios == NULL) |
| 1676 | return false; | 1866 | return false; |
| @@ -1691,26 +1881,40 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1691 | switch (ddc_type) { | 1881 | switch (ddc_type) { |
| 1692 | case DDC_MONID: | 1882 | case DDC_MONID: |
| 1693 | ddc_i2c = | 1883 | ddc_i2c = |
| 1694 | combios_setup_i2c_bus(RADEON_GPIO_MONID); | 1884 | combios_setup_i2c_bus(rdev, RADEON_GPIO_MONID); |
| 1695 | break; | 1885 | break; |
| 1696 | case DDC_DVI: | 1886 | case DDC_DVI: |
| 1697 | ddc_i2c = | 1887 | ddc_i2c = |
| 1698 | combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 1888 | combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 1699 | break; | 1889 | break; |
| 1700 | case DDC_VGA: | 1890 | case DDC_VGA: |
| 1701 | ddc_i2c = | 1891 | ddc_i2c = |
| 1702 | combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 1892 | combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 1703 | break; | 1893 | break; |
| 1704 | case DDC_CRT2: | 1894 | case DDC_CRT2: |
| 1705 | ddc_i2c = | 1895 | ddc_i2c = |
| 1706 | combios_setup_i2c_bus(RADEON_GPIO_CRT2_DDC); | 1896 | combios_setup_i2c_bus(rdev, RADEON_GPIO_CRT2_DDC); |
| 1707 | break; | 1897 | break; |
| 1708 | default: | 1898 | default: |
| 1709 | break; | 1899 | break; |
| 1710 | } | 1900 | } |
| 1711 | 1901 | ||
| 1902 | switch (connector) { | ||
| 1903 | case CONNECTOR_PROPRIETARY_LEGACY: | ||
| 1904 | case CONNECTOR_DVI_I_LEGACY: | ||
| 1905 | case CONNECTOR_DVI_D_LEGACY: | ||
| 1906 | if ((tmp >> 4) & 0x1) | ||
| 1907 | hpd.hpd = RADEON_HPD_2; | ||
| 1908 | else | ||
| 1909 | hpd.hpd = RADEON_HPD_1; | ||
| 1910 | break; | ||
| 1911 | default: | ||
| 1912 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1913 | break; | ||
| 1914 | } | ||
| 1915 | |||
| 1712 | if (!radeon_apply_legacy_quirks(dev, i, &connector, | 1916 | if (!radeon_apply_legacy_quirks(dev, i, &connector, |
| 1713 | &ddc_i2c)) | 1917 | &ddc_i2c, &hpd)) |
| 1714 | continue; | 1918 | continue; |
| 1715 | 1919 | ||
| 1716 | switch (connector) { | 1920 | switch (connector) { |
| @@ -1727,7 +1931,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1727 | legacy_connector_convert | 1931 | legacy_connector_convert |
| 1728 | [connector], | 1932 | [connector], |
| 1729 | &ddc_i2c, | 1933 | &ddc_i2c, |
| 1730 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_D); | 1934 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_D, |
| 1935 | &hpd); | ||
| 1731 | break; | 1936 | break; |
| 1732 | case CONNECTOR_CRT_LEGACY: | 1937 | case CONNECTOR_CRT_LEGACY: |
| 1733 | if (tmp & 0x1) { | 1938 | if (tmp & 0x1) { |
| @@ -1753,7 +1958,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1753 | legacy_connector_convert | 1958 | legacy_connector_convert |
| 1754 | [connector], | 1959 | [connector], |
| 1755 | &ddc_i2c, | 1960 | &ddc_i2c, |
| 1756 | CONNECTOR_OBJECT_ID_VGA); | 1961 | CONNECTOR_OBJECT_ID_VGA, |
| 1962 | &hpd); | ||
| 1757 | break; | 1963 | break; |
| 1758 | case CONNECTOR_DVI_I_LEGACY: | 1964 | case CONNECTOR_DVI_I_LEGACY: |
| 1759 | devices = 0; | 1965 | devices = 0; |
| @@ -1799,7 +2005,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1799 | legacy_connector_convert | 2005 | legacy_connector_convert |
| 1800 | [connector], | 2006 | [connector], |
| 1801 | &ddc_i2c, | 2007 | &ddc_i2c, |
| 1802 | connector_object_id); | 2008 | connector_object_id, |
| 2009 | &hpd); | ||
| 1803 | break; | 2010 | break; |
| 1804 | case CONNECTOR_DVI_D_LEGACY: | 2011 | case CONNECTOR_DVI_D_LEGACY: |
| 1805 | if ((tmp >> 4) & 0x1) { | 2012 | if ((tmp >> 4) & 0x1) { |
| @@ -1817,7 +2024,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1817 | legacy_connector_convert | 2024 | legacy_connector_convert |
| 1818 | [connector], | 2025 | [connector], |
| 1819 | &ddc_i2c, | 2026 | &ddc_i2c, |
| 1820 | connector_object_id); | 2027 | connector_object_id, |
| 2028 | &hpd); | ||
| 1821 | break; | 2029 | break; |
| 1822 | case CONNECTOR_CTV_LEGACY: | 2030 | case CONNECTOR_CTV_LEGACY: |
| 1823 | case CONNECTOR_STV_LEGACY: | 2031 | case CONNECTOR_STV_LEGACY: |
| @@ -1832,7 +2040,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1832 | legacy_connector_convert | 2040 | legacy_connector_convert |
| 1833 | [connector], | 2041 | [connector], |
| 1834 | &ddc_i2c, | 2042 | &ddc_i2c, |
| 1835 | CONNECTOR_OBJECT_ID_SVIDEO); | 2043 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 2044 | &hpd); | ||
| 1836 | break; | 2045 | break; |
| 1837 | default: | 2046 | default: |
| 1838 | DRM_ERROR("Unknown connector type: %d\n", | 2047 | DRM_ERROR("Unknown connector type: %d\n", |
| @@ -1858,14 +2067,16 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1858 | 0), | 2067 | 0), |
| 1859 | ATOM_DEVICE_DFP1_SUPPORT); | 2068 | ATOM_DEVICE_DFP1_SUPPORT); |
| 1860 | 2069 | ||
| 1861 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_DVI_DDC); | 2070 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_DVI_DDC); |
| 2071 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1862 | radeon_add_legacy_connector(dev, | 2072 | radeon_add_legacy_connector(dev, |
| 1863 | 0, | 2073 | 0, |
| 1864 | ATOM_DEVICE_CRT1_SUPPORT | | 2074 | ATOM_DEVICE_CRT1_SUPPORT | |
| 1865 | ATOM_DEVICE_DFP1_SUPPORT, | 2075 | ATOM_DEVICE_DFP1_SUPPORT, |
| 1866 | DRM_MODE_CONNECTOR_DVII, | 2076 | DRM_MODE_CONNECTOR_DVII, |
| 1867 | &ddc_i2c, | 2077 | &ddc_i2c, |
| 1868 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I); | 2078 | CONNECTOR_OBJECT_ID_SINGLE_LINK_DVI_I, |
| 2079 | &hpd); | ||
| 1869 | } else { | 2080 | } else { |
| 1870 | uint16_t crt_info = | 2081 | uint16_t crt_info = |
| 1871 | combios_get_table_offset(dev, COMBIOS_CRT_INFO_TABLE); | 2082 | combios_get_table_offset(dev, COMBIOS_CRT_INFO_TABLE); |
| @@ -1876,13 +2087,15 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1876 | ATOM_DEVICE_CRT1_SUPPORT, | 2087 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1877 | 1), | 2088 | 1), |
| 1878 | ATOM_DEVICE_CRT1_SUPPORT); | 2089 | ATOM_DEVICE_CRT1_SUPPORT); |
| 1879 | ddc_i2c = combios_setup_i2c_bus(RADEON_GPIO_VGA_DDC); | 2090 | ddc_i2c = combios_setup_i2c_bus(rdev, RADEON_GPIO_VGA_DDC); |
| 2091 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1880 | radeon_add_legacy_connector(dev, | 2092 | radeon_add_legacy_connector(dev, |
| 1881 | 0, | 2093 | 0, |
| 1882 | ATOM_DEVICE_CRT1_SUPPORT, | 2094 | ATOM_DEVICE_CRT1_SUPPORT, |
| 1883 | DRM_MODE_CONNECTOR_VGA, | 2095 | DRM_MODE_CONNECTOR_VGA, |
| 1884 | &ddc_i2c, | 2096 | &ddc_i2c, |
| 1885 | CONNECTOR_OBJECT_ID_VGA); | 2097 | CONNECTOR_OBJECT_ID_VGA, |
| 2098 | &hpd); | ||
| 1886 | } else { | 2099 | } else { |
| 1887 | DRM_DEBUG("No connector info found\n"); | 2100 | DRM_DEBUG("No connector info found\n"); |
| 1888 | return false; | 2101 | return false; |
| @@ -1910,27 +2123,27 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1910 | case DDC_MONID: | 2123 | case DDC_MONID: |
| 1911 | ddc_i2c = | 2124 | ddc_i2c = |
| 1912 | combios_setup_i2c_bus | 2125 | combios_setup_i2c_bus |
| 1913 | (RADEON_GPIO_MONID); | 2126 | (rdev, RADEON_GPIO_MONID); |
| 1914 | break; | 2127 | break; |
| 1915 | case DDC_DVI: | 2128 | case DDC_DVI: |
| 1916 | ddc_i2c = | 2129 | ddc_i2c = |
| 1917 | combios_setup_i2c_bus | 2130 | combios_setup_i2c_bus |
| 1918 | (RADEON_GPIO_DVI_DDC); | 2131 | (rdev, RADEON_GPIO_DVI_DDC); |
| 1919 | break; | 2132 | break; |
| 1920 | case DDC_VGA: | 2133 | case DDC_VGA: |
| 1921 | ddc_i2c = | 2134 | ddc_i2c = |
| 1922 | combios_setup_i2c_bus | 2135 | combios_setup_i2c_bus |
| 1923 | (RADEON_GPIO_VGA_DDC); | 2136 | (rdev, RADEON_GPIO_VGA_DDC); |
| 1924 | break; | 2137 | break; |
| 1925 | case DDC_CRT2: | 2138 | case DDC_CRT2: |
| 1926 | ddc_i2c = | 2139 | ddc_i2c = |
| 1927 | combios_setup_i2c_bus | 2140 | combios_setup_i2c_bus |
| 1928 | (RADEON_GPIO_CRT2_DDC); | 2141 | (rdev, RADEON_GPIO_CRT2_DDC); |
| 1929 | break; | 2142 | break; |
| 1930 | case DDC_LCD: | 2143 | case DDC_LCD: |
| 1931 | ddc_i2c = | 2144 | ddc_i2c = |
| 1932 | combios_setup_i2c_bus | 2145 | combios_setup_i2c_bus |
| 1933 | (RADEON_LCD_GPIO_MASK); | 2146 | (rdev, RADEON_GPIOPAD_MASK); |
| 1934 | ddc_i2c.mask_clk_mask = | 2147 | ddc_i2c.mask_clk_mask = |
| 1935 | RBIOS32(lcd_ddc_info + 3); | 2148 | RBIOS32(lcd_ddc_info + 3); |
| 1936 | ddc_i2c.mask_data_mask = | 2149 | ddc_i2c.mask_data_mask = |
| @@ -1939,19 +2152,19 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1939 | RBIOS32(lcd_ddc_info + 3); | 2152 | RBIOS32(lcd_ddc_info + 3); |
| 1940 | ddc_i2c.a_data_mask = | 2153 | ddc_i2c.a_data_mask = |
| 1941 | RBIOS32(lcd_ddc_info + 7); | 2154 | RBIOS32(lcd_ddc_info + 7); |
| 1942 | ddc_i2c.put_clk_mask = | 2155 | ddc_i2c.en_clk_mask = |
| 1943 | RBIOS32(lcd_ddc_info + 3); | 2156 | RBIOS32(lcd_ddc_info + 3); |
| 1944 | ddc_i2c.put_data_mask = | 2157 | ddc_i2c.en_data_mask = |
| 1945 | RBIOS32(lcd_ddc_info + 7); | 2158 | RBIOS32(lcd_ddc_info + 7); |
| 1946 | ddc_i2c.get_clk_mask = | 2159 | ddc_i2c.y_clk_mask = |
| 1947 | RBIOS32(lcd_ddc_info + 3); | 2160 | RBIOS32(lcd_ddc_info + 3); |
| 1948 | ddc_i2c.get_data_mask = | 2161 | ddc_i2c.y_data_mask = |
| 1949 | RBIOS32(lcd_ddc_info + 7); | 2162 | RBIOS32(lcd_ddc_info + 7); |
| 1950 | break; | 2163 | break; |
| 1951 | case DDC_GPIO: | 2164 | case DDC_GPIO: |
| 1952 | ddc_i2c = | 2165 | ddc_i2c = |
| 1953 | combios_setup_i2c_bus | 2166 | combios_setup_i2c_bus |
| 1954 | (RADEON_MDGPIO_EN_REG); | 2167 | (rdev, RADEON_MDGPIO_MASK); |
| 1955 | ddc_i2c.mask_clk_mask = | 2168 | ddc_i2c.mask_clk_mask = |
| 1956 | RBIOS32(lcd_ddc_info + 3); | 2169 | RBIOS32(lcd_ddc_info + 3); |
| 1957 | ddc_i2c.mask_data_mask = | 2170 | ddc_i2c.mask_data_mask = |
| @@ -1960,13 +2173,13 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1960 | RBIOS32(lcd_ddc_info + 3); | 2173 | RBIOS32(lcd_ddc_info + 3); |
| 1961 | ddc_i2c.a_data_mask = | 2174 | ddc_i2c.a_data_mask = |
| 1962 | RBIOS32(lcd_ddc_info + 7); | 2175 | RBIOS32(lcd_ddc_info + 7); |
| 1963 | ddc_i2c.put_clk_mask = | 2176 | ddc_i2c.en_clk_mask = |
| 1964 | RBIOS32(lcd_ddc_info + 3); | 2177 | RBIOS32(lcd_ddc_info + 3); |
| 1965 | ddc_i2c.put_data_mask = | 2178 | ddc_i2c.en_data_mask = |
| 1966 | RBIOS32(lcd_ddc_info + 7); | 2179 | RBIOS32(lcd_ddc_info + 7); |
| 1967 | ddc_i2c.get_clk_mask = | 2180 | ddc_i2c.y_clk_mask = |
| 1968 | RBIOS32(lcd_ddc_info + 3); | 2181 | RBIOS32(lcd_ddc_info + 3); |
| 1969 | ddc_i2c.get_data_mask = | 2182 | ddc_i2c.y_data_mask = |
| 1970 | RBIOS32(lcd_ddc_info + 7); | 2183 | RBIOS32(lcd_ddc_info + 7); |
| 1971 | break; | 2184 | break; |
| 1972 | default: | 2185 | default: |
| @@ -1977,12 +2190,14 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1977 | } else | 2190 | } else |
| 1978 | ddc_i2c.valid = false; | 2191 | ddc_i2c.valid = false; |
| 1979 | 2192 | ||
| 2193 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1980 | radeon_add_legacy_connector(dev, | 2194 | radeon_add_legacy_connector(dev, |
| 1981 | 5, | 2195 | 5, |
| 1982 | ATOM_DEVICE_LCD1_SUPPORT, | 2196 | ATOM_DEVICE_LCD1_SUPPORT, |
| 1983 | DRM_MODE_CONNECTOR_LVDS, | 2197 | DRM_MODE_CONNECTOR_LVDS, |
| 1984 | &ddc_i2c, | 2198 | &ddc_i2c, |
| 1985 | CONNECTOR_OBJECT_ID_LVDS); | 2199 | CONNECTOR_OBJECT_ID_LVDS, |
| 2200 | &hpd); | ||
| 1986 | } | 2201 | } |
| 1987 | } | 2202 | } |
| 1988 | 2203 | ||
| @@ -1993,6 +2208,7 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 1993 | if (tv_info) { | 2208 | if (tv_info) { |
| 1994 | if (RBIOS8(tv_info + 6) == 'T') { | 2209 | if (RBIOS8(tv_info + 6) == 'T') { |
| 1995 | if (radeon_apply_legacy_tv_quirks(dev)) { | 2210 | if (radeon_apply_legacy_tv_quirks(dev)) { |
| 2211 | hpd.hpd = RADEON_HPD_NONE; | ||
| 1996 | radeon_add_legacy_encoder(dev, | 2212 | radeon_add_legacy_encoder(dev, |
| 1997 | radeon_get_encoder_id | 2213 | radeon_get_encoder_id |
| 1998 | (dev, | 2214 | (dev, |
| @@ -2003,7 +2219,8 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 2003 | ATOM_DEVICE_TV1_SUPPORT, | 2219 | ATOM_DEVICE_TV1_SUPPORT, |
| 2004 | DRM_MODE_CONNECTOR_SVIDEO, | 2220 | DRM_MODE_CONNECTOR_SVIDEO, |
| 2005 | &ddc_i2c, | 2221 | &ddc_i2c, |
| 2006 | CONNECTOR_OBJECT_ID_SVIDEO); | 2222 | CONNECTOR_OBJECT_ID_SVIDEO, |
| 2223 | &hpd); | ||
| 2007 | } | 2224 | } |
| 2008 | } | 2225 | } |
| 2009 | } | 2226 | } |
| @@ -2014,6 +2231,193 @@ bool radeon_get_legacy_connector_info_from_bios(struct drm_device *dev) | |||
| 2014 | return true; | 2231 | return true; |
| 2015 | } | 2232 | } |
| 2016 | 2233 | ||
| 2234 | void radeon_external_tmds_setup(struct drm_encoder *encoder) | ||
| 2235 | { | ||
| 2236 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 2237 | struct radeon_encoder_ext_tmds *tmds = radeon_encoder->enc_priv; | ||
| 2238 | |||
| 2239 | if (!tmds) | ||
| 2240 | return; | ||
| 2241 | |||
| 2242 | switch (tmds->dvo_chip) { | ||
| 2243 | case DVO_SIL164: | ||
| 2244 | /* sil 164 */ | ||
| 2245 | radeon_i2c_do_lock(tmds->i2c_bus, 1); | ||
| 2246 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2247 | tmds->slave_addr, | ||
| 2248 | 0x08, 0x30); | ||
| 2249 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2250 | tmds->slave_addr, | ||
| 2251 | 0x09, 0x00); | ||
| 2252 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2253 | tmds->slave_addr, | ||
| 2254 | 0x0a, 0x90); | ||
| 2255 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2256 | tmds->slave_addr, | ||
| 2257 | 0x0c, 0x89); | ||
| 2258 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2259 | tmds->slave_addr, | ||
| 2260 | 0x08, 0x3b); | ||
| 2261 | radeon_i2c_do_lock(tmds->i2c_bus, 0); | ||
| 2262 | break; | ||
| 2263 | case DVO_SIL1178: | ||
| 2264 | /* sil 1178 - untested */ | ||
| 2265 | /* | ||
| 2266 | * 0x0f, 0x44 | ||
| 2267 | * 0x0f, 0x4c | ||
| 2268 | * 0x0e, 0x01 | ||
| 2269 | * 0x0a, 0x80 | ||
| 2270 | * 0x09, 0x30 | ||
| 2271 | * 0x0c, 0xc9 | ||
| 2272 | * 0x0d, 0x70 | ||
| 2273 | * 0x08, 0x32 | ||
| 2274 | * 0x08, 0x33 | ||
| 2275 | */ | ||
| 2276 | break; | ||
| 2277 | default: | ||
| 2278 | break; | ||
| 2279 | } | ||
| 2280 | |||
| 2281 | } | ||
| 2282 | |||
| 2283 | bool radeon_combios_external_tmds_setup(struct drm_encoder *encoder) | ||
| 2284 | { | ||
| 2285 | struct drm_device *dev = encoder->dev; | ||
| 2286 | struct radeon_device *rdev = dev->dev_private; | ||
| 2287 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 2288 | uint16_t offset; | ||
| 2289 | uint8_t blocks, slave_addr, rev; | ||
| 2290 | uint32_t index, id; | ||
| 2291 | uint32_t reg, val, and_mask, or_mask; | ||
| 2292 | struct radeon_encoder_ext_tmds *tmds = radeon_encoder->enc_priv; | ||
| 2293 | |||
| 2294 | if (rdev->bios == NULL) | ||
| 2295 | return false; | ||
| 2296 | |||
| 2297 | if (!tmds) | ||
| 2298 | return false; | ||
| 2299 | |||
| 2300 | if (rdev->flags & RADEON_IS_IGP) { | ||
| 2301 | offset = combios_get_table_offset(dev, COMBIOS_TMDS_POWER_ON_TABLE); | ||
| 2302 | rev = RBIOS8(offset); | ||
| 2303 | if (offset) { | ||
| 2304 | rev = RBIOS8(offset); | ||
| 2305 | if (rev > 1) { | ||
| 2306 | blocks = RBIOS8(offset + 3); | ||
| 2307 | index = offset + 4; | ||
| 2308 | while (blocks > 0) { | ||
| 2309 | id = RBIOS16(index); | ||
| 2310 | index += 2; | ||
| 2311 | switch (id >> 13) { | ||
| 2312 | case 0: | ||
| 2313 | reg = (id & 0x1fff) * 4; | ||
| 2314 | val = RBIOS32(index); | ||
| 2315 | index += 4; | ||
| 2316 | WREG32(reg, val); | ||
| 2317 | break; | ||
| 2318 | case 2: | ||
| 2319 | reg = (id & 0x1fff) * 4; | ||
| 2320 | and_mask = RBIOS32(index); | ||
| 2321 | index += 4; | ||
| 2322 | or_mask = RBIOS32(index); | ||
| 2323 | index += 4; | ||
| 2324 | val = RREG32(reg); | ||
| 2325 | val = (val & and_mask) | or_mask; | ||
| 2326 | WREG32(reg, val); | ||
| 2327 | break; | ||
| 2328 | case 3: | ||
| 2329 | val = RBIOS16(index); | ||
| 2330 | index += 2; | ||
| 2331 | udelay(val); | ||
| 2332 | break; | ||
| 2333 | case 4: | ||
| 2334 | val = RBIOS16(index); | ||
| 2335 | index += 2; | ||
| 2336 | udelay(val * 1000); | ||
| 2337 | break; | ||
| 2338 | case 6: | ||
| 2339 | slave_addr = id & 0xff; | ||
| 2340 | slave_addr >>= 1; /* 7 bit addressing */ | ||
| 2341 | index++; | ||
| 2342 | reg = RBIOS8(index); | ||
| 2343 | index++; | ||
| 2344 | val = RBIOS8(index); | ||
| 2345 | index++; | ||
| 2346 | radeon_i2c_do_lock(tmds->i2c_bus, 1); | ||
| 2347 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2348 | slave_addr, | ||
| 2349 | reg, val); | ||
| 2350 | radeon_i2c_do_lock(tmds->i2c_bus, 0); | ||
| 2351 | break; | ||
| 2352 | default: | ||
| 2353 | DRM_ERROR("Unknown id %d\n", id >> 13); | ||
| 2354 | break; | ||
| 2355 | } | ||
| 2356 | blocks--; | ||
| 2357 | } | ||
| 2358 | return true; | ||
| 2359 | } | ||
| 2360 | } | ||
| 2361 | } else { | ||
| 2362 | offset = combios_get_table_offset(dev, COMBIOS_EXT_TMDS_INFO_TABLE); | ||
| 2363 | if (offset) { | ||
| 2364 | index = offset + 10; | ||
| 2365 | id = RBIOS16(index); | ||
| 2366 | while (id != 0xffff) { | ||
| 2367 | index += 2; | ||
| 2368 | switch (id >> 13) { | ||
| 2369 | case 0: | ||
| 2370 | reg = (id & 0x1fff) * 4; | ||
| 2371 | val = RBIOS32(index); | ||
| 2372 | WREG32(reg, val); | ||
| 2373 | break; | ||
| 2374 | case 2: | ||
| 2375 | reg = (id & 0x1fff) * 4; | ||
| 2376 | and_mask = RBIOS32(index); | ||
| 2377 | index += 4; | ||
| 2378 | or_mask = RBIOS32(index); | ||
| 2379 | index += 4; | ||
| 2380 | val = RREG32(reg); | ||
| 2381 | val = (val & and_mask) | or_mask; | ||
| 2382 | WREG32(reg, val); | ||
| 2383 | break; | ||
| 2384 | case 4: | ||
| 2385 | val = RBIOS16(index); | ||
| 2386 | index += 2; | ||
| 2387 | udelay(val); | ||
| 2388 | break; | ||
| 2389 | case 5: | ||
| 2390 | reg = id & 0x1fff; | ||
| 2391 | and_mask = RBIOS32(index); | ||
| 2392 | index += 4; | ||
| 2393 | or_mask = RBIOS32(index); | ||
| 2394 | index += 4; | ||
| 2395 | val = RREG32_PLL(reg); | ||
| 2396 | val = (val & and_mask) | or_mask; | ||
| 2397 | WREG32_PLL(reg, val); | ||
| 2398 | break; | ||
| 2399 | case 6: | ||
| 2400 | reg = id & 0x1fff; | ||
| 2401 | val = RBIOS8(index); | ||
| 2402 | index += 1; | ||
| 2403 | radeon_i2c_do_lock(tmds->i2c_bus, 1); | ||
| 2404 | radeon_i2c_sw_put_byte(tmds->i2c_bus, | ||
| 2405 | tmds->slave_addr, | ||
| 2406 | reg, val); | ||
| 2407 | radeon_i2c_do_lock(tmds->i2c_bus, 0); | ||
| 2408 | break; | ||
| 2409 | default: | ||
| 2410 | DRM_ERROR("Unknown id %d\n", id >> 13); | ||
| 2411 | break; | ||
| 2412 | } | ||
| 2413 | id = RBIOS16(index); | ||
| 2414 | } | ||
| 2415 | return true; | ||
| 2416 | } | ||
| 2417 | } | ||
| 2418 | return false; | ||
| 2419 | } | ||
| 2420 | |||
| 2017 | static void combios_parse_mmio_table(struct drm_device *dev, uint16_t offset) | 2421 | static void combios_parse_mmio_table(struct drm_device *dev, uint16_t offset) |
| 2018 | { | 2422 | { |
| 2019 | struct radeon_device *rdev = dev->dev_private; | 2423 | struct radeon_device *rdev = dev->dev_private; |
diff --git a/drivers/gpu/drm/radeon/radeon_connectors.c b/drivers/gpu/drm/radeon/radeon_connectors.c index 29763ceae3af..5eece186e03c 100644 --- a/drivers/gpu/drm/radeon/radeon_connectors.c +++ b/drivers/gpu/drm/radeon/radeon_connectors.c | |||
| @@ -40,6 +40,26 @@ radeon_atombios_connected_scratch_regs(struct drm_connector *connector, | |||
| 40 | struct drm_encoder *encoder, | 40 | struct drm_encoder *encoder, |
| 41 | bool connected); | 41 | bool connected); |
| 42 | 42 | ||
| 43 | void radeon_connector_hotplug(struct drm_connector *connector) | ||
| 44 | { | ||
| 45 | struct drm_device *dev = connector->dev; | ||
| 46 | struct radeon_device *rdev = dev->dev_private; | ||
| 47 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 48 | |||
| 49 | if (radeon_connector->hpd.hpd != RADEON_HPD_NONE) | ||
| 50 | radeon_hpd_set_polarity(rdev, radeon_connector->hpd.hpd); | ||
| 51 | |||
| 52 | if (connector->connector_type == DRM_MODE_CONNECTOR_DisplayPort) { | ||
| 53 | if (radeon_dp_getsinktype(radeon_connector) == CONNECTOR_OBJECT_ID_DISPLAYPORT) { | ||
| 54 | if (radeon_dp_needs_link_train(radeon_connector)) { | ||
| 55 | if (connector->encoder) | ||
| 56 | dp_link_train(connector->encoder, connector); | ||
| 57 | } | ||
| 58 | } | ||
| 59 | } | ||
| 60 | |||
| 61 | } | ||
| 62 | |||
| 43 | static void radeon_property_change_mode(struct drm_encoder *encoder) | 63 | static void radeon_property_change_mode(struct drm_encoder *encoder) |
| 44 | { | 64 | { |
| 45 | struct drm_crtc *crtc = encoder->crtc; | 65 | struct drm_crtc *crtc = encoder->crtc; |
| @@ -445,10 +465,10 @@ static enum drm_connector_status radeon_lvds_detect(struct drm_connector *connec | |||
| 445 | ret = connector_status_connected; | 465 | ret = connector_status_connected; |
| 446 | else { | 466 | else { |
| 447 | if (radeon_connector->ddc_bus) { | 467 | if (radeon_connector->ddc_bus) { |
| 448 | radeon_i2c_do_lock(radeon_connector, 1); | 468 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 449 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, | 469 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, |
| 450 | &radeon_connector->ddc_bus->adapter); | 470 | &radeon_connector->ddc_bus->adapter); |
| 451 | radeon_i2c_do_lock(radeon_connector, 0); | 471 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 452 | if (radeon_connector->edid) | 472 | if (radeon_connector->edid) |
| 453 | ret = connector_status_connected; | 473 | ret = connector_status_connected; |
| 454 | } | 474 | } |
| @@ -553,17 +573,17 @@ static enum drm_connector_status radeon_vga_detect(struct drm_connector *connect | |||
| 553 | if (!encoder) | 573 | if (!encoder) |
| 554 | ret = connector_status_disconnected; | 574 | ret = connector_status_disconnected; |
| 555 | 575 | ||
| 556 | radeon_i2c_do_lock(radeon_connector, 1); | 576 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 557 | dret = radeon_ddc_probe(radeon_connector); | 577 | dret = radeon_ddc_probe(radeon_connector); |
| 558 | radeon_i2c_do_lock(radeon_connector, 0); | 578 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 559 | if (dret) { | 579 | if (dret) { |
| 560 | if (radeon_connector->edid) { | 580 | if (radeon_connector->edid) { |
| 561 | kfree(radeon_connector->edid); | 581 | kfree(radeon_connector->edid); |
| 562 | radeon_connector->edid = NULL; | 582 | radeon_connector->edid = NULL; |
| 563 | } | 583 | } |
| 564 | radeon_i2c_do_lock(radeon_connector, 1); | 584 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 565 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); | 585 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); |
| 566 | radeon_i2c_do_lock(radeon_connector, 0); | 586 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 567 | 587 | ||
| 568 | if (!radeon_connector->edid) { | 588 | if (!radeon_connector->edid) { |
| 569 | DRM_ERROR("%s: probed a monitor but no|invalid EDID\n", | 589 | DRM_ERROR("%s: probed a monitor but no|invalid EDID\n", |
| @@ -708,17 +728,17 @@ static enum drm_connector_status radeon_dvi_detect(struct drm_connector *connect | |||
| 708 | enum drm_connector_status ret = connector_status_disconnected; | 728 | enum drm_connector_status ret = connector_status_disconnected; |
| 709 | bool dret; | 729 | bool dret; |
| 710 | 730 | ||
| 711 | radeon_i2c_do_lock(radeon_connector, 1); | 731 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 712 | dret = radeon_ddc_probe(radeon_connector); | 732 | dret = radeon_ddc_probe(radeon_connector); |
| 713 | radeon_i2c_do_lock(radeon_connector, 0); | 733 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 714 | if (dret) { | 734 | if (dret) { |
| 715 | if (radeon_connector->edid) { | 735 | if (radeon_connector->edid) { |
| 716 | kfree(radeon_connector->edid); | 736 | kfree(radeon_connector->edid); |
| 717 | radeon_connector->edid = NULL; | 737 | radeon_connector->edid = NULL; |
| 718 | } | 738 | } |
| 719 | radeon_i2c_do_lock(radeon_connector, 1); | 739 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 720 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); | 740 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); |
| 721 | radeon_i2c_do_lock(radeon_connector, 0); | 741 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 722 | 742 | ||
| 723 | if (!radeon_connector->edid) { | 743 | if (!radeon_connector->edid) { |
| 724 | DRM_ERROR("%s: probed a monitor but no|invalid EDID\n", | 744 | DRM_ERROR("%s: probed a monitor but no|invalid EDID\n", |
| @@ -735,6 +755,39 @@ static enum drm_connector_status radeon_dvi_detect(struct drm_connector *connect | |||
| 735 | ret = connector_status_disconnected; | 755 | ret = connector_status_disconnected; |
| 736 | } else | 756 | } else |
| 737 | ret = connector_status_connected; | 757 | ret = connector_status_connected; |
| 758 | |||
| 759 | /* multiple connectors on the same encoder with the same ddc line | ||
| 760 | * This tends to be HDMI and DVI on the same encoder with the | ||
| 761 | * same ddc line. If the edid says HDMI, consider the HDMI port | ||
| 762 | * connected and the DVI port disconnected. If the edid doesn't | ||
| 763 | * say HDMI, vice versa. | ||
| 764 | */ | ||
| 765 | if (radeon_connector->shared_ddc && connector_status_connected) { | ||
| 766 | struct drm_device *dev = connector->dev; | ||
| 767 | struct drm_connector *list_connector; | ||
| 768 | struct radeon_connector *list_radeon_connector; | ||
| 769 | list_for_each_entry(list_connector, &dev->mode_config.connector_list, head) { | ||
| 770 | if (connector == list_connector) | ||
| 771 | continue; | ||
| 772 | list_radeon_connector = to_radeon_connector(list_connector); | ||
| 773 | if (radeon_connector->devices == list_radeon_connector->devices) { | ||
| 774 | if (drm_detect_hdmi_monitor(radeon_connector->edid)) { | ||
| 775 | if (connector->connector_type == DRM_MODE_CONNECTOR_DVID) { | ||
| 776 | kfree(radeon_connector->edid); | ||
| 777 | radeon_connector->edid = NULL; | ||
| 778 | ret = connector_status_disconnected; | ||
| 779 | } | ||
| 780 | } else { | ||
| 781 | if ((connector->connector_type == DRM_MODE_CONNECTOR_HDMIA) || | ||
| 782 | (connector->connector_type == DRM_MODE_CONNECTOR_HDMIB)) { | ||
| 783 | kfree(radeon_connector->edid); | ||
| 784 | radeon_connector->edid = NULL; | ||
| 785 | ret = connector_status_disconnected; | ||
| 786 | } | ||
| 787 | } | ||
| 788 | } | ||
| 789 | } | ||
| 790 | } | ||
| 738 | } | 791 | } |
| 739 | } | 792 | } |
| 740 | 793 | ||
| @@ -863,6 +916,91 @@ struct drm_connector_funcs radeon_dvi_connector_funcs = { | |||
| 863 | .force = radeon_dvi_force, | 916 | .force = radeon_dvi_force, |
| 864 | }; | 917 | }; |
| 865 | 918 | ||
| 919 | static void radeon_dp_connector_destroy(struct drm_connector *connector) | ||
| 920 | { | ||
| 921 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 922 | struct radeon_connector_atom_dig *radeon_dig_connector = radeon_connector->con_priv; | ||
| 923 | |||
| 924 | if (radeon_connector->ddc_bus) | ||
| 925 | radeon_i2c_destroy(radeon_connector->ddc_bus); | ||
| 926 | if (radeon_connector->edid) | ||
| 927 | kfree(radeon_connector->edid); | ||
| 928 | if (radeon_dig_connector->dp_i2c_bus) | ||
| 929 | radeon_i2c_destroy(radeon_dig_connector->dp_i2c_bus); | ||
| 930 | kfree(radeon_connector->con_priv); | ||
| 931 | drm_sysfs_connector_remove(connector); | ||
| 932 | drm_connector_cleanup(connector); | ||
| 933 | kfree(connector); | ||
| 934 | } | ||
| 935 | |||
| 936 | static int radeon_dp_get_modes(struct drm_connector *connector) | ||
| 937 | { | ||
| 938 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 939 | int ret; | ||
| 940 | |||
| 941 | ret = radeon_ddc_get_modes(radeon_connector); | ||
| 942 | return ret; | ||
| 943 | } | ||
| 944 | |||
| 945 | static enum drm_connector_status radeon_dp_detect(struct drm_connector *connector) | ||
| 946 | { | ||
| 947 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 948 | enum drm_connector_status ret = connector_status_disconnected; | ||
| 949 | struct radeon_connector_atom_dig *radeon_dig_connector = radeon_connector->con_priv; | ||
| 950 | u8 sink_type; | ||
| 951 | |||
| 952 | if (radeon_connector->edid) { | ||
| 953 | kfree(radeon_connector->edid); | ||
| 954 | radeon_connector->edid = NULL; | ||
| 955 | } | ||
| 956 | |||
| 957 | sink_type = radeon_dp_getsinktype(radeon_connector); | ||
| 958 | if (sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) { | ||
| 959 | if (radeon_dp_getdpcd(radeon_connector)) { | ||
| 960 | radeon_dig_connector->dp_sink_type = sink_type; | ||
| 961 | ret = connector_status_connected; | ||
| 962 | } | ||
| 963 | } else { | ||
| 964 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); | ||
| 965 | if (radeon_ddc_probe(radeon_connector)) { | ||
| 966 | radeon_dig_connector->dp_sink_type = sink_type; | ||
| 967 | ret = connector_status_connected; | ||
| 968 | } | ||
| 969 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); | ||
| 970 | } | ||
| 971 | |||
| 972 | return ret; | ||
| 973 | } | ||
| 974 | |||
| 975 | static int radeon_dp_mode_valid(struct drm_connector *connector, | ||
| 976 | struct drm_display_mode *mode) | ||
| 977 | { | ||
| 978 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 979 | struct radeon_connector_atom_dig *radeon_dig_connector = radeon_connector->con_priv; | ||
| 980 | |||
| 981 | /* XXX check mode bandwidth */ | ||
| 982 | |||
| 983 | if (radeon_dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) | ||
| 984 | return radeon_dp_mode_valid_helper(radeon_connector, mode); | ||
| 985 | else | ||
| 986 | return MODE_OK; | ||
| 987 | } | ||
| 988 | |||
| 989 | struct drm_connector_helper_funcs radeon_dp_connector_helper_funcs = { | ||
| 990 | .get_modes = radeon_dp_get_modes, | ||
| 991 | .mode_valid = radeon_dp_mode_valid, | ||
| 992 | .best_encoder = radeon_dvi_encoder, | ||
| 993 | }; | ||
| 994 | |||
| 995 | struct drm_connector_funcs radeon_dp_connector_funcs = { | ||
| 996 | .dpms = drm_helper_connector_dpms, | ||
| 997 | .detect = radeon_dp_detect, | ||
| 998 | .fill_modes = drm_helper_probe_single_connector_modes, | ||
| 999 | .set_property = radeon_connector_set_property, | ||
| 1000 | .destroy = radeon_dp_connector_destroy, | ||
| 1001 | .force = radeon_dvi_force, | ||
| 1002 | }; | ||
| 1003 | |||
| 866 | void | 1004 | void |
| 867 | radeon_add_atom_connector(struct drm_device *dev, | 1005 | radeon_add_atom_connector(struct drm_device *dev, |
| 868 | uint32_t connector_id, | 1006 | uint32_t connector_id, |
| @@ -871,7 +1009,8 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 871 | struct radeon_i2c_bus_rec *i2c_bus, | 1009 | struct radeon_i2c_bus_rec *i2c_bus, |
| 872 | bool linkb, | 1010 | bool linkb, |
| 873 | uint32_t igp_lane_info, | 1011 | uint32_t igp_lane_info, |
| 874 | uint16_t connector_object_id) | 1012 | uint16_t connector_object_id, |
| 1013 | struct radeon_hpd *hpd) | ||
| 875 | { | 1014 | { |
| 876 | struct radeon_device *rdev = dev->dev_private; | 1015 | struct radeon_device *rdev = dev->dev_private; |
| 877 | struct drm_connector *connector; | 1016 | struct drm_connector *connector; |
| @@ -911,6 +1050,7 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 911 | radeon_connector->devices = supported_device; | 1050 | radeon_connector->devices = supported_device; |
| 912 | radeon_connector->shared_ddc = shared_ddc; | 1051 | radeon_connector->shared_ddc = shared_ddc; |
| 913 | radeon_connector->connector_object_id = connector_object_id; | 1052 | radeon_connector->connector_object_id = connector_object_id; |
| 1053 | radeon_connector->hpd = *hpd; | ||
| 914 | switch (connector_type) { | 1054 | switch (connector_type) { |
| 915 | case DRM_MODE_CONNECTOR_VGA: | 1055 | case DRM_MODE_CONNECTOR_VGA: |
| 916 | drm_connector_init(dev, &radeon_connector->base, &radeon_vga_connector_funcs, connector_type); | 1056 | drm_connector_init(dev, &radeon_connector->base, &radeon_vga_connector_funcs, connector_type); |
| @@ -963,10 +1103,12 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 963 | drm_connector_attach_property(&radeon_connector->base, | 1103 | drm_connector_attach_property(&radeon_connector->base, |
| 964 | rdev->mode_info.coherent_mode_property, | 1104 | rdev->mode_info.coherent_mode_property, |
| 965 | 1); | 1105 | 1); |
| 966 | radeon_connector->dac_load_detect = true; | 1106 | if (connector_type == DRM_MODE_CONNECTOR_DVII) { |
| 967 | drm_connector_attach_property(&radeon_connector->base, | 1107 | radeon_connector->dac_load_detect = true; |
| 968 | rdev->mode_info.load_detect_property, | 1108 | drm_connector_attach_property(&radeon_connector->base, |
| 969 | 1); | 1109 | rdev->mode_info.load_detect_property, |
| 1110 | 1); | ||
| 1111 | } | ||
| 970 | break; | 1112 | break; |
| 971 | case DRM_MODE_CONNECTOR_HDMIA: | 1113 | case DRM_MODE_CONNECTOR_HDMIA: |
| 972 | case DRM_MODE_CONNECTOR_HDMIB: | 1114 | case DRM_MODE_CONNECTOR_HDMIB: |
| @@ -997,16 +1139,23 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 997 | radeon_dig_connector->linkb = linkb; | 1139 | radeon_dig_connector->linkb = linkb; |
| 998 | radeon_dig_connector->igp_lane_info = igp_lane_info; | 1140 | radeon_dig_connector->igp_lane_info = igp_lane_info; |
| 999 | radeon_connector->con_priv = radeon_dig_connector; | 1141 | radeon_connector->con_priv = radeon_dig_connector; |
| 1000 | drm_connector_init(dev, &radeon_connector->base, &radeon_dvi_connector_funcs, connector_type); | 1142 | drm_connector_init(dev, &radeon_connector->base, &radeon_dp_connector_funcs, connector_type); |
| 1001 | ret = drm_connector_helper_add(&radeon_connector->base, &radeon_dvi_connector_helper_funcs); | 1143 | ret = drm_connector_helper_add(&radeon_connector->base, &radeon_dp_connector_helper_funcs); |
| 1002 | if (ret) | 1144 | if (ret) |
| 1003 | goto failed; | 1145 | goto failed; |
| 1004 | if (i2c_bus->valid) { | 1146 | if (i2c_bus->valid) { |
| 1147 | /* add DP i2c bus */ | ||
| 1148 | radeon_dig_connector->dp_i2c_bus = radeon_i2c_create_dp(dev, i2c_bus, "DP-auxch"); | ||
| 1149 | if (!radeon_dig_connector->dp_i2c_bus) | ||
| 1150 | goto failed; | ||
| 1005 | radeon_connector->ddc_bus = radeon_i2c_create(dev, i2c_bus, "DP"); | 1151 | radeon_connector->ddc_bus = radeon_i2c_create(dev, i2c_bus, "DP"); |
| 1006 | if (!radeon_connector->ddc_bus) | 1152 | if (!radeon_connector->ddc_bus) |
| 1007 | goto failed; | 1153 | goto failed; |
| 1008 | } | 1154 | } |
| 1009 | subpixel_order = SubPixelHorizontalRGB; | 1155 | subpixel_order = SubPixelHorizontalRGB; |
| 1156 | drm_connector_attach_property(&radeon_connector->base, | ||
| 1157 | rdev->mode_info.coherent_mode_property, | ||
| 1158 | 1); | ||
| 1010 | break; | 1159 | break; |
| 1011 | case DRM_MODE_CONNECTOR_SVIDEO: | 1160 | case DRM_MODE_CONNECTOR_SVIDEO: |
| 1012 | case DRM_MODE_CONNECTOR_Composite: | 1161 | case DRM_MODE_CONNECTOR_Composite: |
| @@ -1020,6 +1169,9 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1020 | drm_connector_attach_property(&radeon_connector->base, | 1169 | drm_connector_attach_property(&radeon_connector->base, |
| 1021 | rdev->mode_info.load_detect_property, | 1170 | rdev->mode_info.load_detect_property, |
| 1022 | 1); | 1171 | 1); |
| 1172 | drm_connector_attach_property(&radeon_connector->base, | ||
| 1173 | rdev->mode_info.tv_std_property, | ||
| 1174 | 1); | ||
| 1023 | } | 1175 | } |
| 1024 | break; | 1176 | break; |
| 1025 | case DRM_MODE_CONNECTOR_LVDS: | 1177 | case DRM_MODE_CONNECTOR_LVDS: |
| @@ -1038,7 +1190,6 @@ radeon_add_atom_connector(struct drm_device *dev, | |||
| 1038 | if (!radeon_connector->ddc_bus) | 1190 | if (!radeon_connector->ddc_bus) |
| 1039 | goto failed; | 1191 | goto failed; |
| 1040 | } | 1192 | } |
| 1041 | drm_mode_create_scaling_mode_property(dev); | ||
| 1042 | drm_connector_attach_property(&radeon_connector->base, | 1193 | drm_connector_attach_property(&radeon_connector->base, |
| 1043 | dev->mode_config.scaling_mode_property, | 1194 | dev->mode_config.scaling_mode_property, |
| 1044 | DRM_MODE_SCALE_FULLSCREEN); | 1195 | DRM_MODE_SCALE_FULLSCREEN); |
| @@ -1063,7 +1214,8 @@ radeon_add_legacy_connector(struct drm_device *dev, | |||
| 1063 | uint32_t supported_device, | 1214 | uint32_t supported_device, |
| 1064 | int connector_type, | 1215 | int connector_type, |
| 1065 | struct radeon_i2c_bus_rec *i2c_bus, | 1216 | struct radeon_i2c_bus_rec *i2c_bus, |
| 1066 | uint16_t connector_object_id) | 1217 | uint16_t connector_object_id, |
| 1218 | struct radeon_hpd *hpd) | ||
| 1067 | { | 1219 | { |
| 1068 | struct radeon_device *rdev = dev->dev_private; | 1220 | struct radeon_device *rdev = dev->dev_private; |
| 1069 | struct drm_connector *connector; | 1221 | struct drm_connector *connector; |
| @@ -1093,6 +1245,7 @@ radeon_add_legacy_connector(struct drm_device *dev, | |||
| 1093 | radeon_connector->connector_id = connector_id; | 1245 | radeon_connector->connector_id = connector_id; |
| 1094 | radeon_connector->devices = supported_device; | 1246 | radeon_connector->devices = supported_device; |
| 1095 | radeon_connector->connector_object_id = connector_object_id; | 1247 | radeon_connector->connector_object_id = connector_object_id; |
| 1248 | radeon_connector->hpd = *hpd; | ||
| 1096 | switch (connector_type) { | 1249 | switch (connector_type) { |
| 1097 | case DRM_MODE_CONNECTOR_VGA: | 1250 | case DRM_MODE_CONNECTOR_VGA: |
| 1098 | drm_connector_init(dev, &radeon_connector->base, &radeon_vga_connector_funcs, connector_type); | 1251 | drm_connector_init(dev, &radeon_connector->base, &radeon_vga_connector_funcs, connector_type); |
| @@ -1160,6 +1313,9 @@ radeon_add_legacy_connector(struct drm_device *dev, | |||
| 1160 | drm_connector_attach_property(&radeon_connector->base, | 1313 | drm_connector_attach_property(&radeon_connector->base, |
| 1161 | rdev->mode_info.load_detect_property, | 1314 | rdev->mode_info.load_detect_property, |
| 1162 | 1); | 1315 | 1); |
| 1316 | drm_connector_attach_property(&radeon_connector->base, | ||
| 1317 | rdev->mode_info.tv_std_property, | ||
| 1318 | 1); | ||
| 1163 | } | 1319 | } |
| 1164 | break; | 1320 | break; |
| 1165 | case DRM_MODE_CONNECTOR_LVDS: | 1321 | case DRM_MODE_CONNECTOR_LVDS: |
diff --git a/drivers/gpu/drm/radeon/radeon_cp.c b/drivers/gpu/drm/radeon/radeon_cp.c index 4f7afc79dd82..0b2f9c2ad2c1 100644 --- a/drivers/gpu/drm/radeon/radeon_cp.c +++ b/drivers/gpu/drm/radeon/radeon_cp.c | |||
| @@ -1941,8 +1941,8 @@ struct drm_buf *radeon_freelist_get(struct drm_device * dev) | |||
| 1941 | for (t = 0; t < dev_priv->usec_timeout; t++) { | 1941 | for (t = 0; t < dev_priv->usec_timeout; t++) { |
| 1942 | u32 done_age = GET_SCRATCH(dev_priv, 1); | 1942 | u32 done_age = GET_SCRATCH(dev_priv, 1); |
| 1943 | DRM_DEBUG("done_age = %d\n", done_age); | 1943 | DRM_DEBUG("done_age = %d\n", done_age); |
| 1944 | for (i = start; i < dma->buf_count; i++) { | 1944 | for (i = 0; i < dma->buf_count; i++) { |
| 1945 | buf = dma->buflist[i]; | 1945 | buf = dma->buflist[start]; |
| 1946 | buf_priv = buf->dev_private; | 1946 | buf_priv = buf->dev_private; |
| 1947 | if (buf->file_priv == NULL || (buf->pending && | 1947 | if (buf->file_priv == NULL || (buf->pending && |
| 1948 | buf_priv->age <= | 1948 | buf_priv->age <= |
| @@ -1951,7 +1951,8 @@ struct drm_buf *radeon_freelist_get(struct drm_device * dev) | |||
| 1951 | buf->pending = 0; | 1951 | buf->pending = 0; |
| 1952 | return buf; | 1952 | return buf; |
| 1953 | } | 1953 | } |
| 1954 | start = 0; | 1954 | if (++start >= dma->buf_count) |
| 1955 | start = 0; | ||
| 1955 | } | 1956 | } |
| 1956 | 1957 | ||
| 1957 | if (t) { | 1958 | if (t) { |
| @@ -1960,47 +1961,9 @@ struct drm_buf *radeon_freelist_get(struct drm_device * dev) | |||
| 1960 | } | 1961 | } |
| 1961 | } | 1962 | } |
| 1962 | 1963 | ||
| 1963 | DRM_DEBUG("returning NULL!\n"); | ||
| 1964 | return NULL; | 1964 | return NULL; |
| 1965 | } | 1965 | } |
| 1966 | 1966 | ||
| 1967 | #if 0 | ||
| 1968 | struct drm_buf *radeon_freelist_get(struct drm_device * dev) | ||
| 1969 | { | ||
| 1970 | struct drm_device_dma *dma = dev->dma; | ||
| 1971 | drm_radeon_private_t *dev_priv = dev->dev_private; | ||
| 1972 | drm_radeon_buf_priv_t *buf_priv; | ||
| 1973 | struct drm_buf *buf; | ||
| 1974 | int i, t; | ||
| 1975 | int start; | ||
| 1976 | u32 done_age; | ||
| 1977 | |||
| 1978 | done_age = radeon_read_ring_rptr(dev_priv, RADEON_SCRATCHOFF(1)); | ||
| 1979 | if (++dev_priv->last_buf >= dma->buf_count) | ||
| 1980 | dev_priv->last_buf = 0; | ||
| 1981 | |||
| 1982 | start = dev_priv->last_buf; | ||
| 1983 | dev_priv->stats.freelist_loops++; | ||
| 1984 | |||
| 1985 | for (t = 0; t < 2; t++) { | ||
| 1986 | for (i = start; i < dma->buf_count; i++) { | ||
| 1987 | buf = dma->buflist[i]; | ||
| 1988 | buf_priv = buf->dev_private; | ||
| 1989 | if (buf->file_priv == 0 || (buf->pending && | ||
| 1990 | buf_priv->age <= | ||
| 1991 | done_age)) { | ||
| 1992 | dev_priv->stats.requested_bufs++; | ||
| 1993 | buf->pending = 0; | ||
| 1994 | return buf; | ||
| 1995 | } | ||
| 1996 | } | ||
| 1997 | start = 0; | ||
| 1998 | } | ||
| 1999 | |||
| 2000 | return NULL; | ||
| 2001 | } | ||
| 2002 | #endif | ||
| 2003 | |||
| 2004 | void radeon_freelist_reset(struct drm_device * dev) | 1967 | void radeon_freelist_reset(struct drm_device * dev) |
| 2005 | { | 1968 | { |
| 2006 | struct drm_device_dma *dma = dev->dma; | 1969 | struct drm_device_dma *dma = dev->dma; |
diff --git a/drivers/gpu/drm/radeon/radeon_cs.c b/drivers/gpu/drm/radeon/radeon_cs.c index 5ab2cf96a264..65590a0f1d93 100644 --- a/drivers/gpu/drm/radeon/radeon_cs.c +++ b/drivers/gpu/drm/radeon/radeon_cs.c | |||
| @@ -76,17 +76,17 @@ int radeon_cs_parser_relocs(struct radeon_cs_parser *p) | |||
| 76 | } | 76 | } |
| 77 | p->relocs_ptr[i] = &p->relocs[i]; | 77 | p->relocs_ptr[i] = &p->relocs[i]; |
| 78 | p->relocs[i].robj = p->relocs[i].gobj->driver_private; | 78 | p->relocs[i].robj = p->relocs[i].gobj->driver_private; |
| 79 | p->relocs[i].lobj.robj = p->relocs[i].robj; | 79 | p->relocs[i].lobj.bo = p->relocs[i].robj; |
| 80 | p->relocs[i].lobj.rdomain = r->read_domains; | 80 | p->relocs[i].lobj.rdomain = r->read_domains; |
| 81 | p->relocs[i].lobj.wdomain = r->write_domain; | 81 | p->relocs[i].lobj.wdomain = r->write_domain; |
| 82 | p->relocs[i].handle = r->handle; | 82 | p->relocs[i].handle = r->handle; |
| 83 | p->relocs[i].flags = r->flags; | 83 | p->relocs[i].flags = r->flags; |
| 84 | INIT_LIST_HEAD(&p->relocs[i].lobj.list); | 84 | INIT_LIST_HEAD(&p->relocs[i].lobj.list); |
| 85 | radeon_object_list_add_object(&p->relocs[i].lobj, | 85 | radeon_bo_list_add_object(&p->relocs[i].lobj, |
| 86 | &p->validated); | 86 | &p->validated); |
| 87 | } | 87 | } |
| 88 | } | 88 | } |
| 89 | return radeon_object_list_validate(&p->validated, p->ib->fence); | 89 | return radeon_bo_list_validate(&p->validated, p->ib->fence); |
| 90 | } | 90 | } |
| 91 | 91 | ||
| 92 | int radeon_cs_parser_init(struct radeon_cs_parser *p, void *data) | 92 | int radeon_cs_parser_init(struct radeon_cs_parser *p, void *data) |
| @@ -190,9 +190,10 @@ static void radeon_cs_parser_fini(struct radeon_cs_parser *parser, int error) | |||
| 190 | unsigned i; | 190 | unsigned i; |
| 191 | 191 | ||
| 192 | if (error) { | 192 | if (error) { |
| 193 | radeon_object_list_unvalidate(&parser->validated); | 193 | radeon_bo_list_unvalidate(&parser->validated, |
| 194 | parser->ib->fence); | ||
| 194 | } else { | 195 | } else { |
| 195 | radeon_object_list_clean(&parser->validated); | 196 | radeon_bo_list_unreserve(&parser->validated); |
| 196 | } | 197 | } |
| 197 | for (i = 0; i < parser->nrelocs; i++) { | 198 | for (i = 0; i < parser->nrelocs; i++) { |
| 198 | if (parser->relocs[i].gobj) { | 199 | if (parser->relocs[i].gobj) { |
diff --git a/drivers/gpu/drm/radeon/radeon_device.c b/drivers/gpu/drm/radeon/radeon_device.c index 41bb76fbe734..02bcdb1240c0 100644 --- a/drivers/gpu/drm/radeon/radeon_device.c +++ b/drivers/gpu/drm/radeon/radeon_device.c | |||
| @@ -44,10 +44,11 @@ void radeon_surface_init(struct radeon_device *rdev) | |||
| 44 | if (rdev->family < CHIP_R600) { | 44 | if (rdev->family < CHIP_R600) { |
| 45 | int i; | 45 | int i; |
| 46 | 46 | ||
| 47 | for (i = 0; i < 8; i++) { | 47 | for (i = 0; i < RADEON_GEM_MAX_SURFACES; i++) { |
| 48 | WREG32(RADEON_SURFACE0_INFO + | 48 | if (rdev->surface_regs[i].bo) |
| 49 | i * (RADEON_SURFACE1_INFO - RADEON_SURFACE0_INFO), | 49 | radeon_bo_get_surface_reg(rdev->surface_regs[i].bo); |
| 50 | 0); | 50 | else |
| 51 | radeon_clear_surface_reg(rdev, i); | ||
| 51 | } | 52 | } |
| 52 | /* enable surfaces */ | 53 | /* enable surfaces */ |
| 53 | WREG32(RADEON_SURFACE_CNTL, 0); | 54 | WREG32(RADEON_SURFACE_CNTL, 0); |
| @@ -208,6 +209,24 @@ bool radeon_card_posted(struct radeon_device *rdev) | |||
| 208 | 209 | ||
| 209 | } | 210 | } |
| 210 | 211 | ||
| 212 | bool radeon_boot_test_post_card(struct radeon_device *rdev) | ||
| 213 | { | ||
| 214 | if (radeon_card_posted(rdev)) | ||
| 215 | return true; | ||
| 216 | |||
| 217 | if (rdev->bios) { | ||
| 218 | DRM_INFO("GPU not posted. posting now...\n"); | ||
| 219 | if (rdev->is_atom_bios) | ||
| 220 | atom_asic_init(rdev->mode_info.atom_context); | ||
| 221 | else | ||
| 222 | radeon_combios_asic_init(rdev->ddev); | ||
| 223 | return true; | ||
| 224 | } else { | ||
| 225 | dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n"); | ||
| 226 | return false; | ||
| 227 | } | ||
| 228 | } | ||
| 229 | |||
| 211 | int radeon_dummy_page_init(struct radeon_device *rdev) | 230 | int radeon_dummy_page_init(struct radeon_device *rdev) |
| 212 | { | 231 | { |
| 213 | rdev->dummy_page.page = alloc_page(GFP_DMA32 | GFP_KERNEL | __GFP_ZERO); | 232 | rdev->dummy_page.page = alloc_page(GFP_DMA32 | GFP_KERNEL | __GFP_ZERO); |
| @@ -463,12 +482,16 @@ int radeon_atombios_init(struct radeon_device *rdev) | |||
| 463 | 482 | ||
| 464 | rdev->mode_info.atom_context = atom_parse(atom_card_info, rdev->bios); | 483 | rdev->mode_info.atom_context = atom_parse(atom_card_info, rdev->bios); |
| 465 | radeon_atom_initialize_bios_scratch_regs(rdev->ddev); | 484 | radeon_atom_initialize_bios_scratch_regs(rdev->ddev); |
| 485 | atom_allocate_fb_scratch(rdev->mode_info.atom_context); | ||
| 466 | return 0; | 486 | return 0; |
| 467 | } | 487 | } |
| 468 | 488 | ||
| 469 | void radeon_atombios_fini(struct radeon_device *rdev) | 489 | void radeon_atombios_fini(struct radeon_device *rdev) |
| 470 | { | 490 | { |
| 471 | kfree(rdev->mode_info.atom_context); | 491 | if (rdev->mode_info.atom_context) { |
| 492 | kfree(rdev->mode_info.atom_context->scratch); | ||
| 493 | kfree(rdev->mode_info.atom_context); | ||
| 494 | } | ||
| 472 | kfree(rdev->mode_info.atom_card_info); | 495 | kfree(rdev->mode_info.atom_card_info); |
| 473 | } | 496 | } |
| 474 | 497 | ||
| @@ -544,16 +567,24 @@ int radeon_device_init(struct radeon_device *rdev, | |||
| 544 | mutex_init(&rdev->cs_mutex); | 567 | mutex_init(&rdev->cs_mutex); |
| 545 | mutex_init(&rdev->ib_pool.mutex); | 568 | mutex_init(&rdev->ib_pool.mutex); |
| 546 | mutex_init(&rdev->cp.mutex); | 569 | mutex_init(&rdev->cp.mutex); |
| 570 | if (rdev->family >= CHIP_R600) | ||
| 571 | spin_lock_init(&rdev->ih.lock); | ||
| 572 | mutex_init(&rdev->gem.mutex); | ||
| 547 | rwlock_init(&rdev->fence_drv.lock); | 573 | rwlock_init(&rdev->fence_drv.lock); |
| 548 | INIT_LIST_HEAD(&rdev->gem.objects); | 574 | INIT_LIST_HEAD(&rdev->gem.objects); |
| 549 | 575 | ||
| 576 | /* setup workqueue */ | ||
| 577 | rdev->wq = create_workqueue("radeon"); | ||
| 578 | if (rdev->wq == NULL) | ||
| 579 | return -ENOMEM; | ||
| 580 | |||
| 550 | /* Set asic functions */ | 581 | /* Set asic functions */ |
| 551 | r = radeon_asic_init(rdev); | 582 | r = radeon_asic_init(rdev); |
| 552 | if (r) { | 583 | if (r) { |
| 553 | return r; | 584 | return r; |
| 554 | } | 585 | } |
| 555 | 586 | ||
| 556 | if (radeon_agpmode == -1) { | 587 | if (rdev->flags & RADEON_IS_AGP && radeon_agpmode == -1) { |
| 557 | radeon_agp_disable(rdev); | 588 | radeon_agp_disable(rdev); |
| 558 | } | 589 | } |
| 559 | 590 | ||
| @@ -620,6 +651,7 @@ void radeon_device_fini(struct radeon_device *rdev) | |||
| 620 | DRM_INFO("radeon: finishing device.\n"); | 651 | DRM_INFO("radeon: finishing device.\n"); |
| 621 | rdev->shutdown = true; | 652 | rdev->shutdown = true; |
| 622 | radeon_fini(rdev); | 653 | radeon_fini(rdev); |
| 654 | destroy_workqueue(rdev->wq); | ||
| 623 | vga_client_register(rdev->pdev, NULL, NULL, NULL); | 655 | vga_client_register(rdev->pdev, NULL, NULL, NULL); |
| 624 | iounmap(rdev->rmmio); | 656 | iounmap(rdev->rmmio); |
| 625 | rdev->rmmio = NULL; | 657 | rdev->rmmio = NULL; |
| @@ -633,6 +665,7 @@ int radeon_suspend_kms(struct drm_device *dev, pm_message_t state) | |||
| 633 | { | 665 | { |
| 634 | struct radeon_device *rdev = dev->dev_private; | 666 | struct radeon_device *rdev = dev->dev_private; |
| 635 | struct drm_crtc *crtc; | 667 | struct drm_crtc *crtc; |
| 668 | int r; | ||
| 636 | 669 | ||
| 637 | if (dev == NULL || rdev == NULL) { | 670 | if (dev == NULL || rdev == NULL) { |
| 638 | return -ENODEV; | 671 | return -ENODEV; |
| @@ -643,26 +676,31 @@ int radeon_suspend_kms(struct drm_device *dev, pm_message_t state) | |||
| 643 | /* unpin the front buffers */ | 676 | /* unpin the front buffers */ |
| 644 | list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { | 677 | list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) { |
| 645 | struct radeon_framebuffer *rfb = to_radeon_framebuffer(crtc->fb); | 678 | struct radeon_framebuffer *rfb = to_radeon_framebuffer(crtc->fb); |
| 646 | struct radeon_object *robj; | 679 | struct radeon_bo *robj; |
| 647 | 680 | ||
| 648 | if (rfb == NULL || rfb->obj == NULL) { | 681 | if (rfb == NULL || rfb->obj == NULL) { |
| 649 | continue; | 682 | continue; |
| 650 | } | 683 | } |
| 651 | robj = rfb->obj->driver_private; | 684 | robj = rfb->obj->driver_private; |
| 652 | if (robj != rdev->fbdev_robj) { | 685 | if (robj != rdev->fbdev_rbo) { |
| 653 | radeon_object_unpin(robj); | 686 | r = radeon_bo_reserve(robj, false); |
| 687 | if (unlikely(r == 0)) { | ||
| 688 | radeon_bo_unpin(robj); | ||
| 689 | radeon_bo_unreserve(robj); | ||
| 690 | } | ||
| 654 | } | 691 | } |
| 655 | } | 692 | } |
| 656 | /* evict vram memory */ | 693 | /* evict vram memory */ |
| 657 | radeon_object_evict_vram(rdev); | 694 | radeon_bo_evict_vram(rdev); |
| 658 | /* wait for gpu to finish processing current batch */ | 695 | /* wait for gpu to finish processing current batch */ |
| 659 | radeon_fence_wait_last(rdev); | 696 | radeon_fence_wait_last(rdev); |
| 660 | 697 | ||
| 661 | radeon_save_bios_scratch_regs(rdev); | 698 | radeon_save_bios_scratch_regs(rdev); |
| 662 | 699 | ||
| 663 | radeon_suspend(rdev); | 700 | radeon_suspend(rdev); |
| 701 | radeon_hpd_fini(rdev); | ||
| 664 | /* evict remaining vram memory */ | 702 | /* evict remaining vram memory */ |
| 665 | radeon_object_evict_vram(rdev); | 703 | radeon_bo_evict_vram(rdev); |
| 666 | 704 | ||
| 667 | pci_save_state(dev->pdev); | 705 | pci_save_state(dev->pdev); |
| 668 | if (state.event == PM_EVENT_SUSPEND) { | 706 | if (state.event == PM_EVENT_SUSPEND) { |
| @@ -695,6 +733,8 @@ int radeon_resume_kms(struct drm_device *dev) | |||
| 695 | fb_set_suspend(rdev->fbdev_info, 0); | 733 | fb_set_suspend(rdev->fbdev_info, 0); |
| 696 | release_console_sem(); | 734 | release_console_sem(); |
| 697 | 735 | ||
| 736 | /* reset hpd state */ | ||
| 737 | radeon_hpd_init(rdev); | ||
| 698 | /* blat the mode back in */ | 738 | /* blat the mode back in */ |
| 699 | drm_helper_resume_force_mode(dev); | 739 | drm_helper_resume_force_mode(dev); |
| 700 | return 0; | 740 | return 0; |
diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c index c85df4afcb7a..a133b833e45d 100644 --- a/drivers/gpu/drm/radeon/radeon_display.c +++ b/drivers/gpu/drm/radeon/radeon_display.c | |||
| @@ -250,6 +250,16 @@ static const char *connector_names[13] = { | |||
| 250 | "HDMI-B", | 250 | "HDMI-B", |
| 251 | }; | 251 | }; |
| 252 | 252 | ||
| 253 | static const char *hpd_names[7] = { | ||
| 254 | "NONE", | ||
| 255 | "HPD1", | ||
| 256 | "HPD2", | ||
| 257 | "HPD3", | ||
| 258 | "HPD4", | ||
| 259 | "HPD5", | ||
| 260 | "HPD6", | ||
| 261 | }; | ||
| 262 | |||
| 253 | static void radeon_print_display_setup(struct drm_device *dev) | 263 | static void radeon_print_display_setup(struct drm_device *dev) |
| 254 | { | 264 | { |
| 255 | struct drm_connector *connector; | 265 | struct drm_connector *connector; |
| @@ -264,16 +274,18 @@ static void radeon_print_display_setup(struct drm_device *dev) | |||
| 264 | radeon_connector = to_radeon_connector(connector); | 274 | radeon_connector = to_radeon_connector(connector); |
| 265 | DRM_INFO("Connector %d:\n", i); | 275 | DRM_INFO("Connector %d:\n", i); |
| 266 | DRM_INFO(" %s\n", connector_names[connector->connector_type]); | 276 | DRM_INFO(" %s\n", connector_names[connector->connector_type]); |
| 277 | if (radeon_connector->hpd.hpd != RADEON_HPD_NONE) | ||
| 278 | DRM_INFO(" %s\n", hpd_names[radeon_connector->hpd.hpd]); | ||
| 267 | if (radeon_connector->ddc_bus) | 279 | if (radeon_connector->ddc_bus) |
| 268 | DRM_INFO(" DDC: 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n", | 280 | DRM_INFO(" DDC: 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n", |
| 269 | radeon_connector->ddc_bus->rec.mask_clk_reg, | 281 | radeon_connector->ddc_bus->rec.mask_clk_reg, |
| 270 | radeon_connector->ddc_bus->rec.mask_data_reg, | 282 | radeon_connector->ddc_bus->rec.mask_data_reg, |
| 271 | radeon_connector->ddc_bus->rec.a_clk_reg, | 283 | radeon_connector->ddc_bus->rec.a_clk_reg, |
| 272 | radeon_connector->ddc_bus->rec.a_data_reg, | 284 | radeon_connector->ddc_bus->rec.a_data_reg, |
| 273 | radeon_connector->ddc_bus->rec.put_clk_reg, | 285 | radeon_connector->ddc_bus->rec.en_clk_reg, |
| 274 | radeon_connector->ddc_bus->rec.put_data_reg, | 286 | radeon_connector->ddc_bus->rec.en_data_reg, |
| 275 | radeon_connector->ddc_bus->rec.get_clk_reg, | 287 | radeon_connector->ddc_bus->rec.y_clk_reg, |
| 276 | radeon_connector->ddc_bus->rec.get_data_reg); | 288 | radeon_connector->ddc_bus->rec.y_data_reg); |
| 277 | DRM_INFO(" Encoders:\n"); | 289 | DRM_INFO(" Encoders:\n"); |
| 278 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { | 290 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { |
| 279 | radeon_encoder = to_radeon_encoder(encoder); | 291 | radeon_encoder = to_radeon_encoder(encoder); |
| @@ -324,6 +336,7 @@ static bool radeon_setup_enc_conn(struct drm_device *dev) | |||
| 324 | ret = radeon_get_legacy_connector_info_from_table(dev); | 336 | ret = radeon_get_legacy_connector_info_from_table(dev); |
| 325 | } | 337 | } |
| 326 | if (ret) { | 338 | if (ret) { |
| 339 | radeon_setup_encoder_clones(dev); | ||
| 327 | radeon_print_display_setup(dev); | 340 | radeon_print_display_setup(dev); |
| 328 | list_for_each_entry(drm_connector, &dev->mode_config.connector_list, head) | 341 | list_for_each_entry(drm_connector, &dev->mode_config.connector_list, head) |
| 329 | radeon_ddc_dump(drm_connector); | 342 | radeon_ddc_dump(drm_connector); |
| @@ -336,12 +349,17 @@ int radeon_ddc_get_modes(struct radeon_connector *radeon_connector) | |||
| 336 | { | 349 | { |
| 337 | int ret = 0; | 350 | int ret = 0; |
| 338 | 351 | ||
| 352 | if (radeon_connector->base.connector_type == DRM_MODE_CONNECTOR_DisplayPort) { | ||
| 353 | struct radeon_connector_atom_dig *dig = radeon_connector->con_priv; | ||
| 354 | if (dig->dp_i2c_bus) | ||
| 355 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &dig->dp_i2c_bus->adapter); | ||
| 356 | } | ||
| 339 | if (!radeon_connector->ddc_bus) | 357 | if (!radeon_connector->ddc_bus) |
| 340 | return -1; | 358 | return -1; |
| 341 | if (!radeon_connector->edid) { | 359 | if (!radeon_connector->edid) { |
| 342 | radeon_i2c_do_lock(radeon_connector, 1); | 360 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 343 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); | 361 | radeon_connector->edid = drm_get_edid(&radeon_connector->base, &radeon_connector->ddc_bus->adapter); |
| 344 | radeon_i2c_do_lock(radeon_connector, 0); | 362 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 345 | } | 363 | } |
| 346 | 364 | ||
| 347 | if (radeon_connector->edid) { | 365 | if (radeon_connector->edid) { |
| @@ -361,9 +379,9 @@ static int radeon_ddc_dump(struct drm_connector *connector) | |||
| 361 | 379 | ||
| 362 | if (!radeon_connector->ddc_bus) | 380 | if (!radeon_connector->ddc_bus) |
| 363 | return -1; | 381 | return -1; |
| 364 | radeon_i2c_do_lock(radeon_connector, 1); | 382 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 1); |
| 365 | edid = drm_get_edid(connector, &radeon_connector->ddc_bus->adapter); | 383 | edid = drm_get_edid(connector, &radeon_connector->ddc_bus->adapter); |
| 366 | radeon_i2c_do_lock(radeon_connector, 0); | 384 | radeon_i2c_do_lock(radeon_connector->ddc_bus, 0); |
| 367 | if (edid) { | 385 | if (edid) { |
| 368 | kfree(edid); | 386 | kfree(edid); |
| 369 | } | 387 | } |
| @@ -542,6 +560,98 @@ void radeon_compute_pll(struct radeon_pll *pll, | |||
| 542 | *post_div_p = best_post_div; | 560 | *post_div_p = best_post_div; |
| 543 | } | 561 | } |
| 544 | 562 | ||
| 563 | void radeon_compute_pll_avivo(struct radeon_pll *pll, | ||
| 564 | uint64_t freq, | ||
| 565 | uint32_t *dot_clock_p, | ||
| 566 | uint32_t *fb_div_p, | ||
| 567 | uint32_t *frac_fb_div_p, | ||
| 568 | uint32_t *ref_div_p, | ||
| 569 | uint32_t *post_div_p, | ||
| 570 | int flags) | ||
| 571 | { | ||
| 572 | fixed20_12 m, n, frac_n, p, f_vco, f_pclk, best_freq; | ||
| 573 | fixed20_12 pll_out_max, pll_out_min; | ||
| 574 | fixed20_12 pll_in_max, pll_in_min; | ||
| 575 | fixed20_12 reference_freq; | ||
| 576 | fixed20_12 error, ffreq, a, b; | ||
| 577 | |||
| 578 | pll_out_max.full = rfixed_const(pll->pll_out_max); | ||
| 579 | pll_out_min.full = rfixed_const(pll->pll_out_min); | ||
| 580 | pll_in_max.full = rfixed_const(pll->pll_in_max); | ||
| 581 | pll_in_min.full = rfixed_const(pll->pll_in_min); | ||
| 582 | reference_freq.full = rfixed_const(pll->reference_freq); | ||
| 583 | do_div(freq, 10); | ||
| 584 | ffreq.full = rfixed_const(freq); | ||
| 585 | error.full = rfixed_const(100 * 100); | ||
| 586 | |||
| 587 | /* max p */ | ||
| 588 | p.full = rfixed_div(pll_out_max, ffreq); | ||
| 589 | p.full = rfixed_floor(p); | ||
| 590 | |||
| 591 | /* min m */ | ||
| 592 | m.full = rfixed_div(reference_freq, pll_in_max); | ||
| 593 | m.full = rfixed_ceil(m); | ||
| 594 | |||
| 595 | while (1) { | ||
| 596 | n.full = rfixed_div(ffreq, reference_freq); | ||
| 597 | n.full = rfixed_mul(n, m); | ||
| 598 | n.full = rfixed_mul(n, p); | ||
| 599 | |||
| 600 | f_vco.full = rfixed_div(n, m); | ||
| 601 | f_vco.full = rfixed_mul(f_vco, reference_freq); | ||
| 602 | |||
| 603 | f_pclk.full = rfixed_div(f_vco, p); | ||
| 604 | |||
| 605 | if (f_pclk.full > ffreq.full) | ||
| 606 | error.full = f_pclk.full - ffreq.full; | ||
| 607 | else | ||
| 608 | error.full = ffreq.full - f_pclk.full; | ||
| 609 | error.full = rfixed_div(error, f_pclk); | ||
| 610 | a.full = rfixed_const(100 * 100); | ||
| 611 | error.full = rfixed_mul(error, a); | ||
| 612 | |||
| 613 | a.full = rfixed_mul(m, p); | ||
| 614 | a.full = rfixed_div(n, a); | ||
| 615 | best_freq.full = rfixed_mul(reference_freq, a); | ||
| 616 | |||
| 617 | if (rfixed_trunc(error) < 25) | ||
| 618 | break; | ||
| 619 | |||
| 620 | a.full = rfixed_const(1); | ||
| 621 | m.full = m.full + a.full; | ||
| 622 | a.full = rfixed_div(reference_freq, m); | ||
| 623 | if (a.full >= pll_in_min.full) | ||
| 624 | continue; | ||
| 625 | |||
| 626 | m.full = rfixed_div(reference_freq, pll_in_max); | ||
| 627 | m.full = rfixed_ceil(m); | ||
| 628 | a.full= rfixed_const(1); | ||
| 629 | p.full = p.full - a.full; | ||
| 630 | a.full = rfixed_mul(p, ffreq); | ||
| 631 | if (a.full >= pll_out_min.full) | ||
| 632 | continue; | ||
| 633 | else { | ||
| 634 | DRM_ERROR("Unable to find pll dividers\n"); | ||
| 635 | break; | ||
| 636 | } | ||
| 637 | } | ||
| 638 | |||
| 639 | a.full = rfixed_const(10); | ||
| 640 | b.full = rfixed_mul(n, a); | ||
| 641 | |||
| 642 | frac_n.full = rfixed_floor(n); | ||
| 643 | frac_n.full = rfixed_mul(frac_n, a); | ||
| 644 | frac_n.full = b.full - frac_n.full; | ||
| 645 | |||
| 646 | *dot_clock_p = rfixed_trunc(best_freq); | ||
| 647 | *fb_div_p = rfixed_trunc(n); | ||
| 648 | *frac_fb_div_p = rfixed_trunc(frac_n); | ||
| 649 | *ref_div_p = rfixed_trunc(m); | ||
| 650 | *post_div_p = rfixed_trunc(p); | ||
| 651 | |||
| 652 | DRM_DEBUG("%u %d.%d, %d, %d\n", *dot_clock_p * 10, *fb_div_p, *frac_fb_div_p, *ref_div_p, *post_div_p); | ||
| 653 | } | ||
| 654 | |||
| 545 | static void radeon_user_framebuffer_destroy(struct drm_framebuffer *fb) | 655 | static void radeon_user_framebuffer_destroy(struct drm_framebuffer *fb) |
| 546 | { | 656 | { |
| 547 | struct radeon_framebuffer *radeon_fb = to_radeon_framebuffer(fb); | 657 | struct radeon_framebuffer *radeon_fb = to_radeon_framebuffer(fb); |
| @@ -642,7 +752,7 @@ int radeon_modeset_create_props(struct radeon_device *rdev) | |||
| 642 | return -ENOMEM; | 752 | return -ENOMEM; |
| 643 | 753 | ||
| 644 | rdev->mode_info.coherent_mode_property->values[0] = 0; | 754 | rdev->mode_info.coherent_mode_property->values[0] = 0; |
| 645 | rdev->mode_info.coherent_mode_property->values[0] = 1; | 755 | rdev->mode_info.coherent_mode_property->values[1] = 1; |
| 646 | } | 756 | } |
| 647 | 757 | ||
| 648 | if (!ASIC_IS_AVIVO(rdev)) { | 758 | if (!ASIC_IS_AVIVO(rdev)) { |
| @@ -666,7 +776,7 @@ int radeon_modeset_create_props(struct radeon_device *rdev) | |||
| 666 | if (!rdev->mode_info.load_detect_property) | 776 | if (!rdev->mode_info.load_detect_property) |
| 667 | return -ENOMEM; | 777 | return -ENOMEM; |
| 668 | rdev->mode_info.load_detect_property->values[0] = 0; | 778 | rdev->mode_info.load_detect_property->values[0] = 0; |
| 669 | rdev->mode_info.load_detect_property->values[0] = 1; | 779 | rdev->mode_info.load_detect_property->values[1] = 1; |
| 670 | 780 | ||
| 671 | drm_mode_create_scaling_mode_property(rdev->ddev); | 781 | drm_mode_create_scaling_mode_property(rdev->ddev); |
| 672 | 782 | ||
| @@ -723,6 +833,8 @@ int radeon_modeset_init(struct radeon_device *rdev) | |||
| 723 | if (!ret) { | 833 | if (!ret) { |
| 724 | return ret; | 834 | return ret; |
| 725 | } | 835 | } |
| 836 | /* initialize hpd */ | ||
| 837 | radeon_hpd_init(rdev); | ||
| 726 | drm_helper_initial_config(rdev->ddev); | 838 | drm_helper_initial_config(rdev->ddev); |
| 727 | return 0; | 839 | return 0; |
| 728 | } | 840 | } |
| @@ -730,6 +842,7 @@ int radeon_modeset_init(struct radeon_device *rdev) | |||
| 730 | void radeon_modeset_fini(struct radeon_device *rdev) | 842 | void radeon_modeset_fini(struct radeon_device *rdev) |
| 731 | { | 843 | { |
| 732 | if (rdev->mode_info.mode_config_initialized) { | 844 | if (rdev->mode_info.mode_config_initialized) { |
| 845 | radeon_hpd_fini(rdev); | ||
| 733 | drm_mode_config_cleanup(rdev->ddev); | 846 | drm_mode_config_cleanup(rdev->ddev); |
| 734 | rdev->mode_info.mode_config_initialized = false; | 847 | rdev->mode_info.mode_config_initialized = false; |
| 735 | } | 848 | } |
| @@ -750,9 +863,17 @@ bool radeon_crtc_scaling_mode_fixup(struct drm_crtc *crtc, | |||
| 750 | if (encoder->crtc != crtc) | 863 | if (encoder->crtc != crtc) |
| 751 | continue; | 864 | continue; |
| 752 | if (first) { | 865 | if (first) { |
| 753 | radeon_crtc->rmx_type = radeon_encoder->rmx_type; | 866 | /* set scaling */ |
| 867 | if (radeon_encoder->rmx_type == RMX_OFF) | ||
| 868 | radeon_crtc->rmx_type = RMX_OFF; | ||
| 869 | else if (mode->hdisplay < radeon_encoder->native_mode.hdisplay || | ||
| 870 | mode->vdisplay < radeon_encoder->native_mode.vdisplay) | ||
| 871 | radeon_crtc->rmx_type = radeon_encoder->rmx_type; | ||
| 872 | else | ||
| 873 | radeon_crtc->rmx_type = RMX_OFF; | ||
| 874 | /* copy native mode */ | ||
| 754 | memcpy(&radeon_crtc->native_mode, | 875 | memcpy(&radeon_crtc->native_mode, |
| 755 | &radeon_encoder->native_mode, | 876 | &radeon_encoder->native_mode, |
| 756 | sizeof(struct drm_display_mode)); | 877 | sizeof(struct drm_display_mode)); |
| 757 | first = false; | 878 | first = false; |
| 758 | } else { | 879 | } else { |
diff --git a/drivers/gpu/drm/radeon/radeon_drv.c b/drivers/gpu/drm/radeon/radeon_drv.c index 7f50fb864af8..28077247f4f3 100644 --- a/drivers/gpu/drm/radeon/radeon_drv.c +++ b/drivers/gpu/drm/radeon/radeon_drv.c | |||
| @@ -86,6 +86,7 @@ int radeon_benchmarking = 0; | |||
| 86 | int radeon_testing = 0; | 86 | int radeon_testing = 0; |
| 87 | int radeon_connector_table = 0; | 87 | int radeon_connector_table = 0; |
| 88 | int radeon_tv = 1; | 88 | int radeon_tv = 1; |
| 89 | int radeon_new_pll = 1; | ||
| 89 | 90 | ||
| 90 | MODULE_PARM_DESC(no_wb, "Disable AGP writeback for scratch registers"); | 91 | MODULE_PARM_DESC(no_wb, "Disable AGP writeback for scratch registers"); |
| 91 | module_param_named(no_wb, radeon_no_wb, int, 0444); | 92 | module_param_named(no_wb, radeon_no_wb, int, 0444); |
| @@ -120,6 +121,9 @@ module_param_named(connector_table, radeon_connector_table, int, 0444); | |||
| 120 | MODULE_PARM_DESC(tv, "TV enable (0 = disable)"); | 121 | MODULE_PARM_DESC(tv, "TV enable (0 = disable)"); |
| 121 | module_param_named(tv, radeon_tv, int, 0444); | 122 | module_param_named(tv, radeon_tv, int, 0444); |
| 122 | 123 | ||
| 124 | MODULE_PARM_DESC(r4xx_atom, "Select new PLL code for AVIVO chips"); | ||
| 125 | module_param_named(new_pll, radeon_new_pll, int, 0444); | ||
| 126 | |||
| 123 | static int radeon_suspend(struct drm_device *dev, pm_message_t state) | 127 | static int radeon_suspend(struct drm_device *dev, pm_message_t state) |
| 124 | { | 128 | { |
| 125 | drm_radeon_private_t *dev_priv = dev->dev_private; | 129 | drm_radeon_private_t *dev_priv = dev->dev_private; |
diff --git a/drivers/gpu/drm/radeon/radeon_drv.h b/drivers/gpu/drm/radeon/radeon_drv.h index 350962e0f346..e13785282a82 100644 --- a/drivers/gpu/drm/radeon/radeon_drv.h +++ b/drivers/gpu/drm/radeon/radeon_drv.h | |||
| @@ -1104,7 +1104,6 @@ extern u32 radeon_get_scratch(drm_radeon_private_t *dev_priv, int index); | |||
| 1104 | # define R600_IT_WAIT_REG_MEM 0x00003C00 | 1104 | # define R600_IT_WAIT_REG_MEM 0x00003C00 |
| 1105 | # define R600_IT_MEM_WRITE 0x00003D00 | 1105 | # define R600_IT_MEM_WRITE 0x00003D00 |
| 1106 | # define R600_IT_INDIRECT_BUFFER 0x00003200 | 1106 | # define R600_IT_INDIRECT_BUFFER 0x00003200 |
| 1107 | # define R600_IT_CP_INTERRUPT 0x00004000 | ||
| 1108 | # define R600_IT_SURFACE_SYNC 0x00004300 | 1107 | # define R600_IT_SURFACE_SYNC 0x00004300 |
| 1109 | # define R600_CB0_DEST_BASE_ENA (1 << 6) | 1108 | # define R600_CB0_DEST_BASE_ENA (1 << 6) |
| 1110 | # define R600_TC_ACTION_ENA (1 << 23) | 1109 | # define R600_TC_ACTION_ENA (1 << 23) |
diff --git a/drivers/gpu/drm/radeon/radeon_encoders.c b/drivers/gpu/drm/radeon/radeon_encoders.c index d42bc512d75a..b4f23ec93201 100644 --- a/drivers/gpu/drm/radeon/radeon_encoders.c +++ b/drivers/gpu/drm/radeon/radeon_encoders.c | |||
| @@ -35,6 +35,51 @@ extern int atom_debug; | |||
| 35 | bool radeon_atom_get_tv_timings(struct radeon_device *rdev, int index, | 35 | bool radeon_atom_get_tv_timings(struct radeon_device *rdev, int index, |
| 36 | struct drm_display_mode *mode); | 36 | struct drm_display_mode *mode); |
| 37 | 37 | ||
| 38 | static uint32_t radeon_encoder_clones(struct drm_encoder *encoder) | ||
| 39 | { | ||
| 40 | struct drm_device *dev = encoder->dev; | ||
| 41 | struct radeon_device *rdev = dev->dev_private; | ||
| 42 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 43 | struct drm_encoder *clone_encoder; | ||
| 44 | uint32_t index_mask = 0; | ||
| 45 | int count; | ||
| 46 | |||
| 47 | /* DIG routing gets problematic */ | ||
| 48 | if (rdev->family >= CHIP_R600) | ||
| 49 | return index_mask; | ||
| 50 | /* LVDS/TV are too wacky */ | ||
| 51 | if (radeon_encoder->devices & ATOM_DEVICE_LCD_SUPPORT) | ||
| 52 | return index_mask; | ||
| 53 | /* DVO requires 2x ppll clocks depending on tmds chip */ | ||
| 54 | if (radeon_encoder->devices & ATOM_DEVICE_DFP2_SUPPORT) | ||
| 55 | return index_mask; | ||
| 56 | |||
| 57 | count = -1; | ||
| 58 | list_for_each_entry(clone_encoder, &dev->mode_config.encoder_list, head) { | ||
| 59 | struct radeon_encoder *radeon_clone = to_radeon_encoder(clone_encoder); | ||
| 60 | count++; | ||
| 61 | |||
| 62 | if (clone_encoder == encoder) | ||
| 63 | continue; | ||
| 64 | if (radeon_clone->devices & (ATOM_DEVICE_LCD_SUPPORT)) | ||
| 65 | continue; | ||
| 66 | if (radeon_clone->devices & ATOM_DEVICE_DFP2_SUPPORT) | ||
| 67 | continue; | ||
| 68 | else | ||
| 69 | index_mask |= (1 << count); | ||
| 70 | } | ||
| 71 | return index_mask; | ||
| 72 | } | ||
| 73 | |||
| 74 | void radeon_setup_encoder_clones(struct drm_device *dev) | ||
| 75 | { | ||
| 76 | struct drm_encoder *encoder; | ||
| 77 | |||
| 78 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) { | ||
| 79 | encoder->possible_clones = radeon_encoder_clones(encoder); | ||
| 80 | } | ||
| 81 | } | ||
| 82 | |||
| 38 | uint32_t | 83 | uint32_t |
| 39 | radeon_get_encoder_id(struct drm_device *dev, uint32_t supported_device, uint8_t dac) | 84 | radeon_get_encoder_id(struct drm_device *dev, uint32_t supported_device, uint8_t dac) |
| 40 | { | 85 | { |
| @@ -163,29 +208,6 @@ radeon_get_connector_for_encoder(struct drm_encoder *encoder) | |||
| 163 | return NULL; | 208 | return NULL; |
| 164 | } | 209 | } |
| 165 | 210 | ||
| 166 | /* used for both atom and legacy */ | ||
| 167 | void radeon_rmx_mode_fixup(struct drm_encoder *encoder, | ||
| 168 | struct drm_display_mode *mode, | ||
| 169 | struct drm_display_mode *adjusted_mode) | ||
| 170 | { | ||
| 171 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 172 | struct drm_device *dev = encoder->dev; | ||
| 173 | struct radeon_device *rdev = dev->dev_private; | ||
| 174 | struct drm_display_mode *native_mode = &radeon_encoder->native_mode; | ||
| 175 | |||
| 176 | if (mode->hdisplay < native_mode->hdisplay || | ||
| 177 | mode->vdisplay < native_mode->vdisplay) { | ||
| 178 | int mode_id = adjusted_mode->base.id; | ||
| 179 | *adjusted_mode = *native_mode; | ||
| 180 | if (!ASIC_IS_AVIVO(rdev)) { | ||
| 181 | adjusted_mode->hdisplay = mode->hdisplay; | ||
| 182 | adjusted_mode->vdisplay = mode->vdisplay; | ||
| 183 | } | ||
| 184 | adjusted_mode->base.id = mode_id; | ||
| 185 | } | ||
| 186 | } | ||
| 187 | |||
| 188 | |||
| 189 | static bool radeon_atom_mode_fixup(struct drm_encoder *encoder, | 211 | static bool radeon_atom_mode_fixup(struct drm_encoder *encoder, |
| 190 | struct drm_display_mode *mode, | 212 | struct drm_display_mode *mode, |
| 191 | struct drm_display_mode *adjusted_mode) | 213 | struct drm_display_mode *adjusted_mode) |
| @@ -198,14 +220,24 @@ static bool radeon_atom_mode_fixup(struct drm_encoder *encoder, | |||
| 198 | radeon_encoder_set_active_device(encoder); | 220 | radeon_encoder_set_active_device(encoder); |
| 199 | drm_mode_set_crtcinfo(adjusted_mode, 0); | 221 | drm_mode_set_crtcinfo(adjusted_mode, 0); |
| 200 | 222 | ||
| 201 | if (radeon_encoder->rmx_type != RMX_OFF) | ||
| 202 | radeon_rmx_mode_fixup(encoder, mode, adjusted_mode); | ||
| 203 | |||
| 204 | /* hw bug */ | 223 | /* hw bug */ |
| 205 | if ((mode->flags & DRM_MODE_FLAG_INTERLACE) | 224 | if ((mode->flags & DRM_MODE_FLAG_INTERLACE) |
| 206 | && (mode->crtc_vsync_start < (mode->crtc_vdisplay + 2))) | 225 | && (mode->crtc_vsync_start < (mode->crtc_vdisplay + 2))) |
| 207 | adjusted_mode->crtc_vsync_start = adjusted_mode->crtc_vdisplay + 2; | 226 | adjusted_mode->crtc_vsync_start = adjusted_mode->crtc_vdisplay + 2; |
| 208 | 227 | ||
| 228 | /* get the native mode for LVDS */ | ||
| 229 | if (radeon_encoder->active_device & (ATOM_DEVICE_LCD_SUPPORT)) { | ||
| 230 | struct drm_display_mode *native_mode = &radeon_encoder->native_mode; | ||
| 231 | int mode_id = adjusted_mode->base.id; | ||
| 232 | *adjusted_mode = *native_mode; | ||
| 233 | if (!ASIC_IS_AVIVO(rdev)) { | ||
| 234 | adjusted_mode->hdisplay = mode->hdisplay; | ||
| 235 | adjusted_mode->vdisplay = mode->vdisplay; | ||
| 236 | } | ||
| 237 | adjusted_mode->base.id = mode_id; | ||
| 238 | } | ||
| 239 | |||
| 240 | /* get the native mode for TV */ | ||
| 209 | if (radeon_encoder->active_device & (ATOM_DEVICE_TV_SUPPORT)) { | 241 | if (radeon_encoder->active_device & (ATOM_DEVICE_TV_SUPPORT)) { |
| 210 | struct radeon_encoder_atom_dac *tv_dac = radeon_encoder->enc_priv; | 242 | struct radeon_encoder_atom_dac *tv_dac = radeon_encoder->enc_priv; |
| 211 | if (tv_dac) { | 243 | if (tv_dac) { |
| @@ -218,6 +250,12 @@ static bool radeon_atom_mode_fixup(struct drm_encoder *encoder, | |||
| 218 | } | 250 | } |
| 219 | } | 251 | } |
| 220 | 252 | ||
| 253 | if (ASIC_IS_DCE3(rdev) && | ||
| 254 | (radeon_encoder->active_device & (ATOM_DEVICE_DFP_SUPPORT))) { | ||
| 255 | struct drm_connector *connector = radeon_get_connector_for_encoder(encoder); | ||
| 256 | radeon_dp_set_link_config(connector, mode); | ||
| 257 | } | ||
| 258 | |||
| 221 | return true; | 259 | return true; |
| 222 | } | 260 | } |
| 223 | 261 | ||
| @@ -392,7 +430,7 @@ union lvds_encoder_control { | |||
| 392 | LVDS_ENCODER_CONTROL_PS_ALLOCATION_V2 v2; | 430 | LVDS_ENCODER_CONTROL_PS_ALLOCATION_V2 v2; |
| 393 | }; | 431 | }; |
| 394 | 432 | ||
| 395 | static void | 433 | void |
| 396 | atombios_digital_setup(struct drm_encoder *encoder, int action) | 434 | atombios_digital_setup(struct drm_encoder *encoder, int action) |
| 397 | { | 435 | { |
| 398 | struct drm_device *dev = encoder->dev; | 436 | struct drm_device *dev = encoder->dev; |
| @@ -522,6 +560,7 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 522 | { | 560 | { |
| 523 | struct drm_connector *connector; | 561 | struct drm_connector *connector; |
| 524 | struct radeon_connector *radeon_connector; | 562 | struct radeon_connector *radeon_connector; |
| 563 | struct radeon_connector_atom_dig *radeon_dig_connector; | ||
| 525 | 564 | ||
| 526 | connector = radeon_get_connector_for_encoder(encoder); | 565 | connector = radeon_get_connector_for_encoder(encoder); |
| 527 | if (!connector) | 566 | if (!connector) |
| @@ -551,10 +590,10 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 551 | return ATOM_ENCODER_MODE_LVDS; | 590 | return ATOM_ENCODER_MODE_LVDS; |
| 552 | break; | 591 | break; |
| 553 | case DRM_MODE_CONNECTOR_DisplayPort: | 592 | case DRM_MODE_CONNECTOR_DisplayPort: |
| 554 | /*if (radeon_output->MonType == MT_DP) | 593 | radeon_dig_connector = radeon_connector->con_priv; |
| 555 | return ATOM_ENCODER_MODE_DP; | 594 | if (radeon_dig_connector->dp_sink_type == CONNECTOR_OBJECT_ID_DISPLAYPORT) |
| 556 | else*/ | 595 | return ATOM_ENCODER_MODE_DP; |
| 557 | if (drm_detect_hdmi_monitor(radeon_connector->edid)) | 596 | else if (drm_detect_hdmi_monitor(radeon_connector->edid)) |
| 558 | return ATOM_ENCODER_MODE_HDMI; | 597 | return ATOM_ENCODER_MODE_HDMI; |
| 559 | else | 598 | else |
| 560 | return ATOM_ENCODER_MODE_DVI; | 599 | return ATOM_ENCODER_MODE_DVI; |
| @@ -573,6 +612,30 @@ atombios_get_encoder_mode(struct drm_encoder *encoder) | |||
| 573 | } | 612 | } |
| 574 | } | 613 | } |
| 575 | 614 | ||
| 615 | /* | ||
| 616 | * DIG Encoder/Transmitter Setup | ||
| 617 | * | ||
| 618 | * DCE 3.0/3.1 | ||
| 619 | * - 2 DIG transmitter blocks. UNIPHY (links A and B) and LVTMA. | ||
| 620 | * Supports up to 3 digital outputs | ||
| 621 | * - 2 DIG encoder blocks. | ||
| 622 | * DIG1 can drive UNIPHY link A or link B | ||
| 623 | * DIG2 can drive UNIPHY link B or LVTMA | ||
| 624 | * | ||
| 625 | * DCE 3.2 | ||
| 626 | * - 3 DIG transmitter blocks. UNIPHY0/1/2 (links A and B). | ||
| 627 | * Supports up to 5 digital outputs | ||
| 628 | * - 2 DIG encoder blocks. | ||
| 629 | * DIG1/2 can drive UNIPHY0/1/2 link A or link B | ||
| 630 | * | ||
| 631 | * Routing | ||
| 632 | * crtc -> dig encoder -> UNIPHY/LVTMA (1 or 2 links) | ||
| 633 | * Examples: | ||
| 634 | * crtc0 -> dig2 -> LVTMA links A+B -> TMDS/HDMI | ||
| 635 | * crtc1 -> dig1 -> UNIPHY0 link B -> DP | ||
| 636 | * crtc0 -> dig1 -> UNIPHY2 link A -> LVDS | ||
| 637 | * crtc1 -> dig2 -> UNIPHY1 link B+A -> TMDS/HDMI | ||
| 638 | */ | ||
| 576 | static void | 639 | static void |
| 577 | atombios_dig_encoder_setup(struct drm_encoder *encoder, int action) | 640 | atombios_dig_encoder_setup(struct drm_encoder *encoder, int action) |
| 578 | { | 641 | { |
| @@ -614,10 +677,17 @@ atombios_dig_encoder_setup(struct drm_encoder *encoder, int action) | |||
| 614 | } else { | 677 | } else { |
| 615 | switch (radeon_encoder->encoder_id) { | 678 | switch (radeon_encoder->encoder_id) { |
| 616 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: | 679 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: |
| 617 | index = GetIndexIntoMasterTable(COMMAND, DIG1EncoderControl); | 680 | /* XXX doesn't really matter which dig encoder we pick as long as it's |
| 681 | * not already in use | ||
| 682 | */ | ||
| 683 | if (dig_connector->linkb) | ||
| 684 | index = GetIndexIntoMasterTable(COMMAND, DIG2EncoderControl); | ||
| 685 | else | ||
| 686 | index = GetIndexIntoMasterTable(COMMAND, DIG1EncoderControl); | ||
| 618 | num = 1; | 687 | num = 1; |
| 619 | break; | 688 | break; |
| 620 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: | 689 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: |
| 690 | /* Only dig2 encoder can drive LVTMA */ | ||
| 621 | index = GetIndexIntoMasterTable(COMMAND, DIG2EncoderControl); | 691 | index = GetIndexIntoMasterTable(COMMAND, DIG2EncoderControl); |
| 622 | num = 2; | 692 | num = 2; |
| 623 | break; | 693 | break; |
| @@ -652,18 +722,21 @@ atombios_dig_encoder_setup(struct drm_encoder *encoder, int action) | |||
| 652 | } | 722 | } |
| 653 | } | 723 | } |
| 654 | 724 | ||
| 655 | if (radeon_encoder->pixel_clock > 165000) { | 725 | args.ucEncoderMode = atombios_get_encoder_mode(encoder); |
| 656 | args.ucConfig |= ATOM_ENCODER_CONFIG_LINKA_B; | 726 | |
| 727 | if (args.ucEncoderMode == ATOM_ENCODER_MODE_DP) { | ||
| 728 | if (dig_connector->dp_clock == 270000) | ||
| 729 | args.ucConfig |= ATOM_ENCODER_CONFIG_DPLINKRATE_2_70GHZ; | ||
| 730 | args.ucLaneNum = dig_connector->dp_lane_count; | ||
| 731 | } else if (radeon_encoder->pixel_clock > 165000) | ||
| 657 | args.ucLaneNum = 8; | 732 | args.ucLaneNum = 8; |
| 658 | } else { | 733 | else |
| 659 | if (dig_connector->linkb) | ||
| 660 | args.ucConfig |= ATOM_ENCODER_CONFIG_LINKB; | ||
| 661 | else | ||
| 662 | args.ucConfig |= ATOM_ENCODER_CONFIG_LINKA; | ||
| 663 | args.ucLaneNum = 4; | 734 | args.ucLaneNum = 4; |
| 664 | } | ||
| 665 | 735 | ||
| 666 | args.ucEncoderMode = atombios_get_encoder_mode(encoder); | 736 | if (dig_connector->linkb) |
| 737 | args.ucConfig |= ATOM_ENCODER_CONFIG_LINKB; | ||
| 738 | else | ||
| 739 | args.ucConfig |= ATOM_ENCODER_CONFIG_LINKA; | ||
| 667 | 740 | ||
| 668 | atom_execute_table(rdev->mode_info.atom_context, index, (uint32_t *)&args); | 741 | atom_execute_table(rdev->mode_info.atom_context, index, (uint32_t *)&args); |
| 669 | 742 | ||
| @@ -674,8 +747,8 @@ union dig_transmitter_control { | |||
| 674 | DIG_TRANSMITTER_CONTROL_PARAMETERS_V2 v2; | 747 | DIG_TRANSMITTER_CONTROL_PARAMETERS_V2 v2; |
| 675 | }; | 748 | }; |
| 676 | 749 | ||
| 677 | static void | 750 | void |
| 678 | atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | 751 | atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action, uint8_t lane_num, uint8_t lane_set) |
| 679 | { | 752 | { |
| 680 | struct drm_device *dev = encoder->dev; | 753 | struct drm_device *dev = encoder->dev; |
| 681 | struct radeon_device *rdev = dev->dev_private; | 754 | struct radeon_device *rdev = dev->dev_private; |
| @@ -687,6 +760,7 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 687 | struct drm_connector *connector; | 760 | struct drm_connector *connector; |
| 688 | struct radeon_connector *radeon_connector; | 761 | struct radeon_connector *radeon_connector; |
| 689 | struct radeon_connector_atom_dig *dig_connector; | 762 | struct radeon_connector_atom_dig *dig_connector; |
| 763 | bool is_dp = false; | ||
| 690 | 764 | ||
| 691 | connector = radeon_get_connector_for_encoder(encoder); | 765 | connector = radeon_get_connector_for_encoder(encoder); |
| 692 | if (!connector) | 766 | if (!connector) |
| @@ -704,6 +778,9 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 704 | 778 | ||
| 705 | dig_connector = radeon_connector->con_priv; | 779 | dig_connector = radeon_connector->con_priv; |
| 706 | 780 | ||
| 781 | if (atombios_get_encoder_mode(encoder) == ATOM_ENCODER_MODE_DP) | ||
| 782 | is_dp = true; | ||
| 783 | |||
| 707 | memset(&args, 0, sizeof(args)); | 784 | memset(&args, 0, sizeof(args)); |
| 708 | 785 | ||
| 709 | if (ASIC_IS_DCE32(rdev)) | 786 | if (ASIC_IS_DCE32(rdev)) |
| @@ -724,17 +801,23 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 724 | args.v1.ucAction = action; | 801 | args.v1.ucAction = action; |
| 725 | if (action == ATOM_TRANSMITTER_ACTION_INIT) { | 802 | if (action == ATOM_TRANSMITTER_ACTION_INIT) { |
| 726 | args.v1.usInitInfo = radeon_connector->connector_object_id; | 803 | args.v1.usInitInfo = radeon_connector->connector_object_id; |
| 804 | } else if (action == ATOM_TRANSMITTER_ACTION_SETUP_VSEMPH) { | ||
| 805 | args.v1.asMode.ucLaneSel = lane_num; | ||
| 806 | args.v1.asMode.ucLaneSet = lane_set; | ||
| 727 | } else { | 807 | } else { |
| 728 | if (radeon_encoder->pixel_clock > 165000) | 808 | if (is_dp) |
| 809 | args.v1.usPixelClock = | ||
| 810 | cpu_to_le16(dig_connector->dp_clock / 10); | ||
| 811 | else if (radeon_encoder->pixel_clock > 165000) | ||
| 729 | args.v1.usPixelClock = cpu_to_le16((radeon_encoder->pixel_clock / 2) / 10); | 812 | args.v1.usPixelClock = cpu_to_le16((radeon_encoder->pixel_clock / 2) / 10); |
| 730 | else | 813 | else |
| 731 | args.v1.usPixelClock = cpu_to_le16(radeon_encoder->pixel_clock / 10); | 814 | args.v1.usPixelClock = cpu_to_le16(radeon_encoder->pixel_clock / 10); |
| 732 | } | 815 | } |
| 733 | if (ASIC_IS_DCE32(rdev)) { | 816 | if (ASIC_IS_DCE32(rdev)) { |
| 734 | if (radeon_encoder->pixel_clock > 165000) | ||
| 735 | args.v2.usPixelClock = cpu_to_le16((radeon_encoder->pixel_clock / 2) / 10); | ||
| 736 | if (dig->dig_block) | 817 | if (dig->dig_block) |
| 737 | args.v2.acConfig.ucEncoderSel = 1; | 818 | args.v2.acConfig.ucEncoderSel = 1; |
| 819 | if (dig_connector->linkb) | ||
| 820 | args.v2.acConfig.ucLinkSel = 1; | ||
| 738 | 821 | ||
| 739 | switch (radeon_encoder->encoder_id) { | 822 | switch (radeon_encoder->encoder_id) { |
| 740 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: | 823 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: |
| @@ -751,7 +834,9 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 751 | break; | 834 | break; |
| 752 | } | 835 | } |
| 753 | 836 | ||
| 754 | if (radeon_encoder->devices & (ATOM_DEVICE_DFP_SUPPORT)) { | 837 | if (is_dp) |
| 838 | args.v2.acConfig.fCoherentMode = 1; | ||
| 839 | else if (radeon_encoder->devices & (ATOM_DEVICE_DFP_SUPPORT)) { | ||
| 755 | if (dig->coherent_mode) | 840 | if (dig->coherent_mode) |
| 756 | args.v2.acConfig.fCoherentMode = 1; | 841 | args.v2.acConfig.fCoherentMode = 1; |
| 757 | } | 842 | } |
| @@ -760,17 +845,20 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 760 | 845 | ||
| 761 | switch (radeon_encoder->encoder_id) { | 846 | switch (radeon_encoder->encoder_id) { |
| 762 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: | 847 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY: |
| 763 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_DIG1_ENCODER; | 848 | /* XXX doesn't really matter which dig encoder we pick as long as it's |
| 849 | * not already in use | ||
| 850 | */ | ||
| 851 | if (dig_connector->linkb) | ||
| 852 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_DIG2_ENCODER; | ||
| 853 | else | ||
| 854 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_DIG1_ENCODER; | ||
| 764 | if (rdev->flags & RADEON_IS_IGP) { | 855 | if (rdev->flags & RADEON_IS_IGP) { |
| 765 | if (radeon_encoder->pixel_clock > 165000) { | 856 | if (radeon_encoder->pixel_clock > 165000) { |
| 766 | args.v1.ucConfig |= (ATOM_TRANSMITTER_CONFIG_8LANE_LINK | | ||
| 767 | ATOM_TRANSMITTER_CONFIG_LINKA_B); | ||
| 768 | if (dig_connector->igp_lane_info & 0x3) | 857 | if (dig_connector->igp_lane_info & 0x3) |
| 769 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_0_7; | 858 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_0_7; |
| 770 | else if (dig_connector->igp_lane_info & 0xc) | 859 | else if (dig_connector->igp_lane_info & 0xc) |
| 771 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_8_15; | 860 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_8_15; |
| 772 | } else { | 861 | } else { |
| 773 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKA; | ||
| 774 | if (dig_connector->igp_lane_info & 0x1) | 862 | if (dig_connector->igp_lane_info & 0x1) |
| 775 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_0_3; | 863 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_0_3; |
| 776 | else if (dig_connector->igp_lane_info & 0x2) | 864 | else if (dig_connector->igp_lane_info & 0x2) |
| @@ -780,35 +868,25 @@ atombios_dig_transmitter_setup(struct drm_encoder *encoder, int action) | |||
| 780 | else if (dig_connector->igp_lane_info & 0x8) | 868 | else if (dig_connector->igp_lane_info & 0x8) |
| 781 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_12_15; | 869 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LANE_12_15; |
| 782 | } | 870 | } |
| 783 | } else { | ||
| 784 | if (radeon_encoder->pixel_clock > 165000) | ||
| 785 | args.v1.ucConfig |= (ATOM_TRANSMITTER_CONFIG_8LANE_LINK | | ||
| 786 | ATOM_TRANSMITTER_CONFIG_LINKA_B | | ||
| 787 | ATOM_TRANSMITTER_CONFIG_LANE_0_7); | ||
| 788 | else { | ||
| 789 | if (dig_connector->linkb) | ||
| 790 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKB | ATOM_TRANSMITTER_CONFIG_LANE_0_3; | ||
| 791 | else | ||
| 792 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKA | ATOM_TRANSMITTER_CONFIG_LANE_0_3; | ||
| 793 | } | ||
| 794 | } | 871 | } |
| 795 | break; | 872 | break; |
| 796 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: | 873 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: |
| 874 | /* Only dig2 encoder can drive LVTMA */ | ||
| 797 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_DIG2_ENCODER; | 875 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_DIG2_ENCODER; |
| 798 | if (radeon_encoder->pixel_clock > 165000) | ||
| 799 | args.v1.ucConfig |= (ATOM_TRANSMITTER_CONFIG_8LANE_LINK | | ||
| 800 | ATOM_TRANSMITTER_CONFIG_LINKA_B | | ||
| 801 | ATOM_TRANSMITTER_CONFIG_LANE_0_7); | ||
| 802 | else { | ||
| 803 | if (dig_connector->linkb) | ||
| 804 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKB | ATOM_TRANSMITTER_CONFIG_LANE_0_3; | ||
| 805 | else | ||
| 806 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKA | ATOM_TRANSMITTER_CONFIG_LANE_0_3; | ||
| 807 | } | ||
| 808 | break; | 876 | break; |
| 809 | } | 877 | } |
| 810 | 878 | ||
| 811 | if (radeon_encoder->devices & (ATOM_DEVICE_DFP_SUPPORT)) { | 879 | if (radeon_encoder->pixel_clock > 165000) |
| 880 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_8LANE_LINK; | ||
| 881 | |||
| 882 | if (dig_connector->linkb) | ||
| 883 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKB; | ||
| 884 | else | ||
| 885 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_LINKA; | ||
| 886 | |||
| 887 | if (is_dp) | ||
| 888 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_COHERENT; | ||
| 889 | else if (radeon_encoder->devices & (ATOM_DEVICE_DFP_SUPPORT)) { | ||
| 812 | if (dig->coherent_mode) | 890 | if (dig->coherent_mode) |
| 813 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_COHERENT; | 891 | args.v1.ucConfig |= ATOM_TRANSMITTER_CONFIG_COHERENT; |
| 814 | } | 892 | } |
| @@ -918,12 +996,16 @@ radeon_atom_encoder_dpms(struct drm_encoder *encoder, int mode) | |||
| 918 | if (is_dig) { | 996 | if (is_dig) { |
| 919 | switch (mode) { | 997 | switch (mode) { |
| 920 | case DRM_MODE_DPMS_ON: | 998 | case DRM_MODE_DPMS_ON: |
| 921 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE); | 999 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE_OUTPUT, 0, 0); |
| 1000 | { | ||
| 1001 | struct drm_connector *connector = radeon_get_connector_for_encoder(encoder); | ||
| 1002 | dp_link_train(encoder, connector); | ||
| 1003 | } | ||
| 922 | break; | 1004 | break; |
| 923 | case DRM_MODE_DPMS_STANDBY: | 1005 | case DRM_MODE_DPMS_STANDBY: |
| 924 | case DRM_MODE_DPMS_SUSPEND: | 1006 | case DRM_MODE_DPMS_SUSPEND: |
| 925 | case DRM_MODE_DPMS_OFF: | 1007 | case DRM_MODE_DPMS_OFF: |
| 926 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_DISABLE); | 1008 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_DISABLE_OUTPUT, 0, 0); |
| 927 | break; | 1009 | break; |
| 928 | } | 1010 | } |
| 929 | } else { | 1011 | } else { |
| @@ -1025,13 +1107,33 @@ atombios_set_encoder_crtc_source(struct drm_encoder *encoder) | |||
| 1025 | args.v2.ucEncoderID = ASIC_INT_DIG2_ENCODER_ID; | 1107 | args.v2.ucEncoderID = ASIC_INT_DIG2_ENCODER_ID; |
| 1026 | else | 1108 | else |
| 1027 | args.v2.ucEncoderID = ASIC_INT_DIG1_ENCODER_ID; | 1109 | args.v2.ucEncoderID = ASIC_INT_DIG1_ENCODER_ID; |
| 1028 | } else | 1110 | } else { |
| 1029 | args.v2.ucEncoderID = ASIC_INT_DIG1_ENCODER_ID; | 1111 | struct drm_connector *connector; |
| 1112 | struct radeon_connector *radeon_connector; | ||
| 1113 | struct radeon_connector_atom_dig *dig_connector; | ||
| 1114 | |||
| 1115 | connector = radeon_get_connector_for_encoder(encoder); | ||
| 1116 | if (!connector) | ||
| 1117 | return; | ||
| 1118 | radeon_connector = to_radeon_connector(connector); | ||
| 1119 | if (!radeon_connector->con_priv) | ||
| 1120 | return; | ||
| 1121 | dig_connector = radeon_connector->con_priv; | ||
| 1122 | |||
| 1123 | /* XXX doesn't really matter which dig encoder we pick as long as it's | ||
| 1124 | * not already in use | ||
| 1125 | */ | ||
| 1126 | if (dig_connector->linkb) | ||
| 1127 | args.v2.ucEncoderID = ASIC_INT_DIG2_ENCODER_ID; | ||
| 1128 | else | ||
| 1129 | args.v2.ucEncoderID = ASIC_INT_DIG1_ENCODER_ID; | ||
| 1130 | } | ||
| 1030 | break; | 1131 | break; |
| 1031 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DVO1: | 1132 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DVO1: |
| 1032 | args.v2.ucEncoderID = ASIC_INT_DVO_ENCODER_ID; | 1133 | args.v2.ucEncoderID = ASIC_INT_DVO_ENCODER_ID; |
| 1033 | break; | 1134 | break; |
| 1034 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: | 1135 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: |
| 1136 | /* Only dig2 encoder can drive LVTMA */ | ||
| 1035 | args.v2.ucEncoderID = ASIC_INT_DIG2_ENCODER_ID; | 1137 | args.v2.ucEncoderID = ASIC_INT_DIG2_ENCODER_ID; |
| 1036 | break; | 1138 | break; |
| 1037 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC1: | 1139 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC1: |
| @@ -1104,11 +1206,14 @@ radeon_atom_encoder_mode_set(struct drm_encoder *encoder, | |||
| 1104 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | 1206 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); |
| 1105 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(encoder->crtc); | 1207 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(encoder->crtc); |
| 1106 | 1208 | ||
| 1107 | if (radeon_encoder->enc_priv) { | 1209 | if (radeon_encoder->active_device & |
| 1108 | struct radeon_encoder_atom_dig *dig; | 1210 | (ATOM_DEVICE_DFP_SUPPORT | ATOM_DEVICE_LCD_SUPPORT)) { |
| 1211 | if (radeon_encoder->enc_priv) { | ||
| 1212 | struct radeon_encoder_atom_dig *dig; | ||
| 1109 | 1213 | ||
| 1110 | dig = radeon_encoder->enc_priv; | 1214 | dig = radeon_encoder->enc_priv; |
| 1111 | dig->dig_block = radeon_crtc->crtc_id; | 1215 | dig->dig_block = radeon_crtc->crtc_id; |
| 1216 | } | ||
| 1112 | } | 1217 | } |
| 1113 | radeon_encoder->pixel_clock = adjusted_mode->clock; | 1218 | radeon_encoder->pixel_clock = adjusted_mode->clock; |
| 1114 | 1219 | ||
| @@ -1134,14 +1239,14 @@ radeon_atom_encoder_mode_set(struct drm_encoder *encoder, | |||
| 1134 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY2: | 1239 | case ENCODER_OBJECT_ID_INTERNAL_UNIPHY2: |
| 1135 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: | 1240 | case ENCODER_OBJECT_ID_INTERNAL_KLDSCP_LVTMA: |
| 1136 | /* disable the encoder and transmitter */ | 1241 | /* disable the encoder and transmitter */ |
| 1137 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_DISABLE); | 1242 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_DISABLE, 0, 0); |
| 1138 | atombios_dig_encoder_setup(encoder, ATOM_DISABLE); | 1243 | atombios_dig_encoder_setup(encoder, ATOM_DISABLE); |
| 1139 | 1244 | ||
| 1140 | /* setup and enable the encoder and transmitter */ | 1245 | /* setup and enable the encoder and transmitter */ |
| 1141 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE); | 1246 | atombios_dig_encoder_setup(encoder, ATOM_ENABLE); |
| 1142 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_INIT); | 1247 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_INIT, 0, 0); |
| 1143 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP); | 1248 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_SETUP, 0, 0); |
| 1144 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE); | 1249 | atombios_dig_transmitter_setup(encoder, ATOM_TRANSMITTER_ACTION_ENABLE, 0, 0); |
| 1145 | break; | 1250 | break; |
| 1146 | case ENCODER_OBJECT_ID_INTERNAL_DDI: | 1251 | case ENCODER_OBJECT_ID_INTERNAL_DDI: |
| 1147 | atombios_ddia_setup(encoder, ATOM_ENABLE); | 1252 | atombios_ddia_setup(encoder, ATOM_ENABLE); |
| @@ -1354,7 +1459,6 @@ radeon_add_atom_encoder(struct drm_device *dev, uint32_t encoder_id, uint32_t su | |||
| 1354 | encoder->possible_crtcs = 0x1; | 1459 | encoder->possible_crtcs = 0x1; |
| 1355 | else | 1460 | else |
| 1356 | encoder->possible_crtcs = 0x3; | 1461 | encoder->possible_crtcs = 0x3; |
| 1357 | encoder->possible_clones = 0; | ||
| 1358 | 1462 | ||
| 1359 | radeon_encoder->enc_priv = NULL; | 1463 | radeon_encoder->enc_priv = NULL; |
| 1360 | 1464 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_fb.c b/drivers/gpu/drm/radeon/radeon_fb.c index d10eb43645c8..3ba213d1b06c 100644 --- a/drivers/gpu/drm/radeon/radeon_fb.c +++ b/drivers/gpu/drm/radeon/radeon_fb.c | |||
| @@ -140,7 +140,7 @@ int radeonfb_create(struct drm_device *dev, | |||
| 140 | struct radeon_framebuffer *rfb; | 140 | struct radeon_framebuffer *rfb; |
| 141 | struct drm_mode_fb_cmd mode_cmd; | 141 | struct drm_mode_fb_cmd mode_cmd; |
| 142 | struct drm_gem_object *gobj = NULL; | 142 | struct drm_gem_object *gobj = NULL; |
| 143 | struct radeon_object *robj = NULL; | 143 | struct radeon_bo *rbo = NULL; |
| 144 | struct device *device = &rdev->pdev->dev; | 144 | struct device *device = &rdev->pdev->dev; |
| 145 | int size, aligned_size, ret; | 145 | int size, aligned_size, ret; |
| 146 | u64 fb_gpuaddr; | 146 | u64 fb_gpuaddr; |
| @@ -168,14 +168,14 @@ int radeonfb_create(struct drm_device *dev, | |||
| 168 | ret = radeon_gem_object_create(rdev, aligned_size, 0, | 168 | ret = radeon_gem_object_create(rdev, aligned_size, 0, |
| 169 | RADEON_GEM_DOMAIN_VRAM, | 169 | RADEON_GEM_DOMAIN_VRAM, |
| 170 | false, ttm_bo_type_kernel, | 170 | false, ttm_bo_type_kernel, |
| 171 | false, &gobj); | 171 | &gobj); |
| 172 | if (ret) { | 172 | if (ret) { |
| 173 | printk(KERN_ERR "failed to allocate framebuffer (%d %d)\n", | 173 | printk(KERN_ERR "failed to allocate framebuffer (%d %d)\n", |
| 174 | surface_width, surface_height); | 174 | surface_width, surface_height); |
| 175 | ret = -ENOMEM; | 175 | ret = -ENOMEM; |
| 176 | goto out; | 176 | goto out; |
| 177 | } | 177 | } |
| 178 | robj = gobj->driver_private; | 178 | rbo = gobj->driver_private; |
| 179 | 179 | ||
| 180 | if (fb_tiled) | 180 | if (fb_tiled) |
| 181 | tiling_flags = RADEON_TILING_MACRO; | 181 | tiling_flags = RADEON_TILING_MACRO; |
| @@ -192,8 +192,13 @@ int radeonfb_create(struct drm_device *dev, | |||
| 192 | } | 192 | } |
| 193 | #endif | 193 | #endif |
| 194 | 194 | ||
| 195 | if (tiling_flags) | 195 | if (tiling_flags) { |
| 196 | radeon_object_set_tiling_flags(robj, tiling_flags | RADEON_TILING_SURFACE, mode_cmd.pitch); | 196 | ret = radeon_bo_set_tiling_flags(rbo, |
| 197 | tiling_flags | RADEON_TILING_SURFACE, | ||
| 198 | mode_cmd.pitch); | ||
| 199 | if (ret) | ||
| 200 | dev_err(rdev->dev, "FB failed to set tiling flags\n"); | ||
| 201 | } | ||
| 197 | mutex_lock(&rdev->ddev->struct_mutex); | 202 | mutex_lock(&rdev->ddev->struct_mutex); |
| 198 | fb = radeon_framebuffer_create(rdev->ddev, &mode_cmd, gobj); | 203 | fb = radeon_framebuffer_create(rdev->ddev, &mode_cmd, gobj); |
| 199 | if (fb == NULL) { | 204 | if (fb == NULL) { |
| @@ -201,10 +206,19 @@ int radeonfb_create(struct drm_device *dev, | |||
| 201 | ret = -ENOMEM; | 206 | ret = -ENOMEM; |
| 202 | goto out_unref; | 207 | goto out_unref; |
| 203 | } | 208 | } |
| 204 | ret = radeon_object_pin(robj, RADEON_GEM_DOMAIN_VRAM, &fb_gpuaddr); | 209 | ret = radeon_bo_reserve(rbo, false); |
| 210 | if (unlikely(ret != 0)) | ||
| 211 | goto out_unref; | ||
| 212 | ret = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_gpuaddr); | ||
| 213 | if (ret) { | ||
| 214 | radeon_bo_unreserve(rbo); | ||
| 215 | goto out_unref; | ||
| 216 | } | ||
| 217 | if (fb_tiled) | ||
| 218 | radeon_bo_check_tiling(rbo, 0, 0); | ||
| 219 | ret = radeon_bo_kmap(rbo, &fbptr); | ||
| 220 | radeon_bo_unreserve(rbo); | ||
| 205 | if (ret) { | 221 | if (ret) { |
| 206 | printk(KERN_ERR "failed to pin framebuffer\n"); | ||
| 207 | ret = -ENOMEM; | ||
| 208 | goto out_unref; | 222 | goto out_unref; |
| 209 | } | 223 | } |
| 210 | 224 | ||
| @@ -213,7 +227,7 @@ int radeonfb_create(struct drm_device *dev, | |||
| 213 | *fb_p = fb; | 227 | *fb_p = fb; |
| 214 | rfb = to_radeon_framebuffer(fb); | 228 | rfb = to_radeon_framebuffer(fb); |
| 215 | rdev->fbdev_rfb = rfb; | 229 | rdev->fbdev_rfb = rfb; |
| 216 | rdev->fbdev_robj = robj; | 230 | rdev->fbdev_rbo = rbo; |
| 217 | 231 | ||
| 218 | info = framebuffer_alloc(sizeof(struct radeon_fb_device), device); | 232 | info = framebuffer_alloc(sizeof(struct radeon_fb_device), device); |
| 219 | if (info == NULL) { | 233 | if (info == NULL) { |
| @@ -234,15 +248,7 @@ int radeonfb_create(struct drm_device *dev, | |||
| 234 | if (ret) | 248 | if (ret) |
| 235 | goto out_unref; | 249 | goto out_unref; |
| 236 | 250 | ||
| 237 | if (fb_tiled) | 251 | memset_io(fbptr, 0xff, aligned_size); |
| 238 | radeon_object_check_tiling(robj, 0, 0); | ||
| 239 | |||
| 240 | ret = radeon_object_kmap(robj, &fbptr); | ||
| 241 | if (ret) { | ||
| 242 | goto out_unref; | ||
| 243 | } | ||
| 244 | |||
| 245 | memset_io(fbptr, 0, aligned_size); | ||
| 246 | 252 | ||
| 247 | strcpy(info->fix.id, "radeondrmfb"); | 253 | strcpy(info->fix.id, "radeondrmfb"); |
| 248 | 254 | ||
| @@ -288,8 +294,12 @@ int radeonfb_create(struct drm_device *dev, | |||
| 288 | return 0; | 294 | return 0; |
| 289 | 295 | ||
| 290 | out_unref: | 296 | out_unref: |
| 291 | if (robj) { | 297 | if (rbo) { |
| 292 | radeon_object_kunmap(robj); | 298 | ret = radeon_bo_reserve(rbo, false); |
| 299 | if (likely(ret == 0)) { | ||
| 300 | radeon_bo_kunmap(rbo); | ||
| 301 | radeon_bo_unreserve(rbo); | ||
| 302 | } | ||
| 293 | } | 303 | } |
| 294 | if (fb && ret) { | 304 | if (fb && ret) { |
| 295 | list_del(&fb->filp_head); | 305 | list_del(&fb->filp_head); |
| @@ -321,14 +331,22 @@ int radeon_parse_options(char *options) | |||
| 321 | 331 | ||
| 322 | int radeonfb_probe(struct drm_device *dev) | 332 | int radeonfb_probe(struct drm_device *dev) |
| 323 | { | 333 | { |
| 324 | return drm_fb_helper_single_fb_probe(dev, 32, &radeonfb_create); | 334 | struct radeon_device *rdev = dev->dev_private; |
| 335 | int bpp_sel = 32; | ||
| 336 | |||
| 337 | /* select 8 bpp console on RN50 or 16MB cards */ | ||
| 338 | if (ASIC_IS_RN50(rdev) || rdev->mc.real_vram_size <= (32*1024*1024)) | ||
| 339 | bpp_sel = 8; | ||
| 340 | |||
| 341 | return drm_fb_helper_single_fb_probe(dev, bpp_sel, &radeonfb_create); | ||
| 325 | } | 342 | } |
| 326 | 343 | ||
| 327 | int radeonfb_remove(struct drm_device *dev, struct drm_framebuffer *fb) | 344 | int radeonfb_remove(struct drm_device *dev, struct drm_framebuffer *fb) |
| 328 | { | 345 | { |
| 329 | struct fb_info *info; | 346 | struct fb_info *info; |
| 330 | struct radeon_framebuffer *rfb = to_radeon_framebuffer(fb); | 347 | struct radeon_framebuffer *rfb = to_radeon_framebuffer(fb); |
| 331 | struct radeon_object *robj; | 348 | struct radeon_bo *rbo; |
| 349 | int r; | ||
| 332 | 350 | ||
| 333 | if (!fb) { | 351 | if (!fb) { |
| 334 | return -EINVAL; | 352 | return -EINVAL; |
| @@ -336,10 +354,14 @@ int radeonfb_remove(struct drm_device *dev, struct drm_framebuffer *fb) | |||
| 336 | info = fb->fbdev; | 354 | info = fb->fbdev; |
| 337 | if (info) { | 355 | if (info) { |
| 338 | struct radeon_fb_device *rfbdev = info->par; | 356 | struct radeon_fb_device *rfbdev = info->par; |
| 339 | robj = rfb->obj->driver_private; | 357 | rbo = rfb->obj->driver_private; |
| 340 | unregister_framebuffer(info); | 358 | unregister_framebuffer(info); |
| 341 | radeon_object_kunmap(robj); | 359 | r = radeon_bo_reserve(rbo, false); |
| 342 | radeon_object_unpin(robj); | 360 | if (likely(r == 0)) { |
| 361 | radeon_bo_kunmap(rbo); | ||
| 362 | radeon_bo_unpin(rbo); | ||
| 363 | radeon_bo_unreserve(rbo); | ||
| 364 | } | ||
| 343 | drm_fb_helper_free(&rfbdev->helper); | 365 | drm_fb_helper_free(&rfbdev->helper); |
| 344 | framebuffer_release(info); | 366 | framebuffer_release(info); |
| 345 | } | 367 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_fence.c b/drivers/gpu/drm/radeon/radeon_fence.c index 3beb26d74719..cb4cd97ae39f 100644 --- a/drivers/gpu/drm/radeon/radeon_fence.c +++ b/drivers/gpu/drm/radeon/radeon_fence.c | |||
| @@ -168,37 +168,6 @@ bool radeon_fence_signaled(struct radeon_fence *fence) | |||
| 168 | return signaled; | 168 | return signaled; |
| 169 | } | 169 | } |
| 170 | 170 | ||
| 171 | int r600_fence_wait(struct radeon_fence *fence, bool intr, bool lazy) | ||
| 172 | { | ||
| 173 | struct radeon_device *rdev; | ||
| 174 | int ret = 0; | ||
| 175 | |||
| 176 | rdev = fence->rdev; | ||
| 177 | |||
| 178 | __set_current_state(intr ? TASK_INTERRUPTIBLE : TASK_UNINTERRUPTIBLE); | ||
| 179 | |||
| 180 | while (1) { | ||
| 181 | if (radeon_fence_signaled(fence)) | ||
| 182 | break; | ||
| 183 | |||
| 184 | if (time_after_eq(jiffies, fence->timeout)) { | ||
| 185 | ret = -EBUSY; | ||
| 186 | break; | ||
| 187 | } | ||
| 188 | |||
| 189 | if (lazy) | ||
| 190 | schedule_timeout(1); | ||
| 191 | |||
| 192 | if (intr && signal_pending(current)) { | ||
| 193 | ret = -ERESTARTSYS; | ||
| 194 | break; | ||
| 195 | } | ||
| 196 | } | ||
| 197 | __set_current_state(TASK_RUNNING); | ||
| 198 | return ret; | ||
| 199 | } | ||
| 200 | |||
| 201 | |||
| 202 | int radeon_fence_wait(struct radeon_fence *fence, bool intr) | 171 | int radeon_fence_wait(struct radeon_fence *fence, bool intr) |
| 203 | { | 172 | { |
| 204 | struct radeon_device *rdev; | 173 | struct radeon_device *rdev; |
| @@ -216,13 +185,6 @@ int radeon_fence_wait(struct radeon_fence *fence, bool intr) | |||
| 216 | return 0; | 185 | return 0; |
| 217 | } | 186 | } |
| 218 | 187 | ||
| 219 | if (rdev->family >= CHIP_R600) { | ||
| 220 | r = r600_fence_wait(fence, intr, 0); | ||
| 221 | if (r == -ERESTARTSYS) | ||
| 222 | return -EBUSY; | ||
| 223 | return r; | ||
| 224 | } | ||
| 225 | |||
| 226 | retry: | 188 | retry: |
| 227 | cur_jiffies = jiffies; | 189 | cur_jiffies = jiffies; |
| 228 | timeout = HZ / 100; | 190 | timeout = HZ / 100; |
| @@ -231,14 +193,17 @@ retry: | |||
| 231 | } | 193 | } |
| 232 | 194 | ||
| 233 | if (intr) { | 195 | if (intr) { |
| 196 | radeon_irq_kms_sw_irq_get(rdev); | ||
| 234 | r = wait_event_interruptible_timeout(rdev->fence_drv.queue, | 197 | r = wait_event_interruptible_timeout(rdev->fence_drv.queue, |
| 235 | radeon_fence_signaled(fence), timeout); | 198 | radeon_fence_signaled(fence), timeout); |
| 236 | if (unlikely(r == -ERESTARTSYS)) { | 199 | radeon_irq_kms_sw_irq_put(rdev); |
| 237 | return -EBUSY; | 200 | if (unlikely(r < 0)) |
| 238 | } | 201 | return r; |
| 239 | } else { | 202 | } else { |
| 203 | radeon_irq_kms_sw_irq_get(rdev); | ||
| 240 | r = wait_event_timeout(rdev->fence_drv.queue, | 204 | r = wait_event_timeout(rdev->fence_drv.queue, |
| 241 | radeon_fence_signaled(fence), timeout); | 205 | radeon_fence_signaled(fence), timeout); |
| 206 | radeon_irq_kms_sw_irq_put(rdev); | ||
| 242 | } | 207 | } |
| 243 | if (unlikely(!radeon_fence_signaled(fence))) { | 208 | if (unlikely(!radeon_fence_signaled(fence))) { |
| 244 | if (unlikely(r == 0)) { | 209 | if (unlikely(r == 0)) { |
diff --git a/drivers/gpu/drm/radeon/radeon_fixed.h b/drivers/gpu/drm/radeon/radeon_fixed.h index 90187d173847..3d4d84e078ac 100644 --- a/drivers/gpu/drm/radeon/radeon_fixed.h +++ b/drivers/gpu/drm/radeon/radeon_fixed.h | |||
| @@ -38,6 +38,23 @@ typedef union rfixed { | |||
| 38 | #define fixed_init_half(A) { .full = rfixed_const_half((A)) } | 38 | #define fixed_init_half(A) { .full = rfixed_const_half((A)) } |
| 39 | #define rfixed_trunc(A) ((A).full >> 12) | 39 | #define rfixed_trunc(A) ((A).full >> 12) |
| 40 | 40 | ||
| 41 | static inline u32 rfixed_floor(fixed20_12 A) | ||
| 42 | { | ||
| 43 | u32 non_frac = rfixed_trunc(A); | ||
| 44 | |||
| 45 | return rfixed_const(non_frac); | ||
| 46 | } | ||
| 47 | |||
| 48 | static inline u32 rfixed_ceil(fixed20_12 A) | ||
| 49 | { | ||
| 50 | u32 non_frac = rfixed_trunc(A); | ||
| 51 | |||
| 52 | if (A.full > rfixed_const(non_frac)) | ||
| 53 | return rfixed_const(non_frac + 1); | ||
| 54 | else | ||
| 55 | return rfixed_const(non_frac); | ||
| 56 | } | ||
| 57 | |||
| 41 | static inline u32 rfixed_div(fixed20_12 A, fixed20_12 B) | 58 | static inline u32 rfixed_div(fixed20_12 A, fixed20_12 B) |
| 42 | { | 59 | { |
| 43 | u64 tmp = ((u64)A.full << 13); | 60 | u64 tmp = ((u64)A.full << 13); |
diff --git a/drivers/gpu/drm/radeon/radeon_gart.c b/drivers/gpu/drm/radeon/radeon_gart.c index a68d7566178c..e73d56e83fa6 100644 --- a/drivers/gpu/drm/radeon/radeon_gart.c +++ b/drivers/gpu/drm/radeon/radeon_gart.c | |||
| @@ -78,11 +78,9 @@ int radeon_gart_table_vram_alloc(struct radeon_device *rdev) | |||
| 78 | int r; | 78 | int r; |
| 79 | 79 | ||
| 80 | if (rdev->gart.table.vram.robj == NULL) { | 80 | if (rdev->gart.table.vram.robj == NULL) { |
| 81 | r = radeon_object_create(rdev, NULL, | 81 | r = radeon_bo_create(rdev, NULL, rdev->gart.table_size, |
| 82 | rdev->gart.table_size, | 82 | true, RADEON_GEM_DOMAIN_VRAM, |
| 83 | true, | 83 | &rdev->gart.table.vram.robj); |
| 84 | RADEON_GEM_DOMAIN_VRAM, | ||
| 85 | false, &rdev->gart.table.vram.robj); | ||
| 86 | if (r) { | 84 | if (r) { |
| 87 | return r; | 85 | return r; |
| 88 | } | 86 | } |
| @@ -95,32 +93,38 @@ int radeon_gart_table_vram_pin(struct radeon_device *rdev) | |||
| 95 | uint64_t gpu_addr; | 93 | uint64_t gpu_addr; |
| 96 | int r; | 94 | int r; |
| 97 | 95 | ||
| 98 | r = radeon_object_pin(rdev->gart.table.vram.robj, | 96 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 99 | RADEON_GEM_DOMAIN_VRAM, &gpu_addr); | 97 | if (unlikely(r != 0)) |
| 100 | if (r) { | ||
| 101 | radeon_object_unref(&rdev->gart.table.vram.robj); | ||
| 102 | return r; | 98 | return r; |
| 103 | } | 99 | r = radeon_bo_pin(rdev->gart.table.vram.robj, |
| 104 | r = radeon_object_kmap(rdev->gart.table.vram.robj, | 100 | RADEON_GEM_DOMAIN_VRAM, &gpu_addr); |
| 105 | (void **)&rdev->gart.table.vram.ptr); | ||
| 106 | if (r) { | 101 | if (r) { |
| 107 | radeon_object_unpin(rdev->gart.table.vram.robj); | 102 | radeon_bo_unreserve(rdev->gart.table.vram.robj); |
| 108 | radeon_object_unref(&rdev->gart.table.vram.robj); | ||
| 109 | DRM_ERROR("radeon: failed to map gart vram table.\n"); | ||
| 110 | return r; | 103 | return r; |
| 111 | } | 104 | } |
| 105 | r = radeon_bo_kmap(rdev->gart.table.vram.robj, | ||
| 106 | (void **)&rdev->gart.table.vram.ptr); | ||
| 107 | if (r) | ||
| 108 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 109 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 112 | rdev->gart.table_addr = gpu_addr; | 110 | rdev->gart.table_addr = gpu_addr; |
| 113 | return 0; | 111 | return r; |
| 114 | } | 112 | } |
| 115 | 113 | ||
| 116 | void radeon_gart_table_vram_free(struct radeon_device *rdev) | 114 | void radeon_gart_table_vram_free(struct radeon_device *rdev) |
| 117 | { | 115 | { |
| 116 | int r; | ||
| 117 | |||
| 118 | if (rdev->gart.table.vram.robj == NULL) { | 118 | if (rdev->gart.table.vram.robj == NULL) { |
| 119 | return; | 119 | return; |
| 120 | } | 120 | } |
| 121 | radeon_object_kunmap(rdev->gart.table.vram.robj); | 121 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 122 | radeon_object_unpin(rdev->gart.table.vram.robj); | 122 | if (likely(r == 0)) { |
| 123 | radeon_object_unref(&rdev->gart.table.vram.robj); | 123 | radeon_bo_kunmap(rdev->gart.table.vram.robj); |
| 124 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 125 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 126 | } | ||
| 127 | radeon_bo_unref(&rdev->gart.table.vram.robj); | ||
| 124 | } | 128 | } |
| 125 | 129 | ||
| 126 | 130 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_gem.c b/drivers/gpu/drm/radeon/radeon_gem.c index d880edf254db..2944486871b0 100644 --- a/drivers/gpu/drm/radeon/radeon_gem.c +++ b/drivers/gpu/drm/radeon/radeon_gem.c | |||
| @@ -38,22 +38,21 @@ int radeon_gem_object_init(struct drm_gem_object *obj) | |||
| 38 | 38 | ||
| 39 | void radeon_gem_object_free(struct drm_gem_object *gobj) | 39 | void radeon_gem_object_free(struct drm_gem_object *gobj) |
| 40 | { | 40 | { |
| 41 | struct radeon_object *robj = gobj->driver_private; | 41 | struct radeon_bo *robj = gobj->driver_private; |
| 42 | 42 | ||
| 43 | gobj->driver_private = NULL; | 43 | gobj->driver_private = NULL; |
| 44 | if (robj) { | 44 | if (robj) { |
| 45 | radeon_object_unref(&robj); | 45 | radeon_bo_unref(&robj); |
| 46 | } | 46 | } |
| 47 | } | 47 | } |
| 48 | 48 | ||
| 49 | int radeon_gem_object_create(struct radeon_device *rdev, int size, | 49 | int radeon_gem_object_create(struct radeon_device *rdev, int size, |
| 50 | int alignment, int initial_domain, | 50 | int alignment, int initial_domain, |
| 51 | bool discardable, bool kernel, | 51 | bool discardable, bool kernel, |
| 52 | bool interruptible, | 52 | struct drm_gem_object **obj) |
| 53 | struct drm_gem_object **obj) | ||
| 54 | { | 53 | { |
| 55 | struct drm_gem_object *gobj; | 54 | struct drm_gem_object *gobj; |
| 56 | struct radeon_object *robj; | 55 | struct radeon_bo *robj; |
| 57 | int r; | 56 | int r; |
| 58 | 57 | ||
| 59 | *obj = NULL; | 58 | *obj = NULL; |
| @@ -65,8 +64,7 @@ int radeon_gem_object_create(struct radeon_device *rdev, int size, | |||
| 65 | if (alignment < PAGE_SIZE) { | 64 | if (alignment < PAGE_SIZE) { |
| 66 | alignment = PAGE_SIZE; | 65 | alignment = PAGE_SIZE; |
| 67 | } | 66 | } |
| 68 | r = radeon_object_create(rdev, gobj, size, kernel, initial_domain, | 67 | r = radeon_bo_create(rdev, gobj, size, kernel, initial_domain, &robj); |
| 69 | interruptible, &robj); | ||
| 70 | if (r) { | 68 | if (r) { |
| 71 | DRM_ERROR("Failed to allocate GEM object (%d, %d, %u)\n", | 69 | DRM_ERROR("Failed to allocate GEM object (%d, %d, %u)\n", |
| 72 | size, initial_domain, alignment); | 70 | size, initial_domain, alignment); |
| @@ -83,33 +81,33 @@ int radeon_gem_object_create(struct radeon_device *rdev, int size, | |||
| 83 | int radeon_gem_object_pin(struct drm_gem_object *obj, uint32_t pin_domain, | 81 | int radeon_gem_object_pin(struct drm_gem_object *obj, uint32_t pin_domain, |
| 84 | uint64_t *gpu_addr) | 82 | uint64_t *gpu_addr) |
| 85 | { | 83 | { |
| 86 | struct radeon_object *robj = obj->driver_private; | 84 | struct radeon_bo *robj = obj->driver_private; |
| 87 | uint32_t flags; | 85 | int r; |
| 88 | 86 | ||
| 89 | switch (pin_domain) { | 87 | r = radeon_bo_reserve(robj, false); |
| 90 | case RADEON_GEM_DOMAIN_VRAM: | 88 | if (unlikely(r != 0)) |
| 91 | flags = TTM_PL_FLAG_VRAM; | 89 | return r; |
| 92 | break; | 90 | r = radeon_bo_pin(robj, pin_domain, gpu_addr); |
| 93 | case RADEON_GEM_DOMAIN_GTT: | 91 | radeon_bo_unreserve(robj); |
| 94 | flags = TTM_PL_FLAG_TT; | 92 | return r; |
| 95 | break; | ||
| 96 | default: | ||
| 97 | flags = TTM_PL_FLAG_SYSTEM; | ||
| 98 | break; | ||
| 99 | } | ||
| 100 | return radeon_object_pin(robj, flags, gpu_addr); | ||
| 101 | } | 93 | } |
| 102 | 94 | ||
| 103 | void radeon_gem_object_unpin(struct drm_gem_object *obj) | 95 | void radeon_gem_object_unpin(struct drm_gem_object *obj) |
| 104 | { | 96 | { |
| 105 | struct radeon_object *robj = obj->driver_private; | 97 | struct radeon_bo *robj = obj->driver_private; |
| 106 | radeon_object_unpin(robj); | 98 | int r; |
| 99 | |||
| 100 | r = radeon_bo_reserve(robj, false); | ||
| 101 | if (likely(r == 0)) { | ||
| 102 | radeon_bo_unpin(robj); | ||
| 103 | radeon_bo_unreserve(robj); | ||
| 104 | } | ||
| 107 | } | 105 | } |
| 108 | 106 | ||
| 109 | int radeon_gem_set_domain(struct drm_gem_object *gobj, | 107 | int radeon_gem_set_domain(struct drm_gem_object *gobj, |
| 110 | uint32_t rdomain, uint32_t wdomain) | 108 | uint32_t rdomain, uint32_t wdomain) |
| 111 | { | 109 | { |
| 112 | struct radeon_object *robj; | 110 | struct radeon_bo *robj; |
| 113 | uint32_t domain; | 111 | uint32_t domain; |
| 114 | int r; | 112 | int r; |
| 115 | 113 | ||
| @@ -127,11 +125,12 @@ int radeon_gem_set_domain(struct drm_gem_object *gobj, | |||
| 127 | } | 125 | } |
| 128 | if (domain == RADEON_GEM_DOMAIN_CPU) { | 126 | if (domain == RADEON_GEM_DOMAIN_CPU) { |
| 129 | /* Asking for cpu access wait for object idle */ | 127 | /* Asking for cpu access wait for object idle */ |
| 130 | r = radeon_object_wait(robj); | 128 | r = radeon_bo_wait(robj, NULL, false); |
| 131 | if (r) { | 129 | if (r) { |
| 132 | printk(KERN_ERR "Failed to wait for object !\n"); | 130 | printk(KERN_ERR "Failed to wait for object !\n"); |
| 133 | return r; | 131 | return r; |
| 134 | } | 132 | } |
| 133 | radeon_hdp_flush(robj->rdev); | ||
| 135 | } | 134 | } |
| 136 | return 0; | 135 | return 0; |
| 137 | } | 136 | } |
| @@ -144,7 +143,7 @@ int radeon_gem_init(struct radeon_device *rdev) | |||
| 144 | 143 | ||
| 145 | void radeon_gem_fini(struct radeon_device *rdev) | 144 | void radeon_gem_fini(struct radeon_device *rdev) |
| 146 | { | 145 | { |
| 147 | radeon_object_force_delete(rdev); | 146 | radeon_bo_force_delete(rdev); |
| 148 | } | 147 | } |
| 149 | 148 | ||
| 150 | 149 | ||
| @@ -158,9 +157,13 @@ int radeon_gem_info_ioctl(struct drm_device *dev, void *data, | |||
| 158 | struct drm_radeon_gem_info *args = data; | 157 | struct drm_radeon_gem_info *args = data; |
| 159 | 158 | ||
| 160 | args->vram_size = rdev->mc.real_vram_size; | 159 | args->vram_size = rdev->mc.real_vram_size; |
| 161 | /* FIXME: report somethings that makes sense */ | 160 | args->vram_visible = rdev->mc.real_vram_size; |
| 162 | args->vram_visible = rdev->mc.real_vram_size - (4 * 1024 * 1024); | 161 | if (rdev->stollen_vga_memory) |
| 163 | args->gart_size = rdev->mc.gtt_size; | 162 | args->vram_visible -= radeon_bo_size(rdev->stollen_vga_memory); |
| 163 | if (rdev->fbdev_rbo) | ||
| 164 | args->vram_visible -= radeon_bo_size(rdev->fbdev_rbo); | ||
| 165 | args->gart_size = rdev->mc.gtt_size - rdev->cp.ring_size - 4096 - | ||
| 166 | RADEON_IB_POOL_SIZE*64*1024; | ||
| 164 | return 0; | 167 | return 0; |
| 165 | } | 168 | } |
| 166 | 169 | ||
| @@ -192,8 +195,8 @@ int radeon_gem_create_ioctl(struct drm_device *dev, void *data, | |||
| 192 | /* create a gem object to contain this object in */ | 195 | /* create a gem object to contain this object in */ |
| 193 | args->size = roundup(args->size, PAGE_SIZE); | 196 | args->size = roundup(args->size, PAGE_SIZE); |
| 194 | r = radeon_gem_object_create(rdev, args->size, args->alignment, | 197 | r = radeon_gem_object_create(rdev, args->size, args->alignment, |
| 195 | args->initial_domain, false, | 198 | args->initial_domain, false, |
| 196 | false, true, &gobj); | 199 | false, &gobj); |
| 197 | if (r) { | 200 | if (r) { |
| 198 | return r; | 201 | return r; |
| 199 | } | 202 | } |
| @@ -218,7 +221,7 @@ int radeon_gem_set_domain_ioctl(struct drm_device *dev, void *data, | |||
| 218 | * just validate the BO into a certain domain */ | 221 | * just validate the BO into a certain domain */ |
| 219 | struct drm_radeon_gem_set_domain *args = data; | 222 | struct drm_radeon_gem_set_domain *args = data; |
| 220 | struct drm_gem_object *gobj; | 223 | struct drm_gem_object *gobj; |
| 221 | struct radeon_object *robj; | 224 | struct radeon_bo *robj; |
| 222 | int r; | 225 | int r; |
| 223 | 226 | ||
| 224 | /* for now if someone requests domain CPU - | 227 | /* for now if someone requests domain CPU - |
| @@ -244,19 +247,18 @@ int radeon_gem_mmap_ioctl(struct drm_device *dev, void *data, | |||
| 244 | { | 247 | { |
| 245 | struct drm_radeon_gem_mmap *args = data; | 248 | struct drm_radeon_gem_mmap *args = data; |
| 246 | struct drm_gem_object *gobj; | 249 | struct drm_gem_object *gobj; |
| 247 | struct radeon_object *robj; | 250 | struct radeon_bo *robj; |
| 248 | int r; | ||
| 249 | 251 | ||
| 250 | gobj = drm_gem_object_lookup(dev, filp, args->handle); | 252 | gobj = drm_gem_object_lookup(dev, filp, args->handle); |
| 251 | if (gobj == NULL) { | 253 | if (gobj == NULL) { |
| 252 | return -EINVAL; | 254 | return -EINVAL; |
| 253 | } | 255 | } |
| 254 | robj = gobj->driver_private; | 256 | robj = gobj->driver_private; |
| 255 | r = radeon_object_mmap(robj, &args->addr_ptr); | 257 | args->addr_ptr = radeon_bo_mmap_offset(robj); |
| 256 | mutex_lock(&dev->struct_mutex); | 258 | mutex_lock(&dev->struct_mutex); |
| 257 | drm_gem_object_unreference(gobj); | 259 | drm_gem_object_unreference(gobj); |
| 258 | mutex_unlock(&dev->struct_mutex); | 260 | mutex_unlock(&dev->struct_mutex); |
| 259 | return r; | 261 | return 0; |
| 260 | } | 262 | } |
| 261 | 263 | ||
| 262 | int radeon_gem_busy_ioctl(struct drm_device *dev, void *data, | 264 | int radeon_gem_busy_ioctl(struct drm_device *dev, void *data, |
| @@ -264,16 +266,16 @@ int radeon_gem_busy_ioctl(struct drm_device *dev, void *data, | |||
| 264 | { | 266 | { |
| 265 | struct drm_radeon_gem_busy *args = data; | 267 | struct drm_radeon_gem_busy *args = data; |
| 266 | struct drm_gem_object *gobj; | 268 | struct drm_gem_object *gobj; |
| 267 | struct radeon_object *robj; | 269 | struct radeon_bo *robj; |
| 268 | int r; | 270 | int r; |
| 269 | uint32_t cur_placement; | 271 | uint32_t cur_placement = 0; |
| 270 | 272 | ||
| 271 | gobj = drm_gem_object_lookup(dev, filp, args->handle); | 273 | gobj = drm_gem_object_lookup(dev, filp, args->handle); |
| 272 | if (gobj == NULL) { | 274 | if (gobj == NULL) { |
| 273 | return -EINVAL; | 275 | return -EINVAL; |
| 274 | } | 276 | } |
| 275 | robj = gobj->driver_private; | 277 | robj = gobj->driver_private; |
| 276 | r = radeon_object_busy_domain(robj, &cur_placement); | 278 | r = radeon_bo_wait(robj, &cur_placement, true); |
| 277 | switch (cur_placement) { | 279 | switch (cur_placement) { |
| 278 | case TTM_PL_VRAM: | 280 | case TTM_PL_VRAM: |
| 279 | args->domain = RADEON_GEM_DOMAIN_VRAM; | 281 | args->domain = RADEON_GEM_DOMAIN_VRAM; |
| @@ -297,7 +299,7 @@ int radeon_gem_wait_idle_ioctl(struct drm_device *dev, void *data, | |||
| 297 | { | 299 | { |
| 298 | struct drm_radeon_gem_wait_idle *args = data; | 300 | struct drm_radeon_gem_wait_idle *args = data; |
| 299 | struct drm_gem_object *gobj; | 301 | struct drm_gem_object *gobj; |
| 300 | struct radeon_object *robj; | 302 | struct radeon_bo *robj; |
| 301 | int r; | 303 | int r; |
| 302 | 304 | ||
| 303 | gobj = drm_gem_object_lookup(dev, filp, args->handle); | 305 | gobj = drm_gem_object_lookup(dev, filp, args->handle); |
| @@ -305,10 +307,11 @@ int radeon_gem_wait_idle_ioctl(struct drm_device *dev, void *data, | |||
| 305 | return -EINVAL; | 307 | return -EINVAL; |
| 306 | } | 308 | } |
| 307 | robj = gobj->driver_private; | 309 | robj = gobj->driver_private; |
| 308 | r = radeon_object_wait(robj); | 310 | r = radeon_bo_wait(robj, NULL, false); |
| 309 | mutex_lock(&dev->struct_mutex); | 311 | mutex_lock(&dev->struct_mutex); |
| 310 | drm_gem_object_unreference(gobj); | 312 | drm_gem_object_unreference(gobj); |
| 311 | mutex_unlock(&dev->struct_mutex); | 313 | mutex_unlock(&dev->struct_mutex); |
| 314 | radeon_hdp_flush(robj->rdev); | ||
| 312 | return r; | 315 | return r; |
| 313 | } | 316 | } |
| 314 | 317 | ||
| @@ -317,7 +320,7 @@ int radeon_gem_set_tiling_ioctl(struct drm_device *dev, void *data, | |||
| 317 | { | 320 | { |
| 318 | struct drm_radeon_gem_set_tiling *args = data; | 321 | struct drm_radeon_gem_set_tiling *args = data; |
| 319 | struct drm_gem_object *gobj; | 322 | struct drm_gem_object *gobj; |
| 320 | struct radeon_object *robj; | 323 | struct radeon_bo *robj; |
| 321 | int r = 0; | 324 | int r = 0; |
| 322 | 325 | ||
| 323 | DRM_DEBUG("%d \n", args->handle); | 326 | DRM_DEBUG("%d \n", args->handle); |
| @@ -325,7 +328,7 @@ int radeon_gem_set_tiling_ioctl(struct drm_device *dev, void *data, | |||
| 325 | if (gobj == NULL) | 328 | if (gobj == NULL) |
| 326 | return -EINVAL; | 329 | return -EINVAL; |
| 327 | robj = gobj->driver_private; | 330 | robj = gobj->driver_private; |
| 328 | radeon_object_set_tiling_flags(robj, args->tiling_flags, args->pitch); | 331 | r = radeon_bo_set_tiling_flags(robj, args->tiling_flags, args->pitch); |
| 329 | mutex_lock(&dev->struct_mutex); | 332 | mutex_lock(&dev->struct_mutex); |
| 330 | drm_gem_object_unreference(gobj); | 333 | drm_gem_object_unreference(gobj); |
| 331 | mutex_unlock(&dev->struct_mutex); | 334 | mutex_unlock(&dev->struct_mutex); |
| @@ -337,16 +340,19 @@ int radeon_gem_get_tiling_ioctl(struct drm_device *dev, void *data, | |||
| 337 | { | 340 | { |
| 338 | struct drm_radeon_gem_get_tiling *args = data; | 341 | struct drm_radeon_gem_get_tiling *args = data; |
| 339 | struct drm_gem_object *gobj; | 342 | struct drm_gem_object *gobj; |
| 340 | struct radeon_object *robj; | 343 | struct radeon_bo *rbo; |
| 341 | int r = 0; | 344 | int r = 0; |
| 342 | 345 | ||
| 343 | DRM_DEBUG("\n"); | 346 | DRM_DEBUG("\n"); |
| 344 | gobj = drm_gem_object_lookup(dev, filp, args->handle); | 347 | gobj = drm_gem_object_lookup(dev, filp, args->handle); |
| 345 | if (gobj == NULL) | 348 | if (gobj == NULL) |
| 346 | return -EINVAL; | 349 | return -EINVAL; |
| 347 | robj = gobj->driver_private; | 350 | rbo = gobj->driver_private; |
| 348 | radeon_object_get_tiling_flags(robj, &args->tiling_flags, | 351 | r = radeon_bo_reserve(rbo, false); |
| 349 | &args->pitch); | 352 | if (unlikely(r != 0)) |
| 353 | return r; | ||
| 354 | radeon_bo_get_tiling_flags(rbo, &args->tiling_flags, &args->pitch); | ||
| 355 | radeon_bo_unreserve(rbo); | ||
| 350 | mutex_lock(&dev->struct_mutex); | 356 | mutex_lock(&dev->struct_mutex); |
| 351 | drm_gem_object_unreference(gobj); | 357 | drm_gem_object_unreference(gobj); |
| 352 | mutex_unlock(&dev->struct_mutex); | 358 | mutex_unlock(&dev->struct_mutex); |
diff --git a/drivers/gpu/drm/radeon/radeon_i2c.c b/drivers/gpu/drm/radeon/radeon_i2c.c index dd438d32e5c0..da3da1e89d00 100644 --- a/drivers/gpu/drm/radeon/radeon_i2c.c +++ b/drivers/gpu/drm/radeon/radeon_i2c.c | |||
| @@ -59,35 +59,43 @@ bool radeon_ddc_probe(struct radeon_connector *radeon_connector) | |||
| 59 | } | 59 | } |
| 60 | 60 | ||
| 61 | 61 | ||
| 62 | void radeon_i2c_do_lock(struct radeon_connector *radeon_connector, int lock_state) | 62 | void radeon_i2c_do_lock(struct radeon_i2c_chan *i2c, int lock_state) |
| 63 | { | 63 | { |
| 64 | struct radeon_device *rdev = radeon_connector->base.dev->dev_private; | 64 | struct radeon_device *rdev = i2c->dev->dev_private; |
| 65 | struct radeon_i2c_bus_rec *rec = &i2c->rec; | ||
| 65 | uint32_t temp; | 66 | uint32_t temp; |
| 66 | struct radeon_i2c_bus_rec *rec = &radeon_connector->ddc_bus->rec; | ||
| 67 | 67 | ||
| 68 | /* RV410 appears to have a bug where the hw i2c in reset | 68 | /* RV410 appears to have a bug where the hw i2c in reset |
| 69 | * holds the i2c port in a bad state - switch hw i2c away before | 69 | * holds the i2c port in a bad state - switch hw i2c away before |
| 70 | * doing DDC - do this for all r200s/r300s/r400s for safety sake | 70 | * doing DDC - do this for all r200s/r300s/r400s for safety sake |
| 71 | */ | 71 | */ |
| 72 | if ((rdev->family >= CHIP_R200) && !ASIC_IS_AVIVO(rdev)) { | 72 | if (rec->hw_capable) { |
| 73 | if (rec->a_clk_reg == RADEON_GPIO_MONID) { | 73 | if ((rdev->family >= CHIP_R200) && !ASIC_IS_AVIVO(rdev)) { |
| 74 | WREG32(RADEON_DVI_I2C_CNTL_0, (RADEON_I2C_SOFT_RST | | 74 | if (rec->a_clk_reg == RADEON_GPIO_MONID) { |
| 75 | R200_DVI_I2C_PIN_SEL(R200_SEL_DDC1))); | 75 | WREG32(RADEON_DVI_I2C_CNTL_0, (RADEON_I2C_SOFT_RST | |
| 76 | } else { | 76 | R200_DVI_I2C_PIN_SEL(R200_SEL_DDC1))); |
| 77 | WREG32(RADEON_DVI_I2C_CNTL_0, (RADEON_I2C_SOFT_RST | | 77 | } else { |
| 78 | R200_DVI_I2C_PIN_SEL(R200_SEL_DDC3))); | 78 | WREG32(RADEON_DVI_I2C_CNTL_0, (RADEON_I2C_SOFT_RST | |
| 79 | R200_DVI_I2C_PIN_SEL(R200_SEL_DDC3))); | ||
| 80 | } | ||
| 79 | } | 81 | } |
| 80 | } | 82 | } |
| 81 | if (lock_state) { | ||
| 82 | temp = RREG32(rec->a_clk_reg); | ||
| 83 | temp &= ~(rec->a_clk_mask); | ||
| 84 | WREG32(rec->a_clk_reg, temp); | ||
| 85 | |||
| 86 | temp = RREG32(rec->a_data_reg); | ||
| 87 | temp &= ~(rec->a_data_mask); | ||
| 88 | WREG32(rec->a_data_reg, temp); | ||
| 89 | } | ||
| 90 | 83 | ||
| 84 | /* clear the output pin values */ | ||
| 85 | temp = RREG32(rec->a_clk_reg) & ~rec->a_clk_mask; | ||
| 86 | WREG32(rec->a_clk_reg, temp); | ||
| 87 | |||
| 88 | temp = RREG32(rec->a_data_reg) & ~rec->a_data_mask; | ||
| 89 | WREG32(rec->a_data_reg, temp); | ||
| 90 | |||
| 91 | /* set the pins to input */ | ||
| 92 | temp = RREG32(rec->en_clk_reg) & ~rec->en_clk_mask; | ||
| 93 | WREG32(rec->en_clk_reg, temp); | ||
| 94 | |||
| 95 | temp = RREG32(rec->en_data_reg) & ~rec->en_data_mask; | ||
| 96 | WREG32(rec->en_data_reg, temp); | ||
| 97 | |||
| 98 | /* mask the gpio pins for software use */ | ||
| 91 | temp = RREG32(rec->mask_clk_reg); | 99 | temp = RREG32(rec->mask_clk_reg); |
| 92 | if (lock_state) | 100 | if (lock_state) |
| 93 | temp |= rec->mask_clk_mask; | 101 | temp |= rec->mask_clk_mask; |
| @@ -112,8 +120,9 @@ static int get_clock(void *i2c_priv) | |||
| 112 | struct radeon_i2c_bus_rec *rec = &i2c->rec; | 120 | struct radeon_i2c_bus_rec *rec = &i2c->rec; |
| 113 | uint32_t val; | 121 | uint32_t val; |
| 114 | 122 | ||
| 115 | val = RREG32(rec->get_clk_reg); | 123 | /* read the value off the pin */ |
| 116 | val &= rec->get_clk_mask; | 124 | val = RREG32(rec->y_clk_reg); |
| 125 | val &= rec->y_clk_mask; | ||
| 117 | 126 | ||
| 118 | return (val != 0); | 127 | return (val != 0); |
| 119 | } | 128 | } |
| @@ -126,8 +135,10 @@ static int get_data(void *i2c_priv) | |||
| 126 | struct radeon_i2c_bus_rec *rec = &i2c->rec; | 135 | struct radeon_i2c_bus_rec *rec = &i2c->rec; |
| 127 | uint32_t val; | 136 | uint32_t val; |
| 128 | 137 | ||
| 129 | val = RREG32(rec->get_data_reg); | 138 | /* read the value off the pin */ |
| 130 | val &= rec->get_data_mask; | 139 | val = RREG32(rec->y_data_reg); |
| 140 | val &= rec->y_data_mask; | ||
| 141 | |||
| 131 | return (val != 0); | 142 | return (val != 0); |
| 132 | } | 143 | } |
| 133 | 144 | ||
| @@ -138,9 +149,10 @@ static void set_clock(void *i2c_priv, int clock) | |||
| 138 | struct radeon_i2c_bus_rec *rec = &i2c->rec; | 149 | struct radeon_i2c_bus_rec *rec = &i2c->rec; |
| 139 | uint32_t val; | 150 | uint32_t val; |
| 140 | 151 | ||
| 141 | val = RREG32(rec->put_clk_reg) & (uint32_t)~(rec->put_clk_mask); | 152 | /* set pin direction */ |
| 142 | val |= clock ? 0 : rec->put_clk_mask; | 153 | val = RREG32(rec->en_clk_reg) & ~rec->en_clk_mask; |
| 143 | WREG32(rec->put_clk_reg, val); | 154 | val |= clock ? 0 : rec->en_clk_mask; |
| 155 | WREG32(rec->en_clk_reg, val); | ||
| 144 | } | 156 | } |
| 145 | 157 | ||
| 146 | static void set_data(void *i2c_priv, int data) | 158 | static void set_data(void *i2c_priv, int data) |
| @@ -150,14 +162,15 @@ static void set_data(void *i2c_priv, int data) | |||
| 150 | struct radeon_i2c_bus_rec *rec = &i2c->rec; | 162 | struct radeon_i2c_bus_rec *rec = &i2c->rec; |
| 151 | uint32_t val; | 163 | uint32_t val; |
| 152 | 164 | ||
| 153 | val = RREG32(rec->put_data_reg) & (uint32_t)~(rec->put_data_mask); | 165 | /* set pin direction */ |
| 154 | val |= data ? 0 : rec->put_data_mask; | 166 | val = RREG32(rec->en_data_reg) & ~rec->en_data_mask; |
| 155 | WREG32(rec->put_data_reg, val); | 167 | val |= data ? 0 : rec->en_data_mask; |
| 168 | WREG32(rec->en_data_reg, val); | ||
| 156 | } | 169 | } |
| 157 | 170 | ||
| 158 | struct radeon_i2c_chan *radeon_i2c_create(struct drm_device *dev, | 171 | struct radeon_i2c_chan *radeon_i2c_create(struct drm_device *dev, |
| 159 | struct radeon_i2c_bus_rec *rec, | 172 | struct radeon_i2c_bus_rec *rec, |
| 160 | const char *name) | 173 | const char *name) |
| 161 | { | 174 | { |
| 162 | struct radeon_i2c_chan *i2c; | 175 | struct radeon_i2c_chan *i2c; |
| 163 | int ret; | 176 | int ret; |
| @@ -167,20 +180,19 @@ struct radeon_i2c_chan *radeon_i2c_create(struct drm_device *dev, | |||
| 167 | return NULL; | 180 | return NULL; |
| 168 | 181 | ||
| 169 | i2c->adapter.owner = THIS_MODULE; | 182 | i2c->adapter.owner = THIS_MODULE; |
| 170 | i2c->adapter.algo_data = &i2c->algo; | ||
| 171 | i2c->dev = dev; | 183 | i2c->dev = dev; |
| 172 | i2c->algo.setsda = set_data; | 184 | i2c_set_adapdata(&i2c->adapter, i2c); |
| 173 | i2c->algo.setscl = set_clock; | 185 | i2c->adapter.algo_data = &i2c->algo.bit; |
| 174 | i2c->algo.getsda = get_data; | 186 | i2c->algo.bit.setsda = set_data; |
| 175 | i2c->algo.getscl = get_clock; | 187 | i2c->algo.bit.setscl = set_clock; |
| 176 | i2c->algo.udelay = 20; | 188 | i2c->algo.bit.getsda = get_data; |
| 189 | i2c->algo.bit.getscl = get_clock; | ||
| 190 | i2c->algo.bit.udelay = 20; | ||
| 177 | /* vesa says 2.2 ms is enough, 1 jiffy doesn't seem to always | 191 | /* vesa says 2.2 ms is enough, 1 jiffy doesn't seem to always |
| 178 | * make this, 2 jiffies is a lot more reliable */ | 192 | * make this, 2 jiffies is a lot more reliable */ |
| 179 | i2c->algo.timeout = 2; | 193 | i2c->algo.bit.timeout = 2; |
| 180 | i2c->algo.data = i2c; | 194 | i2c->algo.bit.data = i2c; |
| 181 | i2c->rec = *rec; | 195 | i2c->rec = *rec; |
| 182 | i2c_set_adapdata(&i2c->adapter, i2c); | ||
| 183 | |||
| 184 | ret = i2c_bit_add_bus(&i2c->adapter); | 196 | ret = i2c_bit_add_bus(&i2c->adapter); |
| 185 | if (ret) { | 197 | if (ret) { |
| 186 | DRM_INFO("Failed to register i2c %s\n", name); | 198 | DRM_INFO("Failed to register i2c %s\n", name); |
| @@ -194,6 +206,38 @@ out_free: | |||
| 194 | 206 | ||
| 195 | } | 207 | } |
| 196 | 208 | ||
| 209 | struct radeon_i2c_chan *radeon_i2c_create_dp(struct drm_device *dev, | ||
| 210 | struct radeon_i2c_bus_rec *rec, | ||
| 211 | const char *name) | ||
| 212 | { | ||
| 213 | struct radeon_i2c_chan *i2c; | ||
| 214 | int ret; | ||
| 215 | |||
| 216 | i2c = kzalloc(sizeof(struct radeon_i2c_chan), GFP_KERNEL); | ||
| 217 | if (i2c == NULL) | ||
| 218 | return NULL; | ||
| 219 | |||
| 220 | i2c->rec = *rec; | ||
| 221 | i2c->adapter.owner = THIS_MODULE; | ||
| 222 | i2c->dev = dev; | ||
| 223 | i2c_set_adapdata(&i2c->adapter, i2c); | ||
| 224 | i2c->adapter.algo_data = &i2c->algo.dp; | ||
| 225 | i2c->algo.dp.aux_ch = radeon_dp_i2c_aux_ch; | ||
| 226 | i2c->algo.dp.address = 0; | ||
| 227 | ret = i2c_dp_aux_add_bus(&i2c->adapter); | ||
| 228 | if (ret) { | ||
| 229 | DRM_INFO("Failed to register i2c %s\n", name); | ||
| 230 | goto out_free; | ||
| 231 | } | ||
| 232 | |||
| 233 | return i2c; | ||
| 234 | out_free: | ||
| 235 | kfree(i2c); | ||
| 236 | return NULL; | ||
| 237 | |||
| 238 | } | ||
| 239 | |||
| 240 | |||
| 197 | void radeon_i2c_destroy(struct radeon_i2c_chan *i2c) | 241 | void radeon_i2c_destroy(struct radeon_i2c_chan *i2c) |
| 198 | { | 242 | { |
| 199 | if (!i2c) | 243 | if (!i2c) |
| @@ -207,3 +251,59 @@ struct drm_encoder *radeon_best_encoder(struct drm_connector *connector) | |||
| 207 | { | 251 | { |
| 208 | return NULL; | 252 | return NULL; |
| 209 | } | 253 | } |
| 254 | |||
| 255 | void radeon_i2c_sw_get_byte(struct radeon_i2c_chan *i2c_bus, | ||
| 256 | u8 slave_addr, | ||
| 257 | u8 addr, | ||
| 258 | u8 *val) | ||
| 259 | { | ||
| 260 | u8 out_buf[2]; | ||
| 261 | u8 in_buf[2]; | ||
| 262 | struct i2c_msg msgs[] = { | ||
| 263 | { | ||
| 264 | .addr = slave_addr, | ||
| 265 | .flags = 0, | ||
| 266 | .len = 1, | ||
| 267 | .buf = out_buf, | ||
| 268 | }, | ||
| 269 | { | ||
| 270 | .addr = slave_addr, | ||
| 271 | .flags = I2C_M_RD, | ||
| 272 | .len = 1, | ||
| 273 | .buf = in_buf, | ||
| 274 | } | ||
| 275 | }; | ||
| 276 | |||
| 277 | out_buf[0] = addr; | ||
| 278 | out_buf[1] = 0; | ||
| 279 | |||
| 280 | if (i2c_transfer(&i2c_bus->adapter, msgs, 2) == 2) { | ||
| 281 | *val = in_buf[0]; | ||
| 282 | DRM_DEBUG("val = 0x%02x\n", *val); | ||
| 283 | } else { | ||
| 284 | DRM_ERROR("i2c 0x%02x 0x%02x read failed\n", | ||
| 285 | addr, *val); | ||
| 286 | } | ||
| 287 | } | ||
| 288 | |||
| 289 | void radeon_i2c_sw_put_byte(struct radeon_i2c_chan *i2c_bus, | ||
| 290 | u8 slave_addr, | ||
| 291 | u8 addr, | ||
| 292 | u8 val) | ||
| 293 | { | ||
| 294 | uint8_t out_buf[2]; | ||
| 295 | struct i2c_msg msg = { | ||
| 296 | .addr = slave_addr, | ||
| 297 | .flags = 0, | ||
| 298 | .len = 2, | ||
| 299 | .buf = out_buf, | ||
| 300 | }; | ||
| 301 | |||
| 302 | out_buf[0] = addr; | ||
| 303 | out_buf[1] = val; | ||
| 304 | |||
| 305 | if (i2c_transfer(&i2c_bus->adapter, &msg, 1) != 1) | ||
| 306 | DRM_ERROR("i2c 0x%02x 0x%02x write failed\n", | ||
| 307 | addr, val); | ||
| 308 | } | ||
| 309 | |||
diff --git a/drivers/gpu/drm/radeon/radeon_irq_kms.c b/drivers/gpu/drm/radeon/radeon_irq_kms.c index a0fe6232dcb6..9223296fe37b 100644 --- a/drivers/gpu/drm/radeon/radeon_irq_kms.c +++ b/drivers/gpu/drm/radeon/radeon_irq_kms.c | |||
| @@ -39,11 +39,32 @@ irqreturn_t radeon_driver_irq_handler_kms(DRM_IRQ_ARGS) | |||
| 39 | return radeon_irq_process(rdev); | 39 | return radeon_irq_process(rdev); |
| 40 | } | 40 | } |
| 41 | 41 | ||
| 42 | /* | ||
| 43 | * Handle hotplug events outside the interrupt handler proper. | ||
| 44 | */ | ||
| 45 | static void radeon_hotplug_work_func(struct work_struct *work) | ||
| 46 | { | ||
| 47 | struct radeon_device *rdev = container_of(work, struct radeon_device, | ||
| 48 | hotplug_work); | ||
| 49 | struct drm_device *dev = rdev->ddev; | ||
| 50 | struct drm_mode_config *mode_config = &dev->mode_config; | ||
| 51 | struct drm_connector *connector; | ||
| 52 | |||
| 53 | if (mode_config->num_connector) { | ||
| 54 | list_for_each_entry(connector, &mode_config->connector_list, head) | ||
| 55 | radeon_connector_hotplug(connector); | ||
| 56 | } | ||
| 57 | /* Just fire off a uevent and let userspace tell us what to do */ | ||
| 58 | drm_sysfs_hotplug_event(dev); | ||
| 59 | } | ||
| 60 | |||
| 42 | void radeon_driver_irq_preinstall_kms(struct drm_device *dev) | 61 | void radeon_driver_irq_preinstall_kms(struct drm_device *dev) |
| 43 | { | 62 | { |
| 44 | struct radeon_device *rdev = dev->dev_private; | 63 | struct radeon_device *rdev = dev->dev_private; |
| 45 | unsigned i; | 64 | unsigned i; |
| 46 | 65 | ||
| 66 | INIT_WORK(&rdev->hotplug_work, radeon_hotplug_work_func); | ||
| 67 | |||
| 47 | /* Disable *all* interrupts */ | 68 | /* Disable *all* interrupts */ |
| 48 | rdev->irq.sw_int = false; | 69 | rdev->irq.sw_int = false; |
| 49 | for (i = 0; i < 2; i++) { | 70 | for (i = 0; i < 2; i++) { |
| @@ -87,17 +108,25 @@ int radeon_irq_kms_init(struct radeon_device *rdev) | |||
| 87 | 108 | ||
| 88 | if (rdev->flags & RADEON_SINGLE_CRTC) | 109 | if (rdev->flags & RADEON_SINGLE_CRTC) |
| 89 | num_crtc = 1; | 110 | num_crtc = 1; |
| 90 | 111 | spin_lock_init(&rdev->irq.sw_lock); | |
| 91 | r = drm_vblank_init(rdev->ddev, num_crtc); | 112 | r = drm_vblank_init(rdev->ddev, num_crtc); |
| 92 | if (r) { | 113 | if (r) { |
| 93 | return r; | 114 | return r; |
| 94 | } | 115 | } |
| 95 | /* enable msi */ | 116 | /* enable msi */ |
| 96 | rdev->msi_enabled = 0; | 117 | rdev->msi_enabled = 0; |
| 97 | if (rdev->family >= CHIP_RV380) { | 118 | /* MSIs don't seem to work on my rs780; |
| 119 | * not sure about rs880 or other rs780s. | ||
| 120 | * Needs more investigation. | ||
| 121 | */ | ||
| 122 | if ((rdev->family >= CHIP_RV380) && | ||
| 123 | (rdev->family != CHIP_RS780) && | ||
| 124 | (rdev->family != CHIP_RS880)) { | ||
| 98 | int ret = pci_enable_msi(rdev->pdev); | 125 | int ret = pci_enable_msi(rdev->pdev); |
| 99 | if (!ret) | 126 | if (!ret) { |
| 100 | rdev->msi_enabled = 1; | 127 | rdev->msi_enabled = 1; |
| 128 | DRM_INFO("radeon: using MSI.\n"); | ||
| 129 | } | ||
| 101 | } | 130 | } |
| 102 | drm_irq_install(rdev->ddev); | 131 | drm_irq_install(rdev->ddev); |
| 103 | rdev->irq.installed = true; | 132 | rdev->irq.installed = true; |
| @@ -114,3 +143,29 @@ void radeon_irq_kms_fini(struct radeon_device *rdev) | |||
| 114 | pci_disable_msi(rdev->pdev); | 143 | pci_disable_msi(rdev->pdev); |
| 115 | } | 144 | } |
| 116 | } | 145 | } |
| 146 | |||
| 147 | void radeon_irq_kms_sw_irq_get(struct radeon_device *rdev) | ||
| 148 | { | ||
| 149 | unsigned long irqflags; | ||
| 150 | |||
| 151 | spin_lock_irqsave(&rdev->irq.sw_lock, irqflags); | ||
| 152 | if (rdev->ddev->irq_enabled && (++rdev->irq.sw_refcount == 1)) { | ||
| 153 | rdev->irq.sw_int = true; | ||
| 154 | radeon_irq_set(rdev); | ||
| 155 | } | ||
| 156 | spin_unlock_irqrestore(&rdev->irq.sw_lock, irqflags); | ||
| 157 | } | ||
| 158 | |||
| 159 | void radeon_irq_kms_sw_irq_put(struct radeon_device *rdev) | ||
| 160 | { | ||
| 161 | unsigned long irqflags; | ||
| 162 | |||
| 163 | spin_lock_irqsave(&rdev->irq.sw_lock, irqflags); | ||
| 164 | BUG_ON(rdev->ddev->irq_enabled && rdev->irq.sw_refcount <= 0); | ||
| 165 | if (rdev->ddev->irq_enabled && (--rdev->irq.sw_refcount == 0)) { | ||
| 166 | rdev->irq.sw_int = false; | ||
| 167 | radeon_irq_set(rdev); | ||
| 168 | } | ||
| 169 | spin_unlock_irqrestore(&rdev->irq.sw_lock, irqflags); | ||
| 170 | } | ||
| 171 | |||
diff --git a/drivers/gpu/drm/radeon/radeon_kms.c b/drivers/gpu/drm/radeon/radeon_kms.c index ba128621057a..f23b05606eb5 100644 --- a/drivers/gpu/drm/radeon/radeon_kms.c +++ b/drivers/gpu/drm/radeon/radeon_kms.c | |||
| @@ -30,10 +30,19 @@ | |||
| 30 | #include "radeon.h" | 30 | #include "radeon.h" |
| 31 | #include "radeon_drm.h" | 31 | #include "radeon_drm.h" |
| 32 | 32 | ||
| 33 | int radeon_driver_unload_kms(struct drm_device *dev) | ||
| 34 | { | ||
| 35 | struct radeon_device *rdev = dev->dev_private; | ||
| 36 | |||
| 37 | if (rdev == NULL) | ||
| 38 | return 0; | ||
| 39 | radeon_modeset_fini(rdev); | ||
| 40 | radeon_device_fini(rdev); | ||
| 41 | kfree(rdev); | ||
| 42 | dev->dev_private = NULL; | ||
| 43 | return 0; | ||
| 44 | } | ||
| 33 | 45 | ||
| 34 | /* | ||
| 35 | * Driver load/unload | ||
| 36 | */ | ||
| 37 | int radeon_driver_load_kms(struct drm_device *dev, unsigned long flags) | 46 | int radeon_driver_load_kms(struct drm_device *dev, unsigned long flags) |
| 38 | { | 47 | { |
| 39 | struct radeon_device *rdev; | 48 | struct radeon_device *rdev; |
| @@ -62,31 +71,20 @@ int radeon_driver_load_kms(struct drm_device *dev, unsigned long flags) | |||
| 62 | */ | 71 | */ |
| 63 | r = radeon_device_init(rdev, dev, dev->pdev, flags); | 72 | r = radeon_device_init(rdev, dev, dev->pdev, flags); |
| 64 | if (r) { | 73 | if (r) { |
| 65 | DRM_ERROR("Fatal error while trying to initialize radeon.\n"); | 74 | dev_err(&dev->pdev->dev, "Fatal error during GPU init\n"); |
| 66 | return r; | 75 | goto out; |
| 67 | } | 76 | } |
| 68 | /* Again modeset_init should fail only on fatal error | 77 | /* Again modeset_init should fail only on fatal error |
| 69 | * otherwise it should provide enough functionalities | 78 | * otherwise it should provide enough functionalities |
| 70 | * for shadowfb to run | 79 | * for shadowfb to run |
| 71 | */ | 80 | */ |
| 72 | r = radeon_modeset_init(rdev); | 81 | r = radeon_modeset_init(rdev); |
| 73 | if (r) { | 82 | if (r) |
| 74 | return r; | 83 | dev_err(&dev->pdev->dev, "Fatal error during modeset init\n"); |
| 75 | } | 84 | out: |
| 76 | return 0; | 85 | if (r) |
| 77 | } | 86 | radeon_driver_unload_kms(dev); |
| 78 | 87 | return r; | |
| 79 | int radeon_driver_unload_kms(struct drm_device *dev) | ||
| 80 | { | ||
| 81 | struct radeon_device *rdev = dev->dev_private; | ||
| 82 | |||
| 83 | if (rdev == NULL) | ||
| 84 | return 0; | ||
| 85 | radeon_modeset_fini(rdev); | ||
| 86 | radeon_device_fini(rdev); | ||
| 87 | kfree(rdev); | ||
| 88 | dev->dev_private = NULL; | ||
| 89 | return 0; | ||
| 90 | } | 88 | } |
| 91 | 89 | ||
| 92 | 90 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_legacy_crtc.c b/drivers/gpu/drm/radeon/radeon_legacy_crtc.c index 8d0b7aa87fa4..b82ede98e152 100644 --- a/drivers/gpu/drm/radeon/radeon_legacy_crtc.c +++ b/drivers/gpu/drm/radeon/radeon_legacy_crtc.c | |||
| @@ -30,6 +30,18 @@ | |||
| 30 | #include "radeon.h" | 30 | #include "radeon.h" |
| 31 | #include "atom.h" | 31 | #include "atom.h" |
| 32 | 32 | ||
| 33 | static void radeon_overscan_setup(struct drm_crtc *crtc, | ||
| 34 | struct drm_display_mode *mode) | ||
| 35 | { | ||
| 36 | struct drm_device *dev = crtc->dev; | ||
| 37 | struct radeon_device *rdev = dev->dev_private; | ||
| 38 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc); | ||
| 39 | |||
| 40 | WREG32(RADEON_OVR_CLR + radeon_crtc->crtc_offset, 0); | ||
| 41 | WREG32(RADEON_OVR_WID_LEFT_RIGHT + radeon_crtc->crtc_offset, 0); | ||
| 42 | WREG32(RADEON_OVR_WID_TOP_BOTTOM + radeon_crtc->crtc_offset, 0); | ||
| 43 | } | ||
| 44 | |||
| 33 | static void radeon_legacy_rmx_mode_set(struct drm_crtc *crtc, | 45 | static void radeon_legacy_rmx_mode_set(struct drm_crtc *crtc, |
| 34 | struct drm_display_mode *mode, | 46 | struct drm_display_mode *mode, |
| 35 | struct drm_display_mode *adjusted_mode) | 47 | struct drm_display_mode *adjusted_mode) |
| @@ -292,8 +304,7 @@ void radeon_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 292 | uint32_t mask; | 304 | uint32_t mask; |
| 293 | 305 | ||
| 294 | if (radeon_crtc->crtc_id) | 306 | if (radeon_crtc->crtc_id) |
| 295 | mask = (RADEON_CRTC2_EN | | 307 | mask = (RADEON_CRTC2_DISP_DIS | |
| 296 | RADEON_CRTC2_DISP_DIS | | ||
| 297 | RADEON_CRTC2_VSYNC_DIS | | 308 | RADEON_CRTC2_VSYNC_DIS | |
| 298 | RADEON_CRTC2_HSYNC_DIS | | 309 | RADEON_CRTC2_HSYNC_DIS | |
| 299 | RADEON_CRTC2_DISP_REQ_EN_B); | 310 | RADEON_CRTC2_DISP_REQ_EN_B); |
| @@ -305,7 +316,7 @@ void radeon_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 305 | switch (mode) { | 316 | switch (mode) { |
| 306 | case DRM_MODE_DPMS_ON: | 317 | case DRM_MODE_DPMS_ON: |
| 307 | if (radeon_crtc->crtc_id) | 318 | if (radeon_crtc->crtc_id) |
| 308 | WREG32_P(RADEON_CRTC2_GEN_CNTL, RADEON_CRTC2_EN, ~mask); | 319 | WREG32_P(RADEON_CRTC2_GEN_CNTL, RADEON_CRTC2_EN, ~(RADEON_CRTC2_EN | mask)); |
| 309 | else { | 320 | else { |
| 310 | WREG32_P(RADEON_CRTC_GEN_CNTL, RADEON_CRTC_EN, ~(RADEON_CRTC_EN | | 321 | WREG32_P(RADEON_CRTC_GEN_CNTL, RADEON_CRTC_EN, ~(RADEON_CRTC_EN | |
| 311 | RADEON_CRTC_DISP_REQ_EN_B)); | 322 | RADEON_CRTC_DISP_REQ_EN_B)); |
| @@ -319,7 +330,7 @@ void radeon_crtc_dpms(struct drm_crtc *crtc, int mode) | |||
| 319 | case DRM_MODE_DPMS_OFF: | 330 | case DRM_MODE_DPMS_OFF: |
| 320 | drm_vblank_pre_modeset(dev, radeon_crtc->crtc_id); | 331 | drm_vblank_pre_modeset(dev, radeon_crtc->crtc_id); |
| 321 | if (radeon_crtc->crtc_id) | 332 | if (radeon_crtc->crtc_id) |
| 322 | WREG32_P(RADEON_CRTC2_GEN_CNTL, mask, ~mask); | 333 | WREG32_P(RADEON_CRTC2_GEN_CNTL, mask, ~(RADEON_CRTC2_EN | mask)); |
| 323 | else { | 334 | else { |
| 324 | WREG32_P(RADEON_CRTC_GEN_CNTL, RADEON_CRTC_DISP_REQ_EN_B, ~(RADEON_CRTC_EN | | 335 | WREG32_P(RADEON_CRTC_GEN_CNTL, RADEON_CRTC_DISP_REQ_EN_B, ~(RADEON_CRTC_EN | |
| 325 | RADEON_CRTC_DISP_REQ_EN_B)); | 336 | RADEON_CRTC_DISP_REQ_EN_B)); |
| @@ -400,14 +411,21 @@ int radeon_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 400 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc); | 411 | struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc); |
| 401 | struct radeon_framebuffer *radeon_fb; | 412 | struct radeon_framebuffer *radeon_fb; |
| 402 | struct drm_gem_object *obj; | 413 | struct drm_gem_object *obj; |
| 414 | struct radeon_bo *rbo; | ||
| 403 | uint64_t base; | 415 | uint64_t base; |
| 404 | uint32_t crtc_offset, crtc_offset_cntl, crtc_tile_x0_y0 = 0; | 416 | uint32_t crtc_offset, crtc_offset_cntl, crtc_tile_x0_y0 = 0; |
| 405 | uint32_t crtc_pitch, pitch_pixels; | 417 | uint32_t crtc_pitch, pitch_pixels; |
| 406 | uint32_t tiling_flags; | 418 | uint32_t tiling_flags; |
| 407 | int format; | 419 | int format; |
| 408 | uint32_t gen_cntl_reg, gen_cntl_val; | 420 | uint32_t gen_cntl_reg, gen_cntl_val; |
| 421 | int r; | ||
| 409 | 422 | ||
| 410 | DRM_DEBUG("\n"); | 423 | DRM_DEBUG("\n"); |
| 424 | /* no fb bound */ | ||
| 425 | if (!crtc->fb) { | ||
| 426 | DRM_DEBUG("No FB bound\n"); | ||
| 427 | return 0; | ||
| 428 | } | ||
| 411 | 429 | ||
| 412 | radeon_fb = to_radeon_framebuffer(crtc->fb); | 430 | radeon_fb = to_radeon_framebuffer(crtc->fb); |
| 413 | 431 | ||
| @@ -431,10 +449,22 @@ int radeon_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 431 | return false; | 449 | return false; |
| 432 | } | 450 | } |
| 433 | 451 | ||
| 452 | /* Pin framebuffer & get tilling informations */ | ||
| 434 | obj = radeon_fb->obj; | 453 | obj = radeon_fb->obj; |
| 435 | if (radeon_gem_object_pin(obj, RADEON_GEM_DOMAIN_VRAM, &base)) { | 454 | rbo = obj->driver_private; |
| 455 | r = radeon_bo_reserve(rbo, false); | ||
| 456 | if (unlikely(r != 0)) | ||
| 457 | return r; | ||
| 458 | r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &base); | ||
| 459 | if (unlikely(r != 0)) { | ||
| 460 | radeon_bo_unreserve(rbo); | ||
| 436 | return -EINVAL; | 461 | return -EINVAL; |
| 437 | } | 462 | } |
| 463 | radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); | ||
| 464 | radeon_bo_unreserve(rbo); | ||
| 465 | if (tiling_flags & RADEON_TILING_MICRO) | ||
| 466 | DRM_ERROR("trying to scanout microtiled buffer\n"); | ||
| 467 | |||
| 438 | /* if scanout was in GTT this really wouldn't work */ | 468 | /* if scanout was in GTT this really wouldn't work */ |
| 439 | /* crtc offset is from display base addr not FB location */ | 469 | /* crtc offset is from display base addr not FB location */ |
| 440 | radeon_crtc->legacy_display_base_addr = rdev->mc.vram_location; | 470 | radeon_crtc->legacy_display_base_addr = rdev->mc.vram_location; |
| @@ -449,10 +479,6 @@ int radeon_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 449 | (crtc->fb->bits_per_pixel * 8)); | 479 | (crtc->fb->bits_per_pixel * 8)); |
| 450 | crtc_pitch |= crtc_pitch << 16; | 480 | crtc_pitch |= crtc_pitch << 16; |
| 451 | 481 | ||
| 452 | radeon_object_get_tiling_flags(obj->driver_private, | ||
| 453 | &tiling_flags, NULL); | ||
| 454 | if (tiling_flags & RADEON_TILING_MICRO) | ||
| 455 | DRM_ERROR("trying to scanout microtiled buffer\n"); | ||
| 456 | 482 | ||
| 457 | if (tiling_flags & RADEON_TILING_MACRO) { | 483 | if (tiling_flags & RADEON_TILING_MACRO) { |
| 458 | if (ASIC_IS_R300(rdev)) | 484 | if (ASIC_IS_R300(rdev)) |
| @@ -530,7 +556,12 @@ int radeon_crtc_set_base(struct drm_crtc *crtc, int x, int y, | |||
| 530 | 556 | ||
| 531 | if (old_fb && old_fb != crtc->fb) { | 557 | if (old_fb && old_fb != crtc->fb) { |
| 532 | radeon_fb = to_radeon_framebuffer(old_fb); | 558 | radeon_fb = to_radeon_framebuffer(old_fb); |
| 533 | radeon_gem_object_unpin(radeon_fb->obj); | 559 | rbo = radeon_fb->obj->driver_private; |
| 560 | r = radeon_bo_reserve(rbo, false); | ||
| 561 | if (unlikely(r != 0)) | ||
| 562 | return r; | ||
| 563 | radeon_bo_unpin(rbo); | ||
| 564 | radeon_bo_unreserve(rbo); | ||
| 534 | } | 565 | } |
| 535 | 566 | ||
| 536 | /* Bytes per pixel may have changed */ | 567 | /* Bytes per pixel may have changed */ |
| @@ -642,12 +673,8 @@ static bool radeon_set_crtc_timing(struct drm_crtc *crtc, struct drm_display_mod | |||
| 642 | uint32_t crtc2_gen_cntl; | 673 | uint32_t crtc2_gen_cntl; |
| 643 | uint32_t disp2_merge_cntl; | 674 | uint32_t disp2_merge_cntl; |
| 644 | 675 | ||
| 645 | /* check to see if TV DAC is enabled for another crtc and keep it enabled */ | 676 | /* if TV DAC is enabled for another crtc and keep it enabled */ |
| 646 | if (RREG32(RADEON_CRTC2_GEN_CNTL) & RADEON_CRTC2_CRT2_ON) | 677 | crtc2_gen_cntl = RREG32(RADEON_CRTC2_GEN_CNTL) & 0x00718080; |
| 647 | crtc2_gen_cntl = RADEON_CRTC2_CRT2_ON; | ||
| 648 | else | ||
| 649 | crtc2_gen_cntl = 0; | ||
| 650 | |||
| 651 | crtc2_gen_cntl |= ((format << 8) | 678 | crtc2_gen_cntl |= ((format << 8) |
| 652 | | RADEON_CRTC2_VSYNC_DIS | 679 | | RADEON_CRTC2_VSYNC_DIS |
| 653 | | RADEON_CRTC2_HSYNC_DIS | 680 | | RADEON_CRTC2_HSYNC_DIS |
| @@ -676,7 +703,8 @@ static bool radeon_set_crtc_timing(struct drm_crtc *crtc, struct drm_display_mod | |||
| 676 | uint32_t crtc_ext_cntl; | 703 | uint32_t crtc_ext_cntl; |
| 677 | uint32_t disp_merge_cntl; | 704 | uint32_t disp_merge_cntl; |
| 678 | 705 | ||
| 679 | crtc_gen_cntl = (RADEON_CRTC_EXT_DISP_EN | 706 | crtc_gen_cntl = RREG32(RADEON_CRTC_GEN_CNTL) & 0x00718000; |
| 707 | crtc_gen_cntl |= (RADEON_CRTC_EXT_DISP_EN | ||
| 680 | | (format << 8) | 708 | | (format << 8) |
| 681 | | RADEON_CRTC_DISP_REQ_EN_B | 709 | | RADEON_CRTC_DISP_REQ_EN_B |
| 682 | | ((mode->flags & DRM_MODE_FLAG_DBLSCAN) | 710 | | ((mode->flags & DRM_MODE_FLAG_DBLSCAN) |
| @@ -779,15 +807,17 @@ static void radeon_set_pll(struct drm_crtc *crtc, struct drm_display_mode *mode) | |||
| 779 | if (encoder->encoder_type != DRM_MODE_ENCODER_DAC) | 807 | if (encoder->encoder_type != DRM_MODE_ENCODER_DAC) |
| 780 | pll_flags |= RADEON_PLL_NO_ODD_POST_DIV; | 808 | pll_flags |= RADEON_PLL_NO_ODD_POST_DIV; |
| 781 | if (encoder->encoder_type == DRM_MODE_ENCODER_LVDS) { | 809 | if (encoder->encoder_type == DRM_MODE_ENCODER_LVDS) { |
| 782 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | 810 | if (!rdev->is_atom_bios) { |
| 783 | struct radeon_encoder_lvds *lvds = (struct radeon_encoder_lvds *)radeon_encoder->enc_priv; | 811 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); |
| 784 | if (lvds) { | 812 | struct radeon_encoder_lvds *lvds = (struct radeon_encoder_lvds *)radeon_encoder->enc_priv; |
| 785 | if (lvds->use_bios_dividers) { | 813 | if (lvds) { |
| 786 | pll_ref_div = lvds->panel_ref_divider; | 814 | if (lvds->use_bios_dividers) { |
| 787 | pll_fb_post_div = (lvds->panel_fb_divider | | 815 | pll_ref_div = lvds->panel_ref_divider; |
| 788 | (lvds->panel_post_divider << 16)); | 816 | pll_fb_post_div = (lvds->panel_fb_divider | |
| 789 | htotal_cntl = 0; | 817 | (lvds->panel_post_divider << 16)); |
| 790 | use_bios_divs = true; | 818 | htotal_cntl = 0; |
| 819 | use_bios_divs = true; | ||
| 820 | } | ||
| 791 | } | 821 | } |
| 792 | } | 822 | } |
| 793 | pll_flags |= RADEON_PLL_USE_REF_DIV; | 823 | pll_flags |= RADEON_PLL_USE_REF_DIV; |
| @@ -1027,6 +1057,7 @@ static int radeon_crtc_mode_set(struct drm_crtc *crtc, | |||
| 1027 | radeon_crtc_set_base(crtc, x, y, old_fb); | 1057 | radeon_crtc_set_base(crtc, x, y, old_fb); |
| 1028 | radeon_set_crtc_timing(crtc, adjusted_mode); | 1058 | radeon_set_crtc_timing(crtc, adjusted_mode); |
| 1029 | radeon_set_pll(crtc, adjusted_mode); | 1059 | radeon_set_pll(crtc, adjusted_mode); |
| 1060 | radeon_overscan_setup(crtc, adjusted_mode); | ||
| 1030 | if (radeon_crtc->crtc_id == 0) { | 1061 | if (radeon_crtc->crtc_id == 0) { |
| 1031 | radeon_legacy_rmx_mode_set(crtc, mode, adjusted_mode); | 1062 | radeon_legacy_rmx_mode_set(crtc, mode, adjusted_mode); |
| 1032 | } else { | 1063 | } else { |
| @@ -1042,12 +1073,29 @@ static int radeon_crtc_mode_set(struct drm_crtc *crtc, | |||
| 1042 | 1073 | ||
| 1043 | static void radeon_crtc_prepare(struct drm_crtc *crtc) | 1074 | static void radeon_crtc_prepare(struct drm_crtc *crtc) |
| 1044 | { | 1075 | { |
| 1045 | radeon_crtc_dpms(crtc, DRM_MODE_DPMS_OFF); | 1076 | struct drm_device *dev = crtc->dev; |
| 1077 | struct drm_crtc *crtci; | ||
| 1078 | |||
| 1079 | /* | ||
| 1080 | * The hardware wedges sometimes if you reconfigure one CRTC | ||
| 1081 | * whilst another is running (see fdo bug #24611). | ||
| 1082 | */ | ||
| 1083 | list_for_each_entry(crtci, &dev->mode_config.crtc_list, head) | ||
| 1084 | radeon_crtc_dpms(crtci, DRM_MODE_DPMS_OFF); | ||
| 1046 | } | 1085 | } |
| 1047 | 1086 | ||
| 1048 | static void radeon_crtc_commit(struct drm_crtc *crtc) | 1087 | static void radeon_crtc_commit(struct drm_crtc *crtc) |
| 1049 | { | 1088 | { |
| 1050 | radeon_crtc_dpms(crtc, DRM_MODE_DPMS_ON); | 1089 | struct drm_device *dev = crtc->dev; |
| 1090 | struct drm_crtc *crtci; | ||
| 1091 | |||
| 1092 | /* | ||
| 1093 | * Reenable the CRTCs that should be running. | ||
| 1094 | */ | ||
| 1095 | list_for_each_entry(crtci, &dev->mode_config.crtc_list, head) { | ||
| 1096 | if (crtci->enabled) | ||
| 1097 | radeon_crtc_dpms(crtci, DRM_MODE_DPMS_ON); | ||
| 1098 | } | ||
| 1051 | } | 1099 | } |
| 1052 | 1100 | ||
| 1053 | static const struct drm_crtc_helper_funcs legacy_helper_funcs = { | 1101 | static const struct drm_crtc_helper_funcs legacy_helper_funcs = { |
diff --git a/drivers/gpu/drm/radeon/radeon_legacy_encoders.c b/drivers/gpu/drm/radeon/radeon_legacy_encoders.c index 00382122869b..df00515e81fa 100644 --- a/drivers/gpu/drm/radeon/radeon_legacy_encoders.c +++ b/drivers/gpu/drm/radeon/radeon_legacy_encoders.c | |||
| @@ -136,7 +136,14 @@ static void radeon_legacy_lvds_mode_set(struct drm_encoder *encoder, | |||
| 136 | lvds_pll_cntl &= ~RADEON_LVDS_PLL_EN; | 136 | lvds_pll_cntl &= ~RADEON_LVDS_PLL_EN; |
| 137 | 137 | ||
| 138 | lvds_ss_gen_cntl = RREG32(RADEON_LVDS_SS_GEN_CNTL); | 138 | lvds_ss_gen_cntl = RREG32(RADEON_LVDS_SS_GEN_CNTL); |
| 139 | if ((!rdev->is_atom_bios)) { | 139 | if (rdev->is_atom_bios) { |
| 140 | /* LVDS_GEN_CNTL parameters are computed in LVDSEncoderControl | ||
| 141 | * need to call that on resume to set up the reg properly. | ||
| 142 | */ | ||
| 143 | radeon_encoder->pixel_clock = adjusted_mode->clock; | ||
| 144 | atombios_digital_setup(encoder, PANEL_ENCODER_ACTION_ENABLE); | ||
| 145 | lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); | ||
| 146 | } else { | ||
| 140 | struct radeon_encoder_lvds *lvds = (struct radeon_encoder_lvds *)radeon_encoder->enc_priv; | 147 | struct radeon_encoder_lvds *lvds = (struct radeon_encoder_lvds *)radeon_encoder->enc_priv; |
| 141 | if (lvds) { | 148 | if (lvds) { |
| 142 | DRM_DEBUG("bios LVDS_GEN_CNTL: 0x%x\n", lvds->lvds_gen_cntl); | 149 | DRM_DEBUG("bios LVDS_GEN_CNTL: 0x%x\n", lvds->lvds_gen_cntl); |
| @@ -147,8 +154,7 @@ static void radeon_legacy_lvds_mode_set(struct drm_encoder *encoder, | |||
| 147 | (lvds->panel_blon_delay << RADEON_LVDS_PWRSEQ_DELAY2_SHIFT)); | 154 | (lvds->panel_blon_delay << RADEON_LVDS_PWRSEQ_DELAY2_SHIFT)); |
| 148 | } else | 155 | } else |
| 149 | lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); | 156 | lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); |
| 150 | } else | 157 | } |
| 151 | lvds_gen_cntl = RREG32(RADEON_LVDS_GEN_CNTL); | ||
| 152 | lvds_gen_cntl |= RADEON_LVDS_DISPLAY_DIS; | 158 | lvds_gen_cntl |= RADEON_LVDS_DISPLAY_DIS; |
| 153 | lvds_gen_cntl &= ~(RADEON_LVDS_ON | | 159 | lvds_gen_cntl &= ~(RADEON_LVDS_ON | |
| 154 | RADEON_LVDS_BLON | | 160 | RADEON_LVDS_BLON | |
| @@ -184,9 +190,9 @@ static void radeon_legacy_lvds_mode_set(struct drm_encoder *encoder, | |||
| 184 | radeon_combios_encoder_crtc_scratch_regs(encoder, radeon_crtc->crtc_id); | 190 | radeon_combios_encoder_crtc_scratch_regs(encoder, radeon_crtc->crtc_id); |
| 185 | } | 191 | } |
| 186 | 192 | ||
| 187 | static bool radeon_legacy_lvds_mode_fixup(struct drm_encoder *encoder, | 193 | static bool radeon_legacy_mode_fixup(struct drm_encoder *encoder, |
| 188 | struct drm_display_mode *mode, | 194 | struct drm_display_mode *mode, |
| 189 | struct drm_display_mode *adjusted_mode) | 195 | struct drm_display_mode *adjusted_mode) |
| 190 | { | 196 | { |
| 191 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | 197 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); |
| 192 | 198 | ||
| @@ -194,15 +200,22 @@ static bool radeon_legacy_lvds_mode_fixup(struct drm_encoder *encoder, | |||
| 194 | radeon_encoder_set_active_device(encoder); | 200 | radeon_encoder_set_active_device(encoder); |
| 195 | drm_mode_set_crtcinfo(adjusted_mode, 0); | 201 | drm_mode_set_crtcinfo(adjusted_mode, 0); |
| 196 | 202 | ||
| 197 | if (radeon_encoder->rmx_type != RMX_OFF) | 203 | /* get the native mode for LVDS */ |
| 198 | radeon_rmx_mode_fixup(encoder, mode, adjusted_mode); | 204 | if (radeon_encoder->active_device & (ATOM_DEVICE_LCD_SUPPORT)) { |
| 205 | struct drm_display_mode *native_mode = &radeon_encoder->native_mode; | ||
| 206 | int mode_id = adjusted_mode->base.id; | ||
| 207 | *adjusted_mode = *native_mode; | ||
| 208 | adjusted_mode->hdisplay = mode->hdisplay; | ||
| 209 | adjusted_mode->vdisplay = mode->vdisplay; | ||
| 210 | adjusted_mode->base.id = mode_id; | ||
| 211 | } | ||
| 199 | 212 | ||
| 200 | return true; | 213 | return true; |
| 201 | } | 214 | } |
| 202 | 215 | ||
| 203 | static const struct drm_encoder_helper_funcs radeon_legacy_lvds_helper_funcs = { | 216 | static const struct drm_encoder_helper_funcs radeon_legacy_lvds_helper_funcs = { |
| 204 | .dpms = radeon_legacy_lvds_dpms, | 217 | .dpms = radeon_legacy_lvds_dpms, |
| 205 | .mode_fixup = radeon_legacy_lvds_mode_fixup, | 218 | .mode_fixup = radeon_legacy_mode_fixup, |
| 206 | .prepare = radeon_legacy_lvds_prepare, | 219 | .prepare = radeon_legacy_lvds_prepare, |
| 207 | .mode_set = radeon_legacy_lvds_mode_set, | 220 | .mode_set = radeon_legacy_lvds_mode_set, |
| 208 | .commit = radeon_legacy_lvds_commit, | 221 | .commit = radeon_legacy_lvds_commit, |
| @@ -214,17 +227,6 @@ static const struct drm_encoder_funcs radeon_legacy_lvds_enc_funcs = { | |||
| 214 | .destroy = radeon_enc_destroy, | 227 | .destroy = radeon_enc_destroy, |
| 215 | }; | 228 | }; |
| 216 | 229 | ||
| 217 | static bool radeon_legacy_primary_dac_mode_fixup(struct drm_encoder *encoder, | ||
| 218 | struct drm_display_mode *mode, | ||
| 219 | struct drm_display_mode *adjusted_mode) | ||
| 220 | { | ||
| 221 | /* set the active encoder to connector routing */ | ||
| 222 | radeon_encoder_set_active_device(encoder); | ||
| 223 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
| 224 | |||
| 225 | return true; | ||
| 226 | } | ||
| 227 | |||
| 228 | static void radeon_legacy_primary_dac_dpms(struct drm_encoder *encoder, int mode) | 230 | static void radeon_legacy_primary_dac_dpms(struct drm_encoder *encoder, int mode) |
| 229 | { | 231 | { |
| 230 | struct drm_device *dev = encoder->dev; | 232 | struct drm_device *dev = encoder->dev; |
| @@ -410,7 +412,7 @@ static enum drm_connector_status radeon_legacy_primary_dac_detect(struct drm_enc | |||
| 410 | 412 | ||
| 411 | static const struct drm_encoder_helper_funcs radeon_legacy_primary_dac_helper_funcs = { | 413 | static const struct drm_encoder_helper_funcs radeon_legacy_primary_dac_helper_funcs = { |
| 412 | .dpms = radeon_legacy_primary_dac_dpms, | 414 | .dpms = radeon_legacy_primary_dac_dpms, |
| 413 | .mode_fixup = radeon_legacy_primary_dac_mode_fixup, | 415 | .mode_fixup = radeon_legacy_mode_fixup, |
| 414 | .prepare = radeon_legacy_primary_dac_prepare, | 416 | .prepare = radeon_legacy_primary_dac_prepare, |
| 415 | .mode_set = radeon_legacy_primary_dac_mode_set, | 417 | .mode_set = radeon_legacy_primary_dac_mode_set, |
| 416 | .commit = radeon_legacy_primary_dac_commit, | 418 | .commit = radeon_legacy_primary_dac_commit, |
| @@ -423,16 +425,6 @@ static const struct drm_encoder_funcs radeon_legacy_primary_dac_enc_funcs = { | |||
| 423 | .destroy = radeon_enc_destroy, | 425 | .destroy = radeon_enc_destroy, |
| 424 | }; | 426 | }; |
| 425 | 427 | ||
| 426 | static bool radeon_legacy_tmds_int_mode_fixup(struct drm_encoder *encoder, | ||
| 427 | struct drm_display_mode *mode, | ||
| 428 | struct drm_display_mode *adjusted_mode) | ||
| 429 | { | ||
| 430 | |||
| 431 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
| 432 | |||
| 433 | return true; | ||
| 434 | } | ||
| 435 | |||
| 436 | static void radeon_legacy_tmds_int_dpms(struct drm_encoder *encoder, int mode) | 428 | static void radeon_legacy_tmds_int_dpms(struct drm_encoder *encoder, int mode) |
| 437 | { | 429 | { |
| 438 | struct drm_device *dev = encoder->dev; | 430 | struct drm_device *dev = encoder->dev; |
| @@ -584,7 +576,7 @@ static void radeon_legacy_tmds_int_mode_set(struct drm_encoder *encoder, | |||
| 584 | 576 | ||
| 585 | static const struct drm_encoder_helper_funcs radeon_legacy_tmds_int_helper_funcs = { | 577 | static const struct drm_encoder_helper_funcs radeon_legacy_tmds_int_helper_funcs = { |
| 586 | .dpms = radeon_legacy_tmds_int_dpms, | 578 | .dpms = radeon_legacy_tmds_int_dpms, |
| 587 | .mode_fixup = radeon_legacy_tmds_int_mode_fixup, | 579 | .mode_fixup = radeon_legacy_mode_fixup, |
| 588 | .prepare = radeon_legacy_tmds_int_prepare, | 580 | .prepare = radeon_legacy_tmds_int_prepare, |
| 589 | .mode_set = radeon_legacy_tmds_int_mode_set, | 581 | .mode_set = radeon_legacy_tmds_int_mode_set, |
| 590 | .commit = radeon_legacy_tmds_int_commit, | 582 | .commit = radeon_legacy_tmds_int_commit, |
| @@ -596,17 +588,6 @@ static const struct drm_encoder_funcs radeon_legacy_tmds_int_enc_funcs = { | |||
| 596 | .destroy = radeon_enc_destroy, | 588 | .destroy = radeon_enc_destroy, |
| 597 | }; | 589 | }; |
| 598 | 590 | ||
| 599 | static bool radeon_legacy_tmds_ext_mode_fixup(struct drm_encoder *encoder, | ||
| 600 | struct drm_display_mode *mode, | ||
| 601 | struct drm_display_mode *adjusted_mode) | ||
| 602 | { | ||
| 603 | /* set the active encoder to connector routing */ | ||
| 604 | radeon_encoder_set_active_device(encoder); | ||
| 605 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
| 606 | |||
| 607 | return true; | ||
| 608 | } | ||
| 609 | |||
| 610 | static void radeon_legacy_tmds_ext_dpms(struct drm_encoder *encoder, int mode) | 591 | static void radeon_legacy_tmds_ext_dpms(struct drm_encoder *encoder, int mode) |
| 611 | { | 592 | { |
| 612 | struct drm_device *dev = encoder->dev; | 593 | struct drm_device *dev = encoder->dev; |
| @@ -697,6 +678,8 @@ static void radeon_legacy_tmds_ext_mode_set(struct drm_encoder *encoder, | |||
| 697 | /*if (mode->clock > 165000) | 678 | /*if (mode->clock > 165000) |
| 698 | fp2_gen_cntl |= R300_FP2_DVO_DUAL_CHANNEL_EN;*/ | 679 | fp2_gen_cntl |= R300_FP2_DVO_DUAL_CHANNEL_EN;*/ |
| 699 | } | 680 | } |
| 681 | if (!radeon_combios_external_tmds_setup(encoder)) | ||
| 682 | radeon_external_tmds_setup(encoder); | ||
| 700 | } | 683 | } |
| 701 | 684 | ||
| 702 | if (radeon_crtc->crtc_id == 0) { | 685 | if (radeon_crtc->crtc_id == 0) { |
| @@ -724,9 +707,22 @@ static void radeon_legacy_tmds_ext_mode_set(struct drm_encoder *encoder, | |||
| 724 | radeon_combios_encoder_crtc_scratch_regs(encoder, radeon_crtc->crtc_id); | 707 | radeon_combios_encoder_crtc_scratch_regs(encoder, radeon_crtc->crtc_id); |
| 725 | } | 708 | } |
| 726 | 709 | ||
| 710 | static void radeon_ext_tmds_enc_destroy(struct drm_encoder *encoder) | ||
| 711 | { | ||
| 712 | struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder); | ||
| 713 | struct radeon_encoder_ext_tmds *tmds = radeon_encoder->enc_priv; | ||
| 714 | if (tmds) { | ||
| 715 | if (tmds->i2c_bus) | ||
| 716 | radeon_i2c_destroy(tmds->i2c_bus); | ||
| 717 | } | ||
| 718 | kfree(radeon_encoder->enc_priv); | ||
| 719 | drm_encoder_cleanup(encoder); | ||
| 720 | kfree(radeon_encoder); | ||
| 721 | } | ||
| 722 | |||
| 727 | static const struct drm_encoder_helper_funcs radeon_legacy_tmds_ext_helper_funcs = { | 723 | static const struct drm_encoder_helper_funcs radeon_legacy_tmds_ext_helper_funcs = { |
| 728 | .dpms = radeon_legacy_tmds_ext_dpms, | 724 | .dpms = radeon_legacy_tmds_ext_dpms, |
| 729 | .mode_fixup = radeon_legacy_tmds_ext_mode_fixup, | 725 | .mode_fixup = radeon_legacy_mode_fixup, |
| 730 | .prepare = radeon_legacy_tmds_ext_prepare, | 726 | .prepare = radeon_legacy_tmds_ext_prepare, |
| 731 | .mode_set = radeon_legacy_tmds_ext_mode_set, | 727 | .mode_set = radeon_legacy_tmds_ext_mode_set, |
| 732 | .commit = radeon_legacy_tmds_ext_commit, | 728 | .commit = radeon_legacy_tmds_ext_commit, |
| @@ -735,20 +731,9 @@ static const struct drm_encoder_helper_funcs radeon_legacy_tmds_ext_helper_funcs | |||
| 735 | 731 | ||
| 736 | 732 | ||
| 737 | static const struct drm_encoder_funcs radeon_legacy_tmds_ext_enc_funcs = { | 733 | static const struct drm_encoder_funcs radeon_legacy_tmds_ext_enc_funcs = { |
| 738 | .destroy = radeon_enc_destroy, | 734 | .destroy = radeon_ext_tmds_enc_destroy, |
| 739 | }; | 735 | }; |
| 740 | 736 | ||
| 741 | static bool radeon_legacy_tv_dac_mode_fixup(struct drm_encoder *encoder, | ||
| 742 | struct drm_display_mode *mode, | ||
| 743 | struct drm_display_mode *adjusted_mode) | ||
| 744 | { | ||
| 745 | /* set the active encoder to connector routing */ | ||
| 746 | radeon_encoder_set_active_device(encoder); | ||
| 747 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
| 748 | |||
| 749 | return true; | ||
| 750 | } | ||
| 751 | |||
| 752 | static void radeon_legacy_tv_dac_dpms(struct drm_encoder *encoder, int mode) | 737 | static void radeon_legacy_tv_dac_dpms(struct drm_encoder *encoder, int mode) |
| 753 | { | 738 | { |
| 754 | struct drm_device *dev = encoder->dev; | 739 | struct drm_device *dev = encoder->dev; |
| @@ -1265,7 +1250,7 @@ static enum drm_connector_status radeon_legacy_tv_dac_detect(struct drm_encoder | |||
| 1265 | 1250 | ||
| 1266 | static const struct drm_encoder_helper_funcs radeon_legacy_tv_dac_helper_funcs = { | 1251 | static const struct drm_encoder_helper_funcs radeon_legacy_tv_dac_helper_funcs = { |
| 1267 | .dpms = radeon_legacy_tv_dac_dpms, | 1252 | .dpms = radeon_legacy_tv_dac_dpms, |
| 1268 | .mode_fixup = radeon_legacy_tv_dac_mode_fixup, | 1253 | .mode_fixup = radeon_legacy_mode_fixup, |
| 1269 | .prepare = radeon_legacy_tv_dac_prepare, | 1254 | .prepare = radeon_legacy_tv_dac_prepare, |
| 1270 | .mode_set = radeon_legacy_tv_dac_mode_set, | 1255 | .mode_set = radeon_legacy_tv_dac_mode_set, |
| 1271 | .commit = radeon_legacy_tv_dac_commit, | 1256 | .commit = radeon_legacy_tv_dac_commit, |
| @@ -1302,6 +1287,29 @@ static struct radeon_encoder_int_tmds *radeon_legacy_get_tmds_info(struct radeon | |||
| 1302 | return tmds; | 1287 | return tmds; |
| 1303 | } | 1288 | } |
| 1304 | 1289 | ||
| 1290 | static struct radeon_encoder_ext_tmds *radeon_legacy_get_ext_tmds_info(struct radeon_encoder *encoder) | ||
| 1291 | { | ||
| 1292 | struct drm_device *dev = encoder->base.dev; | ||
| 1293 | struct radeon_device *rdev = dev->dev_private; | ||
| 1294 | struct radeon_encoder_ext_tmds *tmds = NULL; | ||
| 1295 | bool ret; | ||
| 1296 | |||
| 1297 | if (rdev->is_atom_bios) | ||
| 1298 | return NULL; | ||
| 1299 | |||
| 1300 | tmds = kzalloc(sizeof(struct radeon_encoder_ext_tmds), GFP_KERNEL); | ||
| 1301 | |||
| 1302 | if (!tmds) | ||
| 1303 | return NULL; | ||
| 1304 | |||
| 1305 | ret = radeon_legacy_get_ext_tmds_info_from_combios(encoder, tmds); | ||
| 1306 | |||
| 1307 | if (ret == false) | ||
| 1308 | radeon_legacy_get_ext_tmds_info_from_table(encoder, tmds); | ||
| 1309 | |||
| 1310 | return tmds; | ||
| 1311 | } | ||
| 1312 | |||
| 1305 | void | 1313 | void |
| 1306 | radeon_add_legacy_encoder(struct drm_device *dev, uint32_t encoder_id, uint32_t supported_device) | 1314 | radeon_add_legacy_encoder(struct drm_device *dev, uint32_t encoder_id, uint32_t supported_device) |
| 1307 | { | 1315 | { |
| @@ -1329,7 +1337,6 @@ radeon_add_legacy_encoder(struct drm_device *dev, uint32_t encoder_id, uint32_t | |||
| 1329 | encoder->possible_crtcs = 0x1; | 1337 | encoder->possible_crtcs = 0x1; |
| 1330 | else | 1338 | else |
| 1331 | encoder->possible_crtcs = 0x3; | 1339 | encoder->possible_crtcs = 0x3; |
| 1332 | encoder->possible_clones = 0; | ||
| 1333 | 1340 | ||
| 1334 | radeon_encoder->enc_priv = NULL; | 1341 | radeon_encoder->enc_priv = NULL; |
| 1335 | 1342 | ||
| @@ -1373,7 +1380,7 @@ radeon_add_legacy_encoder(struct drm_device *dev, uint32_t encoder_id, uint32_t | |||
| 1373 | drm_encoder_init(dev, encoder, &radeon_legacy_tmds_ext_enc_funcs, DRM_MODE_ENCODER_TMDS); | 1380 | drm_encoder_init(dev, encoder, &radeon_legacy_tmds_ext_enc_funcs, DRM_MODE_ENCODER_TMDS); |
| 1374 | drm_encoder_helper_add(encoder, &radeon_legacy_tmds_ext_helper_funcs); | 1381 | drm_encoder_helper_add(encoder, &radeon_legacy_tmds_ext_helper_funcs); |
| 1375 | if (!rdev->is_atom_bios) | 1382 | if (!rdev->is_atom_bios) |
| 1376 | radeon_combios_get_ext_tmds_info(radeon_encoder); | 1383 | radeon_encoder->enc_priv = radeon_legacy_get_ext_tmds_info(radeon_encoder); |
| 1377 | break; | 1384 | break; |
| 1378 | } | 1385 | } |
| 1379 | } | 1386 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_mode.h b/drivers/gpu/drm/radeon/radeon_mode.h index ace726aa0d76..44d4b652ea12 100644 --- a/drivers/gpu/drm/radeon/radeon_mode.h +++ b/drivers/gpu/drm/radeon/radeon_mode.h | |||
| @@ -33,6 +33,7 @@ | |||
| 33 | #include <drm_crtc.h> | 33 | #include <drm_crtc.h> |
| 34 | #include <drm_mode.h> | 34 | #include <drm_mode.h> |
| 35 | #include <drm_edid.h> | 35 | #include <drm_edid.h> |
| 36 | #include <drm_dp_helper.h> | ||
| 36 | #include <linux/i2c.h> | 37 | #include <linux/i2c.h> |
| 37 | #include <linux/i2c-id.h> | 38 | #include <linux/i2c-id.h> |
| 38 | #include <linux/i2c-algo-bit.h> | 39 | #include <linux/i2c-algo-bit.h> |
| @@ -89,24 +90,45 @@ enum radeon_tv_std { | |||
| 89 | TV_STD_PAL_CN, | 90 | TV_STD_PAL_CN, |
| 90 | }; | 91 | }; |
| 91 | 92 | ||
| 93 | /* radeon gpio-based i2c | ||
| 94 | * 1. "mask" reg and bits | ||
| 95 | * grabs the gpio pins for software use | ||
| 96 | * 0=not held 1=held | ||
| 97 | * 2. "a" reg and bits | ||
| 98 | * output pin value | ||
| 99 | * 0=low 1=high | ||
| 100 | * 3. "en" reg and bits | ||
| 101 | * sets the pin direction | ||
| 102 | * 0=input 1=output | ||
| 103 | * 4. "y" reg and bits | ||
| 104 | * input pin value | ||
| 105 | * 0=low 1=high | ||
| 106 | */ | ||
| 92 | struct radeon_i2c_bus_rec { | 107 | struct radeon_i2c_bus_rec { |
| 93 | bool valid; | 108 | bool valid; |
| 109 | /* id used by atom */ | ||
| 110 | uint8_t i2c_id; | ||
| 111 | /* can be used with hw i2c engine */ | ||
| 112 | bool hw_capable; | ||
| 113 | /* uses multi-media i2c engine */ | ||
| 114 | bool mm_i2c; | ||
| 115 | /* regs and bits */ | ||
| 94 | uint32_t mask_clk_reg; | 116 | uint32_t mask_clk_reg; |
| 95 | uint32_t mask_data_reg; | 117 | uint32_t mask_data_reg; |
| 96 | uint32_t a_clk_reg; | 118 | uint32_t a_clk_reg; |
| 97 | uint32_t a_data_reg; | 119 | uint32_t a_data_reg; |
| 98 | uint32_t put_clk_reg; | 120 | uint32_t en_clk_reg; |
| 99 | uint32_t put_data_reg; | 121 | uint32_t en_data_reg; |
| 100 | uint32_t get_clk_reg; | 122 | uint32_t y_clk_reg; |
| 101 | uint32_t get_data_reg; | 123 | uint32_t y_data_reg; |
| 102 | uint32_t mask_clk_mask; | 124 | uint32_t mask_clk_mask; |
| 103 | uint32_t mask_data_mask; | 125 | uint32_t mask_data_mask; |
| 104 | uint32_t put_clk_mask; | ||
| 105 | uint32_t put_data_mask; | ||
| 106 | uint32_t get_clk_mask; | ||
| 107 | uint32_t get_data_mask; | ||
| 108 | uint32_t a_clk_mask; | 126 | uint32_t a_clk_mask; |
| 109 | uint32_t a_data_mask; | 127 | uint32_t a_data_mask; |
| 128 | uint32_t en_clk_mask; | ||
| 129 | uint32_t en_data_mask; | ||
| 130 | uint32_t y_clk_mask; | ||
| 131 | uint32_t y_data_mask; | ||
| 110 | }; | 132 | }; |
| 111 | 133 | ||
| 112 | struct radeon_tmds_pll { | 134 | struct radeon_tmds_pll { |
| @@ -150,9 +172,12 @@ struct radeon_pll { | |||
| 150 | }; | 172 | }; |
| 151 | 173 | ||
| 152 | struct radeon_i2c_chan { | 174 | struct radeon_i2c_chan { |
| 153 | struct drm_device *dev; | ||
| 154 | struct i2c_adapter adapter; | 175 | struct i2c_adapter adapter; |
| 155 | struct i2c_algo_bit_data algo; | 176 | struct drm_device *dev; |
| 177 | union { | ||
| 178 | struct i2c_algo_dp_aux_data dp; | ||
| 179 | struct i2c_algo_bit_data bit; | ||
| 180 | } algo; | ||
| 156 | struct radeon_i2c_bus_rec rec; | 181 | struct radeon_i2c_bus_rec rec; |
| 157 | }; | 182 | }; |
| 158 | 183 | ||
| @@ -170,6 +195,11 @@ enum radeon_connector_table { | |||
| 170 | CT_EMAC, | 195 | CT_EMAC, |
| 171 | }; | 196 | }; |
| 172 | 197 | ||
| 198 | enum radeon_dvo_chip { | ||
| 199 | DVO_SIL164, | ||
| 200 | DVO_SIL1178, | ||
| 201 | }; | ||
| 202 | |||
| 173 | struct radeon_mode_info { | 203 | struct radeon_mode_info { |
| 174 | struct atom_context *atom_context; | 204 | struct atom_context *atom_context; |
| 175 | struct card_info *atom_card_info; | 205 | struct card_info *atom_card_info; |
| @@ -261,6 +291,13 @@ struct radeon_encoder_int_tmds { | |||
| 261 | struct radeon_tmds_pll tmds_pll[4]; | 291 | struct radeon_tmds_pll tmds_pll[4]; |
| 262 | }; | 292 | }; |
| 263 | 293 | ||
| 294 | struct radeon_encoder_ext_tmds { | ||
| 295 | /* tmds over dvo */ | ||
| 296 | struct radeon_i2c_chan *i2c_bus; | ||
| 297 | uint8_t slave_addr; | ||
| 298 | enum radeon_dvo_chip dvo_chip; | ||
| 299 | }; | ||
| 300 | |||
| 264 | /* spread spectrum */ | 301 | /* spread spectrum */ |
| 265 | struct radeon_atom_ss { | 302 | struct radeon_atom_ss { |
| 266 | uint16_t percentage; | 303 | uint16_t percentage; |
| @@ -302,6 +339,35 @@ struct radeon_encoder { | |||
| 302 | struct radeon_connector_atom_dig { | 339 | struct radeon_connector_atom_dig { |
| 303 | uint32_t igp_lane_info; | 340 | uint32_t igp_lane_info; |
| 304 | bool linkb; | 341 | bool linkb; |
| 342 | /* displayport */ | ||
| 343 | struct radeon_i2c_chan *dp_i2c_bus; | ||
| 344 | u8 dpcd[8]; | ||
| 345 | u8 dp_sink_type; | ||
| 346 | int dp_clock; | ||
| 347 | int dp_lane_count; | ||
| 348 | }; | ||
| 349 | |||
| 350 | struct radeon_gpio_rec { | ||
| 351 | bool valid; | ||
| 352 | u8 id; | ||
| 353 | u32 reg; | ||
| 354 | u32 mask; | ||
| 355 | }; | ||
| 356 | |||
| 357 | enum radeon_hpd_id { | ||
| 358 | RADEON_HPD_NONE = 0, | ||
| 359 | RADEON_HPD_1, | ||
| 360 | RADEON_HPD_2, | ||
| 361 | RADEON_HPD_3, | ||
| 362 | RADEON_HPD_4, | ||
| 363 | RADEON_HPD_5, | ||
| 364 | RADEON_HPD_6, | ||
| 365 | }; | ||
| 366 | |||
| 367 | struct radeon_hpd { | ||
| 368 | enum radeon_hpd_id hpd; | ||
| 369 | u8 plugged_state; | ||
| 370 | struct radeon_gpio_rec gpio; | ||
| 305 | }; | 371 | }; |
| 306 | 372 | ||
| 307 | struct radeon_connector { | 373 | struct radeon_connector { |
| @@ -318,6 +384,7 @@ struct radeon_connector { | |||
| 318 | void *con_priv; | 384 | void *con_priv; |
| 319 | bool dac_load_detect; | 385 | bool dac_load_detect; |
| 320 | uint16_t connector_object_id; | 386 | uint16_t connector_object_id; |
| 387 | struct radeon_hpd hpd; | ||
| 321 | }; | 388 | }; |
| 322 | 389 | ||
| 323 | struct radeon_framebuffer { | 390 | struct radeon_framebuffer { |
| @@ -325,10 +392,37 @@ struct radeon_framebuffer { | |||
| 325 | struct drm_gem_object *obj; | 392 | struct drm_gem_object *obj; |
| 326 | }; | 393 | }; |
| 327 | 394 | ||
| 395 | extern void radeon_connector_hotplug(struct drm_connector *connector); | ||
| 396 | extern bool radeon_dp_needs_link_train(struct radeon_connector *radeon_connector); | ||
| 397 | extern int radeon_dp_mode_valid_helper(struct radeon_connector *radeon_connector, | ||
| 398 | struct drm_display_mode *mode); | ||
| 399 | extern void radeon_dp_set_link_config(struct drm_connector *connector, | ||
| 400 | struct drm_display_mode *mode); | ||
| 401 | extern void dp_link_train(struct drm_encoder *encoder, | ||
| 402 | struct drm_connector *connector); | ||
| 403 | extern u8 radeon_dp_getsinktype(struct radeon_connector *radeon_connector); | ||
| 404 | extern bool radeon_dp_getdpcd(struct radeon_connector *radeon_connector); | ||
| 405 | extern void atombios_dig_transmitter_setup(struct drm_encoder *encoder, | ||
| 406 | int action, uint8_t lane_num, | ||
| 407 | uint8_t lane_set); | ||
| 408 | extern int radeon_dp_i2c_aux_ch(struct i2c_adapter *adapter, int mode, | ||
| 409 | uint8_t write_byte, uint8_t *read_byte); | ||
| 410 | |||
| 411 | extern struct radeon_i2c_chan *radeon_i2c_create_dp(struct drm_device *dev, | ||
| 412 | struct radeon_i2c_bus_rec *rec, | ||
| 413 | const char *name); | ||
| 328 | extern struct radeon_i2c_chan *radeon_i2c_create(struct drm_device *dev, | 414 | extern struct radeon_i2c_chan *radeon_i2c_create(struct drm_device *dev, |
| 329 | struct radeon_i2c_bus_rec *rec, | 415 | struct radeon_i2c_bus_rec *rec, |
| 330 | const char *name); | 416 | const char *name); |
| 331 | extern void radeon_i2c_destroy(struct radeon_i2c_chan *i2c); | 417 | extern void radeon_i2c_destroy(struct radeon_i2c_chan *i2c); |
| 418 | extern void radeon_i2c_sw_get_byte(struct radeon_i2c_chan *i2c_bus, | ||
| 419 | u8 slave_addr, | ||
| 420 | u8 addr, | ||
| 421 | u8 *val); | ||
| 422 | extern void radeon_i2c_sw_put_byte(struct radeon_i2c_chan *i2c, | ||
| 423 | u8 slave_addr, | ||
| 424 | u8 addr, | ||
| 425 | u8 val); | ||
| 332 | extern bool radeon_ddc_probe(struct radeon_connector *radeon_connector); | 426 | extern bool radeon_ddc_probe(struct radeon_connector *radeon_connector); |
| 333 | extern int radeon_ddc_get_modes(struct radeon_connector *radeon_connector); | 427 | extern int radeon_ddc_get_modes(struct radeon_connector *radeon_connector); |
| 334 | 428 | ||
| @@ -343,12 +437,24 @@ extern void radeon_compute_pll(struct radeon_pll *pll, | |||
| 343 | uint32_t *post_div_p, | 437 | uint32_t *post_div_p, |
| 344 | int flags); | 438 | int flags); |
| 345 | 439 | ||
| 440 | extern void radeon_compute_pll_avivo(struct radeon_pll *pll, | ||
| 441 | uint64_t freq, | ||
| 442 | uint32_t *dot_clock_p, | ||
| 443 | uint32_t *fb_div_p, | ||
| 444 | uint32_t *frac_fb_div_p, | ||
| 445 | uint32_t *ref_div_p, | ||
| 446 | uint32_t *post_div_p, | ||
| 447 | int flags); | ||
| 448 | |||
| 449 | extern void radeon_setup_encoder_clones(struct drm_device *dev); | ||
| 450 | |||
| 346 | struct drm_encoder *radeon_encoder_legacy_lvds_add(struct drm_device *dev, int bios_index); | 451 | struct drm_encoder *radeon_encoder_legacy_lvds_add(struct drm_device *dev, int bios_index); |
| 347 | struct drm_encoder *radeon_encoder_legacy_primary_dac_add(struct drm_device *dev, int bios_index, int with_tv); | 452 | struct drm_encoder *radeon_encoder_legacy_primary_dac_add(struct drm_device *dev, int bios_index, int with_tv); |
| 348 | struct drm_encoder *radeon_encoder_legacy_tv_dac_add(struct drm_device *dev, int bios_index, int with_tv); | 453 | struct drm_encoder *radeon_encoder_legacy_tv_dac_add(struct drm_device *dev, int bios_index, int with_tv); |
| 349 | struct drm_encoder *radeon_encoder_legacy_tmds_int_add(struct drm_device *dev, int bios_index); | 454 | struct drm_encoder *radeon_encoder_legacy_tmds_int_add(struct drm_device *dev, int bios_index); |
| 350 | struct drm_encoder *radeon_encoder_legacy_tmds_ext_add(struct drm_device *dev, int bios_index); | 455 | struct drm_encoder *radeon_encoder_legacy_tmds_ext_add(struct drm_device *dev, int bios_index); |
| 351 | extern void atombios_external_tmds_setup(struct drm_encoder *encoder, int action); | 456 | extern void atombios_external_tmds_setup(struct drm_encoder *encoder, int action); |
| 457 | extern void atombios_digital_setup(struct drm_encoder *encoder, int action); | ||
| 352 | extern int atombios_get_encoder_mode(struct drm_encoder *encoder); | 458 | extern int atombios_get_encoder_mode(struct drm_encoder *encoder); |
| 353 | extern void radeon_encoder_set_active_device(struct drm_encoder *encoder); | 459 | extern void radeon_encoder_set_active_device(struct drm_encoder *encoder); |
| 354 | 460 | ||
| @@ -378,12 +484,16 @@ extern bool radeon_atom_get_clock_info(struct drm_device *dev); | |||
| 378 | extern bool radeon_combios_get_clock_info(struct drm_device *dev); | 484 | extern bool radeon_combios_get_clock_info(struct drm_device *dev); |
| 379 | extern struct radeon_encoder_atom_dig * | 485 | extern struct radeon_encoder_atom_dig * |
| 380 | radeon_atombios_get_lvds_info(struct radeon_encoder *encoder); | 486 | radeon_atombios_get_lvds_info(struct radeon_encoder *encoder); |
| 381 | bool radeon_atombios_get_tmds_info(struct radeon_encoder *encoder, | 487 | extern bool radeon_atombios_get_tmds_info(struct radeon_encoder *encoder, |
| 382 | struct radeon_encoder_int_tmds *tmds); | 488 | struct radeon_encoder_int_tmds *tmds); |
| 383 | bool radeon_legacy_get_tmds_info_from_combios(struct radeon_encoder *encoder, | 489 | extern bool radeon_legacy_get_tmds_info_from_combios(struct radeon_encoder *encoder, |
| 384 | struct radeon_encoder_int_tmds *tmds); | 490 | struct radeon_encoder_int_tmds *tmds); |
| 385 | bool radeon_legacy_get_tmds_info_from_table(struct radeon_encoder *encoder, | 491 | extern bool radeon_legacy_get_tmds_info_from_table(struct radeon_encoder *encoder, |
| 386 | struct radeon_encoder_int_tmds *tmds); | 492 | struct radeon_encoder_int_tmds *tmds); |
| 493 | extern bool radeon_legacy_get_ext_tmds_info_from_combios(struct radeon_encoder *encoder, | ||
| 494 | struct radeon_encoder_ext_tmds *tmds); | ||
| 495 | extern bool radeon_legacy_get_ext_tmds_info_from_table(struct radeon_encoder *encoder, | ||
| 496 | struct radeon_encoder_ext_tmds *tmds); | ||
| 387 | extern struct radeon_encoder_primary_dac * | 497 | extern struct radeon_encoder_primary_dac * |
| 388 | radeon_atombios_get_primary_dac_info(struct radeon_encoder *encoder); | 498 | radeon_atombios_get_primary_dac_info(struct radeon_encoder *encoder); |
| 389 | extern struct radeon_encoder_tv_dac * | 499 | extern struct radeon_encoder_tv_dac * |
| @@ -395,6 +505,8 @@ extern struct radeon_encoder_tv_dac * | |||
| 395 | radeon_combios_get_tv_dac_info(struct radeon_encoder *encoder); | 505 | radeon_combios_get_tv_dac_info(struct radeon_encoder *encoder); |
| 396 | extern struct radeon_encoder_primary_dac * | 506 | extern struct radeon_encoder_primary_dac * |
| 397 | radeon_combios_get_primary_dac_info(struct radeon_encoder *encoder); | 507 | radeon_combios_get_primary_dac_info(struct radeon_encoder *encoder); |
| 508 | extern bool radeon_combios_external_tmds_setup(struct drm_encoder *encoder); | ||
| 509 | extern void radeon_external_tmds_setup(struct drm_encoder *encoder); | ||
| 398 | extern void radeon_combios_output_lock(struct drm_encoder *encoder, bool lock); | 510 | extern void radeon_combios_output_lock(struct drm_encoder *encoder, bool lock); |
| 399 | extern void radeon_combios_initialize_bios_scratch_regs(struct drm_device *dev); | 511 | extern void radeon_combios_initialize_bios_scratch_regs(struct drm_device *dev); |
| 400 | extern void radeon_atom_output_lock(struct drm_encoder *encoder, bool lock); | 512 | extern void radeon_atom_output_lock(struct drm_encoder *encoder, bool lock); |
| @@ -426,16 +538,13 @@ void radeon_atombios_init_crtc(struct drm_device *dev, | |||
| 426 | struct radeon_crtc *radeon_crtc); | 538 | struct radeon_crtc *radeon_crtc); |
| 427 | void radeon_legacy_init_crtc(struct drm_device *dev, | 539 | void radeon_legacy_init_crtc(struct drm_device *dev, |
| 428 | struct radeon_crtc *radeon_crtc); | 540 | struct radeon_crtc *radeon_crtc); |
| 429 | void radeon_i2c_do_lock(struct radeon_connector *radeon_connector, int lock_state); | 541 | extern void radeon_i2c_do_lock(struct radeon_i2c_chan *i2c, int lock_state); |
| 430 | 542 | ||
| 431 | void radeon_get_clock_info(struct drm_device *dev); | 543 | void radeon_get_clock_info(struct drm_device *dev); |
| 432 | 544 | ||
| 433 | extern bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev); | 545 | extern bool radeon_get_atom_connector_info_from_object_table(struct drm_device *dev); |
| 434 | extern bool radeon_get_atom_connector_info_from_supported_devices_table(struct drm_device *dev); | 546 | extern bool radeon_get_atom_connector_info_from_supported_devices_table(struct drm_device *dev); |
| 435 | 547 | ||
| 436 | void radeon_rmx_mode_fixup(struct drm_encoder *encoder, | ||
| 437 | struct drm_display_mode *mode, | ||
| 438 | struct drm_display_mode *adjusted_mode); | ||
| 439 | void radeon_enc_destroy(struct drm_encoder *encoder); | 548 | void radeon_enc_destroy(struct drm_encoder *encoder); |
| 440 | void radeon_copy_fb(struct drm_device *dev, struct drm_gem_object *dst_obj); | 549 | void radeon_copy_fb(struct drm_device *dev, struct drm_gem_object *dst_obj); |
| 441 | void radeon_combios_asic_init(struct drm_device *dev); | 550 | void radeon_combios_asic_init(struct drm_device *dev); |
diff --git a/drivers/gpu/drm/radeon/radeon_object.c b/drivers/gpu/drm/radeon/radeon_object.c index 1f056dadc5c2..2040937682fd 100644 --- a/drivers/gpu/drm/radeon/radeon_object.c +++ b/drivers/gpu/drm/radeon/radeon_object.c | |||
| @@ -34,74 +34,32 @@ | |||
| 34 | #include "radeon_drm.h" | 34 | #include "radeon_drm.h" |
| 35 | #include "radeon.h" | 35 | #include "radeon.h" |
| 36 | 36 | ||
| 37 | struct radeon_object { | ||
| 38 | struct ttm_buffer_object tobj; | ||
| 39 | struct list_head list; | ||
| 40 | struct radeon_device *rdev; | ||
| 41 | struct drm_gem_object *gobj; | ||
| 42 | struct ttm_bo_kmap_obj kmap; | ||
| 43 | unsigned pin_count; | ||
| 44 | uint64_t gpu_addr; | ||
| 45 | void *kptr; | ||
| 46 | bool is_iomem; | ||
| 47 | uint32_t tiling_flags; | ||
| 48 | uint32_t pitch; | ||
| 49 | int surface_reg; | ||
| 50 | }; | ||
| 51 | 37 | ||
| 52 | int radeon_ttm_init(struct radeon_device *rdev); | 38 | int radeon_ttm_init(struct radeon_device *rdev); |
| 53 | void radeon_ttm_fini(struct radeon_device *rdev); | 39 | void radeon_ttm_fini(struct radeon_device *rdev); |
| 40 | static void radeon_bo_clear_surface_reg(struct radeon_bo *bo); | ||
| 54 | 41 | ||
| 55 | /* | 42 | /* |
| 56 | * To exclude mutual BO access we rely on bo_reserve exclusion, as all | 43 | * To exclude mutual BO access we rely on bo_reserve exclusion, as all |
| 57 | * function are calling it. | 44 | * function are calling it. |
| 58 | */ | 45 | */ |
| 59 | 46 | ||
| 60 | static int radeon_object_reserve(struct radeon_object *robj, bool interruptible) | 47 | static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo) |
| 61 | { | 48 | { |
| 62 | return ttm_bo_reserve(&robj->tobj, interruptible, false, false, 0); | 49 | struct radeon_bo *bo; |
| 63 | } | ||
| 64 | |||
| 65 | static void radeon_object_unreserve(struct radeon_object *robj) | ||
| 66 | { | ||
| 67 | ttm_bo_unreserve(&robj->tobj); | ||
| 68 | } | ||
| 69 | |||
| 70 | static void radeon_ttm_object_object_destroy(struct ttm_buffer_object *tobj) | ||
| 71 | { | ||
| 72 | struct radeon_object *robj; | ||
| 73 | 50 | ||
| 74 | robj = container_of(tobj, struct radeon_object, tobj); | 51 | bo = container_of(tbo, struct radeon_bo, tbo); |
| 75 | list_del_init(&robj->list); | 52 | mutex_lock(&bo->rdev->gem.mutex); |
| 76 | radeon_object_clear_surface_reg(robj); | 53 | list_del_init(&bo->list); |
| 77 | kfree(robj); | 54 | mutex_unlock(&bo->rdev->gem.mutex); |
| 55 | radeon_bo_clear_surface_reg(bo); | ||
| 56 | kfree(bo); | ||
| 78 | } | 57 | } |
| 79 | 58 | ||
| 80 | static inline void radeon_object_gpu_addr(struct radeon_object *robj) | 59 | static inline u32 radeon_ttm_flags_from_domain(u32 domain) |
| 81 | { | 60 | { |
| 82 | /* Default gpu address */ | 61 | u32 flags = 0; |
| 83 | robj->gpu_addr = 0xFFFFFFFFFFFFFFFFULL; | ||
| 84 | if (robj->tobj.mem.mm_node == NULL) { | ||
| 85 | return; | ||
| 86 | } | ||
| 87 | robj->gpu_addr = ((u64)robj->tobj.mem.mm_node->start) << PAGE_SHIFT; | ||
| 88 | switch (robj->tobj.mem.mem_type) { | ||
| 89 | case TTM_PL_VRAM: | ||
| 90 | robj->gpu_addr += (u64)robj->rdev->mc.vram_location; | ||
| 91 | break; | ||
| 92 | case TTM_PL_TT: | ||
| 93 | robj->gpu_addr += (u64)robj->rdev->mc.gtt_location; | ||
| 94 | break; | ||
| 95 | default: | ||
| 96 | DRM_ERROR("Unknown placement %d\n", robj->tobj.mem.mem_type); | ||
| 97 | robj->gpu_addr = 0xFFFFFFFFFFFFFFFFULL; | ||
| 98 | return; | ||
| 99 | } | ||
| 100 | } | ||
| 101 | 62 | ||
| 102 | static inline uint32_t radeon_object_flags_from_domain(uint32_t domain) | ||
| 103 | { | ||
| 104 | uint32_t flags = 0; | ||
| 105 | if (domain & RADEON_GEM_DOMAIN_VRAM) { | 63 | if (domain & RADEON_GEM_DOMAIN_VRAM) { |
| 106 | flags |= TTM_PL_FLAG_VRAM | TTM_PL_FLAG_WC | TTM_PL_FLAG_UNCACHED; | 64 | flags |= TTM_PL_FLAG_VRAM | TTM_PL_FLAG_WC | TTM_PL_FLAG_UNCACHED; |
| 107 | } | 65 | } |
| @@ -117,17 +75,32 @@ static inline uint32_t radeon_object_flags_from_domain(uint32_t domain) | |||
| 117 | return flags; | 75 | return flags; |
| 118 | } | 76 | } |
| 119 | 77 | ||
| 120 | int radeon_object_create(struct radeon_device *rdev, | 78 | void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain) |
| 121 | struct drm_gem_object *gobj, | 79 | { |
| 122 | unsigned long size, | 80 | u32 c = 0; |
| 123 | bool kernel, | 81 | |
| 124 | uint32_t domain, | 82 | rbo->placement.fpfn = 0; |
| 125 | bool interruptible, | 83 | rbo->placement.lpfn = 0; |
| 126 | struct radeon_object **robj_ptr) | 84 | rbo->placement.placement = rbo->placements; |
| 85 | rbo->placement.busy_placement = rbo->placements; | ||
| 86 | if (domain & RADEON_GEM_DOMAIN_VRAM) | ||
| 87 | rbo->placements[c++] = TTM_PL_FLAG_WC | TTM_PL_FLAG_UNCACHED | | ||
| 88 | TTM_PL_FLAG_VRAM; | ||
| 89 | if (domain & RADEON_GEM_DOMAIN_GTT) | ||
| 90 | rbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_TT; | ||
| 91 | if (domain & RADEON_GEM_DOMAIN_CPU) | ||
| 92 | rbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_SYSTEM; | ||
| 93 | rbo->placement.num_placement = c; | ||
| 94 | rbo->placement.num_busy_placement = c; | ||
| 95 | } | ||
| 96 | |||
| 97 | int radeon_bo_create(struct radeon_device *rdev, struct drm_gem_object *gobj, | ||
| 98 | unsigned long size, bool kernel, u32 domain, | ||
| 99 | struct radeon_bo **bo_ptr) | ||
| 127 | { | 100 | { |
| 128 | struct radeon_object *robj; | 101 | struct radeon_bo *bo; |
| 129 | enum ttm_bo_type type; | 102 | enum ttm_bo_type type; |
| 130 | uint32_t flags; | 103 | u32 flags; |
| 131 | int r; | 104 | int r; |
| 132 | 105 | ||
| 133 | if (unlikely(rdev->mman.bdev.dev_mapping == NULL)) { | 106 | if (unlikely(rdev->mman.bdev.dev_mapping == NULL)) { |
| @@ -138,206 +111,125 @@ int radeon_object_create(struct radeon_device *rdev, | |||
| 138 | } else { | 111 | } else { |
| 139 | type = ttm_bo_type_device; | 112 | type = ttm_bo_type_device; |
| 140 | } | 113 | } |
| 141 | *robj_ptr = NULL; | 114 | *bo_ptr = NULL; |
| 142 | robj = kzalloc(sizeof(struct radeon_object), GFP_KERNEL); | 115 | bo = kzalloc(sizeof(struct radeon_bo), GFP_KERNEL); |
| 143 | if (robj == NULL) { | 116 | if (bo == NULL) |
| 144 | return -ENOMEM; | 117 | return -ENOMEM; |
| 145 | } | 118 | bo->rdev = rdev; |
| 146 | robj->rdev = rdev; | 119 | bo->gobj = gobj; |
| 147 | robj->gobj = gobj; | 120 | bo->surface_reg = -1; |
| 148 | robj->surface_reg = -1; | 121 | INIT_LIST_HEAD(&bo->list); |
| 149 | INIT_LIST_HEAD(&robj->list); | 122 | |
| 150 | 123 | flags = radeon_ttm_flags_from_domain(domain); | |
| 151 | flags = radeon_object_flags_from_domain(domain); | 124 | /* Kernel allocation are uninterruptible */ |
| 152 | r = ttm_buffer_object_init(&rdev->mman.bdev, &robj->tobj, size, type, flags, | 125 | r = ttm_buffer_object_init(&rdev->mman.bdev, &bo->tbo, size, type, |
| 153 | 0, 0, false, NULL, size, | 126 | flags, 0, 0, !kernel, NULL, size, |
| 154 | &radeon_ttm_object_object_destroy); | 127 | &radeon_ttm_bo_destroy); |
| 155 | if (unlikely(r != 0)) { | 128 | if (unlikely(r != 0)) { |
| 156 | /* ttm call radeon_ttm_object_object_destroy if error happen */ | 129 | if (r != -ERESTARTSYS) |
| 157 | DRM_ERROR("Failed to allocate TTM object (%ld, 0x%08X, %u)\n", | 130 | dev_err(rdev->dev, |
| 158 | size, flags, 0); | 131 | "object_init failed for (%ld, 0x%08X)\n", |
| 132 | size, flags); | ||
| 159 | return r; | 133 | return r; |
| 160 | } | 134 | } |
| 161 | *robj_ptr = robj; | 135 | *bo_ptr = bo; |
| 162 | if (gobj) { | 136 | if (gobj) { |
| 163 | list_add_tail(&robj->list, &rdev->gem.objects); | 137 | mutex_lock(&bo->rdev->gem.mutex); |
| 138 | list_add_tail(&bo->list, &rdev->gem.objects); | ||
| 139 | mutex_unlock(&bo->rdev->gem.mutex); | ||
| 164 | } | 140 | } |
| 165 | return 0; | 141 | return 0; |
| 166 | } | 142 | } |
| 167 | 143 | ||
| 168 | int radeon_object_kmap(struct radeon_object *robj, void **ptr) | 144 | int radeon_bo_kmap(struct radeon_bo *bo, void **ptr) |
| 169 | { | 145 | { |
| 146 | bool is_iomem; | ||
| 170 | int r; | 147 | int r; |
| 171 | 148 | ||
| 172 | spin_lock(&robj->tobj.lock); | 149 | if (bo->kptr) { |
| 173 | if (robj->kptr) { | ||
| 174 | if (ptr) { | 150 | if (ptr) { |
| 175 | *ptr = robj->kptr; | 151 | *ptr = bo->kptr; |
| 176 | } | 152 | } |
| 177 | spin_unlock(&robj->tobj.lock); | ||
| 178 | return 0; | 153 | return 0; |
| 179 | } | 154 | } |
| 180 | spin_unlock(&robj->tobj.lock); | 155 | r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); |
| 181 | r = ttm_bo_kmap(&robj->tobj, 0, robj->tobj.num_pages, &robj->kmap); | ||
| 182 | if (r) { | 156 | if (r) { |
| 183 | return r; | 157 | return r; |
| 184 | } | 158 | } |
| 185 | spin_lock(&robj->tobj.lock); | 159 | bo->kptr = ttm_kmap_obj_virtual(&bo->kmap, &is_iomem); |
| 186 | robj->kptr = ttm_kmap_obj_virtual(&robj->kmap, &robj->is_iomem); | ||
| 187 | spin_unlock(&robj->tobj.lock); | ||
| 188 | if (ptr) { | 160 | if (ptr) { |
| 189 | *ptr = robj->kptr; | 161 | *ptr = bo->kptr; |
| 190 | } | 162 | } |
| 191 | radeon_object_check_tiling(robj, 0, 0); | 163 | radeon_bo_check_tiling(bo, 0, 0); |
| 192 | return 0; | 164 | return 0; |
| 193 | } | 165 | } |
| 194 | 166 | ||
| 195 | void radeon_object_kunmap(struct radeon_object *robj) | 167 | void radeon_bo_kunmap(struct radeon_bo *bo) |
| 196 | { | 168 | { |
| 197 | spin_lock(&robj->tobj.lock); | 169 | if (bo->kptr == NULL) |
| 198 | if (robj->kptr == NULL) { | ||
| 199 | spin_unlock(&robj->tobj.lock); | ||
| 200 | return; | 170 | return; |
| 201 | } | 171 | bo->kptr = NULL; |
| 202 | robj->kptr = NULL; | 172 | radeon_bo_check_tiling(bo, 0, 0); |
| 203 | spin_unlock(&robj->tobj.lock); | 173 | ttm_bo_kunmap(&bo->kmap); |
| 204 | radeon_object_check_tiling(robj, 0, 0); | ||
| 205 | ttm_bo_kunmap(&robj->kmap); | ||
| 206 | } | 174 | } |
| 207 | 175 | ||
| 208 | void radeon_object_unref(struct radeon_object **robj) | 176 | void radeon_bo_unref(struct radeon_bo **bo) |
| 209 | { | 177 | { |
| 210 | struct ttm_buffer_object *tobj; | 178 | struct ttm_buffer_object *tbo; |
| 211 | 179 | ||
| 212 | if ((*robj) == NULL) { | 180 | if ((*bo) == NULL) |
| 213 | return; | 181 | return; |
| 214 | } | 182 | tbo = &((*bo)->tbo); |
| 215 | tobj = &((*robj)->tobj); | 183 | ttm_bo_unref(&tbo); |
| 216 | ttm_bo_unref(&tobj); | 184 | if (tbo == NULL) |
| 217 | if (tobj == NULL) { | 185 | *bo = NULL; |
| 218 | *robj = NULL; | ||
| 219 | } | ||
| 220 | } | ||
| 221 | |||
| 222 | int radeon_object_mmap(struct radeon_object *robj, uint64_t *offset) | ||
| 223 | { | ||
| 224 | *offset = robj->tobj.addr_space_offset; | ||
| 225 | return 0; | ||
| 226 | } | 186 | } |
| 227 | 187 | ||
| 228 | int radeon_object_pin(struct radeon_object *robj, uint32_t domain, | 188 | int radeon_bo_pin(struct radeon_bo *bo, u32 domain, u64 *gpu_addr) |
| 229 | uint64_t *gpu_addr) | ||
| 230 | { | 189 | { |
| 231 | uint32_t flags; | 190 | int r, i; |
| 232 | uint32_t tmp; | ||
| 233 | int r; | ||
| 234 | 191 | ||
| 235 | flags = radeon_object_flags_from_domain(domain); | 192 | radeon_ttm_placement_from_domain(bo, domain); |
| 236 | spin_lock(&robj->tobj.lock); | 193 | if (bo->pin_count) { |
| 237 | if (robj->pin_count) { | 194 | bo->pin_count++; |
| 238 | robj->pin_count++; | 195 | if (gpu_addr) |
| 239 | if (gpu_addr != NULL) { | 196 | *gpu_addr = radeon_bo_gpu_offset(bo); |
| 240 | *gpu_addr = robj->gpu_addr; | ||
| 241 | } | ||
| 242 | spin_unlock(&robj->tobj.lock); | ||
| 243 | return 0; | 197 | return 0; |
| 244 | } | 198 | } |
| 245 | spin_unlock(&robj->tobj.lock); | 199 | radeon_ttm_placement_from_domain(bo, domain); |
| 246 | r = radeon_object_reserve(robj, false); | 200 | for (i = 0; i < bo->placement.num_placement; i++) |
| 247 | if (unlikely(r != 0)) { | 201 | bo->placements[i] |= TTM_PL_FLAG_NO_EVICT; |
| 248 | DRM_ERROR("radeon: failed to reserve object for pinning it.\n"); | 202 | r = ttm_buffer_object_validate(&bo->tbo, &bo->placement, false, false); |
| 249 | return r; | 203 | if (likely(r == 0)) { |
| 250 | } | 204 | bo->pin_count = 1; |
| 251 | tmp = robj->tobj.mem.placement; | 205 | if (gpu_addr != NULL) |
| 252 | ttm_flag_masked(&tmp, flags, TTM_PL_MASK_MEM); | 206 | *gpu_addr = radeon_bo_gpu_offset(bo); |
| 253 | robj->tobj.proposed_placement = tmp | TTM_PL_FLAG_NO_EVICT | TTM_PL_MASK_CACHING; | 207 | } |
| 254 | r = ttm_buffer_object_validate(&robj->tobj, | 208 | if (unlikely(r != 0)) |
| 255 | robj->tobj.proposed_placement, | 209 | dev_err(bo->rdev->dev, "%p pin failed\n", bo); |
| 256 | false, false); | ||
| 257 | radeon_object_gpu_addr(robj); | ||
| 258 | if (gpu_addr != NULL) { | ||
| 259 | *gpu_addr = robj->gpu_addr; | ||
| 260 | } | ||
| 261 | robj->pin_count = 1; | ||
| 262 | if (unlikely(r != 0)) { | ||
| 263 | DRM_ERROR("radeon: failed to pin object.\n"); | ||
| 264 | } | ||
| 265 | radeon_object_unreserve(robj); | ||
| 266 | return r; | 210 | return r; |
| 267 | } | 211 | } |
| 268 | 212 | ||
| 269 | void radeon_object_unpin(struct radeon_object *robj) | 213 | int radeon_bo_unpin(struct radeon_bo *bo) |
| 270 | { | 214 | { |
| 271 | uint32_t flags; | 215 | int r, i; |
| 272 | int r; | ||
| 273 | 216 | ||
| 274 | spin_lock(&robj->tobj.lock); | 217 | if (!bo->pin_count) { |
| 275 | if (!robj->pin_count) { | 218 | dev_warn(bo->rdev->dev, "%p unpin not necessary\n", bo); |
| 276 | spin_unlock(&robj->tobj.lock); | 219 | return 0; |
| 277 | printk(KERN_WARNING "Unpin not necessary for %p !\n", robj); | ||
| 278 | return; | ||
| 279 | } | ||
| 280 | robj->pin_count--; | ||
| 281 | if (robj->pin_count) { | ||
| 282 | spin_unlock(&robj->tobj.lock); | ||
| 283 | return; | ||
| 284 | } | ||
| 285 | spin_unlock(&robj->tobj.lock); | ||
| 286 | r = radeon_object_reserve(robj, false); | ||
| 287 | if (unlikely(r != 0)) { | ||
| 288 | DRM_ERROR("radeon: failed to reserve object for unpinning it.\n"); | ||
| 289 | return; | ||
| 290 | } | ||
| 291 | flags = robj->tobj.mem.placement; | ||
| 292 | robj->tobj.proposed_placement = flags & ~TTM_PL_FLAG_NO_EVICT; | ||
| 293 | r = ttm_buffer_object_validate(&robj->tobj, | ||
| 294 | robj->tobj.proposed_placement, | ||
| 295 | false, false); | ||
| 296 | if (unlikely(r != 0)) { | ||
| 297 | DRM_ERROR("radeon: failed to unpin buffer.\n"); | ||
| 298 | } | ||
| 299 | radeon_object_unreserve(robj); | ||
| 300 | } | ||
| 301 | |||
| 302 | int radeon_object_wait(struct radeon_object *robj) | ||
| 303 | { | ||
| 304 | int r = 0; | ||
| 305 | |||
| 306 | /* FIXME: should use block reservation instead */ | ||
| 307 | r = radeon_object_reserve(robj, true); | ||
| 308 | if (unlikely(r != 0)) { | ||
| 309 | DRM_ERROR("radeon: failed to reserve object for waiting.\n"); | ||
| 310 | return r; | ||
| 311 | } | ||
| 312 | spin_lock(&robj->tobj.lock); | ||
| 313 | if (robj->tobj.sync_obj) { | ||
| 314 | r = ttm_bo_wait(&robj->tobj, true, true, false); | ||
| 315 | } | ||
| 316 | spin_unlock(&robj->tobj.lock); | ||
| 317 | radeon_object_unreserve(robj); | ||
| 318 | return r; | ||
| 319 | } | ||
| 320 | |||
| 321 | int radeon_object_busy_domain(struct radeon_object *robj, uint32_t *cur_placement) | ||
| 322 | { | ||
| 323 | int r = 0; | ||
| 324 | |||
| 325 | r = radeon_object_reserve(robj, true); | ||
| 326 | if (unlikely(r != 0)) { | ||
| 327 | DRM_ERROR("radeon: failed to reserve object for waiting.\n"); | ||
| 328 | return r; | ||
| 329 | } | ||
| 330 | spin_lock(&robj->tobj.lock); | ||
| 331 | *cur_placement = robj->tobj.mem.mem_type; | ||
| 332 | if (robj->tobj.sync_obj) { | ||
| 333 | r = ttm_bo_wait(&robj->tobj, true, true, true); | ||
| 334 | } | 220 | } |
| 335 | spin_unlock(&robj->tobj.lock); | 221 | bo->pin_count--; |
| 336 | radeon_object_unreserve(robj); | 222 | if (bo->pin_count) |
| 223 | return 0; | ||
| 224 | for (i = 0; i < bo->placement.num_placement; i++) | ||
| 225 | bo->placements[i] &= ~TTM_PL_FLAG_NO_EVICT; | ||
| 226 | r = ttm_buffer_object_validate(&bo->tbo, &bo->placement, false, false); | ||
| 227 | if (unlikely(r != 0)) | ||
| 228 | dev_err(bo->rdev->dev, "%p validate failed for unpin\n", bo); | ||
| 337 | return r; | 229 | return r; |
| 338 | } | 230 | } |
| 339 | 231 | ||
| 340 | int radeon_object_evict_vram(struct radeon_device *rdev) | 232 | int radeon_bo_evict_vram(struct radeon_device *rdev) |
| 341 | { | 233 | { |
| 342 | if (rdev->flags & RADEON_IS_IGP) { | 234 | if (rdev->flags & RADEON_IS_IGP) { |
| 343 | /* Useless to evict on IGP chips */ | 235 | /* Useless to evict on IGP chips */ |
| @@ -346,30 +238,32 @@ int radeon_object_evict_vram(struct radeon_device *rdev) | |||
| 346 | return ttm_bo_evict_mm(&rdev->mman.bdev, TTM_PL_VRAM); | 238 | return ttm_bo_evict_mm(&rdev->mman.bdev, TTM_PL_VRAM); |
| 347 | } | 239 | } |
| 348 | 240 | ||
| 349 | void radeon_object_force_delete(struct radeon_device *rdev) | 241 | void radeon_bo_force_delete(struct radeon_device *rdev) |
| 350 | { | 242 | { |
| 351 | struct radeon_object *robj, *n; | 243 | struct radeon_bo *bo, *n; |
| 352 | struct drm_gem_object *gobj; | 244 | struct drm_gem_object *gobj; |
| 353 | 245 | ||
| 354 | if (list_empty(&rdev->gem.objects)) { | 246 | if (list_empty(&rdev->gem.objects)) { |
| 355 | return; | 247 | return; |
| 356 | } | 248 | } |
| 357 | DRM_ERROR("Userspace still has active objects !\n"); | 249 | dev_err(rdev->dev, "Userspace still has active objects !\n"); |
| 358 | list_for_each_entry_safe(robj, n, &rdev->gem.objects, list) { | 250 | list_for_each_entry_safe(bo, n, &rdev->gem.objects, list) { |
| 359 | mutex_lock(&rdev->ddev->struct_mutex); | 251 | mutex_lock(&rdev->ddev->struct_mutex); |
| 360 | gobj = robj->gobj; | 252 | gobj = bo->gobj; |
| 361 | DRM_ERROR("Force free for (%p,%p,%lu,%lu)\n", | 253 | dev_err(rdev->dev, "%p %p %lu %lu force free\n", |
| 362 | gobj, robj, (unsigned long)gobj->size, | 254 | gobj, bo, (unsigned long)gobj->size, |
| 363 | *((unsigned long *)&gobj->refcount)); | 255 | *((unsigned long *)&gobj->refcount)); |
| 364 | list_del_init(&robj->list); | 256 | mutex_lock(&bo->rdev->gem.mutex); |
| 365 | radeon_object_unref(&robj); | 257 | list_del_init(&bo->list); |
| 258 | mutex_unlock(&bo->rdev->gem.mutex); | ||
| 259 | radeon_bo_unref(&bo); | ||
| 366 | gobj->driver_private = NULL; | 260 | gobj->driver_private = NULL; |
| 367 | drm_gem_object_unreference(gobj); | 261 | drm_gem_object_unreference(gobj); |
| 368 | mutex_unlock(&rdev->ddev->struct_mutex); | 262 | mutex_unlock(&rdev->ddev->struct_mutex); |
| 369 | } | 263 | } |
| 370 | } | 264 | } |
| 371 | 265 | ||
| 372 | int radeon_object_init(struct radeon_device *rdev) | 266 | int radeon_bo_init(struct radeon_device *rdev) |
| 373 | { | 267 | { |
| 374 | /* Add an MTRR for the VRAM */ | 268 | /* Add an MTRR for the VRAM */ |
| 375 | rdev->mc.vram_mtrr = mtrr_add(rdev->mc.aper_base, rdev->mc.aper_size, | 269 | rdev->mc.vram_mtrr = mtrr_add(rdev->mc.aper_base, rdev->mc.aper_size, |
| @@ -382,13 +276,13 @@ int radeon_object_init(struct radeon_device *rdev) | |||
| 382 | return radeon_ttm_init(rdev); | 276 | return radeon_ttm_init(rdev); |
| 383 | } | 277 | } |
| 384 | 278 | ||
| 385 | void radeon_object_fini(struct radeon_device *rdev) | 279 | void radeon_bo_fini(struct radeon_device *rdev) |
| 386 | { | 280 | { |
| 387 | radeon_ttm_fini(rdev); | 281 | radeon_ttm_fini(rdev); |
| 388 | } | 282 | } |
| 389 | 283 | ||
| 390 | void radeon_object_list_add_object(struct radeon_object_list *lobj, | 284 | void radeon_bo_list_add_object(struct radeon_bo_list *lobj, |
| 391 | struct list_head *head) | 285 | struct list_head *head) |
| 392 | { | 286 | { |
| 393 | if (lobj->wdomain) { | 287 | if (lobj->wdomain) { |
| 394 | list_add(&lobj->list, head); | 288 | list_add(&lobj->list, head); |
| @@ -397,72 +291,63 @@ void radeon_object_list_add_object(struct radeon_object_list *lobj, | |||
| 397 | } | 291 | } |
| 398 | } | 292 | } |
| 399 | 293 | ||
| 400 | int radeon_object_list_reserve(struct list_head *head) | 294 | int radeon_bo_list_reserve(struct list_head *head) |
| 401 | { | 295 | { |
| 402 | struct radeon_object_list *lobj; | 296 | struct radeon_bo_list *lobj; |
| 403 | int r; | 297 | int r; |
| 404 | 298 | ||
| 405 | list_for_each_entry(lobj, head, list){ | 299 | list_for_each_entry(lobj, head, list){ |
| 406 | if (!lobj->robj->pin_count) { | 300 | r = radeon_bo_reserve(lobj->bo, false); |
| 407 | r = radeon_object_reserve(lobj->robj, true); | 301 | if (unlikely(r != 0)) |
| 408 | if (unlikely(r != 0)) { | 302 | return r; |
| 409 | DRM_ERROR("radeon: failed to reserve object.\n"); | ||
| 410 | return r; | ||
| 411 | } | ||
| 412 | } else { | ||
| 413 | } | ||
| 414 | } | 303 | } |
| 415 | return 0; | 304 | return 0; |
| 416 | } | 305 | } |
| 417 | 306 | ||
| 418 | void radeon_object_list_unreserve(struct list_head *head) | 307 | void radeon_bo_list_unreserve(struct list_head *head) |
| 419 | { | 308 | { |
| 420 | struct radeon_object_list *lobj; | 309 | struct radeon_bo_list *lobj; |
| 421 | 310 | ||
| 422 | list_for_each_entry(lobj, head, list) { | 311 | list_for_each_entry(lobj, head, list) { |
| 423 | if (!lobj->robj->pin_count) { | 312 | /* only unreserve object we successfully reserved */ |
| 424 | radeon_object_unreserve(lobj->robj); | 313 | if (radeon_bo_is_reserved(lobj->bo)) |
| 425 | } | 314 | radeon_bo_unreserve(lobj->bo); |
| 426 | } | 315 | } |
| 427 | } | 316 | } |
| 428 | 317 | ||
| 429 | int radeon_object_list_validate(struct list_head *head, void *fence) | 318 | int radeon_bo_list_validate(struct list_head *head, void *fence) |
| 430 | { | 319 | { |
| 431 | struct radeon_object_list *lobj; | 320 | struct radeon_bo_list *lobj; |
| 432 | struct radeon_object *robj; | 321 | struct radeon_bo *bo; |
| 433 | struct radeon_fence *old_fence = NULL; | 322 | struct radeon_fence *old_fence = NULL; |
| 434 | int r; | 323 | int r; |
| 435 | 324 | ||
| 436 | r = radeon_object_list_reserve(head); | 325 | r = radeon_bo_list_reserve(head); |
| 437 | if (unlikely(r != 0)) { | 326 | if (unlikely(r != 0)) { |
| 438 | radeon_object_list_unreserve(head); | ||
| 439 | return r; | 327 | return r; |
| 440 | } | 328 | } |
| 441 | list_for_each_entry(lobj, head, list) { | 329 | list_for_each_entry(lobj, head, list) { |
| 442 | robj = lobj->robj; | 330 | bo = lobj->bo; |
| 443 | if (!robj->pin_count) { | 331 | if (!bo->pin_count) { |
| 444 | if (lobj->wdomain) { | 332 | if (lobj->wdomain) { |
| 445 | robj->tobj.proposed_placement = | 333 | radeon_ttm_placement_from_domain(bo, |
| 446 | radeon_object_flags_from_domain(lobj->wdomain); | 334 | lobj->wdomain); |
| 447 | } else { | 335 | } else { |
| 448 | robj->tobj.proposed_placement = | 336 | radeon_ttm_placement_from_domain(bo, |
| 449 | radeon_object_flags_from_domain(lobj->rdomain); | 337 | lobj->rdomain); |
| 450 | } | 338 | } |
| 451 | r = ttm_buffer_object_validate(&robj->tobj, | 339 | r = ttm_buffer_object_validate(&bo->tbo, |
| 452 | robj->tobj.proposed_placement, | 340 | &bo->placement, |
| 453 | true, false); | 341 | true, false); |
| 454 | if (unlikely(r)) { | 342 | if (unlikely(r)) |
| 455 | DRM_ERROR("radeon: failed to validate.\n"); | ||
| 456 | return r; | 343 | return r; |
| 457 | } | ||
| 458 | radeon_object_gpu_addr(robj); | ||
| 459 | } | 344 | } |
| 460 | lobj->gpu_offset = robj->gpu_addr; | 345 | lobj->gpu_offset = radeon_bo_gpu_offset(bo); |
| 461 | lobj->tiling_flags = robj->tiling_flags; | 346 | lobj->tiling_flags = bo->tiling_flags; |
| 462 | if (fence) { | 347 | if (fence) { |
| 463 | old_fence = (struct radeon_fence *)robj->tobj.sync_obj; | 348 | old_fence = (struct radeon_fence *)bo->tbo.sync_obj; |
| 464 | robj->tobj.sync_obj = radeon_fence_ref(fence); | 349 | bo->tbo.sync_obj = radeon_fence_ref(fence); |
| 465 | robj->tobj.sync_obj_arg = NULL; | 350 | bo->tbo.sync_obj_arg = NULL; |
| 466 | } | 351 | } |
| 467 | if (old_fence) { | 352 | if (old_fence) { |
| 468 | radeon_fence_unref(&old_fence); | 353 | radeon_fence_unref(&old_fence); |
| @@ -471,51 +356,44 @@ int radeon_object_list_validate(struct list_head *head, void *fence) | |||
| 471 | return 0; | 356 | return 0; |
| 472 | } | 357 | } |
| 473 | 358 | ||
| 474 | void radeon_object_list_unvalidate(struct list_head *head) | 359 | void radeon_bo_list_unvalidate(struct list_head *head, void *fence) |
| 475 | { | 360 | { |
| 476 | struct radeon_object_list *lobj; | 361 | struct radeon_bo_list *lobj; |
| 477 | struct radeon_fence *old_fence = NULL; | 362 | struct radeon_fence *old_fence; |
| 478 | 363 | ||
| 479 | list_for_each_entry(lobj, head, list) { | 364 | if (fence) |
| 480 | old_fence = (struct radeon_fence *)lobj->robj->tobj.sync_obj; | 365 | list_for_each_entry(lobj, head, list) { |
| 481 | lobj->robj->tobj.sync_obj = NULL; | 366 | old_fence = to_radeon_fence(lobj->bo->tbo.sync_obj); |
| 482 | if (old_fence) { | 367 | if (old_fence == fence) { |
| 483 | radeon_fence_unref(&old_fence); | 368 | lobj->bo->tbo.sync_obj = NULL; |
| 369 | radeon_fence_unref(&old_fence); | ||
| 370 | } | ||
| 484 | } | 371 | } |
| 485 | } | 372 | radeon_bo_list_unreserve(head); |
| 486 | radeon_object_list_unreserve(head); | ||
| 487 | } | ||
| 488 | |||
| 489 | void radeon_object_list_clean(struct list_head *head) | ||
| 490 | { | ||
| 491 | radeon_object_list_unreserve(head); | ||
| 492 | } | 373 | } |
| 493 | 374 | ||
| 494 | int radeon_object_fbdev_mmap(struct radeon_object *robj, | 375 | int radeon_bo_fbdev_mmap(struct radeon_bo *bo, |
| 495 | struct vm_area_struct *vma) | 376 | struct vm_area_struct *vma) |
| 496 | { | 377 | { |
| 497 | return ttm_fbdev_mmap(vma, &robj->tobj); | 378 | return ttm_fbdev_mmap(vma, &bo->tbo); |
| 498 | } | 379 | } |
| 499 | 380 | ||
| 500 | unsigned long radeon_object_size(struct radeon_object *robj) | 381 | int radeon_bo_get_surface_reg(struct radeon_bo *bo) |
| 501 | { | 382 | { |
| 502 | return robj->tobj.num_pages << PAGE_SHIFT; | 383 | struct radeon_device *rdev = bo->rdev; |
| 503 | } | ||
| 504 | |||
| 505 | int radeon_object_get_surface_reg(struct radeon_object *robj) | ||
| 506 | { | ||
| 507 | struct radeon_device *rdev = robj->rdev; | ||
| 508 | struct radeon_surface_reg *reg; | 384 | struct radeon_surface_reg *reg; |
| 509 | struct radeon_object *old_object; | 385 | struct radeon_bo *old_object; |
| 510 | int steal; | 386 | int steal; |
| 511 | int i; | 387 | int i; |
| 512 | 388 | ||
| 513 | if (!robj->tiling_flags) | 389 | BUG_ON(!atomic_read(&bo->tbo.reserved)); |
| 390 | |||
| 391 | if (!bo->tiling_flags) | ||
| 514 | return 0; | 392 | return 0; |
| 515 | 393 | ||
| 516 | if (robj->surface_reg >= 0) { | 394 | if (bo->surface_reg >= 0) { |
| 517 | reg = &rdev->surface_regs[robj->surface_reg]; | 395 | reg = &rdev->surface_regs[bo->surface_reg]; |
| 518 | i = robj->surface_reg; | 396 | i = bo->surface_reg; |
| 519 | goto out; | 397 | goto out; |
| 520 | } | 398 | } |
| 521 | 399 | ||
| @@ -523,10 +401,10 @@ int radeon_object_get_surface_reg(struct radeon_object *robj) | |||
| 523 | for (i = 0; i < RADEON_GEM_MAX_SURFACES; i++) { | 401 | for (i = 0; i < RADEON_GEM_MAX_SURFACES; i++) { |
| 524 | 402 | ||
| 525 | reg = &rdev->surface_regs[i]; | 403 | reg = &rdev->surface_regs[i]; |
| 526 | if (!reg->robj) | 404 | if (!reg->bo) |
| 527 | break; | 405 | break; |
| 528 | 406 | ||
| 529 | old_object = reg->robj; | 407 | old_object = reg->bo; |
| 530 | if (old_object->pin_count == 0) | 408 | if (old_object->pin_count == 0) |
| 531 | steal = i; | 409 | steal = i; |
| 532 | } | 410 | } |
| @@ -537,91 +415,101 @@ int radeon_object_get_surface_reg(struct radeon_object *robj) | |||
| 537 | return -ENOMEM; | 415 | return -ENOMEM; |
| 538 | /* find someone with a surface reg and nuke their BO */ | 416 | /* find someone with a surface reg and nuke their BO */ |
| 539 | reg = &rdev->surface_regs[steal]; | 417 | reg = &rdev->surface_regs[steal]; |
| 540 | old_object = reg->robj; | 418 | old_object = reg->bo; |
| 541 | /* blow away the mapping */ | 419 | /* blow away the mapping */ |
| 542 | DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); | 420 | DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); |
| 543 | ttm_bo_unmap_virtual(&old_object->tobj); | 421 | ttm_bo_unmap_virtual(&old_object->tbo); |
| 544 | old_object->surface_reg = -1; | 422 | old_object->surface_reg = -1; |
| 545 | i = steal; | 423 | i = steal; |
| 546 | } | 424 | } |
| 547 | 425 | ||
| 548 | robj->surface_reg = i; | 426 | bo->surface_reg = i; |
| 549 | reg->robj = robj; | 427 | reg->bo = bo; |
| 550 | 428 | ||
| 551 | out: | 429 | out: |
| 552 | radeon_set_surface_reg(rdev, i, robj->tiling_flags, robj->pitch, | 430 | radeon_set_surface_reg(rdev, i, bo->tiling_flags, bo->pitch, |
| 553 | robj->tobj.mem.mm_node->start << PAGE_SHIFT, | 431 | bo->tbo.mem.mm_node->start << PAGE_SHIFT, |
| 554 | robj->tobj.num_pages << PAGE_SHIFT); | 432 | bo->tbo.num_pages << PAGE_SHIFT); |
| 555 | return 0; | 433 | return 0; |
| 556 | } | 434 | } |
| 557 | 435 | ||
| 558 | void radeon_object_clear_surface_reg(struct radeon_object *robj) | 436 | static void radeon_bo_clear_surface_reg(struct radeon_bo *bo) |
| 559 | { | 437 | { |
| 560 | struct radeon_device *rdev = robj->rdev; | 438 | struct radeon_device *rdev = bo->rdev; |
| 561 | struct radeon_surface_reg *reg; | 439 | struct radeon_surface_reg *reg; |
| 562 | 440 | ||
| 563 | if (robj->surface_reg == -1) | 441 | if (bo->surface_reg == -1) |
| 564 | return; | 442 | return; |
| 565 | 443 | ||
| 566 | reg = &rdev->surface_regs[robj->surface_reg]; | 444 | reg = &rdev->surface_regs[bo->surface_reg]; |
| 567 | radeon_clear_surface_reg(rdev, robj->surface_reg); | 445 | radeon_clear_surface_reg(rdev, bo->surface_reg); |
| 568 | 446 | ||
| 569 | reg->robj = NULL; | 447 | reg->bo = NULL; |
| 570 | robj->surface_reg = -1; | 448 | bo->surface_reg = -1; |
| 571 | } | 449 | } |
| 572 | 450 | ||
| 573 | void radeon_object_set_tiling_flags(struct radeon_object *robj, | 451 | int radeon_bo_set_tiling_flags(struct radeon_bo *bo, |
| 574 | uint32_t tiling_flags, uint32_t pitch) | 452 | uint32_t tiling_flags, uint32_t pitch) |
| 575 | { | 453 | { |
| 576 | robj->tiling_flags = tiling_flags; | 454 | int r; |
| 577 | robj->pitch = pitch; | 455 | |
| 456 | r = radeon_bo_reserve(bo, false); | ||
| 457 | if (unlikely(r != 0)) | ||
| 458 | return r; | ||
| 459 | bo->tiling_flags = tiling_flags; | ||
| 460 | bo->pitch = pitch; | ||
| 461 | radeon_bo_unreserve(bo); | ||
| 462 | return 0; | ||
| 578 | } | 463 | } |
| 579 | 464 | ||
| 580 | void radeon_object_get_tiling_flags(struct radeon_object *robj, | 465 | void radeon_bo_get_tiling_flags(struct radeon_bo *bo, |
| 581 | uint32_t *tiling_flags, | 466 | uint32_t *tiling_flags, |
| 582 | uint32_t *pitch) | 467 | uint32_t *pitch) |
| 583 | { | 468 | { |
| 469 | BUG_ON(!atomic_read(&bo->tbo.reserved)); | ||
| 584 | if (tiling_flags) | 470 | if (tiling_flags) |
| 585 | *tiling_flags = robj->tiling_flags; | 471 | *tiling_flags = bo->tiling_flags; |
| 586 | if (pitch) | 472 | if (pitch) |
| 587 | *pitch = robj->pitch; | 473 | *pitch = bo->pitch; |
| 588 | } | 474 | } |
| 589 | 475 | ||
| 590 | int radeon_object_check_tiling(struct radeon_object *robj, bool has_moved, | 476 | int radeon_bo_check_tiling(struct radeon_bo *bo, bool has_moved, |
| 591 | bool force_drop) | 477 | bool force_drop) |
| 592 | { | 478 | { |
| 593 | if (!(robj->tiling_flags & RADEON_TILING_SURFACE)) | 479 | BUG_ON(!atomic_read(&bo->tbo.reserved)); |
| 480 | |||
| 481 | if (!(bo->tiling_flags & RADEON_TILING_SURFACE)) | ||
| 594 | return 0; | 482 | return 0; |
| 595 | 483 | ||
| 596 | if (force_drop) { | 484 | if (force_drop) { |
| 597 | radeon_object_clear_surface_reg(robj); | 485 | radeon_bo_clear_surface_reg(bo); |
| 598 | return 0; | 486 | return 0; |
| 599 | } | 487 | } |
| 600 | 488 | ||
| 601 | if (robj->tobj.mem.mem_type != TTM_PL_VRAM) { | 489 | if (bo->tbo.mem.mem_type != TTM_PL_VRAM) { |
| 602 | if (!has_moved) | 490 | if (!has_moved) |
| 603 | return 0; | 491 | return 0; |
| 604 | 492 | ||
| 605 | if (robj->surface_reg >= 0) | 493 | if (bo->surface_reg >= 0) |
| 606 | radeon_object_clear_surface_reg(robj); | 494 | radeon_bo_clear_surface_reg(bo); |
| 607 | return 0; | 495 | return 0; |
| 608 | } | 496 | } |
| 609 | 497 | ||
| 610 | if ((robj->surface_reg >= 0) && !has_moved) | 498 | if ((bo->surface_reg >= 0) && !has_moved) |
| 611 | return 0; | 499 | return 0; |
| 612 | 500 | ||
| 613 | return radeon_object_get_surface_reg(robj); | 501 | return radeon_bo_get_surface_reg(bo); |
| 614 | } | 502 | } |
| 615 | 503 | ||
| 616 | void radeon_bo_move_notify(struct ttm_buffer_object *bo, | 504 | void radeon_bo_move_notify(struct ttm_buffer_object *bo, |
| 617 | struct ttm_mem_reg *mem) | 505 | struct ttm_mem_reg *mem) |
| 618 | { | 506 | { |
| 619 | struct radeon_object *robj = container_of(bo, struct radeon_object, tobj); | 507 | struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo); |
| 620 | radeon_object_check_tiling(robj, 0, 1); | 508 | radeon_bo_check_tiling(rbo, 0, 1); |
| 621 | } | 509 | } |
| 622 | 510 | ||
| 623 | void radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo) | 511 | void radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo) |
| 624 | { | 512 | { |
| 625 | struct radeon_object *robj = container_of(bo, struct radeon_object, tobj); | 513 | struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo); |
| 626 | radeon_object_check_tiling(robj, 0, 0); | 514 | radeon_bo_check_tiling(rbo, 0, 0); |
| 627 | } | 515 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_object.h b/drivers/gpu/drm/radeon/radeon_object.h index 10e8af6bb456..f6b69c2c0d00 100644 --- a/drivers/gpu/drm/radeon/radeon_object.h +++ b/drivers/gpu/drm/radeon/radeon_object.h | |||
| @@ -28,19 +28,152 @@ | |||
| 28 | #ifndef __RADEON_OBJECT_H__ | 28 | #ifndef __RADEON_OBJECT_H__ |
| 29 | #define __RADEON_OBJECT_H__ | 29 | #define __RADEON_OBJECT_H__ |
| 30 | 30 | ||
| 31 | #include <ttm/ttm_bo_api.h> | 31 | #include <drm/radeon_drm.h> |
| 32 | #include <ttm/ttm_bo_driver.h> | 32 | #include "radeon.h" |
| 33 | #include <ttm/ttm_placement.h> | ||
| 34 | #include <ttm/ttm_module.h> | ||
| 35 | 33 | ||
| 36 | /* | 34 | /** |
| 37 | * TTM. | 35 | * radeon_mem_type_to_domain - return domain corresponding to mem_type |
| 36 | * @mem_type: ttm memory type | ||
| 37 | * | ||
| 38 | * Returns corresponding domain of the ttm mem_type | ||
| 39 | */ | ||
| 40 | static inline unsigned radeon_mem_type_to_domain(u32 mem_type) | ||
| 41 | { | ||
| 42 | switch (mem_type) { | ||
| 43 | case TTM_PL_VRAM: | ||
| 44 | return RADEON_GEM_DOMAIN_VRAM; | ||
| 45 | case TTM_PL_TT: | ||
| 46 | return RADEON_GEM_DOMAIN_GTT; | ||
| 47 | case TTM_PL_SYSTEM: | ||
| 48 | return RADEON_GEM_DOMAIN_CPU; | ||
| 49 | default: | ||
| 50 | break; | ||
| 51 | } | ||
| 52 | return 0; | ||
| 53 | } | ||
| 54 | |||
| 55 | /** | ||
| 56 | * radeon_bo_reserve - reserve bo | ||
| 57 | * @bo: bo structure | ||
| 58 | * @no_wait: don't sleep while trying to reserve (return -EBUSY) | ||
| 59 | * | ||
| 60 | * Returns: | ||
| 61 | * -EBUSY: buffer is busy and @no_wait is true | ||
| 62 | * -ERESTART: A wait for the buffer to become unreserved was interrupted by | ||
| 63 | * a signal. Release all buffer reservations and return to user-space. | ||
| 38 | */ | 64 | */ |
| 39 | struct radeon_mman { | 65 | static inline int radeon_bo_reserve(struct radeon_bo *bo, bool no_wait) |
| 40 | struct ttm_bo_global_ref bo_global_ref; | 66 | { |
| 41 | struct ttm_global_reference mem_global_ref; | 67 | int r; |
| 42 | bool mem_global_referenced; | 68 | |
| 43 | struct ttm_bo_device bdev; | 69 | retry: |
| 44 | }; | 70 | r = ttm_bo_reserve(&bo->tbo, true, no_wait, false, 0); |
| 71 | if (unlikely(r != 0)) { | ||
| 72 | if (r == -ERESTART) | ||
| 73 | goto retry; | ||
| 74 | dev_err(bo->rdev->dev, "%p reserve failed\n", bo); | ||
| 75 | return r; | ||
| 76 | } | ||
| 77 | return 0; | ||
| 78 | } | ||
| 79 | |||
| 80 | static inline void radeon_bo_unreserve(struct radeon_bo *bo) | ||
| 81 | { | ||
| 82 | ttm_bo_unreserve(&bo->tbo); | ||
| 83 | } | ||
| 84 | |||
| 85 | /** | ||
| 86 | * radeon_bo_gpu_offset - return GPU offset of bo | ||
| 87 | * @bo: radeon object for which we query the offset | ||
| 88 | * | ||
| 89 | * Returns current GPU offset of the object. | ||
| 90 | * | ||
| 91 | * Note: object should either be pinned or reserved when calling this | ||
| 92 | * function, it might be usefull to add check for this for debugging. | ||
| 93 | */ | ||
| 94 | static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) | ||
| 95 | { | ||
| 96 | return bo->tbo.offset; | ||
| 97 | } | ||
| 98 | |||
| 99 | static inline unsigned long radeon_bo_size(struct radeon_bo *bo) | ||
| 100 | { | ||
| 101 | return bo->tbo.num_pages << PAGE_SHIFT; | ||
| 102 | } | ||
| 103 | |||
| 104 | static inline bool radeon_bo_is_reserved(struct radeon_bo *bo) | ||
| 105 | { | ||
| 106 | return !!atomic_read(&bo->tbo.reserved); | ||
| 107 | } | ||
| 108 | |||
| 109 | /** | ||
| 110 | * radeon_bo_mmap_offset - return mmap offset of bo | ||
| 111 | * @bo: radeon object for which we query the offset | ||
| 112 | * | ||
| 113 | * Returns mmap offset of the object. | ||
| 114 | * | ||
| 115 | * Note: addr_space_offset is constant after ttm bo init thus isn't protected | ||
| 116 | * by any lock. | ||
| 117 | */ | ||
| 118 | static inline u64 radeon_bo_mmap_offset(struct radeon_bo *bo) | ||
| 119 | { | ||
| 120 | return bo->tbo.addr_space_offset; | ||
| 121 | } | ||
| 122 | |||
| 123 | static inline int radeon_bo_wait(struct radeon_bo *bo, u32 *mem_type, | ||
| 124 | bool no_wait) | ||
| 125 | { | ||
| 126 | int r; | ||
| 127 | |||
| 128 | retry: | ||
| 129 | r = ttm_bo_reserve(&bo->tbo, true, no_wait, false, 0); | ||
| 130 | if (unlikely(r != 0)) { | ||
| 131 | if (r == -ERESTART) | ||
| 132 | goto retry; | ||
| 133 | dev_err(bo->rdev->dev, "%p reserve failed for wait\n", bo); | ||
| 134 | return r; | ||
| 135 | } | ||
| 136 | spin_lock(&bo->tbo.lock); | ||
| 137 | if (mem_type) | ||
| 138 | *mem_type = bo->tbo.mem.mem_type; | ||
| 139 | if (bo->tbo.sync_obj) | ||
| 140 | r = ttm_bo_wait(&bo->tbo, true, true, no_wait); | ||
| 141 | spin_unlock(&bo->tbo.lock); | ||
| 142 | ttm_bo_unreserve(&bo->tbo); | ||
| 143 | if (unlikely(r == -ERESTART)) | ||
| 144 | goto retry; | ||
| 145 | return r; | ||
| 146 | } | ||
| 45 | 147 | ||
| 148 | extern int radeon_bo_create(struct radeon_device *rdev, | ||
| 149 | struct drm_gem_object *gobj, unsigned long size, | ||
| 150 | bool kernel, u32 domain, | ||
| 151 | struct radeon_bo **bo_ptr); | ||
| 152 | extern int radeon_bo_kmap(struct radeon_bo *bo, void **ptr); | ||
| 153 | extern void radeon_bo_kunmap(struct radeon_bo *bo); | ||
| 154 | extern void radeon_bo_unref(struct radeon_bo **bo); | ||
| 155 | extern int radeon_bo_pin(struct radeon_bo *bo, u32 domain, u64 *gpu_addr); | ||
| 156 | extern int radeon_bo_unpin(struct radeon_bo *bo); | ||
| 157 | extern int radeon_bo_evict_vram(struct radeon_device *rdev); | ||
| 158 | extern void radeon_bo_force_delete(struct radeon_device *rdev); | ||
| 159 | extern int radeon_bo_init(struct radeon_device *rdev); | ||
| 160 | extern void radeon_bo_fini(struct radeon_device *rdev); | ||
| 161 | extern void radeon_bo_list_add_object(struct radeon_bo_list *lobj, | ||
| 162 | struct list_head *head); | ||
| 163 | extern int radeon_bo_list_reserve(struct list_head *head); | ||
| 164 | extern void radeon_bo_list_unreserve(struct list_head *head); | ||
| 165 | extern int radeon_bo_list_validate(struct list_head *head, void *fence); | ||
| 166 | extern void radeon_bo_list_unvalidate(struct list_head *head, void *fence); | ||
| 167 | extern int radeon_bo_fbdev_mmap(struct radeon_bo *bo, | ||
| 168 | struct vm_area_struct *vma); | ||
| 169 | extern int radeon_bo_set_tiling_flags(struct radeon_bo *bo, | ||
| 170 | u32 tiling_flags, u32 pitch); | ||
| 171 | extern void radeon_bo_get_tiling_flags(struct radeon_bo *bo, | ||
| 172 | u32 *tiling_flags, u32 *pitch); | ||
| 173 | extern int radeon_bo_check_tiling(struct radeon_bo *bo, bool has_moved, | ||
| 174 | bool force_drop); | ||
| 175 | extern void radeon_bo_move_notify(struct ttm_buffer_object *bo, | ||
| 176 | struct ttm_mem_reg *mem); | ||
| 177 | extern void radeon_bo_fault_reserve_notify(struct ttm_buffer_object *bo); | ||
| 178 | extern int radeon_bo_get_surface_reg(struct radeon_bo *bo); | ||
| 46 | #endif | 179 | #endif |
diff --git a/drivers/gpu/drm/radeon/radeon_pm.c b/drivers/gpu/drm/radeon/radeon_pm.c index 46146c6a2a06..34b08d307c81 100644 --- a/drivers/gpu/drm/radeon/radeon_pm.c +++ b/drivers/gpu/drm/radeon/radeon_pm.c | |||
| @@ -27,7 +27,7 @@ int radeon_debugfs_pm_init(struct radeon_device *rdev); | |||
| 27 | int radeon_pm_init(struct radeon_device *rdev) | 27 | int radeon_pm_init(struct radeon_device *rdev) |
| 28 | { | 28 | { |
| 29 | if (radeon_debugfs_pm_init(rdev)) { | 29 | if (radeon_debugfs_pm_init(rdev)) { |
| 30 | DRM_ERROR("Failed to register debugfs file for CP !\n"); | 30 | DRM_ERROR("Failed to register debugfs file for PM!\n"); |
| 31 | } | 31 | } |
| 32 | 32 | ||
| 33 | return 0; | 33 | return 0; |
| @@ -44,8 +44,8 @@ static int radeon_debugfs_pm_info(struct seq_file *m, void *data) | |||
| 44 | struct drm_device *dev = node->minor->dev; | 44 | struct drm_device *dev = node->minor->dev; |
| 45 | struct radeon_device *rdev = dev->dev_private; | 45 | struct radeon_device *rdev = dev->dev_private; |
| 46 | 46 | ||
| 47 | seq_printf(m, "engine clock: %u0 Hz\n", radeon_get_engine_clock(rdev)); | 47 | seq_printf(m, "engine clock: %u0 kHz\n", radeon_get_engine_clock(rdev)); |
| 48 | seq_printf(m, "memory clock: %u0 Hz\n", radeon_get_memory_clock(rdev)); | 48 | seq_printf(m, "memory clock: %u0 kHz\n", radeon_get_memory_clock(rdev)); |
| 49 | 49 | ||
| 50 | return 0; | 50 | return 0; |
| 51 | } | 51 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_reg.h b/drivers/gpu/drm/radeon/radeon_reg.h index 29ab75903ec1..6d0a009dd4a1 100644 --- a/drivers/gpu/drm/radeon/radeon_reg.h +++ b/drivers/gpu/drm/radeon/radeon_reg.h | |||
| @@ -887,6 +887,7 @@ | |||
| 887 | # define RADEON_FP_PANEL_FORMAT (1 << 3) | 887 | # define RADEON_FP_PANEL_FORMAT (1 << 3) |
| 888 | # define RADEON_FP_EN_TMDS (1 << 7) | 888 | # define RADEON_FP_EN_TMDS (1 << 7) |
| 889 | # define RADEON_FP_DETECT_SENSE (1 << 8) | 889 | # define RADEON_FP_DETECT_SENSE (1 << 8) |
| 890 | # define RADEON_FP_DETECT_INT_POL (1 << 9) | ||
| 890 | # define R200_FP_SOURCE_SEL_MASK (3 << 10) | 891 | # define R200_FP_SOURCE_SEL_MASK (3 << 10) |
| 891 | # define R200_FP_SOURCE_SEL_CRTC1 (0 << 10) | 892 | # define R200_FP_SOURCE_SEL_CRTC1 (0 << 10) |
| 892 | # define R200_FP_SOURCE_SEL_CRTC2 (1 << 10) | 893 | # define R200_FP_SOURCE_SEL_CRTC2 (1 << 10) |
| @@ -894,6 +895,7 @@ | |||
| 894 | # define R200_FP_SOURCE_SEL_TRANS (3 << 10) | 895 | # define R200_FP_SOURCE_SEL_TRANS (3 << 10) |
| 895 | # define RADEON_FP_SEL_CRTC1 (0 << 13) | 896 | # define RADEON_FP_SEL_CRTC1 (0 << 13) |
| 896 | # define RADEON_FP_SEL_CRTC2 (1 << 13) | 897 | # define RADEON_FP_SEL_CRTC2 (1 << 13) |
| 898 | # define R300_HPD_SEL(x) ((x) << 13) | ||
| 897 | # define RADEON_FP_CRTC_DONT_SHADOW_HPAR (1 << 15) | 899 | # define RADEON_FP_CRTC_DONT_SHADOW_HPAR (1 << 15) |
| 898 | # define RADEON_FP_CRTC_DONT_SHADOW_VPAR (1 << 16) | 900 | # define RADEON_FP_CRTC_DONT_SHADOW_VPAR (1 << 16) |
| 899 | # define RADEON_FP_CRTC_DONT_SHADOW_HEND (1 << 17) | 901 | # define RADEON_FP_CRTC_DONT_SHADOW_HEND (1 << 17) |
| @@ -909,6 +911,7 @@ | |||
| 909 | # define RADEON_FP2_ON (1 << 2) | 911 | # define RADEON_FP2_ON (1 << 2) |
| 910 | # define RADEON_FP2_PANEL_FORMAT (1 << 3) | 912 | # define RADEON_FP2_PANEL_FORMAT (1 << 3) |
| 911 | # define RADEON_FP2_DETECT_SENSE (1 << 8) | 913 | # define RADEON_FP2_DETECT_SENSE (1 << 8) |
| 914 | # define RADEON_FP2_DETECT_INT_POL (1 << 9) | ||
| 912 | # define R200_FP2_SOURCE_SEL_MASK (3 << 10) | 915 | # define R200_FP2_SOURCE_SEL_MASK (3 << 10) |
| 913 | # define R200_FP2_SOURCE_SEL_CRTC1 (0 << 10) | 916 | # define R200_FP2_SOURCE_SEL_CRTC1 (0 << 10) |
| 914 | # define R200_FP2_SOURCE_SEL_CRTC2 (1 << 10) | 917 | # define R200_FP2_SOURCE_SEL_CRTC2 (1 << 10) |
| @@ -988,14 +991,20 @@ | |||
| 988 | 991 | ||
| 989 | #define RADEON_GEN_INT_CNTL 0x0040 | 992 | #define RADEON_GEN_INT_CNTL 0x0040 |
| 990 | # define RADEON_CRTC_VBLANK_MASK (1 << 0) | 993 | # define RADEON_CRTC_VBLANK_MASK (1 << 0) |
| 994 | # define RADEON_FP_DETECT_MASK (1 << 4) | ||
| 991 | # define RADEON_CRTC2_VBLANK_MASK (1 << 9) | 995 | # define RADEON_CRTC2_VBLANK_MASK (1 << 9) |
| 996 | # define RADEON_FP2_DETECT_MASK (1 << 10) | ||
| 992 | # define RADEON_SW_INT_ENABLE (1 << 25) | 997 | # define RADEON_SW_INT_ENABLE (1 << 25) |
| 993 | #define RADEON_GEN_INT_STATUS 0x0044 | 998 | #define RADEON_GEN_INT_STATUS 0x0044 |
| 994 | # define AVIVO_DISPLAY_INT_STATUS (1 << 0) | 999 | # define AVIVO_DISPLAY_INT_STATUS (1 << 0) |
| 995 | # define RADEON_CRTC_VBLANK_STAT (1 << 0) | 1000 | # define RADEON_CRTC_VBLANK_STAT (1 << 0) |
| 996 | # define RADEON_CRTC_VBLANK_STAT_ACK (1 << 0) | 1001 | # define RADEON_CRTC_VBLANK_STAT_ACK (1 << 0) |
| 1002 | # define RADEON_FP_DETECT_STAT (1 << 4) | ||
| 1003 | # define RADEON_FP_DETECT_STAT_ACK (1 << 4) | ||
| 997 | # define RADEON_CRTC2_VBLANK_STAT (1 << 9) | 1004 | # define RADEON_CRTC2_VBLANK_STAT (1 << 9) |
| 998 | # define RADEON_CRTC2_VBLANK_STAT_ACK (1 << 9) | 1005 | # define RADEON_CRTC2_VBLANK_STAT_ACK (1 << 9) |
| 1006 | # define RADEON_FP2_DETECT_STAT (1 << 10) | ||
| 1007 | # define RADEON_FP2_DETECT_STAT_ACK (1 << 10) | ||
| 999 | # define RADEON_SW_INT_FIRE (1 << 26) | 1008 | # define RADEON_SW_INT_FIRE (1 << 26) |
| 1000 | # define RADEON_SW_INT_TEST (1 << 25) | 1009 | # define RADEON_SW_INT_TEST (1 << 25) |
| 1001 | # define RADEON_SW_INT_TEST_ACK (1 << 25) | 1010 | # define RADEON_SW_INT_TEST_ACK (1 << 25) |
| @@ -1051,20 +1060,25 @@ | |||
| 1051 | 1060 | ||
| 1052 | /* Multimedia I2C bus */ | 1061 | /* Multimedia I2C bus */ |
| 1053 | #define RADEON_I2C_CNTL_0 0x0090 | 1062 | #define RADEON_I2C_CNTL_0 0x0090 |
| 1054 | #define RADEON_I2C_DONE (1<<0) | 1063 | #define RADEON_I2C_DONE (1 << 0) |
| 1055 | #define RADEON_I2C_NACK (1<<1) | 1064 | #define RADEON_I2C_NACK (1 << 1) |
| 1056 | #define RADEON_I2C_HALT (1<<2) | 1065 | #define RADEON_I2C_HALT (1 << 2) |
| 1057 | #define RADEON_I2C_SOFT_RST (1<<5) | 1066 | #define RADEON_I2C_SOFT_RST (1 << 5) |
| 1058 | #define RADEON_I2C_DRIVE_EN (1<<6) | 1067 | #define RADEON_I2C_DRIVE_EN (1 << 6) |
| 1059 | #define RADEON_I2C_DRIVE_SEL (1<<7) | 1068 | #define RADEON_I2C_DRIVE_SEL (1 << 7) |
| 1060 | #define RADEON_I2C_START (1<<8) | 1069 | #define RADEON_I2C_START (1 << 8) |
| 1061 | #define RADEON_I2C_STOP (1<<9) | 1070 | #define RADEON_I2C_STOP (1 << 9) |
| 1062 | #define RADEON_I2C_RECEIVE (1<<10) | 1071 | #define RADEON_I2C_RECEIVE (1 << 10) |
| 1063 | #define RADEON_I2C_ABORT (1<<11) | 1072 | #define RADEON_I2C_ABORT (1 << 11) |
| 1064 | #define RADEON_I2C_GO (1<<12) | 1073 | #define RADEON_I2C_GO (1 << 12) |
| 1074 | #define RADEON_I2C_PRESCALE_SHIFT 16 | ||
| 1065 | #define RADEON_I2C_CNTL_1 0x0094 | 1075 | #define RADEON_I2C_CNTL_1 0x0094 |
| 1066 | #define RADEON_I2C_SEL (1<<16) | 1076 | #define RADEON_I2C_DATA_COUNT_SHIFT 0 |
| 1067 | #define RADEON_I2C_EN (1<<17) | 1077 | #define RADEON_I2C_ADDR_COUNT_SHIFT 4 |
| 1078 | #define RADEON_I2C_INTRA_BYTE_DELAY_SHIFT 8 | ||
| 1079 | #define RADEON_I2C_SEL (1 << 16) | ||
| 1080 | #define RADEON_I2C_EN (1 << 17) | ||
| 1081 | #define RADEON_I2C_TIME_LIMIT_SHIFT 24 | ||
| 1068 | #define RADEON_I2C_DATA 0x0098 | 1082 | #define RADEON_I2C_DATA 0x0098 |
| 1069 | 1083 | ||
| 1070 | #define RADEON_DVI_I2C_CNTL_0 0x02e0 | 1084 | #define RADEON_DVI_I2C_CNTL_0 0x02e0 |
| @@ -1072,7 +1086,7 @@ | |||
| 1072 | # define R200_SEL_DDC1 0 /* 0x60 - VGA_DDC */ | 1086 | # define R200_SEL_DDC1 0 /* 0x60 - VGA_DDC */ |
| 1073 | # define R200_SEL_DDC2 1 /* 0x64 - DVI_DDC */ | 1087 | # define R200_SEL_DDC2 1 /* 0x64 - DVI_DDC */ |
| 1074 | # define R200_SEL_DDC3 2 /* 0x68 - MONID_DDC */ | 1088 | # define R200_SEL_DDC3 2 /* 0x68 - MONID_DDC */ |
| 1075 | #define RADEON_DVI_I2C_CNTL_1 0x02e4 /* ? */ | 1089 | #define RADEON_DVI_I2C_CNTL_1 0x02e4 |
| 1076 | #define RADEON_DVI_I2C_DATA 0x02e8 | 1090 | #define RADEON_DVI_I2C_DATA 0x02e8 |
| 1077 | 1091 | ||
| 1078 | #define RADEON_INTERRUPT_LINE 0x0f3c /* PCI */ | 1092 | #define RADEON_INTERRUPT_LINE 0x0f3c /* PCI */ |
| @@ -1143,15 +1157,16 @@ | |||
| 1143 | # define RADEON_IO_MCLK_MAX_DYN_STOP_LAT (1 << 13) | 1157 | # define RADEON_IO_MCLK_MAX_DYN_STOP_LAT (1 << 13) |
| 1144 | # define RADEON_MC_MCLK_DYN_ENABLE (1 << 14) | 1158 | # define RADEON_MC_MCLK_DYN_ENABLE (1 << 14) |
| 1145 | # define RADEON_IO_MCLK_DYN_ENABLE (1 << 15) | 1159 | # define RADEON_IO_MCLK_DYN_ENABLE (1 << 15) |
| 1146 | #define RADEON_LCD_GPIO_MASK 0x01a0 | 1160 | |
| 1147 | #define RADEON_GPIOPAD_EN 0x01a0 | ||
| 1148 | #define RADEON_LCD_GPIO_Y_REG 0x01a4 | ||
| 1149 | #define RADEON_MDGPIO_A_REG 0x01ac | ||
| 1150 | #define RADEON_MDGPIO_EN_REG 0x01b0 | ||
| 1151 | #define RADEON_MDGPIO_MASK 0x0198 | ||
| 1152 | #define RADEON_GPIOPAD_MASK 0x0198 | 1161 | #define RADEON_GPIOPAD_MASK 0x0198 |
| 1153 | #define RADEON_GPIOPAD_A 0x019c | 1162 | #define RADEON_GPIOPAD_A 0x019c |
| 1154 | #define RADEON_MDGPIO_Y_REG 0x01b4 | 1163 | #define RADEON_GPIOPAD_EN 0x01a0 |
| 1164 | #define RADEON_GPIOPAD_Y 0x01a4 | ||
| 1165 | #define RADEON_MDGPIO_MASK 0x01a8 | ||
| 1166 | #define RADEON_MDGPIO_A 0x01ac | ||
| 1167 | #define RADEON_MDGPIO_EN 0x01b0 | ||
| 1168 | #define RADEON_MDGPIO_Y 0x01b4 | ||
| 1169 | |||
| 1155 | #define RADEON_MEM_ADDR_CONFIG 0x0148 | 1170 | #define RADEON_MEM_ADDR_CONFIG 0x0148 |
| 1156 | #define RADEON_MEM_BASE 0x0f10 /* PCI */ | 1171 | #define RADEON_MEM_BASE 0x0f10 /* PCI */ |
| 1157 | #define RADEON_MEM_CNTL 0x0140 | 1172 | #define RADEON_MEM_CNTL 0x0140 |
| @@ -1360,6 +1375,9 @@ | |||
| 1360 | #define RADEON_OVR_CLR 0x0230 | 1375 | #define RADEON_OVR_CLR 0x0230 |
| 1361 | #define RADEON_OVR_WID_LEFT_RIGHT 0x0234 | 1376 | #define RADEON_OVR_WID_LEFT_RIGHT 0x0234 |
| 1362 | #define RADEON_OVR_WID_TOP_BOTTOM 0x0238 | 1377 | #define RADEON_OVR_WID_TOP_BOTTOM 0x0238 |
| 1378 | #define RADEON_OVR2_CLR 0x0330 | ||
| 1379 | #define RADEON_OVR2_WID_LEFT_RIGHT 0x0334 | ||
| 1380 | #define RADEON_OVR2_WID_TOP_BOTTOM 0x0338 | ||
| 1363 | 1381 | ||
| 1364 | /* first capture unit */ | 1382 | /* first capture unit */ |
| 1365 | 1383 | ||
diff --git a/drivers/gpu/drm/radeon/radeon_ring.c b/drivers/gpu/drm/radeon/radeon_ring.c index 747b4bffb84b..4d12b2d17b4d 100644 --- a/drivers/gpu/drm/radeon/radeon_ring.c +++ b/drivers/gpu/drm/radeon/radeon_ring.c | |||
| @@ -165,19 +165,24 @@ int radeon_ib_pool_init(struct radeon_device *rdev) | |||
| 165 | return 0; | 165 | return 0; |
| 166 | /* Allocate 1M object buffer */ | 166 | /* Allocate 1M object buffer */ |
| 167 | INIT_LIST_HEAD(&rdev->ib_pool.scheduled_ibs); | 167 | INIT_LIST_HEAD(&rdev->ib_pool.scheduled_ibs); |
| 168 | r = radeon_object_create(rdev, NULL, RADEON_IB_POOL_SIZE*64*1024, | 168 | r = radeon_bo_create(rdev, NULL, RADEON_IB_POOL_SIZE*64*1024, |
| 169 | true, RADEON_GEM_DOMAIN_GTT, | 169 | true, RADEON_GEM_DOMAIN_GTT, |
| 170 | false, &rdev->ib_pool.robj); | 170 | &rdev->ib_pool.robj); |
| 171 | if (r) { | 171 | if (r) { |
| 172 | DRM_ERROR("radeon: failed to ib pool (%d).\n", r); | 172 | DRM_ERROR("radeon: failed to ib pool (%d).\n", r); |
| 173 | return r; | 173 | return r; |
| 174 | } | 174 | } |
| 175 | r = radeon_object_pin(rdev->ib_pool.robj, RADEON_GEM_DOMAIN_GTT, &gpu_addr); | 175 | r = radeon_bo_reserve(rdev->ib_pool.robj, false); |
| 176 | if (unlikely(r != 0)) | ||
| 177 | return r; | ||
| 178 | r = radeon_bo_pin(rdev->ib_pool.robj, RADEON_GEM_DOMAIN_GTT, &gpu_addr); | ||
| 176 | if (r) { | 179 | if (r) { |
| 180 | radeon_bo_unreserve(rdev->ib_pool.robj); | ||
| 177 | DRM_ERROR("radeon: failed to pin ib pool (%d).\n", r); | 181 | DRM_ERROR("radeon: failed to pin ib pool (%d).\n", r); |
| 178 | return r; | 182 | return r; |
| 179 | } | 183 | } |
| 180 | r = radeon_object_kmap(rdev->ib_pool.robj, &ptr); | 184 | r = radeon_bo_kmap(rdev->ib_pool.robj, &ptr); |
| 185 | radeon_bo_unreserve(rdev->ib_pool.robj); | ||
| 181 | if (r) { | 186 | if (r) { |
| 182 | DRM_ERROR("radeon: failed to map ib poll (%d).\n", r); | 187 | DRM_ERROR("radeon: failed to map ib poll (%d).\n", r); |
| 183 | return r; | 188 | return r; |
| @@ -203,14 +208,21 @@ int radeon_ib_pool_init(struct radeon_device *rdev) | |||
| 203 | 208 | ||
| 204 | void radeon_ib_pool_fini(struct radeon_device *rdev) | 209 | void radeon_ib_pool_fini(struct radeon_device *rdev) |
| 205 | { | 210 | { |
| 211 | int r; | ||
| 212 | |||
| 206 | if (!rdev->ib_pool.ready) { | 213 | if (!rdev->ib_pool.ready) { |
| 207 | return; | 214 | return; |
| 208 | } | 215 | } |
| 209 | mutex_lock(&rdev->ib_pool.mutex); | 216 | mutex_lock(&rdev->ib_pool.mutex); |
| 210 | bitmap_zero(rdev->ib_pool.alloc_bm, RADEON_IB_POOL_SIZE); | 217 | bitmap_zero(rdev->ib_pool.alloc_bm, RADEON_IB_POOL_SIZE); |
| 211 | if (rdev->ib_pool.robj) { | 218 | if (rdev->ib_pool.robj) { |
| 212 | radeon_object_kunmap(rdev->ib_pool.robj); | 219 | r = radeon_bo_reserve(rdev->ib_pool.robj, false); |
| 213 | radeon_object_unref(&rdev->ib_pool.robj); | 220 | if (likely(r == 0)) { |
| 221 | radeon_bo_kunmap(rdev->ib_pool.robj); | ||
| 222 | radeon_bo_unpin(rdev->ib_pool.robj); | ||
| 223 | radeon_bo_unreserve(rdev->ib_pool.robj); | ||
| 224 | } | ||
| 225 | radeon_bo_unref(&rdev->ib_pool.robj); | ||
| 214 | rdev->ib_pool.robj = NULL; | 226 | rdev->ib_pool.robj = NULL; |
| 215 | } | 227 | } |
| 216 | mutex_unlock(&rdev->ib_pool.mutex); | 228 | mutex_unlock(&rdev->ib_pool.mutex); |
| @@ -288,29 +300,28 @@ int radeon_ring_init(struct radeon_device *rdev, unsigned ring_size) | |||
| 288 | rdev->cp.ring_size = ring_size; | 300 | rdev->cp.ring_size = ring_size; |
| 289 | /* Allocate ring buffer */ | 301 | /* Allocate ring buffer */ |
| 290 | if (rdev->cp.ring_obj == NULL) { | 302 | if (rdev->cp.ring_obj == NULL) { |
| 291 | r = radeon_object_create(rdev, NULL, rdev->cp.ring_size, | 303 | r = radeon_bo_create(rdev, NULL, rdev->cp.ring_size, true, |
| 292 | true, | 304 | RADEON_GEM_DOMAIN_GTT, |
| 293 | RADEON_GEM_DOMAIN_GTT, | 305 | &rdev->cp.ring_obj); |
| 294 | false, | ||
| 295 | &rdev->cp.ring_obj); | ||
| 296 | if (r) { | 306 | if (r) { |
| 297 | DRM_ERROR("radeon: failed to create ring buffer (%d).\n", r); | 307 | dev_err(rdev->dev, "(%d) ring create failed\n", r); |
| 298 | mutex_unlock(&rdev->cp.mutex); | ||
| 299 | return r; | 308 | return r; |
| 300 | } | 309 | } |
| 301 | r = radeon_object_pin(rdev->cp.ring_obj, | 310 | r = radeon_bo_reserve(rdev->cp.ring_obj, false); |
| 302 | RADEON_GEM_DOMAIN_GTT, | 311 | if (unlikely(r != 0)) |
| 303 | &rdev->cp.gpu_addr); | 312 | return r; |
| 313 | r = radeon_bo_pin(rdev->cp.ring_obj, RADEON_GEM_DOMAIN_GTT, | ||
| 314 | &rdev->cp.gpu_addr); | ||
| 304 | if (r) { | 315 | if (r) { |
| 305 | DRM_ERROR("radeon: failed to pin ring buffer (%d).\n", r); | 316 | radeon_bo_unreserve(rdev->cp.ring_obj); |
| 306 | mutex_unlock(&rdev->cp.mutex); | 317 | dev_err(rdev->dev, "(%d) ring pin failed\n", r); |
| 307 | return r; | 318 | return r; |
| 308 | } | 319 | } |
| 309 | r = radeon_object_kmap(rdev->cp.ring_obj, | 320 | r = radeon_bo_kmap(rdev->cp.ring_obj, |
| 310 | (void **)&rdev->cp.ring); | 321 | (void **)&rdev->cp.ring); |
| 322 | radeon_bo_unreserve(rdev->cp.ring_obj); | ||
| 311 | if (r) { | 323 | if (r) { |
| 312 | DRM_ERROR("radeon: failed to map ring buffer (%d).\n", r); | 324 | dev_err(rdev->dev, "(%d) ring map failed\n", r); |
| 313 | mutex_unlock(&rdev->cp.mutex); | ||
| 314 | return r; | 325 | return r; |
| 315 | } | 326 | } |
| 316 | } | 327 | } |
| @@ -321,11 +332,17 @@ int radeon_ring_init(struct radeon_device *rdev, unsigned ring_size) | |||
| 321 | 332 | ||
| 322 | void radeon_ring_fini(struct radeon_device *rdev) | 333 | void radeon_ring_fini(struct radeon_device *rdev) |
| 323 | { | 334 | { |
| 335 | int r; | ||
| 336 | |||
| 324 | mutex_lock(&rdev->cp.mutex); | 337 | mutex_lock(&rdev->cp.mutex); |
| 325 | if (rdev->cp.ring_obj) { | 338 | if (rdev->cp.ring_obj) { |
| 326 | radeon_object_kunmap(rdev->cp.ring_obj); | 339 | r = radeon_bo_reserve(rdev->cp.ring_obj, false); |
| 327 | radeon_object_unpin(rdev->cp.ring_obj); | 340 | if (likely(r == 0)) { |
| 328 | radeon_object_unref(&rdev->cp.ring_obj); | 341 | radeon_bo_kunmap(rdev->cp.ring_obj); |
| 342 | radeon_bo_unpin(rdev->cp.ring_obj); | ||
| 343 | radeon_bo_unreserve(rdev->cp.ring_obj); | ||
| 344 | } | ||
| 345 | radeon_bo_unref(&rdev->cp.ring_obj); | ||
| 329 | rdev->cp.ring = NULL; | 346 | rdev->cp.ring = NULL; |
| 330 | rdev->cp.ring_obj = NULL; | 347 | rdev->cp.ring_obj = NULL; |
| 331 | } | 348 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_test.c b/drivers/gpu/drm/radeon/radeon_test.c index f8a465d9a1cf..391c973ec4db 100644 --- a/drivers/gpu/drm/radeon/radeon_test.c +++ b/drivers/gpu/drm/radeon/radeon_test.c | |||
| @@ -30,8 +30,8 @@ | |||
| 30 | /* Test BO GTT->VRAM and VRAM->GTT GPU copies across the whole GTT aperture */ | 30 | /* Test BO GTT->VRAM and VRAM->GTT GPU copies across the whole GTT aperture */ |
| 31 | void radeon_test_moves(struct radeon_device *rdev) | 31 | void radeon_test_moves(struct radeon_device *rdev) |
| 32 | { | 32 | { |
| 33 | struct radeon_object *vram_obj = NULL; | 33 | struct radeon_bo *vram_obj = NULL; |
| 34 | struct radeon_object **gtt_obj = NULL; | 34 | struct radeon_bo **gtt_obj = NULL; |
| 35 | struct radeon_fence *fence = NULL; | 35 | struct radeon_fence *fence = NULL; |
| 36 | uint64_t gtt_addr, vram_addr; | 36 | uint64_t gtt_addr, vram_addr; |
| 37 | unsigned i, n, size; | 37 | unsigned i, n, size; |
| @@ -52,38 +52,42 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 52 | goto out_cleanup; | 52 | goto out_cleanup; |
| 53 | } | 53 | } |
| 54 | 54 | ||
| 55 | r = radeon_object_create(rdev, NULL, size, true, RADEON_GEM_DOMAIN_VRAM, | 55 | r = radeon_bo_create(rdev, NULL, size, true, RADEON_GEM_DOMAIN_VRAM, |
| 56 | false, &vram_obj); | 56 | &vram_obj); |
| 57 | if (r) { | 57 | if (r) { |
| 58 | DRM_ERROR("Failed to create VRAM object\n"); | 58 | DRM_ERROR("Failed to create VRAM object\n"); |
| 59 | goto out_cleanup; | 59 | goto out_cleanup; |
| 60 | } | 60 | } |
| 61 | 61 | r = radeon_bo_reserve(vram_obj, false); | |
| 62 | r = radeon_object_pin(vram_obj, RADEON_GEM_DOMAIN_VRAM, &vram_addr); | 62 | if (unlikely(r != 0)) |
| 63 | goto out_cleanup; | ||
| 64 | r = radeon_bo_pin(vram_obj, RADEON_GEM_DOMAIN_VRAM, &vram_addr); | ||
| 63 | if (r) { | 65 | if (r) { |
| 64 | DRM_ERROR("Failed to pin VRAM object\n"); | 66 | DRM_ERROR("Failed to pin VRAM object\n"); |
| 65 | goto out_cleanup; | 67 | goto out_cleanup; |
| 66 | } | 68 | } |
| 67 | |||
| 68 | for (i = 0; i < n; i++) { | 69 | for (i = 0; i < n; i++) { |
| 69 | void *gtt_map, *vram_map; | 70 | void *gtt_map, *vram_map; |
| 70 | void **gtt_start, **gtt_end; | 71 | void **gtt_start, **gtt_end; |
| 71 | void **vram_start, **vram_end; | 72 | void **vram_start, **vram_end; |
| 72 | 73 | ||
| 73 | r = radeon_object_create(rdev, NULL, size, true, | 74 | r = radeon_bo_create(rdev, NULL, size, true, |
| 74 | RADEON_GEM_DOMAIN_GTT, false, gtt_obj + i); | 75 | RADEON_GEM_DOMAIN_GTT, gtt_obj + i); |
| 75 | if (r) { | 76 | if (r) { |
| 76 | DRM_ERROR("Failed to create GTT object %d\n", i); | 77 | DRM_ERROR("Failed to create GTT object %d\n", i); |
| 77 | goto out_cleanup; | 78 | goto out_cleanup; |
| 78 | } | 79 | } |
| 79 | 80 | ||
| 80 | r = radeon_object_pin(gtt_obj[i], RADEON_GEM_DOMAIN_GTT, >t_addr); | 81 | r = radeon_bo_reserve(gtt_obj[i], false); |
| 82 | if (unlikely(r != 0)) | ||
| 83 | goto out_cleanup; | ||
| 84 | r = radeon_bo_pin(gtt_obj[i], RADEON_GEM_DOMAIN_GTT, >t_addr); | ||
| 81 | if (r) { | 85 | if (r) { |
| 82 | DRM_ERROR("Failed to pin GTT object %d\n", i); | 86 | DRM_ERROR("Failed to pin GTT object %d\n", i); |
| 83 | goto out_cleanup; | 87 | goto out_cleanup; |
| 84 | } | 88 | } |
| 85 | 89 | ||
| 86 | r = radeon_object_kmap(gtt_obj[i], >t_map); | 90 | r = radeon_bo_kmap(gtt_obj[i], >t_map); |
| 87 | if (r) { | 91 | if (r) { |
| 88 | DRM_ERROR("Failed to map GTT object %d\n", i); | 92 | DRM_ERROR("Failed to map GTT object %d\n", i); |
| 89 | goto out_cleanup; | 93 | goto out_cleanup; |
| @@ -94,7 +98,7 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 94 | gtt_start++) | 98 | gtt_start++) |
| 95 | *gtt_start = gtt_start; | 99 | *gtt_start = gtt_start; |
| 96 | 100 | ||
| 97 | radeon_object_kunmap(gtt_obj[i]); | 101 | radeon_bo_kunmap(gtt_obj[i]); |
| 98 | 102 | ||
| 99 | r = radeon_fence_create(rdev, &fence); | 103 | r = radeon_fence_create(rdev, &fence); |
| 100 | if (r) { | 104 | if (r) { |
| @@ -116,7 +120,7 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 116 | 120 | ||
| 117 | radeon_fence_unref(&fence); | 121 | radeon_fence_unref(&fence); |
| 118 | 122 | ||
| 119 | r = radeon_object_kmap(vram_obj, &vram_map); | 123 | r = radeon_bo_kmap(vram_obj, &vram_map); |
| 120 | if (r) { | 124 | if (r) { |
| 121 | DRM_ERROR("Failed to map VRAM object after copy %d\n", i); | 125 | DRM_ERROR("Failed to map VRAM object after copy %d\n", i); |
| 122 | goto out_cleanup; | 126 | goto out_cleanup; |
| @@ -131,13 +135,13 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 131 | "expected 0x%p (GTT map 0x%p-0x%p)\n", | 135 | "expected 0x%p (GTT map 0x%p-0x%p)\n", |
| 132 | i, *vram_start, gtt_start, gtt_map, | 136 | i, *vram_start, gtt_start, gtt_map, |
| 133 | gtt_end); | 137 | gtt_end); |
| 134 | radeon_object_kunmap(vram_obj); | 138 | radeon_bo_kunmap(vram_obj); |
| 135 | goto out_cleanup; | 139 | goto out_cleanup; |
| 136 | } | 140 | } |
| 137 | *vram_start = vram_start; | 141 | *vram_start = vram_start; |
| 138 | } | 142 | } |
| 139 | 143 | ||
| 140 | radeon_object_kunmap(vram_obj); | 144 | radeon_bo_kunmap(vram_obj); |
| 141 | 145 | ||
| 142 | r = radeon_fence_create(rdev, &fence); | 146 | r = radeon_fence_create(rdev, &fence); |
| 143 | if (r) { | 147 | if (r) { |
| @@ -159,7 +163,7 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 159 | 163 | ||
| 160 | radeon_fence_unref(&fence); | 164 | radeon_fence_unref(&fence); |
| 161 | 165 | ||
| 162 | r = radeon_object_kmap(gtt_obj[i], >t_map); | 166 | r = radeon_bo_kmap(gtt_obj[i], >t_map); |
| 163 | if (r) { | 167 | if (r) { |
| 164 | DRM_ERROR("Failed to map GTT object after copy %d\n", i); | 168 | DRM_ERROR("Failed to map GTT object after copy %d\n", i); |
| 165 | goto out_cleanup; | 169 | goto out_cleanup; |
| @@ -174,12 +178,12 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 174 | "expected 0x%p (VRAM map 0x%p-0x%p)\n", | 178 | "expected 0x%p (VRAM map 0x%p-0x%p)\n", |
| 175 | i, *gtt_start, vram_start, vram_map, | 179 | i, *gtt_start, vram_start, vram_map, |
| 176 | vram_end); | 180 | vram_end); |
| 177 | radeon_object_kunmap(gtt_obj[i]); | 181 | radeon_bo_kunmap(gtt_obj[i]); |
| 178 | goto out_cleanup; | 182 | goto out_cleanup; |
| 179 | } | 183 | } |
| 180 | } | 184 | } |
| 181 | 185 | ||
| 182 | radeon_object_kunmap(gtt_obj[i]); | 186 | radeon_bo_kunmap(gtt_obj[i]); |
| 183 | 187 | ||
| 184 | DRM_INFO("Tested GTT->VRAM and VRAM->GTT copy for GTT offset 0x%llx\n", | 188 | DRM_INFO("Tested GTT->VRAM and VRAM->GTT copy for GTT offset 0x%llx\n", |
| 185 | gtt_addr - rdev->mc.gtt_location); | 189 | gtt_addr - rdev->mc.gtt_location); |
| @@ -187,14 +191,20 @@ void radeon_test_moves(struct radeon_device *rdev) | |||
| 187 | 191 | ||
| 188 | out_cleanup: | 192 | out_cleanup: |
| 189 | if (vram_obj) { | 193 | if (vram_obj) { |
| 190 | radeon_object_unpin(vram_obj); | 194 | if (radeon_bo_is_reserved(vram_obj)) { |
| 191 | radeon_object_unref(&vram_obj); | 195 | radeon_bo_unpin(vram_obj); |
| 196 | radeon_bo_unreserve(vram_obj); | ||
| 197 | } | ||
| 198 | radeon_bo_unref(&vram_obj); | ||
| 192 | } | 199 | } |
| 193 | if (gtt_obj) { | 200 | if (gtt_obj) { |
| 194 | for (i = 0; i < n; i++) { | 201 | for (i = 0; i < n; i++) { |
| 195 | if (gtt_obj[i]) { | 202 | if (gtt_obj[i]) { |
| 196 | radeon_object_unpin(gtt_obj[i]); | 203 | if (radeon_bo_is_reserved(gtt_obj[i])) { |
| 197 | radeon_object_unref(>t_obj[i]); | 204 | radeon_bo_unpin(gtt_obj[i]); |
| 205 | radeon_bo_unreserve(gtt_obj[i]); | ||
| 206 | } | ||
| 207 | radeon_bo_unref(>t_obj[i]); | ||
| 198 | } | 208 | } |
| 199 | } | 209 | } |
| 200 | kfree(gtt_obj); | 210 | kfree(gtt_obj); |
| @@ -206,4 +216,3 @@ out_cleanup: | |||
| 206 | printk(KERN_WARNING "Error while testing BO move.\n"); | 216 | printk(KERN_WARNING "Error while testing BO move.\n"); |
| 207 | } | 217 | } |
| 208 | } | 218 | } |
| 209 | |||
diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c index eda4ade24c3a..5a19d529d1c0 100644 --- a/drivers/gpu/drm/radeon/radeon_ttm.c +++ b/drivers/gpu/drm/radeon/radeon_ttm.c | |||
| @@ -150,7 +150,7 @@ static int radeon_init_mem_type(struct ttm_bo_device *bdev, uint32_t type, | |||
| 150 | man->default_caching = TTM_PL_FLAG_CACHED; | 150 | man->default_caching = TTM_PL_FLAG_CACHED; |
| 151 | break; | 151 | break; |
| 152 | case TTM_PL_TT: | 152 | case TTM_PL_TT: |
| 153 | man->gpu_offset = 0; | 153 | man->gpu_offset = rdev->mc.gtt_location; |
| 154 | man->available_caching = TTM_PL_MASK_CACHING; | 154 | man->available_caching = TTM_PL_MASK_CACHING; |
| 155 | man->default_caching = TTM_PL_FLAG_CACHED; | 155 | man->default_caching = TTM_PL_FLAG_CACHED; |
| 156 | man->flags = TTM_MEMTYPE_FLAG_MAPPABLE | TTM_MEMTYPE_FLAG_CMA; | 156 | man->flags = TTM_MEMTYPE_FLAG_MAPPABLE | TTM_MEMTYPE_FLAG_CMA; |
| @@ -180,7 +180,7 @@ static int radeon_init_mem_type(struct ttm_bo_device *bdev, uint32_t type, | |||
| 180 | break; | 180 | break; |
| 181 | case TTM_PL_VRAM: | 181 | case TTM_PL_VRAM: |
| 182 | /* "On-card" video ram */ | 182 | /* "On-card" video ram */ |
| 183 | man->gpu_offset = 0; | 183 | man->gpu_offset = rdev->mc.vram_location; |
| 184 | man->flags = TTM_MEMTYPE_FLAG_FIXED | | 184 | man->flags = TTM_MEMTYPE_FLAG_FIXED | |
| 185 | TTM_MEMTYPE_FLAG_NEEDS_IOREMAP | | 185 | TTM_MEMTYPE_FLAG_NEEDS_IOREMAP | |
| 186 | TTM_MEMTYPE_FLAG_MAPPABLE; | 186 | TTM_MEMTYPE_FLAG_MAPPABLE; |
| @@ -197,16 +197,19 @@ static int radeon_init_mem_type(struct ttm_bo_device *bdev, uint32_t type, | |||
| 197 | return 0; | 197 | return 0; |
| 198 | } | 198 | } |
| 199 | 199 | ||
| 200 | static uint32_t radeon_evict_flags(struct ttm_buffer_object *bo) | 200 | static void radeon_evict_flags(struct ttm_buffer_object *bo, |
| 201 | struct ttm_placement *placement) | ||
| 201 | { | 202 | { |
| 202 | uint32_t cur_placement = bo->mem.placement & ~TTM_PL_MASK_MEMTYPE; | 203 | struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo); |
| 203 | |||
| 204 | switch (bo->mem.mem_type) { | 204 | switch (bo->mem.mem_type) { |
| 205 | case TTM_PL_VRAM: | ||
| 206 | radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_GTT); | ||
| 207 | break; | ||
| 208 | case TTM_PL_TT: | ||
| 205 | default: | 209 | default: |
| 206 | return (cur_placement & ~TTM_PL_MASK_CACHING) | | 210 | radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_CPU); |
| 207 | TTM_PL_FLAG_SYSTEM | | ||
| 208 | TTM_PL_FLAG_CACHED; | ||
| 209 | } | 211 | } |
| 212 | *placement = rbo->placement; | ||
| 210 | } | 213 | } |
| 211 | 214 | ||
| 212 | static int radeon_verify_access(struct ttm_buffer_object *bo, struct file *filp) | 215 | static int radeon_verify_access(struct ttm_buffer_object *bo, struct file *filp) |
| @@ -283,14 +286,21 @@ static int radeon_move_vram_ram(struct ttm_buffer_object *bo, | |||
| 283 | struct radeon_device *rdev; | 286 | struct radeon_device *rdev; |
| 284 | struct ttm_mem_reg *old_mem = &bo->mem; | 287 | struct ttm_mem_reg *old_mem = &bo->mem; |
| 285 | struct ttm_mem_reg tmp_mem; | 288 | struct ttm_mem_reg tmp_mem; |
| 286 | uint32_t proposed_placement; | 289 | u32 placements; |
| 290 | struct ttm_placement placement; | ||
| 287 | int r; | 291 | int r; |
| 288 | 292 | ||
| 289 | rdev = radeon_get_rdev(bo->bdev); | 293 | rdev = radeon_get_rdev(bo->bdev); |
| 290 | tmp_mem = *new_mem; | 294 | tmp_mem = *new_mem; |
| 291 | tmp_mem.mm_node = NULL; | 295 | tmp_mem.mm_node = NULL; |
| 292 | proposed_placement = TTM_PL_FLAG_TT | TTM_PL_MASK_CACHING; | 296 | placement.fpfn = 0; |
| 293 | r = ttm_bo_mem_space(bo, proposed_placement, &tmp_mem, | 297 | placement.lpfn = 0; |
| 298 | placement.num_placement = 1; | ||
| 299 | placement.placement = &placements; | ||
| 300 | placement.num_busy_placement = 1; | ||
| 301 | placement.busy_placement = &placements; | ||
| 302 | placements = TTM_PL_MASK_CACHING | TTM_PL_FLAG_TT; | ||
| 303 | r = ttm_bo_mem_space(bo, &placement, &tmp_mem, | ||
| 294 | interruptible, no_wait); | 304 | interruptible, no_wait); |
| 295 | if (unlikely(r)) { | 305 | if (unlikely(r)) { |
| 296 | return r; | 306 | return r; |
| @@ -329,15 +339,21 @@ static int radeon_move_ram_vram(struct ttm_buffer_object *bo, | |||
| 329 | struct radeon_device *rdev; | 339 | struct radeon_device *rdev; |
| 330 | struct ttm_mem_reg *old_mem = &bo->mem; | 340 | struct ttm_mem_reg *old_mem = &bo->mem; |
| 331 | struct ttm_mem_reg tmp_mem; | 341 | struct ttm_mem_reg tmp_mem; |
| 332 | uint32_t proposed_flags; | 342 | struct ttm_placement placement; |
| 343 | u32 placements; | ||
| 333 | int r; | 344 | int r; |
| 334 | 345 | ||
| 335 | rdev = radeon_get_rdev(bo->bdev); | 346 | rdev = radeon_get_rdev(bo->bdev); |
| 336 | tmp_mem = *new_mem; | 347 | tmp_mem = *new_mem; |
| 337 | tmp_mem.mm_node = NULL; | 348 | tmp_mem.mm_node = NULL; |
| 338 | proposed_flags = TTM_PL_FLAG_TT | TTM_PL_MASK_CACHING; | 349 | placement.fpfn = 0; |
| 339 | r = ttm_bo_mem_space(bo, proposed_flags, &tmp_mem, | 350 | placement.lpfn = 0; |
| 340 | interruptible, no_wait); | 351 | placement.num_placement = 1; |
| 352 | placement.placement = &placements; | ||
| 353 | placement.num_busy_placement = 1; | ||
| 354 | placement.busy_placement = &placements; | ||
| 355 | placements = TTM_PL_MASK_CACHING | TTM_PL_FLAG_TT; | ||
| 356 | r = ttm_bo_mem_space(bo, &placement, &tmp_mem, interruptible, no_wait); | ||
| 341 | if (unlikely(r)) { | 357 | if (unlikely(r)) { |
| 342 | return r; | 358 | return r; |
| 343 | } | 359 | } |
| @@ -407,18 +423,6 @@ memcpy: | |||
| 407 | return r; | 423 | return r; |
| 408 | } | 424 | } |
| 409 | 425 | ||
| 410 | const uint32_t radeon_mem_prios[] = { | ||
| 411 | TTM_PL_VRAM, | ||
| 412 | TTM_PL_TT, | ||
| 413 | TTM_PL_SYSTEM, | ||
| 414 | }; | ||
| 415 | |||
| 416 | const uint32_t radeon_busy_prios[] = { | ||
| 417 | TTM_PL_TT, | ||
| 418 | TTM_PL_VRAM, | ||
| 419 | TTM_PL_SYSTEM, | ||
| 420 | }; | ||
| 421 | |||
| 422 | static int radeon_sync_obj_wait(void *sync_obj, void *sync_arg, | 426 | static int radeon_sync_obj_wait(void *sync_obj, void *sync_arg, |
| 423 | bool lazy, bool interruptible) | 427 | bool lazy, bool interruptible) |
| 424 | { | 428 | { |
| @@ -446,10 +450,6 @@ static bool radeon_sync_obj_signaled(void *sync_obj, void *sync_arg) | |||
| 446 | } | 450 | } |
| 447 | 451 | ||
| 448 | static struct ttm_bo_driver radeon_bo_driver = { | 452 | static struct ttm_bo_driver radeon_bo_driver = { |
| 449 | .mem_type_prio = radeon_mem_prios, | ||
| 450 | .mem_busy_prio = radeon_busy_prios, | ||
| 451 | .num_mem_type_prio = ARRAY_SIZE(radeon_mem_prios), | ||
| 452 | .num_mem_busy_prio = ARRAY_SIZE(radeon_busy_prios), | ||
| 453 | .create_ttm_backend_entry = &radeon_create_ttm_backend_entry, | 453 | .create_ttm_backend_entry = &radeon_create_ttm_backend_entry, |
| 454 | .invalidate_caches = &radeon_invalidate_caches, | 454 | .invalidate_caches = &radeon_invalidate_caches, |
| 455 | .init_mem_type = &radeon_init_mem_type, | 455 | .init_mem_type = &radeon_init_mem_type, |
| @@ -482,27 +482,31 @@ int radeon_ttm_init(struct radeon_device *rdev) | |||
| 482 | DRM_ERROR("failed initializing buffer object driver(%d).\n", r); | 482 | DRM_ERROR("failed initializing buffer object driver(%d).\n", r); |
| 483 | return r; | 483 | return r; |
| 484 | } | 484 | } |
| 485 | r = ttm_bo_init_mm(&rdev->mman.bdev, TTM_PL_VRAM, 0, | 485 | r = ttm_bo_init_mm(&rdev->mman.bdev, TTM_PL_VRAM, |
| 486 | ((rdev->mc.real_vram_size) >> PAGE_SHIFT)); | 486 | rdev->mc.real_vram_size >> PAGE_SHIFT); |
| 487 | if (r) { | 487 | if (r) { |
| 488 | DRM_ERROR("Failed initializing VRAM heap.\n"); | 488 | DRM_ERROR("Failed initializing VRAM heap.\n"); |
| 489 | return r; | 489 | return r; |
| 490 | } | 490 | } |
| 491 | r = radeon_object_create(rdev, NULL, 256 * 1024, true, | 491 | r = radeon_bo_create(rdev, NULL, 256 * 1024, true, |
| 492 | RADEON_GEM_DOMAIN_VRAM, false, | 492 | RADEON_GEM_DOMAIN_VRAM, |
| 493 | &rdev->stollen_vga_memory); | 493 | &rdev->stollen_vga_memory); |
| 494 | if (r) { | 494 | if (r) { |
| 495 | return r; | 495 | return r; |
| 496 | } | 496 | } |
| 497 | r = radeon_object_pin(rdev->stollen_vga_memory, RADEON_GEM_DOMAIN_VRAM, NULL); | 497 | r = radeon_bo_reserve(rdev->stollen_vga_memory, false); |
| 498 | if (r) | ||
| 499 | return r; | ||
| 500 | r = radeon_bo_pin(rdev->stollen_vga_memory, RADEON_GEM_DOMAIN_VRAM, NULL); | ||
| 501 | radeon_bo_unreserve(rdev->stollen_vga_memory); | ||
| 498 | if (r) { | 502 | if (r) { |
| 499 | radeon_object_unref(&rdev->stollen_vga_memory); | 503 | radeon_bo_unref(&rdev->stollen_vga_memory); |
| 500 | return r; | 504 | return r; |
| 501 | } | 505 | } |
| 502 | DRM_INFO("radeon: %uM of VRAM memory ready\n", | 506 | DRM_INFO("radeon: %uM of VRAM memory ready\n", |
| 503 | (unsigned)rdev->mc.real_vram_size / (1024 * 1024)); | 507 | (unsigned)rdev->mc.real_vram_size / (1024 * 1024)); |
| 504 | r = ttm_bo_init_mm(&rdev->mman.bdev, TTM_PL_TT, 0, | 508 | r = ttm_bo_init_mm(&rdev->mman.bdev, TTM_PL_TT, |
| 505 | ((rdev->mc.gtt_size) >> PAGE_SHIFT)); | 509 | rdev->mc.gtt_size >> PAGE_SHIFT); |
| 506 | if (r) { | 510 | if (r) { |
| 507 | DRM_ERROR("Failed initializing GTT heap.\n"); | 511 | DRM_ERROR("Failed initializing GTT heap.\n"); |
| 508 | return r; | 512 | return r; |
| @@ -523,9 +527,15 @@ int radeon_ttm_init(struct radeon_device *rdev) | |||
| 523 | 527 | ||
| 524 | void radeon_ttm_fini(struct radeon_device *rdev) | 528 | void radeon_ttm_fini(struct radeon_device *rdev) |
| 525 | { | 529 | { |
| 530 | int r; | ||
| 531 | |||
| 526 | if (rdev->stollen_vga_memory) { | 532 | if (rdev->stollen_vga_memory) { |
| 527 | radeon_object_unpin(rdev->stollen_vga_memory); | 533 | r = radeon_bo_reserve(rdev->stollen_vga_memory, false); |
| 528 | radeon_object_unref(&rdev->stollen_vga_memory); | 534 | if (r == 0) { |
| 535 | radeon_bo_unpin(rdev->stollen_vga_memory); | ||
| 536 | radeon_bo_unreserve(rdev->stollen_vga_memory); | ||
| 537 | } | ||
| 538 | radeon_bo_unref(&rdev->stollen_vga_memory); | ||
| 529 | } | 539 | } |
| 530 | ttm_bo_clean_mm(&rdev->mman.bdev, TTM_PL_VRAM); | 540 | ttm_bo_clean_mm(&rdev->mman.bdev, TTM_PL_VRAM); |
| 531 | ttm_bo_clean_mm(&rdev->mman.bdev, TTM_PL_TT); | 541 | ttm_bo_clean_mm(&rdev->mman.bdev, TTM_PL_TT); |
diff --git a/drivers/gpu/drm/radeon/rs400.c b/drivers/gpu/drm/radeon/rs400.c index ca037160a582..c1fcdddb6be6 100644 --- a/drivers/gpu/drm/radeon/rs400.c +++ b/drivers/gpu/drm/radeon/rs400.c | |||
| @@ -352,7 +352,7 @@ static int rs400_mc_init(struct radeon_device *rdev) | |||
| 352 | u32 tmp; | 352 | u32 tmp; |
| 353 | 353 | ||
| 354 | /* Setup GPU memory space */ | 354 | /* Setup GPU memory space */ |
| 355 | tmp = G_00015C_MC_FB_START(RREG32(R_00015C_NB_TOM)); | 355 | tmp = RREG32(R_00015C_NB_TOM); |
| 356 | rdev->mc.vram_location = G_00015C_MC_FB_START(tmp) << 16; | 356 | rdev->mc.vram_location = G_00015C_MC_FB_START(tmp) << 16; |
| 357 | rdev->mc.gtt_location = 0xFFFFFFFFUL; | 357 | rdev->mc.gtt_location = 0xFFFFFFFFUL; |
| 358 | r = radeon_mc_setup(rdev); | 358 | r = radeon_mc_setup(rdev); |
| @@ -387,13 +387,13 @@ static int rs400_startup(struct radeon_device *rdev) | |||
| 387 | r300_clock_startup(rdev); | 387 | r300_clock_startup(rdev); |
| 388 | /* Initialize GPU configuration (# pipes, ...) */ | 388 | /* Initialize GPU configuration (# pipes, ...) */ |
| 389 | rs400_gpu_init(rdev); | 389 | rs400_gpu_init(rdev); |
| 390 | r100_enable_bm(rdev); | ||
| 390 | /* Initialize GART (initialize after TTM so we can allocate | 391 | /* Initialize GART (initialize after TTM so we can allocate |
| 391 | * memory through TTM but finalize after TTM) */ | 392 | * memory through TTM but finalize after TTM) */ |
| 392 | r = rs400_gart_enable(rdev); | 393 | r = rs400_gart_enable(rdev); |
| 393 | if (r) | 394 | if (r) |
| 394 | return r; | 395 | return r; |
| 395 | /* Enable IRQ */ | 396 | /* Enable IRQ */ |
| 396 | rdev->irq.sw_int = true; | ||
| 397 | r100_irq_set(rdev); | 397 | r100_irq_set(rdev); |
| 398 | /* 1M ring buffer */ | 398 | /* 1M ring buffer */ |
| 399 | r = r100_cp_init(rdev, 1024 * 1024); | 399 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -430,6 +430,8 @@ int rs400_resume(struct radeon_device *rdev) | |||
| 430 | radeon_combios_asic_init(rdev->ddev); | 430 | radeon_combios_asic_init(rdev->ddev); |
| 431 | /* Resume clock after posting */ | 431 | /* Resume clock after posting */ |
| 432 | r300_clock_startup(rdev); | 432 | r300_clock_startup(rdev); |
| 433 | /* Initialize surface registers */ | ||
| 434 | radeon_surface_init(rdev); | ||
| 433 | return rs400_startup(rdev); | 435 | return rs400_startup(rdev); |
| 434 | } | 436 | } |
| 435 | 437 | ||
| @@ -452,7 +454,7 @@ void rs400_fini(struct radeon_device *rdev) | |||
| 452 | rs400_gart_fini(rdev); | 454 | rs400_gart_fini(rdev); |
| 453 | radeon_irq_kms_fini(rdev); | 455 | radeon_irq_kms_fini(rdev); |
| 454 | radeon_fence_driver_fini(rdev); | 456 | radeon_fence_driver_fini(rdev); |
| 455 | radeon_object_fini(rdev); | 457 | radeon_bo_fini(rdev); |
| 456 | radeon_atombios_fini(rdev); | 458 | radeon_atombios_fini(rdev); |
| 457 | kfree(rdev->bios); | 459 | kfree(rdev->bios); |
| 458 | rdev->bios = NULL; | 460 | rdev->bios = NULL; |
| @@ -490,10 +492,9 @@ int rs400_init(struct radeon_device *rdev) | |||
| 490 | RREG32(R_0007C0_CP_STAT)); | 492 | RREG32(R_0007C0_CP_STAT)); |
| 491 | } | 493 | } |
| 492 | /* check if cards are posted or not */ | 494 | /* check if cards are posted or not */ |
| 493 | if (!radeon_card_posted(rdev) && rdev->bios) { | 495 | if (radeon_boot_test_post_card(rdev) == false) |
| 494 | DRM_INFO("GPU not posted. posting now...\n"); | 496 | return -EINVAL; |
| 495 | radeon_combios_asic_init(rdev->ddev); | 497 | |
| 496 | } | ||
| 497 | /* Initialize clocks */ | 498 | /* Initialize clocks */ |
| 498 | radeon_get_clock_info(rdev->ddev); | 499 | radeon_get_clock_info(rdev->ddev); |
| 499 | /* Get vram informations */ | 500 | /* Get vram informations */ |
| @@ -510,7 +511,7 @@ int rs400_init(struct radeon_device *rdev) | |||
| 510 | if (r) | 511 | if (r) |
| 511 | return r; | 512 | return r; |
| 512 | /* Memory manager */ | 513 | /* Memory manager */ |
| 513 | r = radeon_object_init(rdev); | 514 | r = radeon_bo_init(rdev); |
| 514 | if (r) | 515 | if (r) |
| 515 | return r; | 516 | return r; |
| 516 | r = rs400_gart_init(rdev); | 517 | r = rs400_gart_init(rdev); |
diff --git a/drivers/gpu/drm/radeon/rs600.c b/drivers/gpu/drm/radeon/rs600.c index 5f117cd8736a..4f8ea4260572 100644 --- a/drivers/gpu/drm/radeon/rs600.c +++ b/drivers/gpu/drm/radeon/rs600.c | |||
| @@ -45,6 +45,122 @@ | |||
| 45 | void rs600_gpu_init(struct radeon_device *rdev); | 45 | void rs600_gpu_init(struct radeon_device *rdev); |
| 46 | int rs600_mc_wait_for_idle(struct radeon_device *rdev); | 46 | int rs600_mc_wait_for_idle(struct radeon_device *rdev); |
| 47 | 47 | ||
| 48 | int rs600_mc_init(struct radeon_device *rdev) | ||
| 49 | { | ||
| 50 | /* read back the MC value from the hw */ | ||
| 51 | int r; | ||
| 52 | u32 tmp; | ||
| 53 | |||
| 54 | /* Setup GPU memory space */ | ||
| 55 | tmp = RREG32_MC(R_000004_MC_FB_LOCATION); | ||
| 56 | rdev->mc.vram_location = G_000004_MC_FB_START(tmp) << 16; | ||
| 57 | rdev->mc.gtt_location = 0xffffffffUL; | ||
| 58 | r = radeon_mc_setup(rdev); | ||
| 59 | if (r) | ||
| 60 | return r; | ||
| 61 | return 0; | ||
| 62 | } | ||
| 63 | |||
| 64 | /* hpd for digital panel detect/disconnect */ | ||
| 65 | bool rs600_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd) | ||
| 66 | { | ||
| 67 | u32 tmp; | ||
| 68 | bool connected = false; | ||
| 69 | |||
| 70 | switch (hpd) { | ||
| 71 | case RADEON_HPD_1: | ||
| 72 | tmp = RREG32(R_007D04_DC_HOT_PLUG_DETECT1_INT_STATUS); | ||
| 73 | if (G_007D04_DC_HOT_PLUG_DETECT1_SENSE(tmp)) | ||
| 74 | connected = true; | ||
| 75 | break; | ||
| 76 | case RADEON_HPD_2: | ||
| 77 | tmp = RREG32(R_007D14_DC_HOT_PLUG_DETECT2_INT_STATUS); | ||
| 78 | if (G_007D14_DC_HOT_PLUG_DETECT2_SENSE(tmp)) | ||
| 79 | connected = true; | ||
| 80 | break; | ||
| 81 | default: | ||
| 82 | break; | ||
| 83 | } | ||
| 84 | return connected; | ||
| 85 | } | ||
| 86 | |||
| 87 | void rs600_hpd_set_polarity(struct radeon_device *rdev, | ||
| 88 | enum radeon_hpd_id hpd) | ||
| 89 | { | ||
| 90 | u32 tmp; | ||
| 91 | bool connected = rs600_hpd_sense(rdev, hpd); | ||
| 92 | |||
| 93 | switch (hpd) { | ||
| 94 | case RADEON_HPD_1: | ||
| 95 | tmp = RREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL); | ||
| 96 | if (connected) | ||
| 97 | tmp &= ~S_007D08_DC_HOT_PLUG_DETECT1_INT_POLARITY(1); | ||
| 98 | else | ||
| 99 | tmp |= S_007D08_DC_HOT_PLUG_DETECT1_INT_POLARITY(1); | ||
| 100 | WREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL, tmp); | ||
| 101 | break; | ||
| 102 | case RADEON_HPD_2: | ||
| 103 | tmp = RREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL); | ||
| 104 | if (connected) | ||
| 105 | tmp &= ~S_007D18_DC_HOT_PLUG_DETECT2_INT_POLARITY(1); | ||
| 106 | else | ||
| 107 | tmp |= S_007D18_DC_HOT_PLUG_DETECT2_INT_POLARITY(1); | ||
| 108 | WREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL, tmp); | ||
| 109 | break; | ||
| 110 | default: | ||
| 111 | break; | ||
| 112 | } | ||
| 113 | } | ||
| 114 | |||
| 115 | void rs600_hpd_init(struct radeon_device *rdev) | ||
| 116 | { | ||
| 117 | struct drm_device *dev = rdev->ddev; | ||
| 118 | struct drm_connector *connector; | ||
| 119 | |||
| 120 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 121 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 122 | switch (radeon_connector->hpd.hpd) { | ||
| 123 | case RADEON_HPD_1: | ||
| 124 | WREG32(R_007D00_DC_HOT_PLUG_DETECT1_CONTROL, | ||
| 125 | S_007D00_DC_HOT_PLUG_DETECT1_EN(1)); | ||
| 126 | rdev->irq.hpd[0] = true; | ||
| 127 | break; | ||
| 128 | case RADEON_HPD_2: | ||
| 129 | WREG32(R_007D10_DC_HOT_PLUG_DETECT2_CONTROL, | ||
| 130 | S_007D10_DC_HOT_PLUG_DETECT2_EN(1)); | ||
| 131 | rdev->irq.hpd[1] = true; | ||
| 132 | break; | ||
| 133 | default: | ||
| 134 | break; | ||
| 135 | } | ||
| 136 | } | ||
| 137 | rs600_irq_set(rdev); | ||
| 138 | } | ||
| 139 | |||
| 140 | void rs600_hpd_fini(struct radeon_device *rdev) | ||
| 141 | { | ||
| 142 | struct drm_device *dev = rdev->ddev; | ||
| 143 | struct drm_connector *connector; | ||
| 144 | |||
| 145 | list_for_each_entry(connector, &dev->mode_config.connector_list, head) { | ||
| 146 | struct radeon_connector *radeon_connector = to_radeon_connector(connector); | ||
| 147 | switch (radeon_connector->hpd.hpd) { | ||
| 148 | case RADEON_HPD_1: | ||
| 149 | WREG32(R_007D00_DC_HOT_PLUG_DETECT1_CONTROL, | ||
| 150 | S_007D00_DC_HOT_PLUG_DETECT1_EN(0)); | ||
| 151 | rdev->irq.hpd[0] = false; | ||
| 152 | break; | ||
| 153 | case RADEON_HPD_2: | ||
| 154 | WREG32(R_007D10_DC_HOT_PLUG_DETECT2_CONTROL, | ||
| 155 | S_007D10_DC_HOT_PLUG_DETECT2_EN(0)); | ||
| 156 | rdev->irq.hpd[1] = false; | ||
| 157 | break; | ||
| 158 | default: | ||
| 159 | break; | ||
| 160 | } | ||
| 161 | } | ||
| 162 | } | ||
| 163 | |||
| 48 | /* | 164 | /* |
| 49 | * GART. | 165 | * GART. |
| 50 | */ | 166 | */ |
| @@ -100,40 +216,40 @@ int rs600_gart_enable(struct radeon_device *rdev) | |||
| 100 | WREG32(R_00004C_BUS_CNTL, tmp); | 216 | WREG32(R_00004C_BUS_CNTL, tmp); |
| 101 | /* FIXME: setup default page */ | 217 | /* FIXME: setup default page */ |
| 102 | WREG32_MC(R_000100_MC_PT0_CNTL, | 218 | WREG32_MC(R_000100_MC_PT0_CNTL, |
| 103 | (S_000100_EFFECTIVE_L2_CACHE_SIZE(6) | | 219 | (S_000100_EFFECTIVE_L2_CACHE_SIZE(6) | |
| 104 | S_000100_EFFECTIVE_L2_QUEUE_SIZE(6))); | 220 | S_000100_EFFECTIVE_L2_QUEUE_SIZE(6))); |
| 221 | |||
| 105 | for (i = 0; i < 19; i++) { | 222 | for (i = 0; i < 19; i++) { |
| 106 | WREG32_MC(R_00016C_MC_PT0_CLIENT0_CNTL + i, | 223 | WREG32_MC(R_00016C_MC_PT0_CLIENT0_CNTL + i, |
| 107 | S_00016C_ENABLE_TRANSLATION_MODE_OVERRIDE(1) | | 224 | S_00016C_ENABLE_TRANSLATION_MODE_OVERRIDE(1) | |
| 108 | S_00016C_SYSTEM_ACCESS_MODE_MASK( | 225 | S_00016C_SYSTEM_ACCESS_MODE_MASK( |
| 109 | V_00016C_SYSTEM_ACCESS_MODE_IN_SYS) | | 226 | V_00016C_SYSTEM_ACCESS_MODE_NOT_IN_SYS) | |
| 110 | S_00016C_SYSTEM_APERTURE_UNMAPPED_ACCESS( | 227 | S_00016C_SYSTEM_APERTURE_UNMAPPED_ACCESS( |
| 111 | V_00016C_SYSTEM_APERTURE_UNMAPPED_DEFAULT_PAGE) | | 228 | V_00016C_SYSTEM_APERTURE_UNMAPPED_PASSTHROUGH) | |
| 112 | S_00016C_EFFECTIVE_L1_CACHE_SIZE(1) | | 229 | S_00016C_EFFECTIVE_L1_CACHE_SIZE(3) | |
| 113 | S_00016C_ENABLE_FRAGMENT_PROCESSING(1) | | 230 | S_00016C_ENABLE_FRAGMENT_PROCESSING(1) | |
| 114 | S_00016C_EFFECTIVE_L1_QUEUE_SIZE(1)); | 231 | S_00016C_EFFECTIVE_L1_QUEUE_SIZE(3)); |
| 115 | } | 232 | } |
| 116 | |||
| 117 | /* System context map to GART space */ | ||
| 118 | WREG32_MC(R_000112_MC_PT0_SYSTEM_APERTURE_LOW_ADDR, rdev->mc.gtt_start); | ||
| 119 | WREG32_MC(R_000114_MC_PT0_SYSTEM_APERTURE_HIGH_ADDR, rdev->mc.gtt_end); | ||
| 120 | |||
| 121 | /* enable first context */ | 233 | /* enable first context */ |
| 122 | WREG32_MC(R_00013C_MC_PT0_CONTEXT0_FLAT_START_ADDR, rdev->mc.gtt_start); | ||
| 123 | WREG32_MC(R_00014C_MC_PT0_CONTEXT0_FLAT_END_ADDR, rdev->mc.gtt_end); | ||
| 124 | WREG32_MC(R_000102_MC_PT0_CONTEXT0_CNTL, | 234 | WREG32_MC(R_000102_MC_PT0_CONTEXT0_CNTL, |
| 125 | S_000102_ENABLE_PAGE_TABLE(1) | | 235 | S_000102_ENABLE_PAGE_TABLE(1) | |
| 126 | S_000102_PAGE_TABLE_DEPTH(V_000102_PAGE_TABLE_FLAT)); | 236 | S_000102_PAGE_TABLE_DEPTH(V_000102_PAGE_TABLE_FLAT)); |
| 237 | |||
| 127 | /* disable all other contexts */ | 238 | /* disable all other contexts */ |
| 128 | for (i = 1; i < 8; i++) { | 239 | for (i = 1; i < 8; i++) |
| 129 | WREG32_MC(R_000102_MC_PT0_CONTEXT0_CNTL + i, 0); | 240 | WREG32_MC(R_000102_MC_PT0_CONTEXT0_CNTL + i, 0); |
| 130 | } | ||
| 131 | 241 | ||
| 132 | /* setup the page table */ | 242 | /* setup the page table */ |
| 133 | WREG32_MC(R_00012C_MC_PT0_CONTEXT0_FLAT_BASE_ADDR, | 243 | WREG32_MC(R_00012C_MC_PT0_CONTEXT0_FLAT_BASE_ADDR, |
| 134 | rdev->gart.table_addr); | 244 | rdev->gart.table_addr); |
| 245 | WREG32_MC(R_00013C_MC_PT0_CONTEXT0_FLAT_START_ADDR, rdev->mc.gtt_start); | ||
| 246 | WREG32_MC(R_00014C_MC_PT0_CONTEXT0_FLAT_END_ADDR, rdev->mc.gtt_end); | ||
| 135 | WREG32_MC(R_00011C_MC_PT0_CONTEXT0_DEFAULT_READ_ADDR, 0); | 247 | WREG32_MC(R_00011C_MC_PT0_CONTEXT0_DEFAULT_READ_ADDR, 0); |
| 136 | 248 | ||
| 249 | /* System context maps to VRAM space */ | ||
| 250 | WREG32_MC(R_000112_MC_PT0_SYSTEM_APERTURE_LOW_ADDR, rdev->mc.vram_start); | ||
| 251 | WREG32_MC(R_000114_MC_PT0_SYSTEM_APERTURE_HIGH_ADDR, rdev->mc.vram_end); | ||
| 252 | |||
| 137 | /* enable page tables */ | 253 | /* enable page tables */ |
| 138 | tmp = RREG32_MC(R_000100_MC_PT0_CNTL); | 254 | tmp = RREG32_MC(R_000100_MC_PT0_CNTL); |
| 139 | WREG32_MC(R_000100_MC_PT0_CNTL, (tmp | S_000100_ENABLE_PT(1))); | 255 | WREG32_MC(R_000100_MC_PT0_CNTL, (tmp | S_000100_ENABLE_PT(1))); |
| @@ -146,15 +262,20 @@ int rs600_gart_enable(struct radeon_device *rdev) | |||
| 146 | 262 | ||
| 147 | void rs600_gart_disable(struct radeon_device *rdev) | 263 | void rs600_gart_disable(struct radeon_device *rdev) |
| 148 | { | 264 | { |
| 149 | uint32_t tmp; | 265 | u32 tmp; |
| 266 | int r; | ||
| 150 | 267 | ||
| 151 | /* FIXME: disable out of gart access */ | 268 | /* FIXME: disable out of gart access */ |
| 152 | WREG32_MC(R_000100_MC_PT0_CNTL, 0); | 269 | WREG32_MC(R_000100_MC_PT0_CNTL, 0); |
| 153 | tmp = RREG32_MC(R_000009_MC_CNTL1); | 270 | tmp = RREG32_MC(R_000009_MC_CNTL1); |
| 154 | WREG32_MC(R_000009_MC_CNTL1, tmp & C_000009_ENABLE_PAGE_TABLES); | 271 | WREG32_MC(R_000009_MC_CNTL1, tmp & C_000009_ENABLE_PAGE_TABLES); |
| 155 | if (rdev->gart.table.vram.robj) { | 272 | if (rdev->gart.table.vram.robj) { |
| 156 | radeon_object_kunmap(rdev->gart.table.vram.robj); | 273 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 157 | radeon_object_unpin(rdev->gart.table.vram.robj); | 274 | if (r == 0) { |
| 275 | radeon_bo_kunmap(rdev->gart.table.vram.robj); | ||
| 276 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 277 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 278 | } | ||
| 158 | } | 279 | } |
| 159 | } | 280 | } |
| 160 | 281 | ||
| @@ -189,6 +310,10 @@ int rs600_irq_set(struct radeon_device *rdev) | |||
| 189 | { | 310 | { |
| 190 | uint32_t tmp = 0; | 311 | uint32_t tmp = 0; |
| 191 | uint32_t mode_int = 0; | 312 | uint32_t mode_int = 0; |
| 313 | u32 hpd1 = RREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL) & | ||
| 314 | ~S_007D08_DC_HOT_PLUG_DETECT1_INT_EN(1); | ||
| 315 | u32 hpd2 = RREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL) & | ||
| 316 | ~S_007D18_DC_HOT_PLUG_DETECT2_INT_EN(1); | ||
| 192 | 317 | ||
| 193 | if (rdev->irq.sw_int) { | 318 | if (rdev->irq.sw_int) { |
| 194 | tmp |= S_000040_SW_INT_EN(1); | 319 | tmp |= S_000040_SW_INT_EN(1); |
| @@ -199,8 +324,16 @@ int rs600_irq_set(struct radeon_device *rdev) | |||
| 199 | if (rdev->irq.crtc_vblank_int[1]) { | 324 | if (rdev->irq.crtc_vblank_int[1]) { |
| 200 | mode_int |= S_006540_D2MODE_VBLANK_INT_MASK(1); | 325 | mode_int |= S_006540_D2MODE_VBLANK_INT_MASK(1); |
| 201 | } | 326 | } |
| 327 | if (rdev->irq.hpd[0]) { | ||
| 328 | hpd1 |= S_007D08_DC_HOT_PLUG_DETECT1_INT_EN(1); | ||
| 329 | } | ||
| 330 | if (rdev->irq.hpd[1]) { | ||
| 331 | hpd2 |= S_007D18_DC_HOT_PLUG_DETECT2_INT_EN(1); | ||
| 332 | } | ||
| 202 | WREG32(R_000040_GEN_INT_CNTL, tmp); | 333 | WREG32(R_000040_GEN_INT_CNTL, tmp); |
| 203 | WREG32(R_006540_DxMODE_INT_MASK, mode_int); | 334 | WREG32(R_006540_DxMODE_INT_MASK, mode_int); |
| 335 | WREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL, hpd1); | ||
| 336 | WREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL, hpd2); | ||
| 204 | return 0; | 337 | return 0; |
| 205 | } | 338 | } |
| 206 | 339 | ||
| @@ -208,6 +341,7 @@ static inline uint32_t rs600_irq_ack(struct radeon_device *rdev, u32 *r500_disp_ | |||
| 208 | { | 341 | { |
| 209 | uint32_t irqs = RREG32(R_000044_GEN_INT_STATUS); | 342 | uint32_t irqs = RREG32(R_000044_GEN_INT_STATUS); |
| 210 | uint32_t irq_mask = ~C_000044_SW_INT; | 343 | uint32_t irq_mask = ~C_000044_SW_INT; |
| 344 | u32 tmp; | ||
| 211 | 345 | ||
| 212 | if (G_000044_DISPLAY_INT_STAT(irqs)) { | 346 | if (G_000044_DISPLAY_INT_STAT(irqs)) { |
| 213 | *r500_disp_int = RREG32(R_007EDC_DISP_INTERRUPT_STATUS); | 347 | *r500_disp_int = RREG32(R_007EDC_DISP_INTERRUPT_STATUS); |
| @@ -219,6 +353,16 @@ static inline uint32_t rs600_irq_ack(struct radeon_device *rdev, u32 *r500_disp_ | |||
| 219 | WREG32(R_006D34_D2MODE_VBLANK_STATUS, | 353 | WREG32(R_006D34_D2MODE_VBLANK_STATUS, |
| 220 | S_006D34_D2MODE_VBLANK_ACK(1)); | 354 | S_006D34_D2MODE_VBLANK_ACK(1)); |
| 221 | } | 355 | } |
| 356 | if (G_007EDC_DC_HOT_PLUG_DETECT1_INTERRUPT(*r500_disp_int)) { | ||
| 357 | tmp = RREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL); | ||
| 358 | tmp |= S_007D08_DC_HOT_PLUG_DETECT1_INT_ACK(1); | ||
| 359 | WREG32(R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL, tmp); | ||
| 360 | } | ||
| 361 | if (G_007EDC_DC_HOT_PLUG_DETECT2_INTERRUPT(*r500_disp_int)) { | ||
| 362 | tmp = RREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL); | ||
| 363 | tmp |= S_007D18_DC_HOT_PLUG_DETECT2_INT_ACK(1); | ||
| 364 | WREG32(R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL, tmp); | ||
| 365 | } | ||
| 222 | } else { | 366 | } else { |
| 223 | *r500_disp_int = 0; | 367 | *r500_disp_int = 0; |
| 224 | } | 368 | } |
| @@ -244,6 +388,7 @@ int rs600_irq_process(struct radeon_device *rdev) | |||
| 244 | { | 388 | { |
| 245 | uint32_t status, msi_rearm; | 389 | uint32_t status, msi_rearm; |
| 246 | uint32_t r500_disp_int; | 390 | uint32_t r500_disp_int; |
| 391 | bool queue_hotplug = false; | ||
| 247 | 392 | ||
| 248 | status = rs600_irq_ack(rdev, &r500_disp_int); | 393 | status = rs600_irq_ack(rdev, &r500_disp_int); |
| 249 | if (!status && !r500_disp_int) { | 394 | if (!status && !r500_disp_int) { |
| @@ -258,8 +403,18 @@ int rs600_irq_process(struct radeon_device *rdev) | |||
| 258 | drm_handle_vblank(rdev->ddev, 0); | 403 | drm_handle_vblank(rdev->ddev, 0); |
| 259 | if (G_007EDC_LB_D2_VBLANK_INTERRUPT(r500_disp_int)) | 404 | if (G_007EDC_LB_D2_VBLANK_INTERRUPT(r500_disp_int)) |
| 260 | drm_handle_vblank(rdev->ddev, 1); | 405 | drm_handle_vblank(rdev->ddev, 1); |
| 406 | if (G_007EDC_DC_HOT_PLUG_DETECT1_INTERRUPT(r500_disp_int)) { | ||
| 407 | queue_hotplug = true; | ||
| 408 | DRM_DEBUG("HPD1\n"); | ||
| 409 | } | ||
| 410 | if (G_007EDC_DC_HOT_PLUG_DETECT2_INTERRUPT(r500_disp_int)) { | ||
| 411 | queue_hotplug = true; | ||
| 412 | DRM_DEBUG("HPD2\n"); | ||
| 413 | } | ||
| 261 | status = rs600_irq_ack(rdev, &r500_disp_int); | 414 | status = rs600_irq_ack(rdev, &r500_disp_int); |
| 262 | } | 415 | } |
| 416 | if (queue_hotplug) | ||
| 417 | queue_work(rdev->wq, &rdev->hotplug_work); | ||
| 263 | if (rdev->msi_enabled) { | 418 | if (rdev->msi_enabled) { |
| 264 | switch (rdev->family) { | 419 | switch (rdev->family) { |
| 265 | case CHIP_RS600: | 420 | case CHIP_RS600: |
| @@ -301,9 +456,7 @@ int rs600_mc_wait_for_idle(struct radeon_device *rdev) | |||
| 301 | 456 | ||
| 302 | void rs600_gpu_init(struct radeon_device *rdev) | 457 | void rs600_gpu_init(struct radeon_device *rdev) |
| 303 | { | 458 | { |
| 304 | /* FIXME: HDP same place on rs600 ? */ | ||
| 305 | r100_hdp_reset(rdev); | 459 | r100_hdp_reset(rdev); |
| 306 | /* FIXME: is this correct ? */ | ||
| 307 | r420_pipes_init(rdev); | 460 | r420_pipes_init(rdev); |
| 308 | /* Wait for mc idle */ | 461 | /* Wait for mc idle */ |
| 309 | if (rs600_mc_wait_for_idle(rdev)) | 462 | if (rs600_mc_wait_for_idle(rdev)) |
| @@ -312,9 +465,20 @@ void rs600_gpu_init(struct radeon_device *rdev) | |||
| 312 | 465 | ||
| 313 | void rs600_vram_info(struct radeon_device *rdev) | 466 | void rs600_vram_info(struct radeon_device *rdev) |
| 314 | { | 467 | { |
| 315 | /* FIXME: to do or is these values sane ? */ | ||
| 316 | rdev->mc.vram_is_ddr = true; | 468 | rdev->mc.vram_is_ddr = true; |
| 317 | rdev->mc.vram_width = 128; | 469 | rdev->mc.vram_width = 128; |
| 470 | |||
| 471 | rdev->mc.real_vram_size = RREG32(RADEON_CONFIG_MEMSIZE); | ||
| 472 | rdev->mc.mc_vram_size = rdev->mc.real_vram_size; | ||
| 473 | |||
| 474 | rdev->mc.aper_base = drm_get_resource_start(rdev->ddev, 0); | ||
| 475 | rdev->mc.aper_size = drm_get_resource_len(rdev->ddev, 0); | ||
| 476 | |||
| 477 | if (rdev->mc.mc_vram_size > rdev->mc.aper_size) | ||
| 478 | rdev->mc.mc_vram_size = rdev->mc.aper_size; | ||
| 479 | |||
| 480 | if (rdev->mc.real_vram_size > rdev->mc.aper_size) | ||
| 481 | rdev->mc.real_vram_size = rdev->mc.aper_size; | ||
| 318 | } | 482 | } |
| 319 | 483 | ||
| 320 | void rs600_bandwidth_update(struct radeon_device *rdev) | 484 | void rs600_bandwidth_update(struct radeon_device *rdev) |
| @@ -388,7 +552,6 @@ static int rs600_startup(struct radeon_device *rdev) | |||
| 388 | if (r) | 552 | if (r) |
| 389 | return r; | 553 | return r; |
| 390 | /* Enable IRQ */ | 554 | /* Enable IRQ */ |
| 391 | rdev->irq.sw_int = true; | ||
| 392 | rs600_irq_set(rdev); | 555 | rs600_irq_set(rdev); |
| 393 | /* 1M ring buffer */ | 556 | /* 1M ring buffer */ |
| 394 | r = r100_cp_init(rdev, 1024 * 1024); | 557 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -423,6 +586,8 @@ int rs600_resume(struct radeon_device *rdev) | |||
| 423 | atom_asic_init(rdev->mode_info.atom_context); | 586 | atom_asic_init(rdev->mode_info.atom_context); |
| 424 | /* Resume clock after posting */ | 587 | /* Resume clock after posting */ |
| 425 | rv515_clock_startup(rdev); | 588 | rv515_clock_startup(rdev); |
| 589 | /* Initialize surface registers */ | ||
| 590 | radeon_surface_init(rdev); | ||
| 426 | return rs600_startup(rdev); | 591 | return rs600_startup(rdev); |
| 427 | } | 592 | } |
| 428 | 593 | ||
| @@ -445,7 +610,7 @@ void rs600_fini(struct radeon_device *rdev) | |||
| 445 | rs600_gart_fini(rdev); | 610 | rs600_gart_fini(rdev); |
| 446 | radeon_irq_kms_fini(rdev); | 611 | radeon_irq_kms_fini(rdev); |
| 447 | radeon_fence_driver_fini(rdev); | 612 | radeon_fence_driver_fini(rdev); |
| 448 | radeon_object_fini(rdev); | 613 | radeon_bo_fini(rdev); |
| 449 | radeon_atombios_fini(rdev); | 614 | radeon_atombios_fini(rdev); |
| 450 | kfree(rdev->bios); | 615 | kfree(rdev->bios); |
| 451 | rdev->bios = NULL; | 616 | rdev->bios = NULL; |
| @@ -482,10 +647,9 @@ int rs600_init(struct radeon_device *rdev) | |||
| 482 | RREG32(R_0007C0_CP_STAT)); | 647 | RREG32(R_0007C0_CP_STAT)); |
| 483 | } | 648 | } |
| 484 | /* check if cards are posted or not */ | 649 | /* check if cards are posted or not */ |
| 485 | if (!radeon_card_posted(rdev) && rdev->bios) { | 650 | if (radeon_boot_test_post_card(rdev) == false) |
| 486 | DRM_INFO("GPU not posted. posting now...\n"); | 651 | return -EINVAL; |
| 487 | atom_asic_init(rdev->mode_info.atom_context); | 652 | |
| 488 | } | ||
| 489 | /* Initialize clocks */ | 653 | /* Initialize clocks */ |
| 490 | radeon_get_clock_info(rdev->ddev); | 654 | radeon_get_clock_info(rdev->ddev); |
| 491 | /* Initialize power management */ | 655 | /* Initialize power management */ |
| @@ -493,7 +657,7 @@ int rs600_init(struct radeon_device *rdev) | |||
| 493 | /* Get vram informations */ | 657 | /* Get vram informations */ |
| 494 | rs600_vram_info(rdev); | 658 | rs600_vram_info(rdev); |
| 495 | /* Initialize memory controller (also test AGP) */ | 659 | /* Initialize memory controller (also test AGP) */ |
| 496 | r = r420_mc_init(rdev); | 660 | r = rs600_mc_init(rdev); |
| 497 | if (r) | 661 | if (r) |
| 498 | return r; | 662 | return r; |
| 499 | rs600_debugfs(rdev); | 663 | rs600_debugfs(rdev); |
| @@ -505,7 +669,7 @@ int rs600_init(struct radeon_device *rdev) | |||
| 505 | if (r) | 669 | if (r) |
| 506 | return r; | 670 | return r; |
| 507 | /* Memory manager */ | 671 | /* Memory manager */ |
| 508 | r = radeon_object_init(rdev); | 672 | r = radeon_bo_init(rdev); |
| 509 | if (r) | 673 | if (r) |
| 510 | return r; | 674 | return r; |
| 511 | r = rs600_gart_init(rdev); | 675 | r = rs600_gart_init(rdev); |
diff --git a/drivers/gpu/drm/radeon/rs600d.h b/drivers/gpu/drm/radeon/rs600d.h index 81308924859a..c1c8f5885cbb 100644 --- a/drivers/gpu/drm/radeon/rs600d.h +++ b/drivers/gpu/drm/radeon/rs600d.h | |||
| @@ -30,27 +30,12 @@ | |||
| 30 | 30 | ||
| 31 | /* Registers */ | 31 | /* Registers */ |
| 32 | #define R_000040_GEN_INT_CNTL 0x000040 | 32 | #define R_000040_GEN_INT_CNTL 0x000040 |
| 33 | #define S_000040_DISPLAY_INT_STATUS(x) (((x) & 0x1) << 0) | 33 | #define S_000040_SCRATCH_INT_MASK(x) (((x) & 0x1) << 18) |
| 34 | #define G_000040_DISPLAY_INT_STATUS(x) (((x) >> 0) & 0x1) | 34 | #define G_000040_SCRATCH_INT_MASK(x) (((x) >> 18) & 0x1) |
| 35 | #define C_000040_DISPLAY_INT_STATUS 0xFFFFFFFE | 35 | #define C_000040_SCRATCH_INT_MASK 0xFFFBFFFF |
| 36 | #define S_000040_DMA_VIPH0_INT_EN(x) (((x) & 0x1) << 12) | 36 | #define S_000040_GUI_IDLE_MASK(x) (((x) & 0x1) << 19) |
| 37 | #define G_000040_DMA_VIPH0_INT_EN(x) (((x) >> 12) & 0x1) | 37 | #define G_000040_GUI_IDLE_MASK(x) (((x) >> 19) & 0x1) |
| 38 | #define C_000040_DMA_VIPH0_INT_EN 0xFFFFEFFF | 38 | #define C_000040_GUI_IDLE_MASK 0xFFF7FFFF |
| 39 | #define S_000040_CRTC2_VSYNC(x) (((x) & 0x1) << 6) | ||
| 40 | #define G_000040_CRTC2_VSYNC(x) (((x) >> 6) & 0x1) | ||
| 41 | #define C_000040_CRTC2_VSYNC 0xFFFFFFBF | ||
| 42 | #define S_000040_SNAPSHOT2(x) (((x) & 0x1) << 7) | ||
| 43 | #define G_000040_SNAPSHOT2(x) (((x) >> 7) & 0x1) | ||
| 44 | #define C_000040_SNAPSHOT2 0xFFFFFF7F | ||
| 45 | #define S_000040_CRTC2_VBLANK(x) (((x) & 0x1) << 9) | ||
| 46 | #define G_000040_CRTC2_VBLANK(x) (((x) >> 9) & 0x1) | ||
| 47 | #define C_000040_CRTC2_VBLANK 0xFFFFFDFF | ||
| 48 | #define S_000040_FP2_DETECT(x) (((x) & 0x1) << 10) | ||
| 49 | #define G_000040_FP2_DETECT(x) (((x) >> 10) & 0x1) | ||
| 50 | #define C_000040_FP2_DETECT 0xFFFFFBFF | ||
| 51 | #define S_000040_VSYNC_DIFF_OVER_LIMIT(x) (((x) & 0x1) << 11) | ||
| 52 | #define G_000040_VSYNC_DIFF_OVER_LIMIT(x) (((x) >> 11) & 0x1) | ||
| 53 | #define C_000040_VSYNC_DIFF_OVER_LIMIT 0xFFFFF7FF | ||
| 54 | #define S_000040_DMA_VIPH1_INT_EN(x) (((x) & 0x1) << 13) | 39 | #define S_000040_DMA_VIPH1_INT_EN(x) (((x) & 0x1) << 13) |
| 55 | #define G_000040_DMA_VIPH1_INT_EN(x) (((x) >> 13) & 0x1) | 40 | #define G_000040_DMA_VIPH1_INT_EN(x) (((x) >> 13) & 0x1) |
| 56 | #define C_000040_DMA_VIPH1_INT_EN 0xFFFFDFFF | 41 | #define C_000040_DMA_VIPH1_INT_EN 0xFFFFDFFF |
| @@ -370,7 +355,90 @@ | |||
| 370 | #define S_007EDC_LB_D2_VBLANK_INTERRUPT(x) (((x) & 0x1) << 5) | 355 | #define S_007EDC_LB_D2_VBLANK_INTERRUPT(x) (((x) & 0x1) << 5) |
| 371 | #define G_007EDC_LB_D2_VBLANK_INTERRUPT(x) (((x) >> 5) & 0x1) | 356 | #define G_007EDC_LB_D2_VBLANK_INTERRUPT(x) (((x) >> 5) & 0x1) |
| 372 | #define C_007EDC_LB_D2_VBLANK_INTERRUPT 0xFFFFFFDF | 357 | #define C_007EDC_LB_D2_VBLANK_INTERRUPT 0xFFFFFFDF |
| 373 | 358 | #define S_007EDC_DACA_AUTODETECT_INTERRUPT(x) (((x) & 0x1) << 16) | |
| 359 | #define G_007EDC_DACA_AUTODETECT_INTERRUPT(x) (((x) >> 16) & 0x1) | ||
| 360 | #define C_007EDC_DACA_AUTODETECT_INTERRUPT 0xFFFEFFFF | ||
| 361 | #define S_007EDC_DACB_AUTODETECT_INTERRUPT(x) (((x) & 0x1) << 17) | ||
| 362 | #define G_007EDC_DACB_AUTODETECT_INTERRUPT(x) (((x) >> 17) & 0x1) | ||
| 363 | #define C_007EDC_DACB_AUTODETECT_INTERRUPT 0xFFFDFFFF | ||
| 364 | #define S_007EDC_DC_HOT_PLUG_DETECT1_INTERRUPT(x) (((x) & 0x1) << 18) | ||
| 365 | #define G_007EDC_DC_HOT_PLUG_DETECT1_INTERRUPT(x) (((x) >> 18) & 0x1) | ||
| 366 | #define C_007EDC_DC_HOT_PLUG_DETECT1_INTERRUPT 0xFFFBFFFF | ||
| 367 | #define S_007EDC_DC_HOT_PLUG_DETECT2_INTERRUPT(x) (((x) & 0x1) << 19) | ||
| 368 | #define G_007EDC_DC_HOT_PLUG_DETECT2_INTERRUPT(x) (((x) >> 19) & 0x1) | ||
| 369 | #define C_007EDC_DC_HOT_PLUG_DETECT2_INTERRUPT 0xFFF7FFFF | ||
| 370 | #define R_007828_DACA_AUTODETECT_CONTROL 0x007828 | ||
| 371 | #define S_007828_DACA_AUTODETECT_MODE(x) (((x) & 0x3) << 0) | ||
| 372 | #define G_007828_DACA_AUTODETECT_MODE(x) (((x) >> 0) & 0x3) | ||
| 373 | #define C_007828_DACA_AUTODETECT_MODE 0xFFFFFFFC | ||
| 374 | #define S_007828_DACA_AUTODETECT_FRAME_TIME_COUNTER(x) (((x) & 0xff) << 8) | ||
| 375 | #define G_007828_DACA_AUTODETECT_FRAME_TIME_COUNTER(x) (((x) >> 8) & 0xff) | ||
| 376 | #define C_007828_DACA_AUTODETECT_FRAME_TIME_COUNTER 0xFFFF00FF | ||
| 377 | #define S_007828_DACA_AUTODETECT_CHECK_MASK(x) (((x) & 0x3) << 16) | ||
| 378 | #define G_007828_DACA_AUTODETECT_CHECK_MASK(x) (((x) >> 16) & 0x3) | ||
| 379 | #define C_007828_DACA_AUTODETECT_CHECK_MASK 0xFFFCFFFF | ||
| 380 | #define R_007838_DACA_AUTODETECT_INT_CONTROL 0x007838 | ||
| 381 | #define S_007838_DACA_AUTODETECT_ACK(x) (((x) & 0x1) << 0) | ||
| 382 | #define C_007838_DACA_DACA_AUTODETECT_ACK 0xFFFFFFFE | ||
| 383 | #define S_007838_DACA_AUTODETECT_INT_ENABLE(x) (((x) & 0x1) << 16) | ||
| 384 | #define G_007838_DACA_AUTODETECT_INT_ENABLE(x) (((x) >> 16) & 0x1) | ||
| 385 | #define C_007838_DACA_AUTODETECT_INT_ENABLE 0xFFFCFFFF | ||
| 386 | #define R_007A28_DACB_AUTODETECT_CONTROL 0x007A28 | ||
| 387 | #define S_007A28_DACB_AUTODETECT_MODE(x) (((x) & 0x3) << 0) | ||
| 388 | #define G_007A28_DACB_AUTODETECT_MODE(x) (((x) >> 0) & 0x3) | ||
| 389 | #define C_007A28_DACB_AUTODETECT_MODE 0xFFFFFFFC | ||
| 390 | #define S_007A28_DACB_AUTODETECT_FRAME_TIME_COUNTER(x) (((x) & 0xff) << 8) | ||
| 391 | #define G_007A28_DACB_AUTODETECT_FRAME_TIME_COUNTER(x) (((x) >> 8) & 0xff) | ||
| 392 | #define C_007A28_DACB_AUTODETECT_FRAME_TIME_COUNTER 0xFFFF00FF | ||
| 393 | #define S_007A28_DACB_AUTODETECT_CHECK_MASK(x) (((x) & 0x3) << 16) | ||
| 394 | #define G_007A28_DACB_AUTODETECT_CHECK_MASK(x) (((x) >> 16) & 0x3) | ||
| 395 | #define C_007A28_DACB_AUTODETECT_CHECK_MASK 0xFFFCFFFF | ||
| 396 | #define R_007A38_DACB_AUTODETECT_INT_CONTROL 0x007A38 | ||
| 397 | #define S_007A38_DACB_AUTODETECT_ACK(x) (((x) & 0x1) << 0) | ||
| 398 | #define C_007A38_DACB_DACA_AUTODETECT_ACK 0xFFFFFFFE | ||
| 399 | #define S_007A38_DACB_AUTODETECT_INT_ENABLE(x) (((x) & 0x1) << 16) | ||
| 400 | #define G_007A38_DACB_AUTODETECT_INT_ENABLE(x) (((x) >> 16) & 0x1) | ||
| 401 | #define C_007A38_DACB_AUTODETECT_INT_ENABLE 0xFFFCFFFF | ||
| 402 | #define R_007D00_DC_HOT_PLUG_DETECT1_CONTROL 0x007D00 | ||
| 403 | #define S_007D00_DC_HOT_PLUG_DETECT1_EN(x) (((x) & 0x1) << 0) | ||
| 404 | #define G_007D00_DC_HOT_PLUG_DETECT1_EN(x) (((x) >> 0) & 0x1) | ||
| 405 | #define C_007D00_DC_HOT_PLUG_DETECT1_EN 0xFFFFFFFE | ||
| 406 | #define R_007D04_DC_HOT_PLUG_DETECT1_INT_STATUS 0x007D04 | ||
| 407 | #define S_007D04_DC_HOT_PLUG_DETECT1_INT_STATUS(x) (((x) & 0x1) << 0) | ||
| 408 | #define G_007D04_DC_HOT_PLUG_DETECT1_INT_STATUS(x) (((x) >> 0) & 0x1) | ||
| 409 | #define C_007D04_DC_HOT_PLUG_DETECT1_INT_STATUS 0xFFFFFFFE | ||
| 410 | #define S_007D04_DC_HOT_PLUG_DETECT1_SENSE(x) (((x) & 0x1) << 1) | ||
| 411 | #define G_007D04_DC_HOT_PLUG_DETECT1_SENSE(x) (((x) >> 1) & 0x1) | ||
| 412 | #define C_007D04_DC_HOT_PLUG_DETECT1_SENSE 0xFFFFFFFD | ||
| 413 | #define R_007D08_DC_HOT_PLUG_DETECT1_INT_CONTROL 0x007D08 | ||
| 414 | #define S_007D08_DC_HOT_PLUG_DETECT1_INT_ACK(x) (((x) & 0x1) << 0) | ||
| 415 | #define C_007D08_DC_HOT_PLUG_DETECT1_INT_ACK 0xFFFFFFFE | ||
| 416 | #define S_007D08_DC_HOT_PLUG_DETECT1_INT_POLARITY(x) (((x) & 0x1) << 8) | ||
| 417 | #define G_007D08_DC_HOT_PLUG_DETECT1_INT_POLARITY(x) (((x) >> 8) & 0x1) | ||
| 418 | #define C_007D08_DC_HOT_PLUG_DETECT1_INT_POLARITY 0xFFFFFEFF | ||
| 419 | #define S_007D08_DC_HOT_PLUG_DETECT1_INT_EN(x) (((x) & 0x1) << 16) | ||
| 420 | #define G_007D08_DC_HOT_PLUG_DETECT1_INT_EN(x) (((x) >> 16) & 0x1) | ||
| 421 | #define C_007D08_DC_HOT_PLUG_DETECT1_INT_EN 0xFFFEFFFF | ||
| 422 | #define R_007D10_DC_HOT_PLUG_DETECT2_CONTROL 0x007D10 | ||
| 423 | #define S_007D10_DC_HOT_PLUG_DETECT2_EN(x) (((x) & 0x1) << 0) | ||
| 424 | #define G_007D10_DC_HOT_PLUG_DETECT2_EN(x) (((x) >> 0) & 0x1) | ||
| 425 | #define C_007D10_DC_HOT_PLUG_DETECT2_EN 0xFFFFFFFE | ||
| 426 | #define R_007D14_DC_HOT_PLUG_DETECT2_INT_STATUS 0x007D14 | ||
| 427 | #define S_007D14_DC_HOT_PLUG_DETECT2_INT_STATUS(x) (((x) & 0x1) << 0) | ||
| 428 | #define G_007D14_DC_HOT_PLUG_DETECT2_INT_STATUS(x) (((x) >> 0) & 0x1) | ||
| 429 | #define C_007D14_DC_HOT_PLUG_DETECT2_INT_STATUS 0xFFFFFFFE | ||
| 430 | #define S_007D14_DC_HOT_PLUG_DETECT2_SENSE(x) (((x) & 0x1) << 1) | ||
| 431 | #define G_007D14_DC_HOT_PLUG_DETECT2_SENSE(x) (((x) >> 1) & 0x1) | ||
| 432 | #define C_007D14_DC_HOT_PLUG_DETECT2_SENSE 0xFFFFFFFD | ||
| 433 | #define R_007D18_DC_HOT_PLUG_DETECT2_INT_CONTROL 0x007D18 | ||
| 434 | #define S_007D18_DC_HOT_PLUG_DETECT2_INT_ACK(x) (((x) & 0x1) << 0) | ||
| 435 | #define C_007D18_DC_HOT_PLUG_DETECT2_INT_ACK 0xFFFFFFFE | ||
| 436 | #define S_007D18_DC_HOT_PLUG_DETECT2_INT_POLARITY(x) (((x) & 0x1) << 8) | ||
| 437 | #define G_007D18_DC_HOT_PLUG_DETECT2_INT_POLARITY(x) (((x) >> 8) & 0x1) | ||
| 438 | #define C_007D18_DC_HOT_PLUG_DETECT2_INT_POLARITY 0xFFFFFEFF | ||
| 439 | #define S_007D18_DC_HOT_PLUG_DETECT2_INT_EN(x) (((x) & 0x1) << 16) | ||
| 440 | #define G_007D18_DC_HOT_PLUG_DETECT2_INT_EN(x) (((x) >> 16) & 0x1) | ||
| 441 | #define C_007D18_DC_HOT_PLUG_DETECT2_INT_EN 0xFFFEFFFF | ||
| 374 | 442 | ||
| 375 | /* MC registers */ | 443 | /* MC registers */ |
| 376 | #define R_000000_MC_STATUS 0x000000 | 444 | #define R_000000_MC_STATUS 0x000000 |
diff --git a/drivers/gpu/drm/radeon/rs690.c b/drivers/gpu/drm/radeon/rs690.c index 27547175cf93..1e22f52d6039 100644 --- a/drivers/gpu/drm/radeon/rs690.c +++ b/drivers/gpu/drm/radeon/rs690.c | |||
| @@ -131,24 +131,25 @@ void rs690_pm_info(struct radeon_device *rdev) | |||
| 131 | 131 | ||
| 132 | void rs690_vram_info(struct radeon_device *rdev) | 132 | void rs690_vram_info(struct radeon_device *rdev) |
| 133 | { | 133 | { |
| 134 | uint32_t tmp; | ||
| 135 | fixed20_12 a; | 134 | fixed20_12 a; |
| 136 | 135 | ||
| 137 | rs400_gart_adjust_size(rdev); | 136 | rs400_gart_adjust_size(rdev); |
| 138 | /* DDR for all card after R300 & IGP */ | 137 | |
| 139 | rdev->mc.vram_is_ddr = true; | 138 | rdev->mc.vram_is_ddr = true; |
| 140 | /* FIXME: is this correct for RS690/RS740 ? */ | 139 | rdev->mc.vram_width = 128; |
| 141 | tmp = RREG32(RADEON_MEM_CNTL); | 140 | |
| 142 | if (tmp & R300_MEM_NUM_CHANNELS_MASK) { | ||
| 143 | rdev->mc.vram_width = 128; | ||
| 144 | } else { | ||
| 145 | rdev->mc.vram_width = 64; | ||
| 146 | } | ||
| 147 | rdev->mc.real_vram_size = RREG32(RADEON_CONFIG_MEMSIZE); | 141 | rdev->mc.real_vram_size = RREG32(RADEON_CONFIG_MEMSIZE); |
| 148 | rdev->mc.mc_vram_size = rdev->mc.real_vram_size; | 142 | rdev->mc.mc_vram_size = rdev->mc.real_vram_size; |
| 149 | 143 | ||
| 150 | rdev->mc.aper_base = drm_get_resource_start(rdev->ddev, 0); | 144 | rdev->mc.aper_base = drm_get_resource_start(rdev->ddev, 0); |
| 151 | rdev->mc.aper_size = drm_get_resource_len(rdev->ddev, 0); | 145 | rdev->mc.aper_size = drm_get_resource_len(rdev->ddev, 0); |
| 146 | |||
| 147 | if (rdev->mc.mc_vram_size > rdev->mc.aper_size) | ||
| 148 | rdev->mc.mc_vram_size = rdev->mc.aper_size; | ||
| 149 | |||
| 150 | if (rdev->mc.real_vram_size > rdev->mc.aper_size) | ||
| 151 | rdev->mc.real_vram_size = rdev->mc.aper_size; | ||
| 152 | |||
| 152 | rs690_pm_info(rdev); | 153 | rs690_pm_info(rdev); |
| 153 | /* FIXME: we should enforce default clock in case GPU is not in | 154 | /* FIXME: we should enforce default clock in case GPU is not in |
| 154 | * default setup | 155 | * default setup |
| @@ -161,6 +162,21 @@ void rs690_vram_info(struct radeon_device *rdev) | |||
| 161 | rdev->pm.core_bandwidth.full = rfixed_div(rdev->pm.sclk, a); | 162 | rdev->pm.core_bandwidth.full = rfixed_div(rdev->pm.sclk, a); |
| 162 | } | 163 | } |
| 163 | 164 | ||
| 165 | static int rs690_mc_init(struct radeon_device *rdev) | ||
| 166 | { | ||
| 167 | int r; | ||
| 168 | u32 tmp; | ||
| 169 | |||
| 170 | /* Setup GPU memory space */ | ||
| 171 | tmp = RREG32_MC(R_000100_MCCFG_FB_LOCATION); | ||
| 172 | rdev->mc.vram_location = G_000100_MC_FB_START(tmp) << 16; | ||
| 173 | rdev->mc.gtt_location = 0xFFFFFFFFUL; | ||
| 174 | r = radeon_mc_setup(rdev); | ||
| 175 | if (r) | ||
| 176 | return r; | ||
| 177 | return 0; | ||
| 178 | } | ||
| 179 | |||
| 164 | void rs690_line_buffer_adjust(struct radeon_device *rdev, | 180 | void rs690_line_buffer_adjust(struct radeon_device *rdev, |
| 165 | struct drm_display_mode *mode1, | 181 | struct drm_display_mode *mode1, |
| 166 | struct drm_display_mode *mode2) | 182 | struct drm_display_mode *mode2) |
| @@ -244,8 +260,9 @@ void rs690_crtc_bandwidth_compute(struct radeon_device *rdev, | |||
| 244 | 260 | ||
| 245 | b.full = rfixed_const(mode->crtc_hdisplay); | 261 | b.full = rfixed_const(mode->crtc_hdisplay); |
| 246 | c.full = rfixed_const(256); | 262 | c.full = rfixed_const(256); |
| 247 | a.full = rfixed_mul(wm->num_line_pair, b); | 263 | a.full = rfixed_div(b, c); |
| 248 | request_fifo_depth.full = rfixed_div(a, c); | 264 | request_fifo_depth.full = rfixed_mul(a, wm->num_line_pair); |
| 265 | request_fifo_depth.full = rfixed_ceil(request_fifo_depth); | ||
| 249 | if (a.full < rfixed_const(4)) { | 266 | if (a.full < rfixed_const(4)) { |
| 250 | wm->lb_request_fifo_depth = 4; | 267 | wm->lb_request_fifo_depth = 4; |
| 251 | } else { | 268 | } else { |
| @@ -374,6 +391,7 @@ void rs690_crtc_bandwidth_compute(struct radeon_device *rdev, | |||
| 374 | a.full = rfixed_const(16); | 391 | a.full = rfixed_const(16); |
| 375 | wm->priority_mark_max.full = rfixed_const(crtc->base.mode.crtc_hdisplay); | 392 | wm->priority_mark_max.full = rfixed_const(crtc->base.mode.crtc_hdisplay); |
| 376 | wm->priority_mark_max.full = rfixed_div(wm->priority_mark_max, a); | 393 | wm->priority_mark_max.full = rfixed_div(wm->priority_mark_max, a); |
| 394 | wm->priority_mark_max.full = rfixed_ceil(wm->priority_mark_max); | ||
| 377 | 395 | ||
| 378 | /* Determine estimated width */ | 396 | /* Determine estimated width */ |
| 379 | estimated_width.full = tolerable_latency.full - wm->worst_case_latency.full; | 397 | estimated_width.full = tolerable_latency.full - wm->worst_case_latency.full; |
| @@ -383,6 +401,7 @@ void rs690_crtc_bandwidth_compute(struct radeon_device *rdev, | |||
| 383 | } else { | 401 | } else { |
| 384 | a.full = rfixed_const(16); | 402 | a.full = rfixed_const(16); |
| 385 | wm->priority_mark.full = rfixed_div(estimated_width, a); | 403 | wm->priority_mark.full = rfixed_div(estimated_width, a); |
| 404 | wm->priority_mark.full = rfixed_ceil(wm->priority_mark); | ||
| 386 | wm->priority_mark.full = wm->priority_mark_max.full - wm->priority_mark.full; | 405 | wm->priority_mark.full = wm->priority_mark_max.full - wm->priority_mark.full; |
| 387 | } | 406 | } |
| 388 | } | 407 | } |
| @@ -605,7 +624,6 @@ static int rs690_startup(struct radeon_device *rdev) | |||
| 605 | if (r) | 624 | if (r) |
| 606 | return r; | 625 | return r; |
| 607 | /* Enable IRQ */ | 626 | /* Enable IRQ */ |
| 608 | rdev->irq.sw_int = true; | ||
| 609 | rs600_irq_set(rdev); | 627 | rs600_irq_set(rdev); |
| 610 | /* 1M ring buffer */ | 628 | /* 1M ring buffer */ |
| 611 | r = r100_cp_init(rdev, 1024 * 1024); | 629 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -640,6 +658,8 @@ int rs690_resume(struct radeon_device *rdev) | |||
| 640 | atom_asic_init(rdev->mode_info.atom_context); | 658 | atom_asic_init(rdev->mode_info.atom_context); |
| 641 | /* Resume clock after posting */ | 659 | /* Resume clock after posting */ |
| 642 | rv515_clock_startup(rdev); | 660 | rv515_clock_startup(rdev); |
| 661 | /* Initialize surface registers */ | ||
| 662 | radeon_surface_init(rdev); | ||
| 643 | return rs690_startup(rdev); | 663 | return rs690_startup(rdev); |
| 644 | } | 664 | } |
| 645 | 665 | ||
| @@ -662,7 +682,7 @@ void rs690_fini(struct radeon_device *rdev) | |||
| 662 | rs400_gart_fini(rdev); | 682 | rs400_gart_fini(rdev); |
| 663 | radeon_irq_kms_fini(rdev); | 683 | radeon_irq_kms_fini(rdev); |
| 664 | radeon_fence_driver_fini(rdev); | 684 | radeon_fence_driver_fini(rdev); |
| 665 | radeon_object_fini(rdev); | 685 | radeon_bo_fini(rdev); |
| 666 | radeon_atombios_fini(rdev); | 686 | radeon_atombios_fini(rdev); |
| 667 | kfree(rdev->bios); | 687 | kfree(rdev->bios); |
| 668 | rdev->bios = NULL; | 688 | rdev->bios = NULL; |
| @@ -700,10 +720,9 @@ int rs690_init(struct radeon_device *rdev) | |||
| 700 | RREG32(R_0007C0_CP_STAT)); | 720 | RREG32(R_0007C0_CP_STAT)); |
| 701 | } | 721 | } |
| 702 | /* check if cards are posted or not */ | 722 | /* check if cards are posted or not */ |
| 703 | if (!radeon_card_posted(rdev) && rdev->bios) { | 723 | if (radeon_boot_test_post_card(rdev) == false) |
| 704 | DRM_INFO("GPU not posted. posting now...\n"); | 724 | return -EINVAL; |
| 705 | atom_asic_init(rdev->mode_info.atom_context); | 725 | |
| 706 | } | ||
| 707 | /* Initialize clocks */ | 726 | /* Initialize clocks */ |
| 708 | radeon_get_clock_info(rdev->ddev); | 727 | radeon_get_clock_info(rdev->ddev); |
| 709 | /* Initialize power management */ | 728 | /* Initialize power management */ |
| @@ -711,7 +730,7 @@ int rs690_init(struct radeon_device *rdev) | |||
| 711 | /* Get vram informations */ | 730 | /* Get vram informations */ |
| 712 | rs690_vram_info(rdev); | 731 | rs690_vram_info(rdev); |
| 713 | /* Initialize memory controller (also test AGP) */ | 732 | /* Initialize memory controller (also test AGP) */ |
| 714 | r = r420_mc_init(rdev); | 733 | r = rs690_mc_init(rdev); |
| 715 | if (r) | 734 | if (r) |
| 716 | return r; | 735 | return r; |
| 717 | rv515_debugfs(rdev); | 736 | rv515_debugfs(rdev); |
| @@ -723,7 +742,7 @@ int rs690_init(struct radeon_device *rdev) | |||
| 723 | if (r) | 742 | if (r) |
| 724 | return r; | 743 | return r; |
| 725 | /* Memory manager */ | 744 | /* Memory manager */ |
| 726 | r = radeon_object_init(rdev); | 745 | r = radeon_bo_init(rdev); |
| 727 | if (r) | 746 | if (r) |
| 728 | return r; | 747 | return r; |
| 729 | r = rs400_gart_init(rdev); | 748 | r = rs400_gart_init(rdev); |
diff --git a/drivers/gpu/drm/radeon/rv515.c b/drivers/gpu/drm/radeon/rv515.c index ba68c9fe90a1..59632a506b46 100644 --- a/drivers/gpu/drm/radeon/rv515.c +++ b/drivers/gpu/drm/radeon/rv515.c | |||
| @@ -478,7 +478,6 @@ static int rv515_startup(struct radeon_device *rdev) | |||
| 478 | return r; | 478 | return r; |
| 479 | } | 479 | } |
| 480 | /* Enable IRQ */ | 480 | /* Enable IRQ */ |
| 481 | rdev->irq.sw_int = true; | ||
| 482 | rs600_irq_set(rdev); | 481 | rs600_irq_set(rdev); |
| 483 | /* 1M ring buffer */ | 482 | /* 1M ring buffer */ |
| 484 | r = r100_cp_init(rdev, 1024 * 1024); | 483 | r = r100_cp_init(rdev, 1024 * 1024); |
| @@ -514,6 +513,8 @@ int rv515_resume(struct radeon_device *rdev) | |||
| 514 | atom_asic_init(rdev->mode_info.atom_context); | 513 | atom_asic_init(rdev->mode_info.atom_context); |
| 515 | /* Resume clock after posting */ | 514 | /* Resume clock after posting */ |
| 516 | rv515_clock_startup(rdev); | 515 | rv515_clock_startup(rdev); |
| 516 | /* Initialize surface registers */ | ||
| 517 | radeon_surface_init(rdev); | ||
| 517 | return rv515_startup(rdev); | 518 | return rv515_startup(rdev); |
| 518 | } | 519 | } |
| 519 | 520 | ||
| @@ -540,11 +541,11 @@ void rv515_fini(struct radeon_device *rdev) | |||
| 540 | r100_wb_fini(rdev); | 541 | r100_wb_fini(rdev); |
| 541 | r100_ib_fini(rdev); | 542 | r100_ib_fini(rdev); |
| 542 | radeon_gem_fini(rdev); | 543 | radeon_gem_fini(rdev); |
| 543 | rv370_pcie_gart_fini(rdev); | 544 | rv370_pcie_gart_fini(rdev); |
| 544 | radeon_agp_fini(rdev); | 545 | radeon_agp_fini(rdev); |
| 545 | radeon_irq_kms_fini(rdev); | 546 | radeon_irq_kms_fini(rdev); |
| 546 | radeon_fence_driver_fini(rdev); | 547 | radeon_fence_driver_fini(rdev); |
| 547 | radeon_object_fini(rdev); | 548 | radeon_bo_fini(rdev); |
| 548 | radeon_atombios_fini(rdev); | 549 | radeon_atombios_fini(rdev); |
| 549 | kfree(rdev->bios); | 550 | kfree(rdev->bios); |
| 550 | rdev->bios = NULL; | 551 | rdev->bios = NULL; |
| @@ -580,10 +581,8 @@ int rv515_init(struct radeon_device *rdev) | |||
| 580 | RREG32(R_0007C0_CP_STAT)); | 581 | RREG32(R_0007C0_CP_STAT)); |
| 581 | } | 582 | } |
| 582 | /* check if cards are posted or not */ | 583 | /* check if cards are posted or not */ |
| 583 | if (!radeon_card_posted(rdev) && rdev->bios) { | 584 | if (radeon_boot_test_post_card(rdev) == false) |
| 584 | DRM_INFO("GPU not posted. posting now...\n"); | 585 | return -EINVAL; |
| 585 | atom_asic_init(rdev->mode_info.atom_context); | ||
| 586 | } | ||
| 587 | /* Initialize clocks */ | 586 | /* Initialize clocks */ |
| 588 | radeon_get_clock_info(rdev->ddev); | 587 | radeon_get_clock_info(rdev->ddev); |
| 589 | /* Initialize power management */ | 588 | /* Initialize power management */ |
| @@ -603,7 +602,7 @@ int rv515_init(struct radeon_device *rdev) | |||
| 603 | if (r) | 602 | if (r) |
| 604 | return r; | 603 | return r; |
| 605 | /* Memory manager */ | 604 | /* Memory manager */ |
| 606 | r = radeon_object_init(rdev); | 605 | r = radeon_bo_init(rdev); |
| 607 | if (r) | 606 | if (r) |
| 608 | return r; | 607 | return r; |
| 609 | r = rv370_pcie_gart_init(rdev); | 608 | r = rv370_pcie_gart_init(rdev); |
| @@ -892,8 +891,9 @@ void rv515_crtc_bandwidth_compute(struct radeon_device *rdev, | |||
| 892 | 891 | ||
| 893 | b.full = rfixed_const(mode->crtc_hdisplay); | 892 | b.full = rfixed_const(mode->crtc_hdisplay); |
| 894 | c.full = rfixed_const(256); | 893 | c.full = rfixed_const(256); |
| 895 | a.full = rfixed_mul(wm->num_line_pair, b); | 894 | a.full = rfixed_div(b, c); |
| 896 | request_fifo_depth.full = rfixed_div(a, c); | 895 | request_fifo_depth.full = rfixed_mul(a, wm->num_line_pair); |
| 896 | request_fifo_depth.full = rfixed_ceil(request_fifo_depth); | ||
| 897 | if (a.full < rfixed_const(4)) { | 897 | if (a.full < rfixed_const(4)) { |
| 898 | wm->lb_request_fifo_depth = 4; | 898 | wm->lb_request_fifo_depth = 4; |
| 899 | } else { | 899 | } else { |
| @@ -995,15 +995,17 @@ void rv515_crtc_bandwidth_compute(struct radeon_device *rdev, | |||
| 995 | a.full = rfixed_const(16); | 995 | a.full = rfixed_const(16); |
| 996 | wm->priority_mark_max.full = rfixed_const(crtc->base.mode.crtc_hdisplay); | 996 | wm->priority_mark_max.full = rfixed_const(crtc->base.mode.crtc_hdisplay); |
| 997 | wm->priority_mark_max.full = rfixed_div(wm->priority_mark_max, a); | 997 | wm->priority_mark_max.full = rfixed_div(wm->priority_mark_max, a); |
| 998 | wm->priority_mark_max.full = rfixed_ceil(wm->priority_mark_max); | ||
| 998 | 999 | ||
| 999 | /* Determine estimated width */ | 1000 | /* Determine estimated width */ |
| 1000 | estimated_width.full = tolerable_latency.full - wm->worst_case_latency.full; | 1001 | estimated_width.full = tolerable_latency.full - wm->worst_case_latency.full; |
| 1001 | estimated_width.full = rfixed_div(estimated_width, consumption_time); | 1002 | estimated_width.full = rfixed_div(estimated_width, consumption_time); |
| 1002 | if (rfixed_trunc(estimated_width) > crtc->base.mode.crtc_hdisplay) { | 1003 | if (rfixed_trunc(estimated_width) > crtc->base.mode.crtc_hdisplay) { |
| 1003 | wm->priority_mark.full = rfixed_const(10); | 1004 | wm->priority_mark.full = wm->priority_mark_max.full; |
| 1004 | } else { | 1005 | } else { |
| 1005 | a.full = rfixed_const(16); | 1006 | a.full = rfixed_const(16); |
| 1006 | wm->priority_mark.full = rfixed_div(estimated_width, a); | 1007 | wm->priority_mark.full = rfixed_div(estimated_width, a); |
| 1008 | wm->priority_mark.full = rfixed_ceil(wm->priority_mark); | ||
| 1007 | wm->priority_mark.full = wm->priority_mark_max.full - wm->priority_mark.full; | 1009 | wm->priority_mark.full = wm->priority_mark_max.full - wm->priority_mark.full; |
| 1008 | } | 1010 | } |
| 1009 | } | 1011 | } |
diff --git a/drivers/gpu/drm/radeon/rv770.c b/drivers/gpu/drm/radeon/rv770.c index 5e06ee7076f5..fbb0357f1ec3 100644 --- a/drivers/gpu/drm/radeon/rv770.c +++ b/drivers/gpu/drm/radeon/rv770.c | |||
| @@ -92,7 +92,7 @@ int rv770_pcie_gart_enable(struct radeon_device *rdev) | |||
| 92 | void rv770_pcie_gart_disable(struct radeon_device *rdev) | 92 | void rv770_pcie_gart_disable(struct radeon_device *rdev) |
| 93 | { | 93 | { |
| 94 | u32 tmp; | 94 | u32 tmp; |
| 95 | int i; | 95 | int i, r; |
| 96 | 96 | ||
| 97 | /* Disable all tables */ | 97 | /* Disable all tables */ |
| 98 | for (i = 0; i < 7; i++) | 98 | for (i = 0; i < 7; i++) |
| @@ -113,8 +113,12 @@ void rv770_pcie_gart_disable(struct radeon_device *rdev) | |||
| 113 | WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp); | 113 | WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp); |
| 114 | WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp); | 114 | WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp); |
| 115 | if (rdev->gart.table.vram.robj) { | 115 | if (rdev->gart.table.vram.robj) { |
| 116 | radeon_object_kunmap(rdev->gart.table.vram.robj); | 116 | r = radeon_bo_reserve(rdev->gart.table.vram.robj, false); |
| 117 | radeon_object_unpin(rdev->gart.table.vram.robj); | 117 | if (likely(r == 0)) { |
| 118 | radeon_bo_kunmap(rdev->gart.table.vram.robj); | ||
| 119 | radeon_bo_unpin(rdev->gart.table.vram.robj); | ||
| 120 | radeon_bo_unreserve(rdev->gart.table.vram.robj); | ||
| 121 | } | ||
| 118 | } | 122 | } |
| 119 | } | 123 | } |
| 120 | 124 | ||
| @@ -870,6 +874,14 @@ static int rv770_startup(struct radeon_device *rdev) | |||
| 870 | { | 874 | { |
| 871 | int r; | 875 | int r; |
| 872 | 876 | ||
| 877 | if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) { | ||
| 878 | r = r600_init_microcode(rdev); | ||
| 879 | if (r) { | ||
| 880 | DRM_ERROR("Failed to load firmware!\n"); | ||
| 881 | return r; | ||
| 882 | } | ||
| 883 | } | ||
| 884 | |||
| 873 | rv770_mc_program(rdev); | 885 | rv770_mc_program(rdev); |
| 874 | if (rdev->flags & RADEON_IS_AGP) { | 886 | if (rdev->flags & RADEON_IS_AGP) { |
| 875 | rv770_agp_enable(rdev); | 887 | rv770_agp_enable(rdev); |
| @@ -880,13 +892,26 @@ static int rv770_startup(struct radeon_device *rdev) | |||
| 880 | } | 892 | } |
| 881 | rv770_gpu_init(rdev); | 893 | rv770_gpu_init(rdev); |
| 882 | 894 | ||
| 883 | r = radeon_object_pin(rdev->r600_blit.shader_obj, RADEON_GEM_DOMAIN_VRAM, | 895 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); |
| 884 | &rdev->r600_blit.shader_gpu_addr); | 896 | if (unlikely(r != 0)) |
| 897 | return r; | ||
| 898 | r = radeon_bo_pin(rdev->r600_blit.shader_obj, RADEON_GEM_DOMAIN_VRAM, | ||
| 899 | &rdev->r600_blit.shader_gpu_addr); | ||
| 900 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); | ||
| 885 | if (r) { | 901 | if (r) { |
| 886 | DRM_ERROR("failed to pin blit object %d\n", r); | 902 | DRM_ERROR("failed to pin blit object %d\n", r); |
| 887 | return r; | 903 | return r; |
| 888 | } | 904 | } |
| 889 | 905 | ||
| 906 | /* Enable IRQ */ | ||
| 907 | r = r600_irq_init(rdev); | ||
| 908 | if (r) { | ||
| 909 | DRM_ERROR("radeon: IH init failed (%d).\n", r); | ||
| 910 | radeon_irq_kms_fini(rdev); | ||
| 911 | return r; | ||
| 912 | } | ||
| 913 | r600_irq_set(rdev); | ||
| 914 | |||
| 890 | r = radeon_ring_init(rdev, rdev->cp.ring_size); | 915 | r = radeon_ring_init(rdev, rdev->cp.ring_size); |
| 891 | if (r) | 916 | if (r) |
| 892 | return r; | 917 | return r; |
| @@ -934,13 +959,19 @@ int rv770_resume(struct radeon_device *rdev) | |||
| 934 | 959 | ||
| 935 | int rv770_suspend(struct radeon_device *rdev) | 960 | int rv770_suspend(struct radeon_device *rdev) |
| 936 | { | 961 | { |
| 962 | int r; | ||
| 963 | |||
| 937 | /* FIXME: we should wait for ring to be empty */ | 964 | /* FIXME: we should wait for ring to be empty */ |
| 938 | r700_cp_stop(rdev); | 965 | r700_cp_stop(rdev); |
| 939 | rdev->cp.ready = false; | 966 | rdev->cp.ready = false; |
| 940 | r600_wb_disable(rdev); | 967 | r600_wb_disable(rdev); |
| 941 | rv770_pcie_gart_disable(rdev); | 968 | rv770_pcie_gart_disable(rdev); |
| 942 | /* unpin shaders bo */ | 969 | /* unpin shaders bo */ |
| 943 | radeon_object_unpin(rdev->r600_blit.shader_obj); | 970 | r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false); |
| 971 | if (likely(r == 0)) { | ||
| 972 | radeon_bo_unpin(rdev->r600_blit.shader_obj); | ||
| 973 | radeon_bo_unreserve(rdev->r600_blit.shader_obj); | ||
| 974 | } | ||
| 944 | return 0; | 975 | return 0; |
| 945 | } | 976 | } |
| 946 | 977 | ||
| @@ -975,7 +1006,11 @@ int rv770_init(struct radeon_device *rdev) | |||
| 975 | if (r) | 1006 | if (r) |
| 976 | return r; | 1007 | return r; |
| 977 | /* Post card if necessary */ | 1008 | /* Post card if necessary */ |
| 978 | if (!r600_card_posted(rdev) && rdev->bios) { | 1009 | if (!r600_card_posted(rdev)) { |
| 1010 | if (!rdev->bios) { | ||
| 1011 | dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n"); | ||
| 1012 | return -EINVAL; | ||
| 1013 | } | ||
| 979 | DRM_INFO("GPU not posted. posting now...\n"); | 1014 | DRM_INFO("GPU not posted. posting now...\n"); |
| 980 | atom_asic_init(rdev->mode_info.atom_context); | 1015 | atom_asic_init(rdev->mode_info.atom_context); |
| 981 | } | 1016 | } |
| @@ -998,31 +1033,31 @@ int rv770_init(struct radeon_device *rdev) | |||
| 998 | if (r) | 1033 | if (r) |
| 999 | return r; | 1034 | return r; |
| 1000 | /* Memory manager */ | 1035 | /* Memory manager */ |
| 1001 | r = radeon_object_init(rdev); | 1036 | r = radeon_bo_init(rdev); |
| 1037 | if (r) | ||
| 1038 | return r; | ||
| 1039 | |||
| 1040 | r = radeon_irq_kms_init(rdev); | ||
| 1002 | if (r) | 1041 | if (r) |
| 1003 | return r; | 1042 | return r; |
| 1043 | |||
| 1004 | rdev->cp.ring_obj = NULL; | 1044 | rdev->cp.ring_obj = NULL; |
| 1005 | r600_ring_init(rdev, 1024 * 1024); | 1045 | r600_ring_init(rdev, 1024 * 1024); |
| 1006 | 1046 | ||
| 1007 | if (!rdev->me_fw || !rdev->pfp_fw) { | 1047 | rdev->ih.ring_obj = NULL; |
| 1008 | r = r600_cp_init_microcode(rdev); | 1048 | r600_ih_ring_init(rdev, 64 * 1024); |
| 1009 | if (r) { | ||
| 1010 | DRM_ERROR("Failed to load firmware!\n"); | ||
| 1011 | return r; | ||
| 1012 | } | ||
| 1013 | } | ||
| 1014 | 1049 | ||
| 1015 | r = r600_pcie_gart_init(rdev); | 1050 | r = r600_pcie_gart_init(rdev); |
| 1016 | if (r) | 1051 | if (r) |
| 1017 | return r; | 1052 | return r; |
| 1018 | 1053 | ||
| 1019 | rdev->accel_working = true; | ||
| 1020 | r = r600_blit_init(rdev); | 1054 | r = r600_blit_init(rdev); |
| 1021 | if (r) { | 1055 | if (r) { |
| 1022 | DRM_ERROR("radeon: failled blitter (%d).\n", r); | 1056 | DRM_ERROR("radeon: failed blitter (%d).\n", r); |
| 1023 | rdev->accel_working = false; | 1057 | return r; |
| 1024 | } | 1058 | } |
| 1025 | 1059 | ||
| 1060 | rdev->accel_working = true; | ||
| 1026 | r = rv770_startup(rdev); | 1061 | r = rv770_startup(rdev); |
| 1027 | if (r) { | 1062 | if (r) { |
| 1028 | rv770_suspend(rdev); | 1063 | rv770_suspend(rdev); |
| @@ -1034,12 +1069,12 @@ int rv770_init(struct radeon_device *rdev) | |||
| 1034 | if (rdev->accel_working) { | 1069 | if (rdev->accel_working) { |
| 1035 | r = radeon_ib_pool_init(rdev); | 1070 | r = radeon_ib_pool_init(rdev); |
| 1036 | if (r) { | 1071 | if (r) { |
| 1037 | DRM_ERROR("radeon: failled initializing IB pool (%d).\n", r); | 1072 | DRM_ERROR("radeon: failed initializing IB pool (%d).\n", r); |
| 1038 | rdev->accel_working = false; | 1073 | rdev->accel_working = false; |
| 1039 | } | 1074 | } |
| 1040 | r = r600_ib_test(rdev); | 1075 | r = r600_ib_test(rdev); |
| 1041 | if (r) { | 1076 | if (r) { |
| 1042 | DRM_ERROR("radeon: failled testing IB (%d).\n", r); | 1077 | DRM_ERROR("radeon: failed testing IB (%d).\n", r); |
| 1043 | rdev->accel_working = false; | 1078 | rdev->accel_working = false; |
| 1044 | } | 1079 | } |
| 1045 | } | 1080 | } |
| @@ -1051,6 +1086,8 @@ void rv770_fini(struct radeon_device *rdev) | |||
| 1051 | rv770_suspend(rdev); | 1086 | rv770_suspend(rdev); |
| 1052 | 1087 | ||
| 1053 | r600_blit_fini(rdev); | 1088 | r600_blit_fini(rdev); |
| 1089 | r600_irq_fini(rdev); | ||
| 1090 | radeon_irq_kms_fini(rdev); | ||
| 1054 | radeon_ring_fini(rdev); | 1091 | radeon_ring_fini(rdev); |
| 1055 | r600_wb_fini(rdev); | 1092 | r600_wb_fini(rdev); |
| 1056 | rv770_pcie_gart_fini(rdev); | 1093 | rv770_pcie_gart_fini(rdev); |
| @@ -1059,7 +1096,7 @@ void rv770_fini(struct radeon_device *rdev) | |||
| 1059 | radeon_clocks_fini(rdev); | 1096 | radeon_clocks_fini(rdev); |
| 1060 | if (rdev->flags & RADEON_IS_AGP) | 1097 | if (rdev->flags & RADEON_IS_AGP) |
| 1061 | radeon_agp_fini(rdev); | 1098 | radeon_agp_fini(rdev); |
| 1062 | radeon_object_fini(rdev); | 1099 | radeon_bo_fini(rdev); |
| 1063 | radeon_atombios_fini(rdev); | 1100 | radeon_atombios_fini(rdev); |
| 1064 | kfree(rdev->bios); | 1101 | kfree(rdev->bios); |
| 1065 | rdev->bios = NULL; | 1102 | rdev->bios = NULL; |
diff --git a/drivers/gpu/drm/ttm/Makefile b/drivers/gpu/drm/ttm/Makefile index b0a9de7a57c2..1e138f5bae09 100644 --- a/drivers/gpu/drm/ttm/Makefile +++ b/drivers/gpu/drm/ttm/Makefile | |||
| @@ -3,6 +3,7 @@ | |||
| 3 | 3 | ||
| 4 | ccflags-y := -Iinclude/drm | 4 | ccflags-y := -Iinclude/drm |
| 5 | ttm-y := ttm_agp_backend.o ttm_memory.o ttm_tt.o ttm_bo.o \ | 5 | ttm-y := ttm_agp_backend.o ttm_memory.o ttm_tt.o ttm_bo.o \ |
| 6 | ttm_bo_util.o ttm_bo_vm.o ttm_module.o ttm_global.o | 6 | ttm_bo_util.o ttm_bo_vm.o ttm_module.o ttm_global.o \ |
| 7 | ttm_object.o ttm_lock.o ttm_execbuf_util.o | ||
| 7 | 8 | ||
| 8 | obj-$(CONFIG_DRM_TTM) += ttm.o | 9 | obj-$(CONFIG_DRM_TTM) += ttm.o |
diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c index 87c06252d464..a835b6fe42a1 100644 --- a/drivers/gpu/drm/ttm/ttm_bo.c +++ b/drivers/gpu/drm/ttm/ttm_bo.c | |||
| @@ -27,6 +27,14 @@ | |||
| 27 | /* | 27 | /* |
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> |
| 29 | */ | 29 | */ |
| 30 | /* Notes: | ||
| 31 | * | ||
| 32 | * We store bo pointer in drm_mm_node struct so we know which bo own a | ||
| 33 | * specific node. There is no protection on the pointer, thus to make | ||
| 34 | * sure things don't go berserk you have to access this pointer while | ||
| 35 | * holding the global lru lock and make sure anytime you free a node you | ||
| 36 | * reset the pointer to NULL. | ||
| 37 | */ | ||
| 30 | 38 | ||
| 31 | #include "ttm/ttm_module.h" | 39 | #include "ttm/ttm_module.h" |
| 32 | #include "ttm/ttm_bo_driver.h" | 40 | #include "ttm/ttm_bo_driver.h" |
| @@ -51,6 +59,60 @@ static struct attribute ttm_bo_count = { | |||
| 51 | .mode = S_IRUGO | 59 | .mode = S_IRUGO |
| 52 | }; | 60 | }; |
| 53 | 61 | ||
| 62 | static inline int ttm_mem_type_from_flags(uint32_t flags, uint32_t *mem_type) | ||
| 63 | { | ||
| 64 | int i; | ||
| 65 | |||
| 66 | for (i = 0; i <= TTM_PL_PRIV5; i++) | ||
| 67 | if (flags & (1 << i)) { | ||
| 68 | *mem_type = i; | ||
| 69 | return 0; | ||
| 70 | } | ||
| 71 | return -EINVAL; | ||
| 72 | } | ||
| 73 | |||
| 74 | static void ttm_mem_type_manager_debug(struct ttm_bo_global *glob, | ||
| 75 | struct ttm_mem_type_manager *man) | ||
| 76 | { | ||
| 77 | printk(KERN_ERR TTM_PFX " has_type: %d\n", man->has_type); | ||
| 78 | printk(KERN_ERR TTM_PFX " use_type: %d\n", man->use_type); | ||
| 79 | printk(KERN_ERR TTM_PFX " flags: 0x%08X\n", man->flags); | ||
| 80 | printk(KERN_ERR TTM_PFX " gpu_offset: 0x%08lX\n", man->gpu_offset); | ||
| 81 | printk(KERN_ERR TTM_PFX " io_offset: 0x%08lX\n", man->io_offset); | ||
| 82 | printk(KERN_ERR TTM_PFX " io_size: %ld\n", man->io_size); | ||
| 83 | printk(KERN_ERR TTM_PFX " size: %ld\n", (unsigned long)man->size); | ||
| 84 | printk(KERN_ERR TTM_PFX " available_caching: 0x%08X\n", | ||
| 85 | man->available_caching); | ||
| 86 | printk(KERN_ERR TTM_PFX " default_caching: 0x%08X\n", | ||
| 87 | man->default_caching); | ||
| 88 | spin_lock(&glob->lru_lock); | ||
| 89 | drm_mm_debug_table(&man->manager, TTM_PFX); | ||
| 90 | spin_unlock(&glob->lru_lock); | ||
| 91 | } | ||
| 92 | |||
| 93 | static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, | ||
| 94 | struct ttm_placement *placement) | ||
| 95 | { | ||
| 96 | struct ttm_bo_device *bdev = bo->bdev; | ||
| 97 | struct ttm_bo_global *glob = bo->glob; | ||
| 98 | struct ttm_mem_type_manager *man; | ||
| 99 | int i, ret, mem_type; | ||
| 100 | |||
| 101 | printk(KERN_ERR TTM_PFX "No space for %p (%ld pages, %ldK, %ldM)\n", | ||
| 102 | bo, bo->mem.num_pages, bo->mem.size >> 10, | ||
| 103 | bo->mem.size >> 20); | ||
| 104 | for (i = 0; i < placement->num_placement; i++) { | ||
| 105 | ret = ttm_mem_type_from_flags(placement->placement[i], | ||
| 106 | &mem_type); | ||
| 107 | if (ret) | ||
| 108 | return; | ||
| 109 | man = &bdev->man[mem_type]; | ||
| 110 | printk(KERN_ERR TTM_PFX " placement[%d]=0x%08X (%d)\n", | ||
| 111 | i, placement->placement[i], mem_type); | ||
| 112 | ttm_mem_type_manager_debug(glob, man); | ||
| 113 | } | ||
| 114 | } | ||
| 115 | |||
| 54 | static ssize_t ttm_bo_global_show(struct kobject *kobj, | 116 | static ssize_t ttm_bo_global_show(struct kobject *kobj, |
| 55 | struct attribute *attr, | 117 | struct attribute *attr, |
| 56 | char *buffer) | 118 | char *buffer) |
| @@ -117,7 +179,7 @@ int ttm_bo_wait_unreserved(struct ttm_buffer_object *bo, bool interruptible) | |||
| 117 | ret = wait_event_interruptible(bo->event_queue, | 179 | ret = wait_event_interruptible(bo->event_queue, |
| 118 | atomic_read(&bo->reserved) == 0); | 180 | atomic_read(&bo->reserved) == 0); |
| 119 | if (unlikely(ret != 0)) | 181 | if (unlikely(ret != 0)) |
| 120 | return -ERESTART; | 182 | return ret; |
| 121 | } else { | 183 | } else { |
| 122 | wait_event(bo->event_queue, atomic_read(&bo->reserved) == 0); | 184 | wait_event(bo->event_queue, atomic_read(&bo->reserved) == 0); |
| 123 | } | 185 | } |
| @@ -247,7 +309,6 @@ EXPORT_SYMBOL(ttm_bo_unreserve); | |||
| 247 | /* | 309 | /* |
| 248 | * Call bo->mutex locked. | 310 | * Call bo->mutex locked. |
| 249 | */ | 311 | */ |
| 250 | |||
| 251 | static int ttm_bo_add_ttm(struct ttm_buffer_object *bo, bool zero_alloc) | 312 | static int ttm_bo_add_ttm(struct ttm_buffer_object *bo, bool zero_alloc) |
| 252 | { | 313 | { |
| 253 | struct ttm_bo_device *bdev = bo->bdev; | 314 | struct ttm_bo_device *bdev = bo->bdev; |
| @@ -275,9 +336,10 @@ static int ttm_bo_add_ttm(struct ttm_buffer_object *bo, bool zero_alloc) | |||
| 275 | bo->ttm = ttm_tt_create(bdev, bo->num_pages << PAGE_SHIFT, | 336 | bo->ttm = ttm_tt_create(bdev, bo->num_pages << PAGE_SHIFT, |
| 276 | page_flags | TTM_PAGE_FLAG_USER, | 337 | page_flags | TTM_PAGE_FLAG_USER, |
| 277 | glob->dummy_read_page); | 338 | glob->dummy_read_page); |
| 278 | if (unlikely(bo->ttm == NULL)) | 339 | if (unlikely(bo->ttm == NULL)) { |
| 279 | ret = -ENOMEM; | 340 | ret = -ENOMEM; |
| 280 | break; | 341 | break; |
| 342 | } | ||
| 281 | 343 | ||
| 282 | ret = ttm_tt_set_user(bo->ttm, current, | 344 | ret = ttm_tt_set_user(bo->ttm, current, |
| 283 | bo->buffer_start, bo->num_pages); | 345 | bo->buffer_start, bo->num_pages); |
| @@ -328,14 +390,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, | |||
| 328 | } | 390 | } |
| 329 | 391 | ||
| 330 | if (bo->mem.mem_type == TTM_PL_SYSTEM) { | 392 | if (bo->mem.mem_type == TTM_PL_SYSTEM) { |
| 331 | 393 | bo->mem = *mem; | |
| 332 | struct ttm_mem_reg *old_mem = &bo->mem; | ||
| 333 | uint32_t save_flags = old_mem->placement; | ||
| 334 | |||
| 335 | *old_mem = *mem; | ||
| 336 | mem->mm_node = NULL; | 394 | mem->mm_node = NULL; |
| 337 | ttm_flag_masked(&save_flags, mem->placement, | ||
| 338 | TTM_PL_MASK_MEMTYPE); | ||
| 339 | goto moved; | 395 | goto moved; |
| 340 | } | 396 | } |
| 341 | 397 | ||
| @@ -418,6 +474,7 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo, bool remove_all) | |||
| 418 | kref_put(&bo->list_kref, ttm_bo_ref_bug); | 474 | kref_put(&bo->list_kref, ttm_bo_ref_bug); |
| 419 | } | 475 | } |
| 420 | if (bo->mem.mm_node) { | 476 | if (bo->mem.mm_node) { |
| 477 | bo->mem.mm_node->private = NULL; | ||
| 421 | drm_mm_put_block(bo->mem.mm_node); | 478 | drm_mm_put_block(bo->mem.mm_node); |
| 422 | bo->mem.mm_node = NULL; | 479 | bo->mem.mm_node = NULL; |
| 423 | } | 480 | } |
| @@ -554,24 +611,21 @@ void ttm_bo_unref(struct ttm_buffer_object **p_bo) | |||
| 554 | } | 611 | } |
| 555 | EXPORT_SYMBOL(ttm_bo_unref); | 612 | EXPORT_SYMBOL(ttm_bo_unref); |
| 556 | 613 | ||
| 557 | static int ttm_bo_evict(struct ttm_buffer_object *bo, unsigned mem_type, | 614 | static int ttm_bo_evict(struct ttm_buffer_object *bo, bool interruptible, |
| 558 | bool interruptible, bool no_wait) | 615 | bool no_wait) |
| 559 | { | 616 | { |
| 560 | int ret = 0; | ||
| 561 | struct ttm_bo_device *bdev = bo->bdev; | 617 | struct ttm_bo_device *bdev = bo->bdev; |
| 562 | struct ttm_bo_global *glob = bo->glob; | 618 | struct ttm_bo_global *glob = bo->glob; |
| 563 | struct ttm_mem_reg evict_mem; | 619 | struct ttm_mem_reg evict_mem; |
| 564 | uint32_t proposed_placement; | 620 | struct ttm_placement placement; |
| 565 | 621 | int ret = 0; | |
| 566 | if (bo->mem.mem_type != mem_type) | ||
| 567 | goto out; | ||
| 568 | 622 | ||
| 569 | spin_lock(&bo->lock); | 623 | spin_lock(&bo->lock); |
| 570 | ret = ttm_bo_wait(bo, false, interruptible, no_wait); | 624 | ret = ttm_bo_wait(bo, false, interruptible, no_wait); |
| 571 | spin_unlock(&bo->lock); | 625 | spin_unlock(&bo->lock); |
| 572 | 626 | ||
| 573 | if (unlikely(ret != 0)) { | 627 | if (unlikely(ret != 0)) { |
| 574 | if (ret != -ERESTART) { | 628 | if (ret != -ERESTARTSYS) { |
| 575 | printk(KERN_ERR TTM_PFX | 629 | printk(KERN_ERR TTM_PFX |
| 576 | "Failed to expire sync object before " | 630 | "Failed to expire sync object before " |
| 577 | "buffer eviction.\n"); | 631 | "buffer eviction.\n"); |
| @@ -584,116 +638,139 @@ static int ttm_bo_evict(struct ttm_buffer_object *bo, unsigned mem_type, | |||
| 584 | evict_mem = bo->mem; | 638 | evict_mem = bo->mem; |
| 585 | evict_mem.mm_node = NULL; | 639 | evict_mem.mm_node = NULL; |
| 586 | 640 | ||
| 587 | proposed_placement = bdev->driver->evict_flags(bo); | 641 | placement.fpfn = 0; |
| 588 | 642 | placement.lpfn = 0; | |
| 589 | ret = ttm_bo_mem_space(bo, proposed_placement, | 643 | placement.num_placement = 0; |
| 590 | &evict_mem, interruptible, no_wait); | 644 | placement.num_busy_placement = 0; |
| 591 | if (unlikely(ret != 0 && ret != -ERESTART)) | 645 | bdev->driver->evict_flags(bo, &placement); |
| 592 | ret = ttm_bo_mem_space(bo, TTM_PL_FLAG_SYSTEM, | 646 | ret = ttm_bo_mem_space(bo, &placement, &evict_mem, interruptible, |
| 593 | &evict_mem, interruptible, no_wait); | 647 | no_wait); |
| 594 | |||
| 595 | if (ret) { | 648 | if (ret) { |
| 596 | if (ret != -ERESTART) | 649 | if (ret != -ERESTARTSYS) { |
| 597 | printk(KERN_ERR TTM_PFX | 650 | printk(KERN_ERR TTM_PFX |
| 598 | "Failed to find memory space for " | 651 | "Failed to find memory space for " |
| 599 | "buffer 0x%p eviction.\n", bo); | 652 | "buffer 0x%p eviction.\n", bo); |
| 653 | ttm_bo_mem_space_debug(bo, &placement); | ||
| 654 | } | ||
| 600 | goto out; | 655 | goto out; |
| 601 | } | 656 | } |
| 602 | 657 | ||
| 603 | ret = ttm_bo_handle_move_mem(bo, &evict_mem, true, interruptible, | 658 | ret = ttm_bo_handle_move_mem(bo, &evict_mem, true, interruptible, |
| 604 | no_wait); | 659 | no_wait); |
| 605 | if (ret) { | 660 | if (ret) { |
| 606 | if (ret != -ERESTART) | 661 | if (ret != -ERESTARTSYS) |
| 607 | printk(KERN_ERR TTM_PFX "Buffer eviction failed\n"); | 662 | printk(KERN_ERR TTM_PFX "Buffer eviction failed\n"); |
| 663 | spin_lock(&glob->lru_lock); | ||
| 664 | if (evict_mem.mm_node) { | ||
| 665 | evict_mem.mm_node->private = NULL; | ||
| 666 | drm_mm_put_block(evict_mem.mm_node); | ||
| 667 | evict_mem.mm_node = NULL; | ||
| 668 | } | ||
| 669 | spin_unlock(&glob->lru_lock); | ||
| 608 | goto out; | 670 | goto out; |
| 609 | } | 671 | } |
| 672 | bo->evicted = true; | ||
| 673 | out: | ||
| 674 | return ret; | ||
| 675 | } | ||
| 676 | |||
| 677 | static int ttm_mem_evict_first(struct ttm_bo_device *bdev, | ||
| 678 | uint32_t mem_type, | ||
| 679 | bool interruptible, bool no_wait) | ||
| 680 | { | ||
| 681 | struct ttm_bo_global *glob = bdev->glob; | ||
| 682 | struct ttm_mem_type_manager *man = &bdev->man[mem_type]; | ||
| 683 | struct ttm_buffer_object *bo; | ||
| 684 | int ret, put_count = 0; | ||
| 610 | 685 | ||
| 611 | spin_lock(&glob->lru_lock); | 686 | spin_lock(&glob->lru_lock); |
| 612 | if (evict_mem.mm_node) { | 687 | bo = list_first_entry(&man->lru, struct ttm_buffer_object, lru); |
| 613 | drm_mm_put_block(evict_mem.mm_node); | 688 | kref_get(&bo->list_kref); |
| 614 | evict_mem.mm_node = NULL; | 689 | ret = ttm_bo_reserve_locked(bo, interruptible, no_wait, false, 0); |
| 615 | } | 690 | if (likely(ret == 0)) |
| 691 | put_count = ttm_bo_del_from_lru(bo); | ||
| 616 | spin_unlock(&glob->lru_lock); | 692 | spin_unlock(&glob->lru_lock); |
| 617 | bo->evicted = true; | 693 | if (unlikely(ret != 0)) |
| 618 | out: | 694 | return ret; |
| 695 | while (put_count--) | ||
| 696 | kref_put(&bo->list_kref, ttm_bo_ref_bug); | ||
| 697 | ret = ttm_bo_evict(bo, interruptible, no_wait); | ||
| 698 | ttm_bo_unreserve(bo); | ||
| 699 | kref_put(&bo->list_kref, ttm_bo_release_list); | ||
| 619 | return ret; | 700 | return ret; |
| 620 | } | 701 | } |
| 621 | 702 | ||
| 703 | static int ttm_bo_man_get_node(struct ttm_buffer_object *bo, | ||
| 704 | struct ttm_mem_type_manager *man, | ||
| 705 | struct ttm_placement *placement, | ||
| 706 | struct ttm_mem_reg *mem, | ||
| 707 | struct drm_mm_node **node) | ||
| 708 | { | ||
| 709 | struct ttm_bo_global *glob = bo->glob; | ||
| 710 | unsigned long lpfn; | ||
| 711 | int ret; | ||
| 712 | |||
| 713 | lpfn = placement->lpfn; | ||
| 714 | if (!lpfn) | ||
| 715 | lpfn = man->size; | ||
| 716 | *node = NULL; | ||
| 717 | do { | ||
| 718 | ret = drm_mm_pre_get(&man->manager); | ||
| 719 | if (unlikely(ret)) | ||
| 720 | return ret; | ||
| 721 | |||
| 722 | spin_lock(&glob->lru_lock); | ||
| 723 | *node = drm_mm_search_free_in_range(&man->manager, | ||
| 724 | mem->num_pages, mem->page_alignment, | ||
| 725 | placement->fpfn, lpfn, 1); | ||
| 726 | if (unlikely(*node == NULL)) { | ||
| 727 | spin_unlock(&glob->lru_lock); | ||
| 728 | return 0; | ||
| 729 | } | ||
| 730 | *node = drm_mm_get_block_atomic_range(*node, mem->num_pages, | ||
| 731 | mem->page_alignment, | ||
| 732 | placement->fpfn, | ||
| 733 | lpfn); | ||
| 734 | spin_unlock(&glob->lru_lock); | ||
| 735 | } while (*node == NULL); | ||
| 736 | return 0; | ||
| 737 | } | ||
| 738 | |||
| 622 | /** | 739 | /** |
| 623 | * Repeatedly evict memory from the LRU for @mem_type until we create enough | 740 | * Repeatedly evict memory from the LRU for @mem_type until we create enough |
| 624 | * space, or we've evicted everything and there isn't enough space. | 741 | * space, or we've evicted everything and there isn't enough space. |
| 625 | */ | 742 | */ |
| 626 | static int ttm_bo_mem_force_space(struct ttm_bo_device *bdev, | 743 | static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo, |
| 627 | struct ttm_mem_reg *mem, | 744 | uint32_t mem_type, |
| 628 | uint32_t mem_type, | 745 | struct ttm_placement *placement, |
| 629 | bool interruptible, bool no_wait) | 746 | struct ttm_mem_reg *mem, |
| 747 | bool interruptible, bool no_wait) | ||
| 630 | { | 748 | { |
| 749 | struct ttm_bo_device *bdev = bo->bdev; | ||
| 631 | struct ttm_bo_global *glob = bdev->glob; | 750 | struct ttm_bo_global *glob = bdev->glob; |
| 632 | struct drm_mm_node *node; | ||
| 633 | struct ttm_buffer_object *entry; | ||
| 634 | struct ttm_mem_type_manager *man = &bdev->man[mem_type]; | 751 | struct ttm_mem_type_manager *man = &bdev->man[mem_type]; |
| 635 | struct list_head *lru; | 752 | struct drm_mm_node *node; |
| 636 | unsigned long num_pages = mem->num_pages; | ||
| 637 | int put_count = 0; | ||
| 638 | int ret; | 753 | int ret; |
| 639 | 754 | ||
| 640 | retry_pre_get: | ||
| 641 | ret = drm_mm_pre_get(&man->manager); | ||
| 642 | if (unlikely(ret != 0)) | ||
| 643 | return ret; | ||
| 644 | |||
| 645 | spin_lock(&glob->lru_lock); | ||
| 646 | do { | 755 | do { |
| 647 | node = drm_mm_search_free(&man->manager, num_pages, | 756 | ret = ttm_bo_man_get_node(bo, man, placement, mem, &node); |
| 648 | mem->page_alignment, 1); | 757 | if (unlikely(ret != 0)) |
| 758 | return ret; | ||
| 649 | if (node) | 759 | if (node) |
| 650 | break; | 760 | break; |
| 651 | 761 | spin_lock(&glob->lru_lock); | |
| 652 | lru = &man->lru; | 762 | if (list_empty(&man->lru)) { |
| 653 | if (list_empty(lru)) | 763 | spin_unlock(&glob->lru_lock); |
| 654 | break; | 764 | break; |
| 655 | 765 | } | |
| 656 | entry = list_first_entry(lru, struct ttm_buffer_object, lru); | ||
| 657 | kref_get(&entry->list_kref); | ||
| 658 | |||
| 659 | ret = | ||
| 660 | ttm_bo_reserve_locked(entry, interruptible, no_wait, | ||
| 661 | false, 0); | ||
| 662 | |||
| 663 | if (likely(ret == 0)) | ||
| 664 | put_count = ttm_bo_del_from_lru(entry); | ||
| 665 | |||
| 666 | spin_unlock(&glob->lru_lock); | 766 | spin_unlock(&glob->lru_lock); |
| 667 | 767 | ret = ttm_mem_evict_first(bdev, mem_type, interruptible, | |
| 768 | no_wait); | ||
| 668 | if (unlikely(ret != 0)) | 769 | if (unlikely(ret != 0)) |
| 669 | return ret; | 770 | return ret; |
| 670 | |||
| 671 | while (put_count--) | ||
| 672 | kref_put(&entry->list_kref, ttm_bo_ref_bug); | ||
| 673 | |||
| 674 | ret = ttm_bo_evict(entry, mem_type, interruptible, no_wait); | ||
| 675 | |||
| 676 | ttm_bo_unreserve(entry); | ||
| 677 | |||
| 678 | kref_put(&entry->list_kref, ttm_bo_release_list); | ||
| 679 | if (ret) | ||
| 680 | return ret; | ||
| 681 | |||
| 682 | spin_lock(&glob->lru_lock); | ||
| 683 | } while (1); | 771 | } while (1); |
| 684 | 772 | if (node == NULL) | |
| 685 | if (!node) { | ||
| 686 | spin_unlock(&glob->lru_lock); | ||
| 687 | return -ENOMEM; | 773 | return -ENOMEM; |
| 688 | } | ||
| 689 | |||
| 690 | node = drm_mm_get_block_atomic(node, num_pages, mem->page_alignment); | ||
| 691 | if (unlikely(!node)) { | ||
| 692 | spin_unlock(&glob->lru_lock); | ||
| 693 | goto retry_pre_get; | ||
| 694 | } | ||
| 695 | |||
| 696 | spin_unlock(&glob->lru_lock); | ||
| 697 | mem->mm_node = node; | 774 | mem->mm_node = node; |
| 698 | mem->mem_type = mem_type; | 775 | mem->mem_type = mem_type; |
| 699 | return 0; | 776 | return 0; |
| @@ -724,7 +801,6 @@ static uint32_t ttm_bo_select_caching(struct ttm_mem_type_manager *man, | |||
| 724 | return result; | 801 | return result; |
| 725 | } | 802 | } |
| 726 | 803 | ||
| 727 | |||
| 728 | static bool ttm_bo_mt_compatible(struct ttm_mem_type_manager *man, | 804 | static bool ttm_bo_mt_compatible(struct ttm_mem_type_manager *man, |
| 729 | bool disallow_fixed, | 805 | bool disallow_fixed, |
| 730 | uint32_t mem_type, | 806 | uint32_t mem_type, |
| @@ -757,66 +833,55 @@ static bool ttm_bo_mt_compatible(struct ttm_mem_type_manager *man, | |||
| 757 | * space. | 833 | * space. |
| 758 | */ | 834 | */ |
| 759 | int ttm_bo_mem_space(struct ttm_buffer_object *bo, | 835 | int ttm_bo_mem_space(struct ttm_buffer_object *bo, |
| 760 | uint32_t proposed_placement, | 836 | struct ttm_placement *placement, |
| 761 | struct ttm_mem_reg *mem, | 837 | struct ttm_mem_reg *mem, |
| 762 | bool interruptible, bool no_wait) | 838 | bool interruptible, bool no_wait) |
| 763 | { | 839 | { |
| 764 | struct ttm_bo_device *bdev = bo->bdev; | 840 | struct ttm_bo_device *bdev = bo->bdev; |
| 765 | struct ttm_bo_global *glob = bo->glob; | ||
| 766 | struct ttm_mem_type_manager *man; | 841 | struct ttm_mem_type_manager *man; |
| 767 | |||
| 768 | uint32_t num_prios = bdev->driver->num_mem_type_prio; | ||
| 769 | const uint32_t *prios = bdev->driver->mem_type_prio; | ||
| 770 | uint32_t i; | ||
| 771 | uint32_t mem_type = TTM_PL_SYSTEM; | 842 | uint32_t mem_type = TTM_PL_SYSTEM; |
| 772 | uint32_t cur_flags = 0; | 843 | uint32_t cur_flags = 0; |
| 773 | bool type_found = false; | 844 | bool type_found = false; |
| 774 | bool type_ok = false; | 845 | bool type_ok = false; |
| 775 | bool has_eagain = false; | 846 | bool has_erestartsys = false; |
| 776 | struct drm_mm_node *node = NULL; | 847 | struct drm_mm_node *node = NULL; |
| 777 | int ret; | 848 | int i, ret; |
| 778 | 849 | ||
| 779 | mem->mm_node = NULL; | 850 | mem->mm_node = NULL; |
| 780 | for (i = 0; i < num_prios; ++i) { | 851 | for (i = 0; i <= placement->num_placement; ++i) { |
| 781 | mem_type = prios[i]; | 852 | ret = ttm_mem_type_from_flags(placement->placement[i], |
| 853 | &mem_type); | ||
| 854 | if (ret) | ||
| 855 | return ret; | ||
| 782 | man = &bdev->man[mem_type]; | 856 | man = &bdev->man[mem_type]; |
| 783 | 857 | ||
| 784 | type_ok = ttm_bo_mt_compatible(man, | 858 | type_ok = ttm_bo_mt_compatible(man, |
| 785 | bo->type == ttm_bo_type_user, | 859 | bo->type == ttm_bo_type_user, |
| 786 | mem_type, proposed_placement, | 860 | mem_type, |
| 787 | &cur_flags); | 861 | placement->placement[i], |
| 862 | &cur_flags); | ||
| 788 | 863 | ||
| 789 | if (!type_ok) | 864 | if (!type_ok) |
| 790 | continue; | 865 | continue; |
| 791 | 866 | ||
| 792 | cur_flags = ttm_bo_select_caching(man, bo->mem.placement, | 867 | cur_flags = ttm_bo_select_caching(man, bo->mem.placement, |
| 793 | cur_flags); | 868 | cur_flags); |
| 869 | /* | ||
| 870 | * Use the access and other non-mapping-related flag bits from | ||
| 871 | * the memory placement flags to the current flags | ||
| 872 | */ | ||
| 873 | ttm_flag_masked(&cur_flags, placement->placement[i], | ||
| 874 | ~TTM_PL_MASK_MEMTYPE); | ||
| 794 | 875 | ||
| 795 | if (mem_type == TTM_PL_SYSTEM) | 876 | if (mem_type == TTM_PL_SYSTEM) |
| 796 | break; | 877 | break; |
| 797 | 878 | ||
| 798 | if (man->has_type && man->use_type) { | 879 | if (man->has_type && man->use_type) { |
| 799 | type_found = true; | 880 | type_found = true; |
| 800 | do { | 881 | ret = ttm_bo_man_get_node(bo, man, placement, mem, |
| 801 | ret = drm_mm_pre_get(&man->manager); | 882 | &node); |
| 802 | if (unlikely(ret)) | 883 | if (unlikely(ret)) |
| 803 | return ret; | 884 | return ret; |
| 804 | |||
| 805 | spin_lock(&glob->lru_lock); | ||
| 806 | node = drm_mm_search_free(&man->manager, | ||
| 807 | mem->num_pages, | ||
| 808 | mem->page_alignment, | ||
| 809 | 1); | ||
| 810 | if (unlikely(!node)) { | ||
| 811 | spin_unlock(&glob->lru_lock); | ||
| 812 | break; | ||
| 813 | } | ||
| 814 | node = drm_mm_get_block_atomic(node, | ||
| 815 | mem->num_pages, | ||
| 816 | mem-> | ||
| 817 | page_alignment); | ||
| 818 | spin_unlock(&glob->lru_lock); | ||
| 819 | } while (!node); | ||
| 820 | } | 885 | } |
| 821 | if (node) | 886 | if (node) |
| 822 | break; | 887 | break; |
| @@ -826,67 +891,65 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo, | |||
| 826 | mem->mm_node = node; | 891 | mem->mm_node = node; |
| 827 | mem->mem_type = mem_type; | 892 | mem->mem_type = mem_type; |
| 828 | mem->placement = cur_flags; | 893 | mem->placement = cur_flags; |
| 894 | if (node) | ||
| 895 | node->private = bo; | ||
| 829 | return 0; | 896 | return 0; |
| 830 | } | 897 | } |
| 831 | 898 | ||
| 832 | if (!type_found) | 899 | if (!type_found) |
| 833 | return -EINVAL; | 900 | return -EINVAL; |
| 834 | 901 | ||
| 835 | num_prios = bdev->driver->num_mem_busy_prio; | 902 | for (i = 0; i <= placement->num_busy_placement; ++i) { |
| 836 | prios = bdev->driver->mem_busy_prio; | 903 | ret = ttm_mem_type_from_flags(placement->placement[i], |
| 837 | 904 | &mem_type); | |
| 838 | for (i = 0; i < num_prios; ++i) { | 905 | if (ret) |
| 839 | mem_type = prios[i]; | 906 | return ret; |
| 840 | man = &bdev->man[mem_type]; | 907 | man = &bdev->man[mem_type]; |
| 841 | |||
| 842 | if (!man->has_type) | 908 | if (!man->has_type) |
| 843 | continue; | 909 | continue; |
| 844 | |||
| 845 | if (!ttm_bo_mt_compatible(man, | 910 | if (!ttm_bo_mt_compatible(man, |
| 846 | bo->type == ttm_bo_type_user, | 911 | bo->type == ttm_bo_type_user, |
| 847 | mem_type, | 912 | mem_type, |
| 848 | proposed_placement, &cur_flags)) | 913 | placement->placement[i], |
| 914 | &cur_flags)) | ||
| 849 | continue; | 915 | continue; |
| 850 | 916 | ||
| 851 | cur_flags = ttm_bo_select_caching(man, bo->mem.placement, | 917 | cur_flags = ttm_bo_select_caching(man, bo->mem.placement, |
| 852 | cur_flags); | 918 | cur_flags); |
| 919 | /* | ||
| 920 | * Use the access and other non-mapping-related flag bits from | ||
| 921 | * the memory placement flags to the current flags | ||
| 922 | */ | ||
| 923 | ttm_flag_masked(&cur_flags, placement->placement[i], | ||
| 924 | ~TTM_PL_MASK_MEMTYPE); | ||
| 853 | 925 | ||
| 854 | ret = ttm_bo_mem_force_space(bdev, mem, mem_type, | 926 | ret = ttm_bo_mem_force_space(bo, mem_type, placement, mem, |
| 855 | interruptible, no_wait); | 927 | interruptible, no_wait); |
| 856 | |||
| 857 | if (ret == 0 && mem->mm_node) { | 928 | if (ret == 0 && mem->mm_node) { |
| 858 | mem->placement = cur_flags; | 929 | mem->placement = cur_flags; |
| 930 | mem->mm_node->private = bo; | ||
| 859 | return 0; | 931 | return 0; |
| 860 | } | 932 | } |
| 861 | 933 | if (ret == -ERESTARTSYS) | |
| 862 | if (ret == -ERESTART) | 934 | has_erestartsys = true; |
| 863 | has_eagain = true; | ||
| 864 | } | 935 | } |
| 865 | 936 | ret = (has_erestartsys) ? -ERESTARTSYS : -ENOMEM; | |
| 866 | ret = (has_eagain) ? -ERESTART : -ENOMEM; | ||
| 867 | return ret; | 937 | return ret; |
| 868 | } | 938 | } |
| 869 | EXPORT_SYMBOL(ttm_bo_mem_space); | 939 | EXPORT_SYMBOL(ttm_bo_mem_space); |
| 870 | 940 | ||
| 871 | int ttm_bo_wait_cpu(struct ttm_buffer_object *bo, bool no_wait) | 941 | int ttm_bo_wait_cpu(struct ttm_buffer_object *bo, bool no_wait) |
| 872 | { | 942 | { |
| 873 | int ret = 0; | ||
| 874 | |||
| 875 | if ((atomic_read(&bo->cpu_writers) > 0) && no_wait) | 943 | if ((atomic_read(&bo->cpu_writers) > 0) && no_wait) |
| 876 | return -EBUSY; | 944 | return -EBUSY; |
| 877 | 945 | ||
| 878 | ret = wait_event_interruptible(bo->event_queue, | 946 | return wait_event_interruptible(bo->event_queue, |
| 879 | atomic_read(&bo->cpu_writers) == 0); | 947 | atomic_read(&bo->cpu_writers) == 0); |
| 880 | |||
| 881 | if (ret == -ERESTARTSYS) | ||
| 882 | ret = -ERESTART; | ||
| 883 | |||
| 884 | return ret; | ||
| 885 | } | 948 | } |
| 886 | 949 | ||
| 887 | int ttm_bo_move_buffer(struct ttm_buffer_object *bo, | 950 | int ttm_bo_move_buffer(struct ttm_buffer_object *bo, |
| 888 | uint32_t proposed_placement, | 951 | struct ttm_placement *placement, |
| 889 | bool interruptible, bool no_wait) | 952 | bool interruptible, bool no_wait) |
| 890 | { | 953 | { |
| 891 | struct ttm_bo_global *glob = bo->glob; | 954 | struct ttm_bo_global *glob = bo->glob; |
| 892 | int ret = 0; | 955 | int ret = 0; |
| @@ -899,101 +962,82 @@ int ttm_bo_move_buffer(struct ttm_buffer_object *bo, | |||
| 899 | * Have the driver move function wait for idle when necessary, | 962 | * Have the driver move function wait for idle when necessary, |
| 900 | * instead of doing it here. | 963 | * instead of doing it here. |
| 901 | */ | 964 | */ |
| 902 | |||
| 903 | spin_lock(&bo->lock); | 965 | spin_lock(&bo->lock); |
| 904 | ret = ttm_bo_wait(bo, false, interruptible, no_wait); | 966 | ret = ttm_bo_wait(bo, false, interruptible, no_wait); |
| 905 | spin_unlock(&bo->lock); | 967 | spin_unlock(&bo->lock); |
| 906 | |||
| 907 | if (ret) | 968 | if (ret) |
| 908 | return ret; | 969 | return ret; |
| 909 | |||
| 910 | mem.num_pages = bo->num_pages; | 970 | mem.num_pages = bo->num_pages; |
| 911 | mem.size = mem.num_pages << PAGE_SHIFT; | 971 | mem.size = mem.num_pages << PAGE_SHIFT; |
| 912 | mem.page_alignment = bo->mem.page_alignment; | 972 | mem.page_alignment = bo->mem.page_alignment; |
| 913 | |||
| 914 | /* | 973 | /* |
| 915 | * Determine where to move the buffer. | 974 | * Determine where to move the buffer. |
| 916 | */ | 975 | */ |
| 917 | 976 | ret = ttm_bo_mem_space(bo, placement, &mem, interruptible, no_wait); | |
| 918 | ret = ttm_bo_mem_space(bo, proposed_placement, &mem, | ||
| 919 | interruptible, no_wait); | ||
| 920 | if (ret) | 977 | if (ret) |
| 921 | goto out_unlock; | 978 | goto out_unlock; |
| 922 | |||
| 923 | ret = ttm_bo_handle_move_mem(bo, &mem, false, interruptible, no_wait); | 979 | ret = ttm_bo_handle_move_mem(bo, &mem, false, interruptible, no_wait); |
| 924 | |||
| 925 | out_unlock: | 980 | out_unlock: |
| 926 | if (ret && mem.mm_node) { | 981 | if (ret && mem.mm_node) { |
| 927 | spin_lock(&glob->lru_lock); | 982 | spin_lock(&glob->lru_lock); |
| 983 | mem.mm_node->private = NULL; | ||
| 928 | drm_mm_put_block(mem.mm_node); | 984 | drm_mm_put_block(mem.mm_node); |
| 929 | spin_unlock(&glob->lru_lock); | 985 | spin_unlock(&glob->lru_lock); |
| 930 | } | 986 | } |
| 931 | return ret; | 987 | return ret; |
| 932 | } | 988 | } |
| 933 | 989 | ||
| 934 | static int ttm_bo_mem_compat(uint32_t proposed_placement, | 990 | static int ttm_bo_mem_compat(struct ttm_placement *placement, |
| 935 | struct ttm_mem_reg *mem) | 991 | struct ttm_mem_reg *mem) |
| 936 | { | 992 | { |
| 937 | if ((proposed_placement & mem->placement & TTM_PL_MASK_MEM) == 0) | 993 | int i; |
| 938 | return 0; | 994 | |
| 939 | if ((proposed_placement & mem->placement & TTM_PL_MASK_CACHING) == 0) | 995 | for (i = 0; i < placement->num_placement; i++) { |
| 940 | return 0; | 996 | if ((placement->placement[i] & mem->placement & |
| 941 | 997 | TTM_PL_MASK_CACHING) && | |
| 942 | return 1; | 998 | (placement->placement[i] & mem->placement & |
| 999 | TTM_PL_MASK_MEM)) | ||
| 1000 | return i; | ||
| 1001 | } | ||
| 1002 | return -1; | ||
| 943 | } | 1003 | } |
| 944 | 1004 | ||
| 945 | int ttm_buffer_object_validate(struct ttm_buffer_object *bo, | 1005 | int ttm_buffer_object_validate(struct ttm_buffer_object *bo, |
| 946 | uint32_t proposed_placement, | 1006 | struct ttm_placement *placement, |
| 947 | bool interruptible, bool no_wait) | 1007 | bool interruptible, bool no_wait) |
| 948 | { | 1008 | { |
| 949 | int ret; | 1009 | int ret; |
| 950 | 1010 | ||
| 951 | BUG_ON(!atomic_read(&bo->reserved)); | 1011 | BUG_ON(!atomic_read(&bo->reserved)); |
| 952 | bo->proposed_placement = proposed_placement; | 1012 | /* Check that range is valid */ |
| 953 | 1013 | if (placement->lpfn || placement->fpfn) | |
| 954 | TTM_DEBUG("Proposed placement 0x%08lx, Old flags 0x%08lx\n", | 1014 | if (placement->fpfn > placement->lpfn || |
| 955 | (unsigned long)proposed_placement, | 1015 | (placement->lpfn - placement->fpfn) < bo->num_pages) |
| 956 | (unsigned long)bo->mem.placement); | 1016 | return -EINVAL; |
| 957 | |||
| 958 | /* | 1017 | /* |
| 959 | * Check whether we need to move buffer. | 1018 | * Check whether we need to move buffer. |
| 960 | */ | 1019 | */ |
| 961 | 1020 | ret = ttm_bo_mem_compat(placement, &bo->mem); | |
| 962 | if (!ttm_bo_mem_compat(bo->proposed_placement, &bo->mem)) { | 1021 | if (ret < 0) { |
| 963 | ret = ttm_bo_move_buffer(bo, bo->proposed_placement, | 1022 | ret = ttm_bo_move_buffer(bo, placement, interruptible, no_wait); |
| 964 | interruptible, no_wait); | 1023 | if (ret) |
| 965 | if (ret) { | ||
| 966 | if (ret != -ERESTART) | ||
| 967 | printk(KERN_ERR TTM_PFX | ||
| 968 | "Failed moving buffer. " | ||
| 969 | "Proposed placement 0x%08x\n", | ||
| 970 | bo->proposed_placement); | ||
| 971 | if (ret == -ENOMEM) | ||
| 972 | printk(KERN_ERR TTM_PFX | ||
| 973 | "Out of aperture space or " | ||
| 974 | "DRM memory quota.\n"); | ||
| 975 | return ret; | 1024 | return ret; |
| 976 | } | 1025 | } else { |
| 1026 | /* | ||
| 1027 | * Use the access and other non-mapping-related flag bits from | ||
| 1028 | * the compatible memory placement flags to the active flags | ||
| 1029 | */ | ||
| 1030 | ttm_flag_masked(&bo->mem.placement, placement->placement[ret], | ||
| 1031 | ~TTM_PL_MASK_MEMTYPE); | ||
| 977 | } | 1032 | } |
| 978 | |||
| 979 | /* | 1033 | /* |
| 980 | * We might need to add a TTM. | 1034 | * We might need to add a TTM. |
| 981 | */ | 1035 | */ |
| 982 | |||
| 983 | if (bo->mem.mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { | 1036 | if (bo->mem.mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { |
| 984 | ret = ttm_bo_add_ttm(bo, true); | 1037 | ret = ttm_bo_add_ttm(bo, true); |
| 985 | if (ret) | 1038 | if (ret) |
| 986 | return ret; | 1039 | return ret; |
| 987 | } | 1040 | } |
| 988 | /* | ||
| 989 | * Validation has succeeded, move the access and other | ||
| 990 | * non-mapping-related flag bits from the proposed flags to | ||
| 991 | * the active flags | ||
| 992 | */ | ||
| 993 | |||
| 994 | ttm_flag_masked(&bo->mem.placement, bo->proposed_placement, | ||
| 995 | ~TTM_PL_MASK_MEMTYPE); | ||
| 996 | |||
| 997 | return 0; | 1041 | return 0; |
| 998 | } | 1042 | } |
| 999 | EXPORT_SYMBOL(ttm_buffer_object_validate); | 1043 | EXPORT_SYMBOL(ttm_buffer_object_validate); |
| @@ -1041,8 +1085,10 @@ int ttm_buffer_object_init(struct ttm_bo_device *bdev, | |||
| 1041 | size_t acc_size, | 1085 | size_t acc_size, |
| 1042 | void (*destroy) (struct ttm_buffer_object *)) | 1086 | void (*destroy) (struct ttm_buffer_object *)) |
| 1043 | { | 1087 | { |
| 1044 | int ret = 0; | 1088 | int i, c, ret = 0; |
| 1045 | unsigned long num_pages; | 1089 | unsigned long num_pages; |
| 1090 | uint32_t placements[8]; | ||
| 1091 | struct ttm_placement placement; | ||
| 1046 | 1092 | ||
| 1047 | size += buffer_start & ~PAGE_MASK; | 1093 | size += buffer_start & ~PAGE_MASK; |
| 1048 | num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; | 1094 | num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; |
| @@ -1099,7 +1145,16 @@ int ttm_buffer_object_init(struct ttm_bo_device *bdev, | |||
| 1099 | goto out_err; | 1145 | goto out_err; |
| 1100 | } | 1146 | } |
| 1101 | 1147 | ||
| 1102 | ret = ttm_buffer_object_validate(bo, flags, interruptible, false); | 1148 | placement.fpfn = 0; |
| 1149 | placement.lpfn = 0; | ||
| 1150 | for (i = 0, c = 0; i <= TTM_PL_PRIV5; i++) | ||
| 1151 | if (flags & (1 << i)) | ||
| 1152 | placements[c++] = (flags & ~TTM_PL_MASK_MEM) | (1 << i); | ||
| 1153 | placement.placement = placements; | ||
| 1154 | placement.num_placement = c; | ||
| 1155 | placement.busy_placement = placements; | ||
| 1156 | placement.num_busy_placement = c; | ||
| 1157 | ret = ttm_buffer_object_validate(bo, &placement, interruptible, false); | ||
| 1103 | if (ret) | 1158 | if (ret) |
| 1104 | goto out_err; | 1159 | goto out_err; |
| 1105 | 1160 | ||
| @@ -1134,8 +1189,8 @@ int ttm_buffer_object_create(struct ttm_bo_device *bdev, | |||
| 1134 | struct ttm_buffer_object **p_bo) | 1189 | struct ttm_buffer_object **p_bo) |
| 1135 | { | 1190 | { |
| 1136 | struct ttm_buffer_object *bo; | 1191 | struct ttm_buffer_object *bo; |
| 1137 | int ret; | ||
| 1138 | struct ttm_mem_global *mem_glob = bdev->glob->mem_glob; | 1192 | struct ttm_mem_global *mem_glob = bdev->glob->mem_glob; |
| 1193 | int ret; | ||
| 1139 | 1194 | ||
| 1140 | size_t acc_size = | 1195 | size_t acc_size = |
| 1141 | ttm_bo_size(bdev->glob, (size + PAGE_SIZE - 1) >> PAGE_SHIFT); | 1196 | ttm_bo_size(bdev->glob, (size + PAGE_SIZE - 1) >> PAGE_SHIFT); |
| @@ -1160,66 +1215,32 @@ int ttm_buffer_object_create(struct ttm_bo_device *bdev, | |||
| 1160 | return ret; | 1215 | return ret; |
| 1161 | } | 1216 | } |
| 1162 | 1217 | ||
| 1163 | static int ttm_bo_leave_list(struct ttm_buffer_object *bo, | ||
| 1164 | uint32_t mem_type, bool allow_errors) | ||
| 1165 | { | ||
| 1166 | int ret; | ||
| 1167 | |||
| 1168 | spin_lock(&bo->lock); | ||
| 1169 | ret = ttm_bo_wait(bo, false, false, false); | ||
| 1170 | spin_unlock(&bo->lock); | ||
| 1171 | |||
| 1172 | if (ret && allow_errors) | ||
| 1173 | goto out; | ||
| 1174 | |||
| 1175 | if (bo->mem.mem_type == mem_type) | ||
| 1176 | ret = ttm_bo_evict(bo, mem_type, false, false); | ||
| 1177 | |||
| 1178 | if (ret) { | ||
| 1179 | if (allow_errors) { | ||
| 1180 | goto out; | ||
| 1181 | } else { | ||
| 1182 | ret = 0; | ||
| 1183 | printk(KERN_ERR TTM_PFX "Cleanup eviction failed\n"); | ||
| 1184 | } | ||
| 1185 | } | ||
| 1186 | |||
| 1187 | out: | ||
| 1188 | return ret; | ||
| 1189 | } | ||
| 1190 | |||
| 1191 | static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev, | 1218 | static int ttm_bo_force_list_clean(struct ttm_bo_device *bdev, |
| 1192 | struct list_head *head, | 1219 | unsigned mem_type, bool allow_errors) |
| 1193 | unsigned mem_type, bool allow_errors) | ||
| 1194 | { | 1220 | { |
| 1221 | struct ttm_mem_type_manager *man = &bdev->man[mem_type]; | ||
| 1195 | struct ttm_bo_global *glob = bdev->glob; | 1222 | struct ttm_bo_global *glob = bdev->glob; |
| 1196 | struct ttm_buffer_object *entry; | ||
| 1197 | int ret; | 1223 | int ret; |
| 1198 | int put_count; | ||
| 1199 | 1224 | ||
| 1200 | /* | 1225 | /* |
| 1201 | * Can't use standard list traversal since we're unlocking. | 1226 | * Can't use standard list traversal since we're unlocking. |
| 1202 | */ | 1227 | */ |
| 1203 | 1228 | ||
| 1204 | spin_lock(&glob->lru_lock); | 1229 | spin_lock(&glob->lru_lock); |
| 1205 | 1230 | while (!list_empty(&man->lru)) { | |
| 1206 | while (!list_empty(head)) { | ||
| 1207 | entry = list_first_entry(head, struct ttm_buffer_object, lru); | ||
| 1208 | kref_get(&entry->list_kref); | ||
| 1209 | ret = ttm_bo_reserve_locked(entry, false, false, false, 0); | ||
| 1210 | put_count = ttm_bo_del_from_lru(entry); | ||
| 1211 | spin_unlock(&glob->lru_lock); | 1231 | spin_unlock(&glob->lru_lock); |
| 1212 | while (put_count--) | 1232 | ret = ttm_mem_evict_first(bdev, mem_type, false, false); |
| 1213 | kref_put(&entry->list_kref, ttm_bo_ref_bug); | 1233 | if (ret) { |
| 1214 | BUG_ON(ret); | 1234 | if (allow_errors) { |
| 1215 | ret = ttm_bo_leave_list(entry, mem_type, allow_errors); | 1235 | return ret; |
| 1216 | ttm_bo_unreserve(entry); | 1236 | } else { |
| 1217 | kref_put(&entry->list_kref, ttm_bo_release_list); | 1237 | printk(KERN_ERR TTM_PFX |
| 1238 | "Cleanup eviction failed\n"); | ||
| 1239 | } | ||
| 1240 | } | ||
| 1218 | spin_lock(&glob->lru_lock); | 1241 | spin_lock(&glob->lru_lock); |
| 1219 | } | 1242 | } |
| 1220 | |||
| 1221 | spin_unlock(&glob->lru_lock); | 1243 | spin_unlock(&glob->lru_lock); |
| 1222 | |||
| 1223 | return 0; | 1244 | return 0; |
| 1224 | } | 1245 | } |
| 1225 | 1246 | ||
| @@ -1246,7 +1267,7 @@ int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type) | |||
| 1246 | 1267 | ||
| 1247 | ret = 0; | 1268 | ret = 0; |
| 1248 | if (mem_type > 0) { | 1269 | if (mem_type > 0) { |
| 1249 | ttm_bo_force_list_clean(bdev, &man->lru, mem_type, false); | 1270 | ttm_bo_force_list_clean(bdev, mem_type, false); |
| 1250 | 1271 | ||
| 1251 | spin_lock(&glob->lru_lock); | 1272 | spin_lock(&glob->lru_lock); |
| 1252 | if (drm_mm_clean(&man->manager)) | 1273 | if (drm_mm_clean(&man->manager)) |
| @@ -1279,12 +1300,12 @@ int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) | |||
| 1279 | return 0; | 1300 | return 0; |
| 1280 | } | 1301 | } |
| 1281 | 1302 | ||
| 1282 | return ttm_bo_force_list_clean(bdev, &man->lru, mem_type, true); | 1303 | return ttm_bo_force_list_clean(bdev, mem_type, true); |
| 1283 | } | 1304 | } |
| 1284 | EXPORT_SYMBOL(ttm_bo_evict_mm); | 1305 | EXPORT_SYMBOL(ttm_bo_evict_mm); |
| 1285 | 1306 | ||
| 1286 | int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, | 1307 | int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, |
| 1287 | unsigned long p_offset, unsigned long p_size) | 1308 | unsigned long p_size) |
| 1288 | { | 1309 | { |
| 1289 | int ret = -EINVAL; | 1310 | int ret = -EINVAL; |
| 1290 | struct ttm_mem_type_manager *man; | 1311 | struct ttm_mem_type_manager *man; |
| @@ -1314,7 +1335,7 @@ int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, | |||
| 1314 | type); | 1335 | type); |
| 1315 | return ret; | 1336 | return ret; |
| 1316 | } | 1337 | } |
| 1317 | ret = drm_mm_init(&man->manager, p_offset, p_size); | 1338 | ret = drm_mm_init(&man->manager, 0, p_size); |
| 1318 | if (ret) | 1339 | if (ret) |
| 1319 | return ret; | 1340 | return ret; |
| 1320 | } | 1341 | } |
| @@ -1463,7 +1484,7 @@ int ttm_bo_device_init(struct ttm_bo_device *bdev, | |||
| 1463 | * Initialize the system memory buffer type. | 1484 | * Initialize the system memory buffer type. |
| 1464 | * Other types need to be driver / IOCTL initialized. | 1485 | * Other types need to be driver / IOCTL initialized. |
| 1465 | */ | 1486 | */ |
| 1466 | ret = ttm_bo_init_mm(bdev, TTM_PL_SYSTEM, 0, 0); | 1487 | ret = ttm_bo_init_mm(bdev, TTM_PL_SYSTEM, 0); |
| 1467 | if (unlikely(ret != 0)) | 1488 | if (unlikely(ret != 0)) |
| 1468 | goto out_no_sys; | 1489 | goto out_no_sys; |
| 1469 | 1490 | ||
| @@ -1693,7 +1714,7 @@ int ttm_bo_block_reservation(struct ttm_buffer_object *bo, bool interruptible, | |||
| 1693 | ret = wait_event_interruptible | 1714 | ret = wait_event_interruptible |
| 1694 | (bo->event_queue, atomic_read(&bo->reserved) == 0); | 1715 | (bo->event_queue, atomic_read(&bo->reserved) == 0); |
| 1695 | if (unlikely(ret != 0)) | 1716 | if (unlikely(ret != 0)) |
| 1696 | return -ERESTART; | 1717 | return ret; |
| 1697 | } else { | 1718 | } else { |
| 1698 | wait_event(bo->event_queue, | 1719 | wait_event(bo->event_queue, |
| 1699 | atomic_read(&bo->reserved) == 0); | 1720 | atomic_read(&bo->reserved) == 0); |
diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c index 61c5572d2b91..2ecf7d0c64f6 100644 --- a/drivers/gpu/drm/ttm/ttm_bo_util.c +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c | |||
| @@ -369,6 +369,7 @@ pgprot_t ttm_io_prot(uint32_t caching_flags, pgprot_t tmp) | |||
| 369 | #endif | 369 | #endif |
| 370 | return tmp; | 370 | return tmp; |
| 371 | } | 371 | } |
| 372 | EXPORT_SYMBOL(ttm_io_prot); | ||
| 372 | 373 | ||
| 373 | static int ttm_bo_ioremap(struct ttm_buffer_object *bo, | 374 | static int ttm_bo_ioremap(struct ttm_buffer_object *bo, |
| 374 | unsigned long bus_base, | 375 | unsigned long bus_base, |
diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c index 1c040d040338..609a85a4d855 100644 --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c | |||
| @@ -114,7 +114,7 @@ static int ttm_bo_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
| 114 | ret = ttm_bo_wait(bo, false, true, false); | 114 | ret = ttm_bo_wait(bo, false, true, false); |
| 115 | spin_unlock(&bo->lock); | 115 | spin_unlock(&bo->lock); |
| 116 | if (unlikely(ret != 0)) { | 116 | if (unlikely(ret != 0)) { |
| 117 | retval = (ret != -ERESTART) ? | 117 | retval = (ret != -ERESTARTSYS) ? |
| 118 | VM_FAULT_SIGBUS : VM_FAULT_NOPAGE; | 118 | VM_FAULT_SIGBUS : VM_FAULT_NOPAGE; |
| 119 | goto out_unlock; | 119 | goto out_unlock; |
| 120 | } | 120 | } |
| @@ -349,9 +349,6 @@ ssize_t ttm_bo_io(struct ttm_bo_device *bdev, struct file *filp, | |||
| 349 | switch (ret) { | 349 | switch (ret) { |
| 350 | case 0: | 350 | case 0: |
| 351 | break; | 351 | break; |
| 352 | case -ERESTART: | ||
| 353 | ret = -EINTR; | ||
| 354 | goto out_unref; | ||
| 355 | case -EBUSY: | 352 | case -EBUSY: |
| 356 | ret = -EAGAIN; | 353 | ret = -EAGAIN; |
| 357 | goto out_unref; | 354 | goto out_unref; |
| @@ -421,8 +418,6 @@ ssize_t ttm_bo_fbdev_io(struct ttm_buffer_object *bo, const char __user *wbuf, | |||
| 421 | switch (ret) { | 418 | switch (ret) { |
| 422 | case 0: | 419 | case 0: |
| 423 | break; | 420 | break; |
| 424 | case -ERESTART: | ||
| 425 | return -EINTR; | ||
| 426 | case -EBUSY: | 421 | case -EBUSY: |
| 427 | return -EAGAIN; | 422 | return -EAGAIN; |
| 428 | default: | 423 | default: |
diff --git a/drivers/gpu/drm/ttm/ttm_execbuf_util.c b/drivers/gpu/drm/ttm/ttm_execbuf_util.c new file mode 100644 index 000000000000..c285c2902d15 --- /dev/null +++ b/drivers/gpu/drm/ttm/ttm_execbuf_util.c | |||
| @@ -0,0 +1,117 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2006-2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | |||
| 28 | #include "ttm/ttm_execbuf_util.h" | ||
| 29 | #include "ttm/ttm_bo_driver.h" | ||
| 30 | #include "ttm/ttm_placement.h" | ||
| 31 | #include <linux/wait.h> | ||
| 32 | #include <linux/sched.h> | ||
| 33 | #include <linux/module.h> | ||
| 34 | |||
| 35 | void ttm_eu_backoff_reservation(struct list_head *list) | ||
| 36 | { | ||
| 37 | struct ttm_validate_buffer *entry; | ||
| 38 | |||
| 39 | list_for_each_entry(entry, list, head) { | ||
| 40 | struct ttm_buffer_object *bo = entry->bo; | ||
| 41 | if (!entry->reserved) | ||
| 42 | continue; | ||
| 43 | |||
| 44 | entry->reserved = false; | ||
| 45 | ttm_bo_unreserve(bo); | ||
| 46 | } | ||
| 47 | } | ||
| 48 | EXPORT_SYMBOL(ttm_eu_backoff_reservation); | ||
| 49 | |||
| 50 | /* | ||
| 51 | * Reserve buffers for validation. | ||
| 52 | * | ||
| 53 | * If a buffer in the list is marked for CPU access, we back off and | ||
| 54 | * wait for that buffer to become free for GPU access. | ||
| 55 | * | ||
| 56 | * If a buffer is reserved for another validation, the validator with | ||
| 57 | * the highest validation sequence backs off and waits for that buffer | ||
| 58 | * to become unreserved. This prevents deadlocks when validating multiple | ||
| 59 | * buffers in different orders. | ||
| 60 | */ | ||
| 61 | |||
| 62 | int ttm_eu_reserve_buffers(struct list_head *list, uint32_t val_seq) | ||
| 63 | { | ||
| 64 | struct ttm_validate_buffer *entry; | ||
| 65 | int ret; | ||
| 66 | |||
| 67 | retry: | ||
| 68 | list_for_each_entry(entry, list, head) { | ||
| 69 | struct ttm_buffer_object *bo = entry->bo; | ||
| 70 | |||
| 71 | entry->reserved = false; | ||
| 72 | ret = ttm_bo_reserve(bo, true, false, true, val_seq); | ||
| 73 | if (ret != 0) { | ||
| 74 | ttm_eu_backoff_reservation(list); | ||
| 75 | if (ret == -EAGAIN) { | ||
| 76 | ret = ttm_bo_wait_unreserved(bo, true); | ||
| 77 | if (unlikely(ret != 0)) | ||
| 78 | return ret; | ||
| 79 | goto retry; | ||
| 80 | } else | ||
| 81 | return ret; | ||
| 82 | } | ||
| 83 | |||
| 84 | entry->reserved = true; | ||
| 85 | if (unlikely(atomic_read(&bo->cpu_writers) > 0)) { | ||
| 86 | ttm_eu_backoff_reservation(list); | ||
| 87 | ret = ttm_bo_wait_cpu(bo, false); | ||
| 88 | if (ret) | ||
| 89 | return ret; | ||
| 90 | goto retry; | ||
| 91 | } | ||
| 92 | } | ||
| 93 | return 0; | ||
| 94 | } | ||
| 95 | EXPORT_SYMBOL(ttm_eu_reserve_buffers); | ||
| 96 | |||
| 97 | void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj) | ||
| 98 | { | ||
| 99 | struct ttm_validate_buffer *entry; | ||
| 100 | |||
| 101 | list_for_each_entry(entry, list, head) { | ||
| 102 | struct ttm_buffer_object *bo = entry->bo; | ||
| 103 | struct ttm_bo_driver *driver = bo->bdev->driver; | ||
| 104 | void *old_sync_obj; | ||
| 105 | |||
| 106 | spin_lock(&bo->lock); | ||
| 107 | old_sync_obj = bo->sync_obj; | ||
| 108 | bo->sync_obj = driver->sync_obj_ref(sync_obj); | ||
| 109 | bo->sync_obj_arg = entry->new_sync_obj_arg; | ||
| 110 | spin_unlock(&bo->lock); | ||
| 111 | ttm_bo_unreserve(bo); | ||
| 112 | entry->reserved = false; | ||
| 113 | if (old_sync_obj) | ||
| 114 | driver->sync_obj_unref(&old_sync_obj); | ||
| 115 | } | ||
| 116 | } | ||
| 117 | EXPORT_SYMBOL(ttm_eu_fence_buffer_objects); | ||
diff --git a/drivers/gpu/drm/ttm/ttm_lock.c b/drivers/gpu/drm/ttm/ttm_lock.c new file mode 100644 index 000000000000..f619ebcaa4ec --- /dev/null +++ b/drivers/gpu/drm/ttm/ttm_lock.c | |||
| @@ -0,0 +1,311 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2007-2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | /* | ||
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | ||
| 29 | */ | ||
| 30 | |||
| 31 | #include "ttm/ttm_lock.h" | ||
| 32 | #include "ttm/ttm_module.h" | ||
| 33 | #include <asm/atomic.h> | ||
| 34 | #include <linux/errno.h> | ||
| 35 | #include <linux/wait.h> | ||
| 36 | #include <linux/sched.h> | ||
| 37 | #include <linux/module.h> | ||
| 38 | |||
| 39 | #define TTM_WRITE_LOCK_PENDING (1 << 0) | ||
| 40 | #define TTM_VT_LOCK_PENDING (1 << 1) | ||
| 41 | #define TTM_SUSPEND_LOCK_PENDING (1 << 2) | ||
| 42 | #define TTM_VT_LOCK (1 << 3) | ||
| 43 | #define TTM_SUSPEND_LOCK (1 << 4) | ||
| 44 | |||
| 45 | void ttm_lock_init(struct ttm_lock *lock) | ||
| 46 | { | ||
| 47 | spin_lock_init(&lock->lock); | ||
| 48 | init_waitqueue_head(&lock->queue); | ||
| 49 | lock->rw = 0; | ||
| 50 | lock->flags = 0; | ||
| 51 | lock->kill_takers = false; | ||
| 52 | lock->signal = SIGKILL; | ||
| 53 | } | ||
| 54 | EXPORT_SYMBOL(ttm_lock_init); | ||
| 55 | |||
| 56 | void ttm_read_unlock(struct ttm_lock *lock) | ||
| 57 | { | ||
| 58 | spin_lock(&lock->lock); | ||
| 59 | if (--lock->rw == 0) | ||
| 60 | wake_up_all(&lock->queue); | ||
| 61 | spin_unlock(&lock->lock); | ||
| 62 | } | ||
| 63 | EXPORT_SYMBOL(ttm_read_unlock); | ||
| 64 | |||
| 65 | static bool __ttm_read_lock(struct ttm_lock *lock) | ||
| 66 | { | ||
| 67 | bool locked = false; | ||
| 68 | |||
| 69 | spin_lock(&lock->lock); | ||
| 70 | if (unlikely(lock->kill_takers)) { | ||
| 71 | send_sig(lock->signal, current, 0); | ||
| 72 | spin_unlock(&lock->lock); | ||
| 73 | return false; | ||
| 74 | } | ||
| 75 | if (lock->rw >= 0 && lock->flags == 0) { | ||
| 76 | ++lock->rw; | ||
| 77 | locked = true; | ||
| 78 | } | ||
| 79 | spin_unlock(&lock->lock); | ||
| 80 | return locked; | ||
| 81 | } | ||
| 82 | |||
| 83 | int ttm_read_lock(struct ttm_lock *lock, bool interruptible) | ||
| 84 | { | ||
| 85 | int ret = 0; | ||
| 86 | |||
| 87 | if (interruptible) | ||
| 88 | ret = wait_event_interruptible(lock->queue, | ||
| 89 | __ttm_read_lock(lock)); | ||
| 90 | else | ||
| 91 | wait_event(lock->queue, __ttm_read_lock(lock)); | ||
| 92 | return ret; | ||
| 93 | } | ||
| 94 | EXPORT_SYMBOL(ttm_read_lock); | ||
| 95 | |||
| 96 | static bool __ttm_read_trylock(struct ttm_lock *lock, bool *locked) | ||
| 97 | { | ||
| 98 | bool block = true; | ||
| 99 | |||
| 100 | *locked = false; | ||
| 101 | |||
| 102 | spin_lock(&lock->lock); | ||
| 103 | if (unlikely(lock->kill_takers)) { | ||
| 104 | send_sig(lock->signal, current, 0); | ||
| 105 | spin_unlock(&lock->lock); | ||
| 106 | return false; | ||
| 107 | } | ||
| 108 | if (lock->rw >= 0 && lock->flags == 0) { | ||
| 109 | ++lock->rw; | ||
| 110 | block = false; | ||
| 111 | *locked = true; | ||
| 112 | } else if (lock->flags == 0) { | ||
| 113 | block = false; | ||
| 114 | } | ||
| 115 | spin_unlock(&lock->lock); | ||
| 116 | |||
| 117 | return !block; | ||
| 118 | } | ||
| 119 | |||
| 120 | int ttm_read_trylock(struct ttm_lock *lock, bool interruptible) | ||
| 121 | { | ||
| 122 | int ret = 0; | ||
| 123 | bool locked; | ||
| 124 | |||
| 125 | if (interruptible) | ||
| 126 | ret = wait_event_interruptible | ||
| 127 | (lock->queue, __ttm_read_trylock(lock, &locked)); | ||
| 128 | else | ||
| 129 | wait_event(lock->queue, __ttm_read_trylock(lock, &locked)); | ||
| 130 | |||
| 131 | if (unlikely(ret != 0)) { | ||
| 132 | BUG_ON(locked); | ||
| 133 | return ret; | ||
| 134 | } | ||
| 135 | |||
| 136 | return (locked) ? 0 : -EBUSY; | ||
| 137 | } | ||
| 138 | |||
| 139 | void ttm_write_unlock(struct ttm_lock *lock) | ||
| 140 | { | ||
| 141 | spin_lock(&lock->lock); | ||
| 142 | lock->rw = 0; | ||
| 143 | wake_up_all(&lock->queue); | ||
| 144 | spin_unlock(&lock->lock); | ||
| 145 | } | ||
| 146 | EXPORT_SYMBOL(ttm_write_unlock); | ||
| 147 | |||
| 148 | static bool __ttm_write_lock(struct ttm_lock *lock) | ||
| 149 | { | ||
| 150 | bool locked = false; | ||
| 151 | |||
| 152 | spin_lock(&lock->lock); | ||
| 153 | if (unlikely(lock->kill_takers)) { | ||
| 154 | send_sig(lock->signal, current, 0); | ||
| 155 | spin_unlock(&lock->lock); | ||
| 156 | return false; | ||
| 157 | } | ||
| 158 | if (lock->rw == 0 && ((lock->flags & ~TTM_WRITE_LOCK_PENDING) == 0)) { | ||
| 159 | lock->rw = -1; | ||
| 160 | lock->flags &= ~TTM_WRITE_LOCK_PENDING; | ||
| 161 | locked = true; | ||
| 162 | } else { | ||
| 163 | lock->flags |= TTM_WRITE_LOCK_PENDING; | ||
| 164 | } | ||
| 165 | spin_unlock(&lock->lock); | ||
| 166 | return locked; | ||
| 167 | } | ||
| 168 | |||
| 169 | int ttm_write_lock(struct ttm_lock *lock, bool interruptible) | ||
| 170 | { | ||
| 171 | int ret = 0; | ||
| 172 | |||
| 173 | if (interruptible) { | ||
| 174 | ret = wait_event_interruptible(lock->queue, | ||
| 175 | __ttm_write_lock(lock)); | ||
| 176 | if (unlikely(ret != 0)) { | ||
| 177 | spin_lock(&lock->lock); | ||
| 178 | lock->flags &= ~TTM_WRITE_LOCK_PENDING; | ||
| 179 | wake_up_all(&lock->queue); | ||
| 180 | spin_unlock(&lock->lock); | ||
| 181 | } | ||
| 182 | } else | ||
| 183 | wait_event(lock->queue, __ttm_read_lock(lock)); | ||
| 184 | |||
| 185 | return ret; | ||
| 186 | } | ||
| 187 | EXPORT_SYMBOL(ttm_write_lock); | ||
| 188 | |||
| 189 | void ttm_write_lock_downgrade(struct ttm_lock *lock) | ||
| 190 | { | ||
| 191 | spin_lock(&lock->lock); | ||
| 192 | lock->rw = 1; | ||
| 193 | wake_up_all(&lock->queue); | ||
| 194 | spin_unlock(&lock->lock); | ||
| 195 | } | ||
| 196 | |||
| 197 | static int __ttm_vt_unlock(struct ttm_lock *lock) | ||
| 198 | { | ||
| 199 | int ret = 0; | ||
| 200 | |||
| 201 | spin_lock(&lock->lock); | ||
| 202 | if (unlikely(!(lock->flags & TTM_VT_LOCK))) | ||
| 203 | ret = -EINVAL; | ||
| 204 | lock->flags &= ~TTM_VT_LOCK; | ||
| 205 | wake_up_all(&lock->queue); | ||
| 206 | spin_unlock(&lock->lock); | ||
| 207 | printk(KERN_INFO TTM_PFX "vt unlock.\n"); | ||
| 208 | |||
| 209 | return ret; | ||
| 210 | } | ||
| 211 | |||
| 212 | static void ttm_vt_lock_remove(struct ttm_base_object **p_base) | ||
| 213 | { | ||
| 214 | struct ttm_base_object *base = *p_base; | ||
| 215 | struct ttm_lock *lock = container_of(base, struct ttm_lock, base); | ||
| 216 | int ret; | ||
| 217 | |||
| 218 | *p_base = NULL; | ||
| 219 | ret = __ttm_vt_unlock(lock); | ||
| 220 | BUG_ON(ret != 0); | ||
| 221 | } | ||
| 222 | |||
| 223 | static bool __ttm_vt_lock(struct ttm_lock *lock) | ||
| 224 | { | ||
| 225 | bool locked = false; | ||
| 226 | |||
| 227 | spin_lock(&lock->lock); | ||
| 228 | if (lock->rw == 0) { | ||
| 229 | lock->flags &= ~TTM_VT_LOCK_PENDING; | ||
| 230 | lock->flags |= TTM_VT_LOCK; | ||
| 231 | locked = true; | ||
| 232 | } else { | ||
| 233 | lock->flags |= TTM_VT_LOCK_PENDING; | ||
| 234 | } | ||
| 235 | spin_unlock(&lock->lock); | ||
| 236 | return locked; | ||
| 237 | } | ||
| 238 | |||
| 239 | int ttm_vt_lock(struct ttm_lock *lock, | ||
| 240 | bool interruptible, | ||
| 241 | struct ttm_object_file *tfile) | ||
| 242 | { | ||
| 243 | int ret = 0; | ||
| 244 | |||
| 245 | if (interruptible) { | ||
| 246 | ret = wait_event_interruptible(lock->queue, | ||
| 247 | __ttm_vt_lock(lock)); | ||
| 248 | if (unlikely(ret != 0)) { | ||
| 249 | spin_lock(&lock->lock); | ||
| 250 | lock->flags &= ~TTM_VT_LOCK_PENDING; | ||
| 251 | wake_up_all(&lock->queue); | ||
| 252 | spin_unlock(&lock->lock); | ||
| 253 | return ret; | ||
| 254 | } | ||
| 255 | } else | ||
| 256 | wait_event(lock->queue, __ttm_vt_lock(lock)); | ||
| 257 | |||
| 258 | /* | ||
| 259 | * Add a base-object, the destructor of which will | ||
| 260 | * make sure the lock is released if the client dies | ||
| 261 | * while holding it. | ||
| 262 | */ | ||
| 263 | |||
| 264 | ret = ttm_base_object_init(tfile, &lock->base, false, | ||
| 265 | ttm_lock_type, &ttm_vt_lock_remove, NULL); | ||
| 266 | if (ret) | ||
| 267 | (void)__ttm_vt_unlock(lock); | ||
| 268 | else { | ||
| 269 | lock->vt_holder = tfile; | ||
| 270 | printk(KERN_INFO TTM_PFX "vt lock.\n"); | ||
| 271 | } | ||
| 272 | |||
| 273 | return ret; | ||
| 274 | } | ||
| 275 | EXPORT_SYMBOL(ttm_vt_lock); | ||
| 276 | |||
| 277 | int ttm_vt_unlock(struct ttm_lock *lock) | ||
| 278 | { | ||
| 279 | return ttm_ref_object_base_unref(lock->vt_holder, | ||
| 280 | lock->base.hash.key, TTM_REF_USAGE); | ||
| 281 | } | ||
| 282 | EXPORT_SYMBOL(ttm_vt_unlock); | ||
| 283 | |||
| 284 | void ttm_suspend_unlock(struct ttm_lock *lock) | ||
| 285 | { | ||
| 286 | spin_lock(&lock->lock); | ||
| 287 | lock->flags &= ~TTM_SUSPEND_LOCK; | ||
| 288 | wake_up_all(&lock->queue); | ||
| 289 | spin_unlock(&lock->lock); | ||
| 290 | } | ||
| 291 | |||
| 292 | static bool __ttm_suspend_lock(struct ttm_lock *lock) | ||
| 293 | { | ||
| 294 | bool locked = false; | ||
| 295 | |||
| 296 | spin_lock(&lock->lock); | ||
| 297 | if (lock->rw == 0) { | ||
| 298 | lock->flags &= ~TTM_SUSPEND_LOCK_PENDING; | ||
| 299 | lock->flags |= TTM_SUSPEND_LOCK; | ||
| 300 | locked = true; | ||
| 301 | } else { | ||
| 302 | lock->flags |= TTM_SUSPEND_LOCK_PENDING; | ||
| 303 | } | ||
| 304 | spin_unlock(&lock->lock); | ||
| 305 | return locked; | ||
| 306 | } | ||
| 307 | |||
| 308 | void ttm_suspend_lock(struct ttm_lock *lock) | ||
| 309 | { | ||
| 310 | wait_event(lock->queue, __ttm_suspend_lock(lock)); | ||
| 311 | } | ||
diff --git a/drivers/gpu/drm/ttm/ttm_memory.c b/drivers/gpu/drm/ttm/ttm_memory.c index 072c281a6bb5..f5245c02b8fd 100644 --- a/drivers/gpu/drm/ttm/ttm_memory.c +++ b/drivers/gpu/drm/ttm/ttm_memory.c | |||
| @@ -274,16 +274,17 @@ static int ttm_mem_init_kernel_zone(struct ttm_mem_global *glob, | |||
| 274 | static int ttm_mem_init_highmem_zone(struct ttm_mem_global *glob, | 274 | static int ttm_mem_init_highmem_zone(struct ttm_mem_global *glob, |
| 275 | const struct sysinfo *si) | 275 | const struct sysinfo *si) |
| 276 | { | 276 | { |
| 277 | struct ttm_mem_zone *zone = kzalloc(sizeof(*zone), GFP_KERNEL); | 277 | struct ttm_mem_zone *zone; |
| 278 | uint64_t mem; | 278 | uint64_t mem; |
| 279 | int ret; | 279 | int ret; |
| 280 | 280 | ||
| 281 | if (unlikely(!zone)) | ||
| 282 | return -ENOMEM; | ||
| 283 | |||
| 284 | if (si->totalhigh == 0) | 281 | if (si->totalhigh == 0) |
| 285 | return 0; | 282 | return 0; |
| 286 | 283 | ||
| 284 | zone = kzalloc(sizeof(*zone), GFP_KERNEL); | ||
| 285 | if (unlikely(!zone)) | ||
| 286 | return -ENOMEM; | ||
| 287 | |||
| 287 | mem = si->totalram; | 288 | mem = si->totalram; |
| 288 | mem *= si->mem_unit; | 289 | mem *= si->mem_unit; |
| 289 | 290 | ||
| @@ -322,8 +323,10 @@ static int ttm_mem_init_dma32_zone(struct ttm_mem_global *glob, | |||
| 322 | * No special dma32 zone needed. | 323 | * No special dma32 zone needed. |
| 323 | */ | 324 | */ |
| 324 | 325 | ||
| 325 | if (mem <= ((uint64_t) 1ULL << 32)) | 326 | if (mem <= ((uint64_t) 1ULL << 32)) { |
| 327 | kfree(zone); | ||
| 326 | return 0; | 328 | return 0; |
| 329 | } | ||
| 327 | 330 | ||
| 328 | /* | 331 | /* |
| 329 | * Limit max dma32 memory to 4GB for now | 332 | * Limit max dma32 memory to 4GB for now |
| @@ -460,6 +463,7 @@ void ttm_mem_global_free(struct ttm_mem_global *glob, | |||
| 460 | { | 463 | { |
| 461 | return ttm_mem_global_free_zone(glob, NULL, amount); | 464 | return ttm_mem_global_free_zone(glob, NULL, amount); |
| 462 | } | 465 | } |
| 466 | EXPORT_SYMBOL(ttm_mem_global_free); | ||
| 463 | 467 | ||
| 464 | static int ttm_mem_global_reserve(struct ttm_mem_global *glob, | 468 | static int ttm_mem_global_reserve(struct ttm_mem_global *glob, |
| 465 | struct ttm_mem_zone *single_zone, | 469 | struct ttm_mem_zone *single_zone, |
| @@ -533,6 +537,7 @@ int ttm_mem_global_alloc(struct ttm_mem_global *glob, uint64_t memory, | |||
| 533 | return ttm_mem_global_alloc_zone(glob, NULL, memory, no_wait, | 537 | return ttm_mem_global_alloc_zone(glob, NULL, memory, no_wait, |
| 534 | interruptible); | 538 | interruptible); |
| 535 | } | 539 | } |
| 540 | EXPORT_SYMBOL(ttm_mem_global_alloc); | ||
| 536 | 541 | ||
| 537 | int ttm_mem_global_alloc_page(struct ttm_mem_global *glob, | 542 | int ttm_mem_global_alloc_page(struct ttm_mem_global *glob, |
| 538 | struct page *page, | 543 | struct page *page, |
| @@ -588,3 +593,4 @@ size_t ttm_round_pot(size_t size) | |||
| 588 | } | 593 | } |
| 589 | return 0; | 594 | return 0; |
| 590 | } | 595 | } |
| 596 | EXPORT_SYMBOL(ttm_round_pot); | ||
diff --git a/drivers/gpu/drm/ttm/ttm_object.c b/drivers/gpu/drm/ttm/ttm_object.c new file mode 100644 index 000000000000..1099abac824b --- /dev/null +++ b/drivers/gpu/drm/ttm/ttm_object.c | |||
| @@ -0,0 +1,452 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | /* | ||
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | ||
| 29 | */ | ||
| 30 | /** @file ttm_ref_object.c | ||
| 31 | * | ||
| 32 | * Base- and reference object implementation for the various | ||
| 33 | * ttm objects. Implements reference counting, minimal security checks | ||
| 34 | * and release on file close. | ||
| 35 | */ | ||
| 36 | |||
| 37 | /** | ||
| 38 | * struct ttm_object_file | ||
| 39 | * | ||
| 40 | * @tdev: Pointer to the ttm_object_device. | ||
| 41 | * | ||
| 42 | * @lock: Lock that protects the ref_list list and the | ||
| 43 | * ref_hash hash tables. | ||
| 44 | * | ||
| 45 | * @ref_list: List of ttm_ref_objects to be destroyed at | ||
| 46 | * file release. | ||
| 47 | * | ||
| 48 | * @ref_hash: Hash tables of ref objects, one per ttm_ref_type, | ||
| 49 | * for fast lookup of ref objects given a base object. | ||
| 50 | */ | ||
| 51 | |||
| 52 | #include "ttm/ttm_object.h" | ||
| 53 | #include "ttm/ttm_module.h" | ||
| 54 | #include <linux/list.h> | ||
| 55 | #include <linux/spinlock.h> | ||
| 56 | #include <linux/slab.h> | ||
| 57 | #include <linux/module.h> | ||
| 58 | #include <asm/atomic.h> | ||
| 59 | |||
| 60 | struct ttm_object_file { | ||
| 61 | struct ttm_object_device *tdev; | ||
| 62 | rwlock_t lock; | ||
| 63 | struct list_head ref_list; | ||
| 64 | struct drm_open_hash ref_hash[TTM_REF_NUM]; | ||
| 65 | struct kref refcount; | ||
| 66 | }; | ||
| 67 | |||
| 68 | /** | ||
| 69 | * struct ttm_object_device | ||
| 70 | * | ||
| 71 | * @object_lock: lock that protects the object_hash hash table. | ||
| 72 | * | ||
| 73 | * @object_hash: hash table for fast lookup of object global names. | ||
| 74 | * | ||
| 75 | * @object_count: Per device object count. | ||
| 76 | * | ||
| 77 | * This is the per-device data structure needed for ttm object management. | ||
| 78 | */ | ||
| 79 | |||
| 80 | struct ttm_object_device { | ||
| 81 | rwlock_t object_lock; | ||
| 82 | struct drm_open_hash object_hash; | ||
| 83 | atomic_t object_count; | ||
| 84 | struct ttm_mem_global *mem_glob; | ||
| 85 | }; | ||
| 86 | |||
| 87 | /** | ||
| 88 | * struct ttm_ref_object | ||
| 89 | * | ||
| 90 | * @hash: Hash entry for the per-file object reference hash. | ||
| 91 | * | ||
| 92 | * @head: List entry for the per-file list of ref-objects. | ||
| 93 | * | ||
| 94 | * @kref: Ref count. | ||
| 95 | * | ||
| 96 | * @obj: Base object this ref object is referencing. | ||
| 97 | * | ||
| 98 | * @ref_type: Type of ref object. | ||
| 99 | * | ||
| 100 | * This is similar to an idr object, but it also has a hash table entry | ||
| 101 | * that allows lookup with a pointer to the referenced object as a key. In | ||
| 102 | * that way, one can easily detect whether a base object is referenced by | ||
| 103 | * a particular ttm_object_file. It also carries a ref count to avoid creating | ||
| 104 | * multiple ref objects if a ttm_object_file references the same base | ||
| 105 | * object more than once. | ||
| 106 | */ | ||
| 107 | |||
| 108 | struct ttm_ref_object { | ||
| 109 | struct drm_hash_item hash; | ||
| 110 | struct list_head head; | ||
| 111 | struct kref kref; | ||
| 112 | struct ttm_base_object *obj; | ||
| 113 | enum ttm_ref_type ref_type; | ||
| 114 | struct ttm_object_file *tfile; | ||
| 115 | }; | ||
| 116 | |||
| 117 | static inline struct ttm_object_file * | ||
| 118 | ttm_object_file_ref(struct ttm_object_file *tfile) | ||
| 119 | { | ||
| 120 | kref_get(&tfile->refcount); | ||
| 121 | return tfile; | ||
| 122 | } | ||
| 123 | |||
| 124 | static void ttm_object_file_destroy(struct kref *kref) | ||
| 125 | { | ||
| 126 | struct ttm_object_file *tfile = | ||
| 127 | container_of(kref, struct ttm_object_file, refcount); | ||
| 128 | |||
| 129 | kfree(tfile); | ||
| 130 | } | ||
| 131 | |||
| 132 | |||
| 133 | static inline void ttm_object_file_unref(struct ttm_object_file **p_tfile) | ||
| 134 | { | ||
| 135 | struct ttm_object_file *tfile = *p_tfile; | ||
| 136 | |||
| 137 | *p_tfile = NULL; | ||
| 138 | kref_put(&tfile->refcount, ttm_object_file_destroy); | ||
| 139 | } | ||
| 140 | |||
| 141 | |||
| 142 | int ttm_base_object_init(struct ttm_object_file *tfile, | ||
| 143 | struct ttm_base_object *base, | ||
| 144 | bool shareable, | ||
| 145 | enum ttm_object_type object_type, | ||
| 146 | void (*refcount_release) (struct ttm_base_object **), | ||
| 147 | void (*ref_obj_release) (struct ttm_base_object *, | ||
| 148 | enum ttm_ref_type ref_type)) | ||
| 149 | { | ||
| 150 | struct ttm_object_device *tdev = tfile->tdev; | ||
| 151 | int ret; | ||
| 152 | |||
| 153 | base->shareable = shareable; | ||
| 154 | base->tfile = ttm_object_file_ref(tfile); | ||
| 155 | base->refcount_release = refcount_release; | ||
| 156 | base->ref_obj_release = ref_obj_release; | ||
| 157 | base->object_type = object_type; | ||
| 158 | write_lock(&tdev->object_lock); | ||
| 159 | kref_init(&base->refcount); | ||
| 160 | ret = drm_ht_just_insert_please(&tdev->object_hash, | ||
| 161 | &base->hash, | ||
| 162 | (unsigned long)base, 31, 0, 0); | ||
| 163 | write_unlock(&tdev->object_lock); | ||
| 164 | if (unlikely(ret != 0)) | ||
| 165 | goto out_err0; | ||
| 166 | |||
| 167 | ret = ttm_ref_object_add(tfile, base, TTM_REF_USAGE, NULL); | ||
| 168 | if (unlikely(ret != 0)) | ||
| 169 | goto out_err1; | ||
| 170 | |||
| 171 | ttm_base_object_unref(&base); | ||
| 172 | |||
| 173 | return 0; | ||
| 174 | out_err1: | ||
| 175 | (void)drm_ht_remove_item(&tdev->object_hash, &base->hash); | ||
| 176 | out_err0: | ||
| 177 | return ret; | ||
| 178 | } | ||
| 179 | EXPORT_SYMBOL(ttm_base_object_init); | ||
| 180 | |||
| 181 | static void ttm_release_base(struct kref *kref) | ||
| 182 | { | ||
| 183 | struct ttm_base_object *base = | ||
| 184 | container_of(kref, struct ttm_base_object, refcount); | ||
| 185 | struct ttm_object_device *tdev = base->tfile->tdev; | ||
| 186 | |||
| 187 | (void)drm_ht_remove_item(&tdev->object_hash, &base->hash); | ||
| 188 | write_unlock(&tdev->object_lock); | ||
| 189 | if (base->refcount_release) { | ||
| 190 | ttm_object_file_unref(&base->tfile); | ||
| 191 | base->refcount_release(&base); | ||
| 192 | } | ||
| 193 | write_lock(&tdev->object_lock); | ||
| 194 | } | ||
| 195 | |||
| 196 | void ttm_base_object_unref(struct ttm_base_object **p_base) | ||
| 197 | { | ||
| 198 | struct ttm_base_object *base = *p_base; | ||
| 199 | struct ttm_object_device *tdev = base->tfile->tdev; | ||
| 200 | |||
| 201 | *p_base = NULL; | ||
| 202 | |||
| 203 | /* | ||
| 204 | * Need to take the lock here to avoid racing with | ||
| 205 | * users trying to look up the object. | ||
| 206 | */ | ||
| 207 | |||
| 208 | write_lock(&tdev->object_lock); | ||
| 209 | (void)kref_put(&base->refcount, &ttm_release_base); | ||
| 210 | write_unlock(&tdev->object_lock); | ||
| 211 | } | ||
| 212 | EXPORT_SYMBOL(ttm_base_object_unref); | ||
| 213 | |||
| 214 | struct ttm_base_object *ttm_base_object_lookup(struct ttm_object_file *tfile, | ||
| 215 | uint32_t key) | ||
| 216 | { | ||
| 217 | struct ttm_object_device *tdev = tfile->tdev; | ||
| 218 | struct ttm_base_object *base; | ||
| 219 | struct drm_hash_item *hash; | ||
| 220 | int ret; | ||
| 221 | |||
| 222 | read_lock(&tdev->object_lock); | ||
| 223 | ret = drm_ht_find_item(&tdev->object_hash, key, &hash); | ||
| 224 | |||
| 225 | if (likely(ret == 0)) { | ||
| 226 | base = drm_hash_entry(hash, struct ttm_base_object, hash); | ||
| 227 | kref_get(&base->refcount); | ||
| 228 | } | ||
| 229 | read_unlock(&tdev->object_lock); | ||
| 230 | |||
| 231 | if (unlikely(ret != 0)) | ||
| 232 | return NULL; | ||
| 233 | |||
| 234 | if (tfile != base->tfile && !base->shareable) { | ||
| 235 | printk(KERN_ERR TTM_PFX | ||
| 236 | "Attempted access of non-shareable object.\n"); | ||
| 237 | ttm_base_object_unref(&base); | ||
| 238 | return NULL; | ||
| 239 | } | ||
| 240 | |||
| 241 | return base; | ||
| 242 | } | ||
| 243 | EXPORT_SYMBOL(ttm_base_object_lookup); | ||
| 244 | |||
| 245 | int ttm_ref_object_add(struct ttm_object_file *tfile, | ||
| 246 | struct ttm_base_object *base, | ||
| 247 | enum ttm_ref_type ref_type, bool *existed) | ||
| 248 | { | ||
| 249 | struct drm_open_hash *ht = &tfile->ref_hash[ref_type]; | ||
| 250 | struct ttm_ref_object *ref; | ||
| 251 | struct drm_hash_item *hash; | ||
| 252 | struct ttm_mem_global *mem_glob = tfile->tdev->mem_glob; | ||
| 253 | int ret = -EINVAL; | ||
| 254 | |||
| 255 | if (existed != NULL) | ||
| 256 | *existed = true; | ||
| 257 | |||
| 258 | while (ret == -EINVAL) { | ||
| 259 | read_lock(&tfile->lock); | ||
| 260 | ret = drm_ht_find_item(ht, base->hash.key, &hash); | ||
| 261 | |||
| 262 | if (ret == 0) { | ||
| 263 | ref = drm_hash_entry(hash, struct ttm_ref_object, hash); | ||
| 264 | kref_get(&ref->kref); | ||
| 265 | read_unlock(&tfile->lock); | ||
| 266 | break; | ||
| 267 | } | ||
| 268 | |||
| 269 | read_unlock(&tfile->lock); | ||
| 270 | ret = ttm_mem_global_alloc(mem_glob, sizeof(*ref), | ||
| 271 | false, false); | ||
| 272 | if (unlikely(ret != 0)) | ||
| 273 | return ret; | ||
| 274 | ref = kmalloc(sizeof(*ref), GFP_KERNEL); | ||
| 275 | if (unlikely(ref == NULL)) { | ||
| 276 | ttm_mem_global_free(mem_glob, sizeof(*ref)); | ||
| 277 | return -ENOMEM; | ||
| 278 | } | ||
| 279 | |||
| 280 | ref->hash.key = base->hash.key; | ||
| 281 | ref->obj = base; | ||
| 282 | ref->tfile = tfile; | ||
| 283 | ref->ref_type = ref_type; | ||
| 284 | kref_init(&ref->kref); | ||
| 285 | |||
| 286 | write_lock(&tfile->lock); | ||
| 287 | ret = drm_ht_insert_item(ht, &ref->hash); | ||
| 288 | |||
| 289 | if (likely(ret == 0)) { | ||
| 290 | list_add_tail(&ref->head, &tfile->ref_list); | ||
| 291 | kref_get(&base->refcount); | ||
| 292 | write_unlock(&tfile->lock); | ||
| 293 | if (existed != NULL) | ||
| 294 | *existed = false; | ||
| 295 | break; | ||
| 296 | } | ||
| 297 | |||
| 298 | write_unlock(&tfile->lock); | ||
| 299 | BUG_ON(ret != -EINVAL); | ||
| 300 | |||
| 301 | ttm_mem_global_free(mem_glob, sizeof(*ref)); | ||
| 302 | kfree(ref); | ||
| 303 | } | ||
| 304 | |||
| 305 | return ret; | ||
| 306 | } | ||
| 307 | EXPORT_SYMBOL(ttm_ref_object_add); | ||
| 308 | |||
| 309 | static void ttm_ref_object_release(struct kref *kref) | ||
| 310 | { | ||
| 311 | struct ttm_ref_object *ref = | ||
| 312 | container_of(kref, struct ttm_ref_object, kref); | ||
| 313 | struct ttm_base_object *base = ref->obj; | ||
| 314 | struct ttm_object_file *tfile = ref->tfile; | ||
| 315 | struct drm_open_hash *ht; | ||
| 316 | struct ttm_mem_global *mem_glob = tfile->tdev->mem_glob; | ||
| 317 | |||
| 318 | ht = &tfile->ref_hash[ref->ref_type]; | ||
| 319 | (void)drm_ht_remove_item(ht, &ref->hash); | ||
| 320 | list_del(&ref->head); | ||
| 321 | write_unlock(&tfile->lock); | ||
| 322 | |||
| 323 | if (ref->ref_type != TTM_REF_USAGE && base->ref_obj_release) | ||
| 324 | base->ref_obj_release(base, ref->ref_type); | ||
| 325 | |||
| 326 | ttm_base_object_unref(&ref->obj); | ||
| 327 | ttm_mem_global_free(mem_glob, sizeof(*ref)); | ||
| 328 | kfree(ref); | ||
| 329 | write_lock(&tfile->lock); | ||
| 330 | } | ||
| 331 | |||
| 332 | int ttm_ref_object_base_unref(struct ttm_object_file *tfile, | ||
| 333 | unsigned long key, enum ttm_ref_type ref_type) | ||
| 334 | { | ||
| 335 | struct drm_open_hash *ht = &tfile->ref_hash[ref_type]; | ||
| 336 | struct ttm_ref_object *ref; | ||
| 337 | struct drm_hash_item *hash; | ||
| 338 | int ret; | ||
| 339 | |||
| 340 | write_lock(&tfile->lock); | ||
| 341 | ret = drm_ht_find_item(ht, key, &hash); | ||
| 342 | if (unlikely(ret != 0)) { | ||
| 343 | write_unlock(&tfile->lock); | ||
| 344 | return -EINVAL; | ||
| 345 | } | ||
| 346 | ref = drm_hash_entry(hash, struct ttm_ref_object, hash); | ||
| 347 | kref_put(&ref->kref, ttm_ref_object_release); | ||
| 348 | write_unlock(&tfile->lock); | ||
| 349 | return 0; | ||
| 350 | } | ||
| 351 | EXPORT_SYMBOL(ttm_ref_object_base_unref); | ||
| 352 | |||
| 353 | void ttm_object_file_release(struct ttm_object_file **p_tfile) | ||
| 354 | { | ||
| 355 | struct ttm_ref_object *ref; | ||
| 356 | struct list_head *list; | ||
| 357 | unsigned int i; | ||
| 358 | struct ttm_object_file *tfile = *p_tfile; | ||
| 359 | |||
| 360 | *p_tfile = NULL; | ||
| 361 | write_lock(&tfile->lock); | ||
| 362 | |||
| 363 | /* | ||
| 364 | * Since we release the lock within the loop, we have to | ||
| 365 | * restart it from the beginning each time. | ||
| 366 | */ | ||
| 367 | |||
| 368 | while (!list_empty(&tfile->ref_list)) { | ||
| 369 | list = tfile->ref_list.next; | ||
| 370 | ref = list_entry(list, struct ttm_ref_object, head); | ||
| 371 | ttm_ref_object_release(&ref->kref); | ||
| 372 | } | ||
| 373 | |||
| 374 | for (i = 0; i < TTM_REF_NUM; ++i) | ||
| 375 | drm_ht_remove(&tfile->ref_hash[i]); | ||
| 376 | |||
| 377 | write_unlock(&tfile->lock); | ||
| 378 | ttm_object_file_unref(&tfile); | ||
| 379 | } | ||
| 380 | EXPORT_SYMBOL(ttm_object_file_release); | ||
| 381 | |||
| 382 | struct ttm_object_file *ttm_object_file_init(struct ttm_object_device *tdev, | ||
| 383 | unsigned int hash_order) | ||
| 384 | { | ||
| 385 | struct ttm_object_file *tfile = kmalloc(sizeof(*tfile), GFP_KERNEL); | ||
| 386 | unsigned int i; | ||
| 387 | unsigned int j = 0; | ||
| 388 | int ret; | ||
| 389 | |||
| 390 | if (unlikely(tfile == NULL)) | ||
| 391 | return NULL; | ||
| 392 | |||
| 393 | rwlock_init(&tfile->lock); | ||
| 394 | tfile->tdev = tdev; | ||
| 395 | kref_init(&tfile->refcount); | ||
| 396 | INIT_LIST_HEAD(&tfile->ref_list); | ||
| 397 | |||
| 398 | for (i = 0; i < TTM_REF_NUM; ++i) { | ||
| 399 | ret = drm_ht_create(&tfile->ref_hash[i], hash_order); | ||
| 400 | if (ret) { | ||
| 401 | j = i; | ||
| 402 | goto out_err; | ||
| 403 | } | ||
| 404 | } | ||
| 405 | |||
| 406 | return tfile; | ||
| 407 | out_err: | ||
| 408 | for (i = 0; i < j; ++i) | ||
| 409 | drm_ht_remove(&tfile->ref_hash[i]); | ||
| 410 | |||
| 411 | kfree(tfile); | ||
| 412 | |||
| 413 | return NULL; | ||
| 414 | } | ||
| 415 | EXPORT_SYMBOL(ttm_object_file_init); | ||
| 416 | |||
| 417 | struct ttm_object_device *ttm_object_device_init(struct ttm_mem_global | ||
| 418 | *mem_glob, | ||
| 419 | unsigned int hash_order) | ||
| 420 | { | ||
| 421 | struct ttm_object_device *tdev = kmalloc(sizeof(*tdev), GFP_KERNEL); | ||
| 422 | int ret; | ||
| 423 | |||
| 424 | if (unlikely(tdev == NULL)) | ||
| 425 | return NULL; | ||
| 426 | |||
| 427 | tdev->mem_glob = mem_glob; | ||
| 428 | rwlock_init(&tdev->object_lock); | ||
| 429 | atomic_set(&tdev->object_count, 0); | ||
| 430 | ret = drm_ht_create(&tdev->object_hash, hash_order); | ||
| 431 | |||
| 432 | if (likely(ret == 0)) | ||
| 433 | return tdev; | ||
| 434 | |||
| 435 | kfree(tdev); | ||
| 436 | return NULL; | ||
| 437 | } | ||
| 438 | EXPORT_SYMBOL(ttm_object_device_init); | ||
| 439 | |||
| 440 | void ttm_object_device_release(struct ttm_object_device **p_tdev) | ||
| 441 | { | ||
| 442 | struct ttm_object_device *tdev = *p_tdev; | ||
| 443 | |||
| 444 | *p_tdev = NULL; | ||
| 445 | |||
| 446 | write_lock(&tdev->object_lock); | ||
| 447 | drm_ht_remove(&tdev->object_hash); | ||
| 448 | write_unlock(&tdev->object_lock); | ||
| 449 | |||
| 450 | kfree(tdev); | ||
| 451 | } | ||
| 452 | EXPORT_SYMBOL(ttm_object_device_release); | ||
diff --git a/drivers/gpu/drm/ttm/ttm_tt.c b/drivers/gpu/drm/ttm/ttm_tt.c index 7bcb89f39ce8..9c2b1cc5dba5 100644 --- a/drivers/gpu/drm/ttm/ttm_tt.c +++ b/drivers/gpu/drm/ttm/ttm_tt.c | |||
| @@ -192,6 +192,7 @@ int ttm_tt_populate(struct ttm_tt *ttm) | |||
| 192 | ttm->state = tt_unbound; | 192 | ttm->state = tt_unbound; |
| 193 | return 0; | 193 | return 0; |
| 194 | } | 194 | } |
| 195 | EXPORT_SYMBOL(ttm_tt_populate); | ||
| 195 | 196 | ||
| 196 | #ifdef CONFIG_X86 | 197 | #ifdef CONFIG_X86 |
| 197 | static inline int ttm_tt_set_page_caching(struct page *p, | 198 | static inline int ttm_tt_set_page_caching(struct page *p, |
diff --git a/include/drm/drm.h b/include/drm/drm.h index 7cb50bdde46d..e3f46e0cb7dc 100644 --- a/include/drm/drm.h +++ b/include/drm/drm.h | |||
| @@ -36,17 +36,27 @@ | |||
| 36 | #ifndef _DRM_H_ | 36 | #ifndef _DRM_H_ |
| 37 | #define _DRM_H_ | 37 | #define _DRM_H_ |
| 38 | 38 | ||
| 39 | #if defined(__linux__) | ||
| 40 | |||
| 39 | #include <linux/types.h> | 41 | #include <linux/types.h> |
| 40 | #include <asm/ioctl.h> /* For _IO* macros */ | 42 | #include <asm/ioctl.h> |
| 41 | #define DRM_IOCTL_NR(n) _IOC_NR(n) | 43 | typedef unsigned int drm_handle_t; |
| 42 | #define DRM_IOC_VOID _IOC_NONE | ||
| 43 | #define DRM_IOC_READ _IOC_READ | ||
| 44 | #define DRM_IOC_WRITE _IOC_WRITE | ||
| 45 | #define DRM_IOC_READWRITE _IOC_READ|_IOC_WRITE | ||
| 46 | #define DRM_IOC(dir, group, nr, size) _IOC(dir, group, nr, size) | ||
| 47 | 44 | ||
| 48 | #define DRM_MAJOR 226 | 45 | #else /* One of the BSDs */ |
| 49 | #define DRM_MAX_MINOR 15 | 46 | |
| 47 | #include <sys/ioccom.h> | ||
| 48 | #include <sys/types.h> | ||
| 49 | typedef int8_t __s8; | ||
| 50 | typedef uint8_t __u8; | ||
| 51 | typedef int16_t __s16; | ||
| 52 | typedef uint16_t __u16; | ||
| 53 | typedef int32_t __s32; | ||
| 54 | typedef uint32_t __u32; | ||
| 55 | typedef int64_t __s64; | ||
| 56 | typedef uint64_t __u64; | ||
| 57 | typedef unsigned long drm_handle_t; | ||
| 58 | |||
| 59 | #endif | ||
| 50 | 60 | ||
| 51 | #define DRM_NAME "drm" /**< Name in kernel, /dev, and /proc */ | 61 | #define DRM_NAME "drm" /**< Name in kernel, /dev, and /proc */ |
| 52 | #define DRM_MIN_ORDER 5 /**< At least 2^5 bytes = 32 bytes */ | 62 | #define DRM_MIN_ORDER 5 /**< At least 2^5 bytes = 32 bytes */ |
| @@ -59,7 +69,6 @@ | |||
| 59 | #define _DRM_LOCK_IS_CONT(lock) ((lock) & _DRM_LOCK_CONT) | 69 | #define _DRM_LOCK_IS_CONT(lock) ((lock) & _DRM_LOCK_CONT) |
| 60 | #define _DRM_LOCKING_CONTEXT(lock) ((lock) & ~(_DRM_LOCK_HELD|_DRM_LOCK_CONT)) | 70 | #define _DRM_LOCKING_CONTEXT(lock) ((lock) & ~(_DRM_LOCK_HELD|_DRM_LOCK_CONT)) |
| 61 | 71 | ||
| 62 | typedef unsigned int drm_handle_t; | ||
| 63 | typedef unsigned int drm_context_t; | 72 | typedef unsigned int drm_context_t; |
| 64 | typedef unsigned int drm_drawable_t; | 73 | typedef unsigned int drm_drawable_t; |
| 65 | typedef unsigned int drm_magic_t; | 74 | typedef unsigned int drm_magic_t; |
| @@ -454,6 +463,7 @@ struct drm_irq_busid { | |||
| 454 | enum drm_vblank_seq_type { | 463 | enum drm_vblank_seq_type { |
| 455 | _DRM_VBLANK_ABSOLUTE = 0x0, /**< Wait for specific vblank sequence number */ | 464 | _DRM_VBLANK_ABSOLUTE = 0x0, /**< Wait for specific vblank sequence number */ |
| 456 | _DRM_VBLANK_RELATIVE = 0x1, /**< Wait for given number of vblanks */ | 465 | _DRM_VBLANK_RELATIVE = 0x1, /**< Wait for given number of vblanks */ |
| 466 | _DRM_VBLANK_EVENT = 0x4000000, /**< Send event instead of blocking */ | ||
| 457 | _DRM_VBLANK_FLIP = 0x8000000, /**< Scheduled buffer swap should flip */ | 467 | _DRM_VBLANK_FLIP = 0x8000000, /**< Scheduled buffer swap should flip */ |
| 458 | _DRM_VBLANK_NEXTONMISS = 0x10000000, /**< If missed, wait for next vblank */ | 468 | _DRM_VBLANK_NEXTONMISS = 0x10000000, /**< If missed, wait for next vblank */ |
| 459 | _DRM_VBLANK_SECONDARY = 0x20000000, /**< Secondary display controller */ | 469 | _DRM_VBLANK_SECONDARY = 0x20000000, /**< Secondary display controller */ |
| @@ -461,8 +471,8 @@ enum drm_vblank_seq_type { | |||
| 461 | }; | 471 | }; |
| 462 | 472 | ||
| 463 | #define _DRM_VBLANK_TYPES_MASK (_DRM_VBLANK_ABSOLUTE | _DRM_VBLANK_RELATIVE) | 473 | #define _DRM_VBLANK_TYPES_MASK (_DRM_VBLANK_ABSOLUTE | _DRM_VBLANK_RELATIVE) |
| 464 | #define _DRM_VBLANK_FLAGS_MASK (_DRM_VBLANK_SIGNAL | _DRM_VBLANK_SECONDARY | \ | 474 | #define _DRM_VBLANK_FLAGS_MASK (_DRM_VBLANK_EVENT | _DRM_VBLANK_SIGNAL | \ |
| 465 | _DRM_VBLANK_NEXTONMISS) | 475 | _DRM_VBLANK_SECONDARY | _DRM_VBLANK_NEXTONMISS) |
| 466 | 476 | ||
| 467 | struct drm_wait_vblank_request { | 477 | struct drm_wait_vblank_request { |
| 468 | enum drm_vblank_seq_type type; | 478 | enum drm_vblank_seq_type type; |
| @@ -686,6 +696,8 @@ struct drm_gem_open { | |||
| 686 | #define DRM_IOCTL_MODE_GETFB DRM_IOWR(0xAD, struct drm_mode_fb_cmd) | 696 | #define DRM_IOCTL_MODE_GETFB DRM_IOWR(0xAD, struct drm_mode_fb_cmd) |
| 687 | #define DRM_IOCTL_MODE_ADDFB DRM_IOWR(0xAE, struct drm_mode_fb_cmd) | 697 | #define DRM_IOCTL_MODE_ADDFB DRM_IOWR(0xAE, struct drm_mode_fb_cmd) |
| 688 | #define DRM_IOCTL_MODE_RMFB DRM_IOWR(0xAF, unsigned int) | 698 | #define DRM_IOCTL_MODE_RMFB DRM_IOWR(0xAF, unsigned int) |
| 699 | #define DRM_IOCTL_MODE_PAGE_FLIP DRM_IOWR(0xB0, struct drm_mode_crtc_page_flip) | ||
| 700 | #define DRM_IOCTL_MODE_DIRTYFB DRM_IOWR(0xB1, struct drm_mode_fb_dirty_cmd) | ||
| 689 | 701 | ||
| 690 | /** | 702 | /** |
| 691 | * Device specific ioctls should only be in their respective headers | 703 | * Device specific ioctls should only be in their respective headers |
| @@ -698,6 +710,35 @@ struct drm_gem_open { | |||
| 698 | #define DRM_COMMAND_BASE 0x40 | 710 | #define DRM_COMMAND_BASE 0x40 |
| 699 | #define DRM_COMMAND_END 0xA0 | 711 | #define DRM_COMMAND_END 0xA0 |
| 700 | 712 | ||
| 713 | /** | ||
| 714 | * Header for events written back to userspace on the drm fd. The | ||
| 715 | * type defines the type of event, the length specifies the total | ||
| 716 | * length of the event (including the header), and user_data is | ||
| 717 | * typically a 64 bit value passed with the ioctl that triggered the | ||
| 718 | * event. A read on the drm fd will always only return complete | ||
| 719 | * events, that is, if for example the read buffer is 100 bytes, and | ||
| 720 | * there are two 64 byte events pending, only one will be returned. | ||
| 721 | * | ||
| 722 | * Event types 0 - 0x7fffffff are generic drm events, 0x80000000 and | ||
| 723 | * up are chipset specific. | ||
| 724 | */ | ||
| 725 | struct drm_event { | ||
| 726 | __u32 type; | ||
| 727 | __u32 length; | ||
| 728 | }; | ||
| 729 | |||
| 730 | #define DRM_EVENT_VBLANK 0x01 | ||
| 731 | #define DRM_EVENT_FLIP_COMPLETE 0x02 | ||
| 732 | |||
| 733 | struct drm_event_vblank { | ||
| 734 | struct drm_event base; | ||
| 735 | __u64 user_data; | ||
| 736 | __u32 tv_sec; | ||
| 737 | __u32 tv_usec; | ||
| 738 | __u32 sequence; | ||
| 739 | __u32 reserved; | ||
| 740 | }; | ||
| 741 | |||
| 701 | /* typedef area */ | 742 | /* typedef area */ |
| 702 | #ifndef __KERNEL__ | 743 | #ifndef __KERNEL__ |
| 703 | typedef struct drm_clip_rect drm_clip_rect_t; | 744 | typedef struct drm_clip_rect drm_clip_rect_t; |
diff --git a/include/drm/drmP.h b/include/drm/drmP.h index c8e64bbadbcf..19ef8ebdc662 100644 --- a/include/drm/drmP.h +++ b/include/drm/drmP.h | |||
| @@ -245,16 +245,6 @@ extern void drm_ut_debug_printk(unsigned int request_level, | |||
| 245 | 245 | ||
| 246 | #endif | 246 | #endif |
| 247 | 247 | ||
| 248 | #define DRM_PROC_LIMIT (PAGE_SIZE-80) | ||
| 249 | |||
| 250 | #define DRM_PROC_PRINT(fmt, arg...) \ | ||
| 251 | len += sprintf(&buf[len], fmt , ##arg); \ | ||
| 252 | if (len > DRM_PROC_LIMIT) { *eof = 1; return len - offset; } | ||
| 253 | |||
| 254 | #define DRM_PROC_PRINT_RET(ret, fmt, arg...) \ | ||
| 255 | len += sprintf(&buf[len], fmt , ##arg); \ | ||
| 256 | if (len > DRM_PROC_LIMIT) { ret; *eof = 1; return len - offset; } | ||
| 257 | |||
| 258 | /*@}*/ | 248 | /*@}*/ |
| 259 | 249 | ||
| 260 | /***********************************************************************/ | 250 | /***********************************************************************/ |
| @@ -265,19 +255,8 @@ extern void drm_ut_debug_printk(unsigned int request_level, | |||
| 265 | 255 | ||
| 266 | #define DRM_LEFTCOUNT(x) (((x)->rp + (x)->count - (x)->wp) % ((x)->count + 1)) | 256 | #define DRM_LEFTCOUNT(x) (((x)->rp + (x)->count - (x)->wp) % ((x)->count + 1)) |
| 267 | #define DRM_BUFCOUNT(x) ((x)->count - DRM_LEFTCOUNT(x)) | 257 | #define DRM_BUFCOUNT(x) ((x)->count - DRM_LEFTCOUNT(x)) |
| 268 | #define DRM_WAITCOUNT(dev,idx) DRM_BUFCOUNT(&dev->queuelist[idx]->waitlist) | ||
| 269 | 258 | ||
| 270 | #define DRM_IF_VERSION(maj, min) (maj << 16 | min) | 259 | #define DRM_IF_VERSION(maj, min) (maj << 16 | min) |
| 271 | /** | ||
| 272 | * Get the private SAREA mapping. | ||
| 273 | * | ||
| 274 | * \param _dev DRM device. | ||
| 275 | * \param _ctx context number. | ||
| 276 | * \param _map output mapping. | ||
| 277 | */ | ||
| 278 | #define DRM_GET_PRIV_SAREA(_dev, _ctx, _map) do { \ | ||
| 279 | (_map) = (_dev)->context_sareas[_ctx]; \ | ||
| 280 | } while(0) | ||
| 281 | 260 | ||
| 282 | /** | 261 | /** |
| 283 | * Test that the hardware lock is held by the caller, returning otherwise. | 262 | * Test that the hardware lock is held by the caller, returning otherwise. |
| @@ -297,18 +276,6 @@ do { \ | |||
| 297 | } while (0) | 276 | } while (0) |
| 298 | 277 | ||
| 299 | /** | 278 | /** |
| 300 | * Copy and IOCTL return string to user space | ||
| 301 | */ | ||
| 302 | #define DRM_COPY( name, value ) \ | ||
| 303 | len = strlen( value ); \ | ||
| 304 | if ( len > name##_len ) len = name##_len; \ | ||
| 305 | name##_len = strlen( value ); \ | ||
| 306 | if ( len && name ) { \ | ||
| 307 | if ( copy_to_user( name, value, len ) ) \ | ||
| 308 | return -EFAULT; \ | ||
| 309 | } | ||
| 310 | |||
| 311 | /** | ||
| 312 | * Ioctl function type. | 279 | * Ioctl function type. |
| 313 | * | 280 | * |
| 314 | * \param inode device inode. | 281 | * \param inode device inode. |
| @@ -322,6 +289,9 @@ typedef int drm_ioctl_t(struct drm_device *dev, void *data, | |||
| 322 | typedef int drm_ioctl_compat_t(struct file *filp, unsigned int cmd, | 289 | typedef int drm_ioctl_compat_t(struct file *filp, unsigned int cmd, |
| 323 | unsigned long arg); | 290 | unsigned long arg); |
| 324 | 291 | ||
| 292 | #define DRM_IOCTL_NR(n) _IOC_NR(n) | ||
| 293 | #define DRM_MAJOR 226 | ||
| 294 | |||
| 325 | #define DRM_AUTH 0x1 | 295 | #define DRM_AUTH 0x1 |
| 326 | #define DRM_MASTER 0x2 | 296 | #define DRM_MASTER 0x2 |
| 327 | #define DRM_ROOT_ONLY 0x4 | 297 | #define DRM_ROOT_ONLY 0x4 |
| @@ -426,6 +396,14 @@ struct drm_buf_entry { | |||
| 426 | struct drm_freelist freelist; | 396 | struct drm_freelist freelist; |
| 427 | }; | 397 | }; |
| 428 | 398 | ||
| 399 | /* Event queued up for userspace to read */ | ||
| 400 | struct drm_pending_event { | ||
| 401 | struct drm_event *event; | ||
| 402 | struct list_head link; | ||
| 403 | struct drm_file *file_priv; | ||
| 404 | void (*destroy)(struct drm_pending_event *event); | ||
| 405 | }; | ||
| 406 | |||
| 429 | /** File private data */ | 407 | /** File private data */ |
| 430 | struct drm_file { | 408 | struct drm_file { |
| 431 | int authenticated; | 409 | int authenticated; |
| @@ -449,6 +427,10 @@ struct drm_file { | |||
| 449 | struct drm_master *master; /* master this node is currently associated with | 427 | struct drm_master *master; /* master this node is currently associated with |
| 450 | N.B. not always minor->master */ | 428 | N.B. not always minor->master */ |
| 451 | struct list_head fbs; | 429 | struct list_head fbs; |
| 430 | |||
| 431 | wait_queue_head_t event_wait; | ||
| 432 | struct list_head event_list; | ||
| 433 | int event_space; | ||
| 452 | }; | 434 | }; |
| 453 | 435 | ||
| 454 | /** Wait queue */ | 436 | /** Wait queue */ |
| @@ -795,6 +777,15 @@ struct drm_driver { | |||
| 795 | /* Master routines */ | 777 | /* Master routines */ |
| 796 | int (*master_create)(struct drm_device *dev, struct drm_master *master); | 778 | int (*master_create)(struct drm_device *dev, struct drm_master *master); |
| 797 | void (*master_destroy)(struct drm_device *dev, struct drm_master *master); | 779 | void (*master_destroy)(struct drm_device *dev, struct drm_master *master); |
| 780 | /** | ||
| 781 | * master_set is called whenever the minor master is set. | ||
| 782 | * master_drop is called whenever the minor master is dropped. | ||
| 783 | */ | ||
| 784 | |||
| 785 | int (*master_set)(struct drm_device *dev, struct drm_file *file_priv, | ||
| 786 | bool from_open); | ||
| 787 | void (*master_drop)(struct drm_device *dev, struct drm_file *file_priv, | ||
| 788 | bool from_release); | ||
| 798 | 789 | ||
| 799 | int (*proc_init)(struct drm_minor *minor); | 790 | int (*proc_init)(struct drm_minor *minor); |
| 800 | void (*proc_cleanup)(struct drm_minor *minor); | 791 | void (*proc_cleanup)(struct drm_minor *minor); |
| @@ -900,6 +891,12 @@ struct drm_minor { | |||
| 900 | struct drm_mode_group mode_group; | 891 | struct drm_mode_group mode_group; |
| 901 | }; | 892 | }; |
| 902 | 893 | ||
| 894 | struct drm_pending_vblank_event { | ||
| 895 | struct drm_pending_event base; | ||
| 896 | int pipe; | ||
| 897 | struct drm_event_vblank event; | ||
| 898 | }; | ||
| 899 | |||
| 903 | /** | 900 | /** |
| 904 | * DRM device structure. This structure represent a complete card that | 901 | * DRM device structure. This structure represent a complete card that |
| 905 | * may contain multiple heads. | 902 | * may contain multiple heads. |
| @@ -999,6 +996,12 @@ struct drm_device { | |||
| 999 | 996 | ||
| 1000 | u32 max_vblank_count; /**< size of vblank counter register */ | 997 | u32 max_vblank_count; /**< size of vblank counter register */ |
| 1001 | 998 | ||
| 999 | /** | ||
| 1000 | * List of events | ||
| 1001 | */ | ||
| 1002 | struct list_head vblank_event_list; | ||
| 1003 | spinlock_t event_lock; | ||
| 1004 | |||
| 1002 | /*@} */ | 1005 | /*@} */ |
| 1003 | cycles_t ctx_start; | 1006 | cycles_t ctx_start; |
| 1004 | cycles_t lck_start; | 1007 | cycles_t lck_start; |
| @@ -1135,6 +1138,8 @@ extern int drm_lastclose(struct drm_device *dev); | |||
| 1135 | extern int drm_open(struct inode *inode, struct file *filp); | 1138 | extern int drm_open(struct inode *inode, struct file *filp); |
| 1136 | extern int drm_stub_open(struct inode *inode, struct file *filp); | 1139 | extern int drm_stub_open(struct inode *inode, struct file *filp); |
| 1137 | extern int drm_fasync(int fd, struct file *filp, int on); | 1140 | extern int drm_fasync(int fd, struct file *filp, int on); |
| 1141 | extern ssize_t drm_read(struct file *filp, char __user *buffer, | ||
| 1142 | size_t count, loff_t *offset); | ||
| 1138 | extern int drm_release(struct inode *inode, struct file *filp); | 1143 | extern int drm_release(struct inode *inode, struct file *filp); |
| 1139 | 1144 | ||
| 1140 | /* Mapping support (drm_vm.h) */ | 1145 | /* Mapping support (drm_vm.h) */ |
| @@ -1295,6 +1300,7 @@ extern u32 drm_vblank_count(struct drm_device *dev, int crtc); | |||
| 1295 | extern void drm_handle_vblank(struct drm_device *dev, int crtc); | 1300 | extern void drm_handle_vblank(struct drm_device *dev, int crtc); |
| 1296 | extern int drm_vblank_get(struct drm_device *dev, int crtc); | 1301 | extern int drm_vblank_get(struct drm_device *dev, int crtc); |
| 1297 | extern void drm_vblank_put(struct drm_device *dev, int crtc); | 1302 | extern void drm_vblank_put(struct drm_device *dev, int crtc); |
| 1303 | extern void drm_vblank_off(struct drm_device *dev, int crtc); | ||
| 1298 | extern void drm_vblank_cleanup(struct drm_device *dev); | 1304 | extern void drm_vblank_cleanup(struct drm_device *dev); |
| 1299 | /* Modesetting support */ | 1305 | /* Modesetting support */ |
| 1300 | extern void drm_vblank_pre_modeset(struct drm_device *dev, int crtc); | 1306 | extern void drm_vblank_pre_modeset(struct drm_device *dev, int crtc); |
| @@ -1519,14 +1525,27 @@ static __inline__ void drm_core_dropmap(struct drm_local_map *map) | |||
| 1519 | 1525 | ||
| 1520 | static __inline__ void *drm_calloc_large(size_t nmemb, size_t size) | 1526 | static __inline__ void *drm_calloc_large(size_t nmemb, size_t size) |
| 1521 | { | 1527 | { |
| 1528 | if (size != 0 && nmemb > ULONG_MAX / size) | ||
| 1529 | return NULL; | ||
| 1530 | |||
| 1522 | if (size * nmemb <= PAGE_SIZE) | 1531 | if (size * nmemb <= PAGE_SIZE) |
| 1523 | return kcalloc(nmemb, size, GFP_KERNEL); | 1532 | return kcalloc(nmemb, size, GFP_KERNEL); |
| 1524 | 1533 | ||
| 1534 | return __vmalloc(size * nmemb, | ||
| 1535 | GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO, PAGE_KERNEL); | ||
| 1536 | } | ||
| 1537 | |||
| 1538 | /* Modeled after cairo's malloc_ab, it's like calloc but without the zeroing. */ | ||
| 1539 | static __inline__ void *drm_malloc_ab(size_t nmemb, size_t size) | ||
| 1540 | { | ||
| 1525 | if (size != 0 && nmemb > ULONG_MAX / size) | 1541 | if (size != 0 && nmemb > ULONG_MAX / size) |
| 1526 | return NULL; | 1542 | return NULL; |
| 1527 | 1543 | ||
| 1544 | if (size * nmemb <= PAGE_SIZE) | ||
| 1545 | return kmalloc(nmemb * size, GFP_KERNEL); | ||
| 1546 | |||
| 1528 | return __vmalloc(size * nmemb, | 1547 | return __vmalloc(size * nmemb, |
| 1529 | GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO, PAGE_KERNEL); | 1548 | GFP_KERNEL | __GFP_HIGHMEM, PAGE_KERNEL); |
| 1530 | } | 1549 | } |
| 1531 | 1550 | ||
| 1532 | static __inline void drm_free_large(void *ptr) | 1551 | static __inline void drm_free_large(void *ptr) |
diff --git a/include/drm/drm_crtc.h b/include/drm/drm_crtc.h index b69347b8904f..fdf43abc36db 100644 --- a/include/drm/drm_crtc.h +++ b/include/drm/drm_crtc.h | |||
| @@ -123,7 +123,7 @@ struct drm_display_mode { | |||
| 123 | int type; | 123 | int type; |
| 124 | 124 | ||
| 125 | /* Proposed mode values */ | 125 | /* Proposed mode values */ |
| 126 | int clock; | 126 | int clock; /* in kHz */ |
| 127 | int hdisplay; | 127 | int hdisplay; |
| 128 | int hsync_start; | 128 | int hsync_start; |
| 129 | int hsync_end; | 129 | int hsync_end; |
| @@ -164,8 +164,8 @@ struct drm_display_mode { | |||
| 164 | int *private; | 164 | int *private; |
| 165 | int private_flags; | 165 | int private_flags; |
| 166 | 166 | ||
| 167 | int vrefresh; | 167 | int vrefresh; /* in Hz */ |
| 168 | float hsync; | 168 | int hsync; /* in kHz */ |
| 169 | }; | 169 | }; |
| 170 | 170 | ||
| 171 | enum drm_connector_status { | 171 | enum drm_connector_status { |
| @@ -242,6 +242,21 @@ struct drm_framebuffer_funcs { | |||
| 242 | int (*create_handle)(struct drm_framebuffer *fb, | 242 | int (*create_handle)(struct drm_framebuffer *fb, |
| 243 | struct drm_file *file_priv, | 243 | struct drm_file *file_priv, |
| 244 | unsigned int *handle); | 244 | unsigned int *handle); |
| 245 | /** | ||
| 246 | * Optinal callback for the dirty fb ioctl. | ||
| 247 | * | ||
| 248 | * Userspace can notify the driver via this callback | ||
| 249 | * that a area of the framebuffer has changed and should | ||
| 250 | * be flushed to the display hardware. | ||
| 251 | * | ||
| 252 | * See documentation in drm_mode.h for the struct | ||
| 253 | * drm_mode_fb_dirty_cmd for more information as all | ||
| 254 | * the semantics and arguments have a one to one mapping | ||
| 255 | * on this function. | ||
| 256 | */ | ||
| 257 | int (*dirty)(struct drm_framebuffer *framebuffer, unsigned flags, | ||
| 258 | unsigned color, struct drm_clip_rect *clips, | ||
| 259 | unsigned num_clips); | ||
| 245 | }; | 260 | }; |
| 246 | 261 | ||
| 247 | struct drm_framebuffer { | 262 | struct drm_framebuffer { |
| @@ -256,7 +271,7 @@ struct drm_framebuffer { | |||
| 256 | unsigned int depth; | 271 | unsigned int depth; |
| 257 | int bits_per_pixel; | 272 | int bits_per_pixel; |
| 258 | int flags; | 273 | int flags; |
| 259 | void *fbdev; | 274 | struct fb_info *fbdev; |
| 260 | u32 pseudo_palette[17]; | 275 | u32 pseudo_palette[17]; |
| 261 | struct list_head filp_head; | 276 | struct list_head filp_head; |
| 262 | /* if you are using the helper */ | 277 | /* if you are using the helper */ |
| @@ -290,6 +305,7 @@ struct drm_property { | |||
| 290 | struct drm_crtc; | 305 | struct drm_crtc; |
| 291 | struct drm_connector; | 306 | struct drm_connector; |
| 292 | struct drm_encoder; | 307 | struct drm_encoder; |
| 308 | struct drm_pending_vblank_event; | ||
| 293 | 309 | ||
| 294 | /** | 310 | /** |
| 295 | * drm_crtc_funcs - control CRTCs for a given device | 311 | * drm_crtc_funcs - control CRTCs for a given device |
| @@ -333,6 +349,19 @@ struct drm_crtc_funcs { | |||
| 333 | void (*destroy)(struct drm_crtc *crtc); | 349 | void (*destroy)(struct drm_crtc *crtc); |
| 334 | 350 | ||
| 335 | int (*set_config)(struct drm_mode_set *set); | 351 | int (*set_config)(struct drm_mode_set *set); |
| 352 | |||
| 353 | /* | ||
| 354 | * Flip to the given framebuffer. This implements the page | ||
| 355 | * flip ioctl descibed in drm_mode.h, specifically, the | ||
| 356 | * implementation must return immediately and block all | ||
| 357 | * rendering to the current fb until the flip has completed. | ||
| 358 | * If userspace set the event flag in the ioctl, the event | ||
| 359 | * argument will point to an event to send back when the flip | ||
| 360 | * completes, otherwise it will be NULL. | ||
| 361 | */ | ||
| 362 | int (*page_flip)(struct drm_crtc *crtc, | ||
| 363 | struct drm_framebuffer *fb, | ||
| 364 | struct drm_pending_vblank_event *event); | ||
| 336 | }; | 365 | }; |
| 337 | 366 | ||
| 338 | /** | 367 | /** |
| @@ -596,6 +625,7 @@ struct drm_mode_config { | |||
| 596 | /* Optional properties */ | 625 | /* Optional properties */ |
| 597 | struct drm_property *scaling_mode_property; | 626 | struct drm_property *scaling_mode_property; |
| 598 | struct drm_property *dithering_mode_property; | 627 | struct drm_property *dithering_mode_property; |
| 628 | struct drm_property *dirty_info_property; | ||
| 599 | }; | 629 | }; |
| 600 | 630 | ||
| 601 | #define obj_to_crtc(x) container_of(x, struct drm_crtc, base) | 631 | #define obj_to_crtc(x) container_of(x, struct drm_crtc, base) |
| @@ -667,6 +697,7 @@ extern void drm_mode_validate_size(struct drm_device *dev, | |||
| 667 | extern void drm_mode_prune_invalid(struct drm_device *dev, | 697 | extern void drm_mode_prune_invalid(struct drm_device *dev, |
| 668 | struct list_head *mode_list, bool verbose); | 698 | struct list_head *mode_list, bool verbose); |
| 669 | extern void drm_mode_sort(struct list_head *mode_list); | 699 | extern void drm_mode_sort(struct list_head *mode_list); |
| 700 | extern int drm_mode_hsync(struct drm_display_mode *mode); | ||
| 670 | extern int drm_mode_vrefresh(struct drm_display_mode *mode); | 701 | extern int drm_mode_vrefresh(struct drm_display_mode *mode); |
| 671 | extern void drm_mode_set_crtcinfo(struct drm_display_mode *p, | 702 | extern void drm_mode_set_crtcinfo(struct drm_display_mode *p, |
| 672 | int adjust_flags); | 703 | int adjust_flags); |
| @@ -703,6 +734,7 @@ extern int drm_mode_create_tv_properties(struct drm_device *dev, int num_formats | |||
| 703 | char *formats[]); | 734 | char *formats[]); |
| 704 | extern int drm_mode_create_scaling_mode_property(struct drm_device *dev); | 735 | extern int drm_mode_create_scaling_mode_property(struct drm_device *dev); |
| 705 | extern int drm_mode_create_dithering_property(struct drm_device *dev); | 736 | extern int drm_mode_create_dithering_property(struct drm_device *dev); |
| 737 | extern int drm_mode_create_dirty_info_property(struct drm_device *dev); | ||
| 706 | extern char *drm_get_encoder_name(struct drm_encoder *encoder); | 738 | extern char *drm_get_encoder_name(struct drm_encoder *encoder); |
| 707 | 739 | ||
| 708 | extern int drm_mode_connector_attach_encoder(struct drm_connector *connector, | 740 | extern int drm_mode_connector_attach_encoder(struct drm_connector *connector, |
| @@ -711,7 +743,8 @@ extern void drm_mode_connector_detach_encoder(struct drm_connector *connector, | |||
| 711 | struct drm_encoder *encoder); | 743 | struct drm_encoder *encoder); |
| 712 | extern bool drm_mode_crtc_set_gamma_size(struct drm_crtc *crtc, | 744 | extern bool drm_mode_crtc_set_gamma_size(struct drm_crtc *crtc, |
| 713 | int gamma_size); | 745 | int gamma_size); |
| 714 | extern void *drm_mode_object_find(struct drm_device *dev, uint32_t id, uint32_t type); | 746 | extern struct drm_mode_object *drm_mode_object_find(struct drm_device *dev, |
| 747 | uint32_t id, uint32_t type); | ||
| 715 | /* IOCTLs */ | 748 | /* IOCTLs */ |
| 716 | extern int drm_mode_getresources(struct drm_device *dev, | 749 | extern int drm_mode_getresources(struct drm_device *dev, |
| 717 | void *data, struct drm_file *file_priv); | 750 | void *data, struct drm_file *file_priv); |
| @@ -730,6 +763,8 @@ extern int drm_mode_rmfb(struct drm_device *dev, | |||
| 730 | void *data, struct drm_file *file_priv); | 763 | void *data, struct drm_file *file_priv); |
| 731 | extern int drm_mode_getfb(struct drm_device *dev, | 764 | extern int drm_mode_getfb(struct drm_device *dev, |
| 732 | void *data, struct drm_file *file_priv); | 765 | void *data, struct drm_file *file_priv); |
| 766 | extern int drm_mode_dirtyfb_ioctl(struct drm_device *dev, | ||
| 767 | void *data, struct drm_file *file_priv); | ||
| 733 | extern int drm_mode_addmode_ioctl(struct drm_device *dev, | 768 | extern int drm_mode_addmode_ioctl(struct drm_device *dev, |
| 734 | void *data, struct drm_file *file_priv); | 769 | void *data, struct drm_file *file_priv); |
| 735 | extern int drm_mode_rmmode_ioctl(struct drm_device *dev, | 770 | extern int drm_mode_rmmode_ioctl(struct drm_device *dev, |
| @@ -756,6 +791,8 @@ extern int drm_mode_gamma_get_ioctl(struct drm_device *dev, | |||
| 756 | extern int drm_mode_gamma_set_ioctl(struct drm_device *dev, | 791 | extern int drm_mode_gamma_set_ioctl(struct drm_device *dev, |
| 757 | void *data, struct drm_file *file_priv); | 792 | void *data, struct drm_file *file_priv); |
| 758 | extern bool drm_detect_hdmi_monitor(struct edid *edid); | 793 | extern bool drm_detect_hdmi_monitor(struct edid *edid); |
| 794 | extern int drm_mode_page_flip_ioctl(struct drm_device *dev, | ||
| 795 | void *data, struct drm_file *file_priv); | ||
| 759 | extern struct drm_display_mode *drm_cvt_mode(struct drm_device *dev, | 796 | extern struct drm_display_mode *drm_cvt_mode(struct drm_device *dev, |
| 760 | int hdisplay, int vdisplay, int vrefresh, | 797 | int hdisplay, int vdisplay, int vrefresh, |
| 761 | bool reduced, bool interlaced, bool margins); | 798 | bool reduced, bool interlaced, bool margins); |
diff --git a/drivers/gpu/drm/i915/intel_dp.h b/include/drm/drm_dp_helper.h index 2b38054d3b6d..a49e791db0b0 100644 --- a/drivers/gpu/drm/i915/intel_dp.h +++ b/include/drm/drm_dp_helper.h | |||
| @@ -20,8 +20,8 @@ | |||
| 20 | * OF THIS SOFTWARE. | 20 | * OF THIS SOFTWARE. |
| 21 | */ | 21 | */ |
| 22 | 22 | ||
| 23 | #ifndef _INTEL_DP_H_ | 23 | #ifndef _DRM_DP_HELPER_H_ |
| 24 | #define _INTEL_DP_H_ | 24 | #define _DRM_DP_HELPER_H_ |
| 25 | 25 | ||
| 26 | /* From the VESA DisplayPort spec */ | 26 | /* From the VESA DisplayPort spec */ |
| 27 | 27 | ||
| @@ -43,16 +43,41 @@ | |||
| 43 | #define AUX_I2C_REPLY_MASK (0x3 << 6) | 43 | #define AUX_I2C_REPLY_MASK (0x3 << 6) |
| 44 | 44 | ||
| 45 | /* AUX CH addresses */ | 45 | /* AUX CH addresses */ |
| 46 | #define DP_LINK_BW_SET 0x100 | 46 | /* DPCD */ |
| 47 | #define DP_DPCD_REV 0x000 | ||
| 48 | |||
| 49 | #define DP_MAX_LINK_RATE 0x001 | ||
| 50 | |||
| 51 | #define DP_MAX_LANE_COUNT 0x002 | ||
| 52 | # define DP_MAX_LANE_COUNT_MASK 0x1f | ||
| 53 | # define DP_ENHANCED_FRAME_CAP (1 << 7) | ||
| 54 | |||
| 55 | #define DP_MAX_DOWNSPREAD 0x003 | ||
| 56 | # define DP_NO_AUX_HANDSHAKE_LINK_TRAINING (1 << 6) | ||
| 57 | |||
| 58 | #define DP_NORP 0x004 | ||
| 59 | |||
| 60 | #define DP_DOWNSTREAMPORT_PRESENT 0x005 | ||
| 61 | # define DP_DWN_STRM_PORT_PRESENT (1 << 0) | ||
| 62 | # define DP_DWN_STRM_PORT_TYPE_MASK 0x06 | ||
| 63 | /* 00b = DisplayPort */ | ||
| 64 | /* 01b = Analog */ | ||
| 65 | /* 10b = TMDS or HDMI */ | ||
| 66 | /* 11b = Other */ | ||
| 67 | # define DP_FORMAT_CONVERSION (1 << 3) | ||
| 68 | |||
| 69 | #define DP_MAIN_LINK_CHANNEL_CODING 0x006 | ||
| 70 | |||
| 71 | /* link configuration */ | ||
| 72 | #define DP_LINK_BW_SET 0x100 | ||
| 47 | # define DP_LINK_BW_1_62 0x06 | 73 | # define DP_LINK_BW_1_62 0x06 |
| 48 | # define DP_LINK_BW_2_7 0x0a | 74 | # define DP_LINK_BW_2_7 0x0a |
| 49 | 75 | ||
| 50 | #define DP_LANE_COUNT_SET 0x101 | 76 | #define DP_LANE_COUNT_SET 0x101 |
| 51 | # define DP_LANE_COUNT_MASK 0x0f | 77 | # define DP_LANE_COUNT_MASK 0x0f |
| 52 | # define DP_LANE_COUNT_ENHANCED_FRAME_EN (1 << 7) | 78 | # define DP_LANE_COUNT_ENHANCED_FRAME_EN (1 << 7) |
| 53 | 79 | ||
| 54 | #define DP_TRAINING_PATTERN_SET 0x102 | 80 | #define DP_TRAINING_PATTERN_SET 0x102 |
| 55 | |||
| 56 | # define DP_TRAINING_PATTERN_DISABLE 0 | 81 | # define DP_TRAINING_PATTERN_DISABLE 0 |
| 57 | # define DP_TRAINING_PATTERN_1 1 | 82 | # define DP_TRAINING_PATTERN_1 1 |
| 58 | # define DP_TRAINING_PATTERN_2 2 | 83 | # define DP_TRAINING_PATTERN_2 2 |
| @@ -102,11 +127,14 @@ | |||
| 102 | 127 | ||
| 103 | #define DP_LANE0_1_STATUS 0x202 | 128 | #define DP_LANE0_1_STATUS 0x202 |
| 104 | #define DP_LANE2_3_STATUS 0x203 | 129 | #define DP_LANE2_3_STATUS 0x203 |
| 105 | |||
| 106 | # define DP_LANE_CR_DONE (1 << 0) | 130 | # define DP_LANE_CR_DONE (1 << 0) |
| 107 | # define DP_LANE_CHANNEL_EQ_DONE (1 << 1) | 131 | # define DP_LANE_CHANNEL_EQ_DONE (1 << 1) |
| 108 | # define DP_LANE_SYMBOL_LOCKED (1 << 2) | 132 | # define DP_LANE_SYMBOL_LOCKED (1 << 2) |
| 109 | 133 | ||
| 134 | #define DP_CHANNEL_EQ_BITS (DP_LANE_CR_DONE | \ | ||
| 135 | DP_LANE_CHANNEL_EQ_DONE | \ | ||
| 136 | DP_LANE_SYMBOL_LOCKED) | ||
| 137 | |||
| 110 | #define DP_LANE_ALIGN_STATUS_UPDATED 0x204 | 138 | #define DP_LANE_ALIGN_STATUS_UPDATED 0x204 |
| 111 | 139 | ||
| 112 | #define DP_INTERLANE_ALIGN_DONE (1 << 0) | 140 | #define DP_INTERLANE_ALIGN_DONE (1 << 0) |
| @@ -120,25 +148,33 @@ | |||
| 120 | 148 | ||
| 121 | #define DP_ADJUST_REQUEST_LANE0_1 0x206 | 149 | #define DP_ADJUST_REQUEST_LANE0_1 0x206 |
| 122 | #define DP_ADJUST_REQUEST_LANE2_3 0x207 | 150 | #define DP_ADJUST_REQUEST_LANE2_3 0x207 |
| 123 | 151 | # define DP_ADJUST_VOLTAGE_SWING_LANE0_MASK 0x03 | |
| 124 | #define DP_ADJUST_VOLTAGE_SWING_LANE0_MASK 0x03 | 152 | # define DP_ADJUST_VOLTAGE_SWING_LANE0_SHIFT 0 |
| 125 | #define DP_ADJUST_VOLTAGE_SWING_LANE0_SHIFT 0 | 153 | # define DP_ADJUST_PRE_EMPHASIS_LANE0_MASK 0x0c |
| 126 | #define DP_ADJUST_PRE_EMPHASIS_LANE0_MASK 0x0c | 154 | # define DP_ADJUST_PRE_EMPHASIS_LANE0_SHIFT 2 |
| 127 | #define DP_ADJUST_PRE_EMPHASIS_LANE0_SHIFT 2 | 155 | # define DP_ADJUST_VOLTAGE_SWING_LANE1_MASK 0x30 |
| 128 | #define DP_ADJUST_VOLTAGE_SWING_LANE1_MASK 0x30 | 156 | # define DP_ADJUST_VOLTAGE_SWING_LANE1_SHIFT 4 |
| 129 | #define DP_ADJUST_VOLTAGE_SWING_LANE1_SHIFT 4 | 157 | # define DP_ADJUST_PRE_EMPHASIS_LANE1_MASK 0xc0 |
| 130 | #define DP_ADJUST_PRE_EMPHASIS_LANE1_MASK 0xc0 | 158 | # define DP_ADJUST_PRE_EMPHASIS_LANE1_SHIFT 6 |
| 131 | #define DP_ADJUST_PRE_EMPHASIS_LANE1_SHIFT 6 | 159 | |
| 160 | #define DP_SET_POWER 0x600 | ||
| 161 | # define DP_SET_POWER_D0 0x1 | ||
| 162 | # define DP_SET_POWER_D3 0x2 | ||
| 163 | |||
| 164 | #define MODE_I2C_START 1 | ||
| 165 | #define MODE_I2C_WRITE 2 | ||
| 166 | #define MODE_I2C_READ 4 | ||
| 167 | #define MODE_I2C_STOP 8 | ||
| 132 | 168 | ||
| 133 | struct i2c_algo_dp_aux_data { | 169 | struct i2c_algo_dp_aux_data { |
| 134 | bool running; | 170 | bool running; |
| 135 | u16 address; | 171 | u16 address; |
| 136 | int (*aux_ch) (struct i2c_adapter *adapter, | 172 | int (*aux_ch) (struct i2c_adapter *adapter, |
| 137 | uint8_t *send, int send_bytes, | 173 | int mode, uint8_t write_byte, |
| 138 | uint8_t *recv, int recv_bytes); | 174 | uint8_t *read_byte); |
| 139 | }; | 175 | }; |
| 140 | 176 | ||
| 141 | int | 177 | int |
| 142 | i2c_dp_aux_add_bus(struct i2c_adapter *adapter); | 178 | i2c_dp_aux_add_bus(struct i2c_adapter *adapter); |
| 143 | 179 | ||
| 144 | #endif /* _INTEL_DP_H_ */ | 180 | #endif /* _DRM_DP_HELPER_H_ */ |
diff --git a/include/drm/drm_edid.h b/include/drm/drm_edid.h index 7d6c9a2dfcbb..d33c3e038606 100644 --- a/include/drm/drm_edid.h +++ b/include/drm/drm_edid.h | |||
| @@ -106,6 +106,10 @@ struct detailed_data_color_point { | |||
| 106 | u8 wpindex2[3]; | 106 | u8 wpindex2[3]; |
| 107 | } __attribute__((packed)); | 107 | } __attribute__((packed)); |
| 108 | 108 | ||
| 109 | struct cvt_timing { | ||
| 110 | u8 code[3]; | ||
| 111 | } __attribute__((packed)); | ||
| 112 | |||
| 109 | struct detailed_non_pixel { | 113 | struct detailed_non_pixel { |
| 110 | u8 pad1; | 114 | u8 pad1; |
| 111 | u8 type; /* ff=serial, fe=string, fd=monitor range, fc=monitor name | 115 | u8 type; /* ff=serial, fe=string, fd=monitor range, fc=monitor name |
| @@ -117,9 +121,13 @@ struct detailed_non_pixel { | |||
| 117 | struct detailed_data_monitor_range range; | 121 | struct detailed_data_monitor_range range; |
| 118 | struct detailed_data_wpindex color; | 122 | struct detailed_data_wpindex color; |
| 119 | struct std_timing timings[5]; | 123 | struct std_timing timings[5]; |
| 124 | struct cvt_timing cvt[4]; | ||
| 120 | } data; | 125 | } data; |
| 121 | } __attribute__((packed)); | 126 | } __attribute__((packed)); |
| 122 | 127 | ||
| 128 | #define EDID_DETAIL_EST_TIMINGS 0xf7 | ||
| 129 | #define EDID_DETAIL_CVT_3BYTE 0xf8 | ||
| 130 | #define EDID_DETAIL_COLOR_MGMT_DATA 0xf9 | ||
| 123 | #define EDID_DETAIL_STD_MODES 0xfa | 131 | #define EDID_DETAIL_STD_MODES 0xfa |
| 124 | #define EDID_DETAIL_MONITOR_CPDATA 0xfb | 132 | #define EDID_DETAIL_MONITOR_CPDATA 0xfb |
| 125 | #define EDID_DETAIL_MONITOR_NAME 0xfc | 133 | #define EDID_DETAIL_MONITOR_NAME 0xfc |
diff --git a/include/drm/drm_mm.h b/include/drm/drm_mm.h index 62329f9a42cb..4c10be39a43b 100644 --- a/include/drm/drm_mm.h +++ b/include/drm/drm_mm.h | |||
| @@ -66,6 +66,13 @@ extern struct drm_mm_node *drm_mm_get_block_generic(struct drm_mm_node *node, | |||
| 66 | unsigned long size, | 66 | unsigned long size, |
| 67 | unsigned alignment, | 67 | unsigned alignment, |
| 68 | int atomic); | 68 | int atomic); |
| 69 | extern struct drm_mm_node *drm_mm_get_block_range_generic( | ||
| 70 | struct drm_mm_node *node, | ||
| 71 | unsigned long size, | ||
| 72 | unsigned alignment, | ||
| 73 | unsigned long start, | ||
| 74 | unsigned long end, | ||
| 75 | int atomic); | ||
| 69 | static inline struct drm_mm_node *drm_mm_get_block(struct drm_mm_node *parent, | 76 | static inline struct drm_mm_node *drm_mm_get_block(struct drm_mm_node *parent, |
| 70 | unsigned long size, | 77 | unsigned long size, |
| 71 | unsigned alignment) | 78 | unsigned alignment) |
| @@ -78,11 +85,38 @@ static inline struct drm_mm_node *drm_mm_get_block_atomic(struct drm_mm_node *pa | |||
| 78 | { | 85 | { |
| 79 | return drm_mm_get_block_generic(parent, size, alignment, 1); | 86 | return drm_mm_get_block_generic(parent, size, alignment, 1); |
| 80 | } | 87 | } |
| 88 | static inline struct drm_mm_node *drm_mm_get_block_range( | ||
| 89 | struct drm_mm_node *parent, | ||
| 90 | unsigned long size, | ||
| 91 | unsigned alignment, | ||
| 92 | unsigned long start, | ||
| 93 | unsigned long end) | ||
| 94 | { | ||
| 95 | return drm_mm_get_block_range_generic(parent, size, alignment, | ||
| 96 | start, end, 0); | ||
| 97 | } | ||
| 98 | static inline struct drm_mm_node *drm_mm_get_block_atomic_range( | ||
| 99 | struct drm_mm_node *parent, | ||
| 100 | unsigned long size, | ||
| 101 | unsigned alignment, | ||
| 102 | unsigned long start, | ||
| 103 | unsigned long end) | ||
| 104 | { | ||
| 105 | return drm_mm_get_block_range_generic(parent, size, alignment, | ||
| 106 | start, end, 1); | ||
| 107 | } | ||
| 81 | extern void drm_mm_put_block(struct drm_mm_node *cur); | 108 | extern void drm_mm_put_block(struct drm_mm_node *cur); |
| 82 | extern struct drm_mm_node *drm_mm_search_free(const struct drm_mm *mm, | 109 | extern struct drm_mm_node *drm_mm_search_free(const struct drm_mm *mm, |
| 83 | unsigned long size, | 110 | unsigned long size, |
| 84 | unsigned alignment, | 111 | unsigned alignment, |
| 85 | int best_match); | 112 | int best_match); |
| 113 | extern struct drm_mm_node *drm_mm_search_free_in_range( | ||
| 114 | const struct drm_mm *mm, | ||
| 115 | unsigned long size, | ||
| 116 | unsigned alignment, | ||
| 117 | unsigned long start, | ||
| 118 | unsigned long end, | ||
| 119 | int best_match); | ||
| 86 | extern int drm_mm_init(struct drm_mm *mm, unsigned long start, | 120 | extern int drm_mm_init(struct drm_mm *mm, unsigned long start, |
| 87 | unsigned long size); | 121 | unsigned long size); |
| 88 | extern void drm_mm_takedown(struct drm_mm *mm); | 122 | extern void drm_mm_takedown(struct drm_mm *mm); |
| @@ -99,6 +133,7 @@ static inline struct drm_mm *drm_get_mm(struct drm_mm_node *block) | |||
| 99 | return block->mm; | 133 | return block->mm; |
| 100 | } | 134 | } |
| 101 | 135 | ||
| 136 | extern void drm_mm_debug_table(struct drm_mm *mm, const char *prefix); | ||
| 102 | #ifdef CONFIG_DEBUG_FS | 137 | #ifdef CONFIG_DEBUG_FS |
| 103 | int drm_mm_dump_table(struct seq_file *m, struct drm_mm *mm); | 138 | int drm_mm_dump_table(struct seq_file *m, struct drm_mm *mm); |
| 104 | #endif | 139 | #endif |
diff --git a/include/drm/drm_mode.h b/include/drm/drm_mode.h index 1f908416aedb..43009bc2e757 100644 --- a/include/drm/drm_mode.h +++ b/include/drm/drm_mode.h | |||
| @@ -27,9 +27,6 @@ | |||
| 27 | #ifndef _DRM_MODE_H | 27 | #ifndef _DRM_MODE_H |
| 28 | #define _DRM_MODE_H | 28 | #define _DRM_MODE_H |
| 29 | 29 | ||
| 30 | #include <linux/kernel.h> | ||
| 31 | #include <linux/types.h> | ||
| 32 | |||
| 33 | #define DRM_DISPLAY_INFO_LEN 32 | 30 | #define DRM_DISPLAY_INFO_LEN 32 |
| 34 | #define DRM_CONNECTOR_NAME_LEN 32 | 31 | #define DRM_CONNECTOR_NAME_LEN 32 |
| 35 | #define DRM_DISPLAY_MODE_LEN 32 | 32 | #define DRM_DISPLAY_MODE_LEN 32 |
| @@ -78,6 +75,11 @@ | |||
| 78 | #define DRM_MODE_DITHERING_OFF 0 | 75 | #define DRM_MODE_DITHERING_OFF 0 |
| 79 | #define DRM_MODE_DITHERING_ON 1 | 76 | #define DRM_MODE_DITHERING_ON 1 |
| 80 | 77 | ||
| 78 | /* Dirty info options */ | ||
| 79 | #define DRM_MODE_DIRTY_OFF 0 | ||
| 80 | #define DRM_MODE_DIRTY_ON 1 | ||
| 81 | #define DRM_MODE_DIRTY_ANNOTATE 2 | ||
| 82 | |||
| 81 | struct drm_mode_modeinfo { | 83 | struct drm_mode_modeinfo { |
| 82 | __u32 clock; | 84 | __u32 clock; |
| 83 | __u16 hdisplay, hsync_start, hsync_end, htotal, hskew; | 85 | __u16 hdisplay, hsync_start, hsync_end, htotal, hskew; |
| @@ -225,6 +227,45 @@ struct drm_mode_fb_cmd { | |||
| 225 | __u32 handle; | 227 | __u32 handle; |
| 226 | }; | 228 | }; |
| 227 | 229 | ||
| 230 | #define DRM_MODE_FB_DIRTY_ANNOTATE_COPY 0x01 | ||
| 231 | #define DRM_MODE_FB_DIRTY_ANNOTATE_FILL 0x02 | ||
| 232 | #define DRM_MODE_FB_DIRTY_FLAGS 0x03 | ||
| 233 | |||
| 234 | /* | ||
| 235 | * Mark a region of a framebuffer as dirty. | ||
| 236 | * | ||
| 237 | * Some hardware does not automatically update display contents | ||
| 238 | * as a hardware or software draw to a framebuffer. This ioctl | ||
| 239 | * allows userspace to tell the kernel and the hardware what | ||
| 240 | * regions of the framebuffer have changed. | ||
| 241 | * | ||
| 242 | * The kernel or hardware is free to update more then just the | ||
| 243 | * region specified by the clip rects. The kernel or hardware | ||
| 244 | * may also delay and/or coalesce several calls to dirty into a | ||
| 245 | * single update. | ||
| 246 | * | ||
| 247 | * Userspace may annotate the updates, the annotates are a | ||
| 248 | * promise made by the caller that the change is either a copy | ||
| 249 | * of pixels or a fill of a single color in the region specified. | ||
| 250 | * | ||
| 251 | * If the DRM_MODE_FB_DIRTY_ANNOTATE_COPY flag is given then | ||
| 252 | * the number of updated regions are half of num_clips given, | ||
| 253 | * where the clip rects are paired in src and dst. The width and | ||
| 254 | * height of each one of the pairs must match. | ||
| 255 | * | ||
| 256 | * If the DRM_MODE_FB_DIRTY_ANNOTATE_FILL flag is given the caller | ||
| 257 | * promises that the region specified of the clip rects is filled | ||
| 258 | * completely with a single color as given in the color argument. | ||
| 259 | */ | ||
| 260 | |||
| 261 | struct drm_mode_fb_dirty_cmd { | ||
| 262 | __u32 fb_id; | ||
| 263 | __u32 flags; | ||
| 264 | __u32 color; | ||
| 265 | __u32 num_clips; | ||
| 266 | __u64 clips_ptr; | ||
| 267 | }; | ||
| 268 | |||
| 228 | struct drm_mode_mode_cmd { | 269 | struct drm_mode_mode_cmd { |
| 229 | __u32 connector_id; | 270 | __u32 connector_id; |
| 230 | struct drm_mode_modeinfo mode; | 271 | struct drm_mode_modeinfo mode; |
| @@ -268,4 +309,37 @@ struct drm_mode_crtc_lut { | |||
| 268 | __u64 blue; | 309 | __u64 blue; |
| 269 | }; | 310 | }; |
| 270 | 311 | ||
| 312 | #define DRM_MODE_PAGE_FLIP_EVENT 0x01 | ||
| 313 | #define DRM_MODE_PAGE_FLIP_FLAGS DRM_MODE_PAGE_FLIP_EVENT | ||
| 314 | |||
| 315 | /* | ||
| 316 | * Request a page flip on the specified crtc. | ||
| 317 | * | ||
| 318 | * This ioctl will ask KMS to schedule a page flip for the specified | ||
| 319 | * crtc. Once any pending rendering targeting the specified fb (as of | ||
| 320 | * ioctl time) has completed, the crtc will be reprogrammed to display | ||
| 321 | * that fb after the next vertical refresh. The ioctl returns | ||
| 322 | * immediately, but subsequent rendering to the current fb will block | ||
| 323 | * in the execbuffer ioctl until the page flip happens. If a page | ||
| 324 | * flip is already pending as the ioctl is called, EBUSY will be | ||
| 325 | * returned. | ||
| 326 | * | ||
| 327 | * The ioctl supports one flag, DRM_MODE_PAGE_FLIP_EVENT, which will | ||
| 328 | * request that drm sends back a vblank event (see drm.h: struct | ||
| 329 | * drm_event_vblank) when the page flip is done. The user_data field | ||
| 330 | * passed in with this ioctl will be returned as the user_data field | ||
| 331 | * in the vblank event struct. | ||
| 332 | * | ||
| 333 | * The reserved field must be zero until we figure out something | ||
| 334 | * clever to use it for. | ||
| 335 | */ | ||
| 336 | |||
| 337 | struct drm_mode_crtc_page_flip { | ||
| 338 | __u32 crtc_id; | ||
| 339 | __u32 fb_id; | ||
| 340 | __u32 flags; | ||
| 341 | __u32 reserved; | ||
| 342 | __u64 user_data; | ||
| 343 | }; | ||
| 344 | |||
| 271 | #endif | 345 | #endif |
diff --git a/include/drm/drm_os_linux.h b/include/drm/drm_os_linux.h index 26641e95e0a4..393369147a2d 100644 --- a/include/drm/drm_os_linux.h +++ b/include/drm/drm_os_linux.h | |||
| @@ -123,5 +123,5 @@ do { \ | |||
| 123 | remove_wait_queue(&(queue), &entry); \ | 123 | remove_wait_queue(&(queue), &entry); \ |
| 124 | } while (0) | 124 | } while (0) |
| 125 | 125 | ||
| 126 | #define DRM_WAKEUP( queue ) wake_up_interruptible( queue ) | 126 | #define DRM_WAKEUP( queue ) wake_up( queue ) |
| 127 | #define DRM_INIT_WAITQUEUE( queue ) init_waitqueue_head( queue ) | 127 | #define DRM_INIT_WAITQUEUE( queue ) init_waitqueue_head( queue ) |
diff --git a/include/drm/i915_drm.h b/include/drm/i915_drm.h index 7e0cb1da92e6..ec3f5e80a5df 100644 --- a/include/drm/i915_drm.h +++ b/include/drm/i915_drm.h | |||
| @@ -27,11 +27,11 @@ | |||
| 27 | #ifndef _I915_DRM_H_ | 27 | #ifndef _I915_DRM_H_ |
| 28 | #define _I915_DRM_H_ | 28 | #define _I915_DRM_H_ |
| 29 | 29 | ||
| 30 | #include "drm.h" | ||
| 31 | |||
| 30 | /* Please note that modifications to all structs defined here are | 32 | /* Please note that modifications to all structs defined here are |
| 31 | * subject to backwards-compatibility constraints. | 33 | * subject to backwards-compatibility constraints. |
| 32 | */ | 34 | */ |
| 33 | #include <linux/types.h> | ||
| 34 | #include "drm.h" | ||
| 35 | 35 | ||
| 36 | /* Each region is a minimum of 16k, and there are at most 255 of them. | 36 | /* Each region is a minimum of 16k, and there are at most 255 of them. |
| 37 | */ | 37 | */ |
| @@ -186,6 +186,8 @@ typedef struct _drm_i915_sarea { | |||
| 186 | #define DRM_I915_GEM_MMAP_GTT 0x24 | 186 | #define DRM_I915_GEM_MMAP_GTT 0x24 |
| 187 | #define DRM_I915_GET_PIPE_FROM_CRTC_ID 0x25 | 187 | #define DRM_I915_GET_PIPE_FROM_CRTC_ID 0x25 |
| 188 | #define DRM_I915_GEM_MADVISE 0x26 | 188 | #define DRM_I915_GEM_MADVISE 0x26 |
| 189 | #define DRM_I915_OVERLAY_PUT_IMAGE 0x27 | ||
| 190 | #define DRM_I915_OVERLAY_ATTRS 0x28 | ||
| 189 | 191 | ||
| 190 | #define DRM_IOCTL_I915_INIT DRM_IOW( DRM_COMMAND_BASE + DRM_I915_INIT, drm_i915_init_t) | 192 | #define DRM_IOCTL_I915_INIT DRM_IOW( DRM_COMMAND_BASE + DRM_I915_INIT, drm_i915_init_t) |
| 191 | #define DRM_IOCTL_I915_FLUSH DRM_IO ( DRM_COMMAND_BASE + DRM_I915_FLUSH) | 193 | #define DRM_IOCTL_I915_FLUSH DRM_IO ( DRM_COMMAND_BASE + DRM_I915_FLUSH) |
| @@ -221,8 +223,10 @@ typedef struct _drm_i915_sarea { | |||
| 221 | #define DRM_IOCTL_I915_GEM_SET_TILING DRM_IOWR (DRM_COMMAND_BASE + DRM_I915_GEM_SET_TILING, struct drm_i915_gem_set_tiling) | 223 | #define DRM_IOCTL_I915_GEM_SET_TILING DRM_IOWR (DRM_COMMAND_BASE + DRM_I915_GEM_SET_TILING, struct drm_i915_gem_set_tiling) |
| 222 | #define DRM_IOCTL_I915_GEM_GET_TILING DRM_IOWR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_TILING, struct drm_i915_gem_get_tiling) | 224 | #define DRM_IOCTL_I915_GEM_GET_TILING DRM_IOWR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_TILING, struct drm_i915_gem_get_tiling) |
| 223 | #define DRM_IOCTL_I915_GEM_GET_APERTURE DRM_IOR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_APERTURE, struct drm_i915_gem_get_aperture) | 225 | #define DRM_IOCTL_I915_GEM_GET_APERTURE DRM_IOR (DRM_COMMAND_BASE + DRM_I915_GEM_GET_APERTURE, struct drm_i915_gem_get_aperture) |
| 224 | #define DRM_IOCTL_I915_GET_PIPE_FROM_CRTC_ID DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GET_PIPE_FROM_CRTC_ID, struct drm_intel_get_pipe_from_crtc_id) | 226 | #define DRM_IOCTL_I915_GET_PIPE_FROM_CRTC_ID DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GET_PIPE_FROM_CRTC_ID, struct drm_i915_get_pipe_from_crtc_id) |
| 225 | #define DRM_IOCTL_I915_GEM_MADVISE DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_MADVISE, struct drm_i915_gem_madvise) | 227 | #define DRM_IOCTL_I915_GEM_MADVISE DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_MADVISE, struct drm_i915_gem_madvise) |
| 228 | #define DRM_IOCTL_I915_OVERLAY_PUT_IMAGE DRM_IOW(DRM_COMMAND_BASE + DRM_IOCTL_I915_OVERLAY_ATTRS, struct drm_intel_overlay_put_image) | ||
| 229 | #define DRM_IOCTL_I915_OVERLAY_ATTRS DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_OVERLAY_ATTRS, struct drm_intel_overlay_attrs) | ||
| 226 | 230 | ||
| 227 | /* Allow drivers to submit batchbuffers directly to hardware, relying | 231 | /* Allow drivers to submit batchbuffers directly to hardware, relying |
| 228 | * on the security mechanisms provided by hardware. | 232 | * on the security mechanisms provided by hardware. |
| @@ -266,6 +270,8 @@ typedef struct drm_i915_irq_wait { | |||
| 266 | #define I915_PARAM_CHIPSET_ID 4 | 270 | #define I915_PARAM_CHIPSET_ID 4 |
| 267 | #define I915_PARAM_HAS_GEM 5 | 271 | #define I915_PARAM_HAS_GEM 5 |
| 268 | #define I915_PARAM_NUM_FENCES_AVAIL 6 | 272 | #define I915_PARAM_NUM_FENCES_AVAIL 6 |
| 273 | #define I915_PARAM_HAS_OVERLAY 7 | ||
| 274 | #define I915_PARAM_HAS_PAGEFLIPPING 8 | ||
| 269 | 275 | ||
| 270 | typedef struct drm_i915_getparam { | 276 | typedef struct drm_i915_getparam { |
| 271 | int param; | 277 | int param; |
| @@ -686,4 +692,70 @@ struct drm_i915_gem_madvise { | |||
| 686 | __u32 retained; | 692 | __u32 retained; |
| 687 | }; | 693 | }; |
| 688 | 694 | ||
| 695 | /* flags */ | ||
| 696 | #define I915_OVERLAY_TYPE_MASK 0xff | ||
| 697 | #define I915_OVERLAY_YUV_PLANAR 0x01 | ||
| 698 | #define I915_OVERLAY_YUV_PACKED 0x02 | ||
| 699 | #define I915_OVERLAY_RGB 0x03 | ||
| 700 | |||
| 701 | #define I915_OVERLAY_DEPTH_MASK 0xff00 | ||
| 702 | #define I915_OVERLAY_RGB24 0x1000 | ||
| 703 | #define I915_OVERLAY_RGB16 0x2000 | ||
| 704 | #define I915_OVERLAY_RGB15 0x3000 | ||
| 705 | #define I915_OVERLAY_YUV422 0x0100 | ||
| 706 | #define I915_OVERLAY_YUV411 0x0200 | ||
| 707 | #define I915_OVERLAY_YUV420 0x0300 | ||
| 708 | #define I915_OVERLAY_YUV410 0x0400 | ||
| 709 | |||
| 710 | #define I915_OVERLAY_SWAP_MASK 0xff0000 | ||
| 711 | #define I915_OVERLAY_NO_SWAP 0x000000 | ||
| 712 | #define I915_OVERLAY_UV_SWAP 0x010000 | ||
| 713 | #define I915_OVERLAY_Y_SWAP 0x020000 | ||
| 714 | #define I915_OVERLAY_Y_AND_UV_SWAP 0x030000 | ||
| 715 | |||
| 716 | #define I915_OVERLAY_FLAGS_MASK 0xff000000 | ||
| 717 | #define I915_OVERLAY_ENABLE 0x01000000 | ||
| 718 | |||
| 719 | struct drm_intel_overlay_put_image { | ||
| 720 | /* various flags and src format description */ | ||
| 721 | __u32 flags; | ||
| 722 | /* source picture description */ | ||
| 723 | __u32 bo_handle; | ||
| 724 | /* stride values and offsets are in bytes, buffer relative */ | ||
| 725 | __u16 stride_Y; /* stride for packed formats */ | ||
| 726 | __u16 stride_UV; | ||
| 727 | __u32 offset_Y; /* offset for packet formats */ | ||
| 728 | __u32 offset_U; | ||
| 729 | __u32 offset_V; | ||
| 730 | /* in pixels */ | ||
| 731 | __u16 src_width; | ||
| 732 | __u16 src_height; | ||
| 733 | /* to compensate the scaling factors for partially covered surfaces */ | ||
| 734 | __u16 src_scan_width; | ||
| 735 | __u16 src_scan_height; | ||
| 736 | /* output crtc description */ | ||
| 737 | __u32 crtc_id; | ||
| 738 | __u16 dst_x; | ||
| 739 | __u16 dst_y; | ||
| 740 | __u16 dst_width; | ||
| 741 | __u16 dst_height; | ||
| 742 | }; | ||
| 743 | |||
| 744 | /* flags */ | ||
| 745 | #define I915_OVERLAY_UPDATE_ATTRS (1<<0) | ||
| 746 | #define I915_OVERLAY_UPDATE_GAMMA (1<<1) | ||
| 747 | struct drm_intel_overlay_attrs { | ||
| 748 | __u32 flags; | ||
| 749 | __u32 color_key; | ||
| 750 | __s32 brightness; | ||
| 751 | __u32 contrast; | ||
| 752 | __u32 saturation; | ||
| 753 | __u32 gamma0; | ||
| 754 | __u32 gamma1; | ||
| 755 | __u32 gamma2; | ||
| 756 | __u32 gamma3; | ||
| 757 | __u32 gamma4; | ||
| 758 | __u32 gamma5; | ||
| 759 | }; | ||
| 760 | |||
| 689 | #endif /* _I915_DRM_H_ */ | 761 | #endif /* _I915_DRM_H_ */ |
diff --git a/include/drm/mga_drm.h b/include/drm/mga_drm.h index 325fd6fb4a42..3ffbc4798afa 100644 --- a/include/drm/mga_drm.h +++ b/include/drm/mga_drm.h | |||
| @@ -35,7 +35,7 @@ | |||
| 35 | #ifndef __MGA_DRM_H__ | 35 | #ifndef __MGA_DRM_H__ |
| 36 | #define __MGA_DRM_H__ | 36 | #define __MGA_DRM_H__ |
| 37 | 37 | ||
| 38 | #include <linux/types.h> | 38 | #include "drm.h" |
| 39 | 39 | ||
| 40 | /* WARNING: If you change any of these defines, make sure to change the | 40 | /* WARNING: If you change any of these defines, make sure to change the |
| 41 | * defines in the Xserver file (mga_sarea.h) | 41 | * defines in the Xserver file (mga_sarea.h) |
diff --git a/include/drm/radeon_drm.h b/include/drm/radeon_drm.h index 3b9932ab1756..39537f3cf98a 100644 --- a/include/drm/radeon_drm.h +++ b/include/drm/radeon_drm.h | |||
| @@ -33,7 +33,7 @@ | |||
| 33 | #ifndef __RADEON_DRM_H__ | 33 | #ifndef __RADEON_DRM_H__ |
| 34 | #define __RADEON_DRM_H__ | 34 | #define __RADEON_DRM_H__ |
| 35 | 35 | ||
| 36 | #include <linux/types.h> | 36 | #include "drm.h" |
| 37 | 37 | ||
| 38 | /* WARNING: If you change any of these defines, make sure to change the | 38 | /* WARNING: If you change any of these defines, make sure to change the |
| 39 | * defines in the X server file (radeon_sarea.h) | 39 | * defines in the X server file (radeon_sarea.h) |
diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h index 491146170522..4fd498523ce3 100644 --- a/include/drm/ttm/ttm_bo_api.h +++ b/include/drm/ttm/ttm_bo_api.h | |||
| @@ -44,6 +44,29 @@ struct ttm_bo_device; | |||
| 44 | 44 | ||
| 45 | struct drm_mm_node; | 45 | struct drm_mm_node; |
| 46 | 46 | ||
| 47 | |||
| 48 | /** | ||
| 49 | * struct ttm_placement | ||
| 50 | * | ||
| 51 | * @fpfn: first valid page frame number to put the object | ||
| 52 | * @lpfn: last valid page frame number to put the object | ||
| 53 | * @num_placement: number of prefered placements | ||
| 54 | * @placement: prefered placements | ||
| 55 | * @num_busy_placement: number of prefered placements when need to evict buffer | ||
| 56 | * @busy_placement: prefered placements when need to evict buffer | ||
| 57 | * | ||
| 58 | * Structure indicating the placement you request for an object. | ||
| 59 | */ | ||
| 60 | struct ttm_placement { | ||
| 61 | unsigned fpfn; | ||
| 62 | unsigned lpfn; | ||
| 63 | unsigned num_placement; | ||
| 64 | const uint32_t *placement; | ||
| 65 | unsigned num_busy_placement; | ||
| 66 | const uint32_t *busy_placement; | ||
| 67 | }; | ||
| 68 | |||
| 69 | |||
| 47 | /** | 70 | /** |
| 48 | * struct ttm_mem_reg | 71 | * struct ttm_mem_reg |
| 49 | * | 72 | * |
| @@ -109,10 +132,6 @@ struct ttm_tt; | |||
| 109 | * the object is destroyed. | 132 | * the object is destroyed. |
| 110 | * @event_queue: Queue for processes waiting on buffer object status change. | 133 | * @event_queue: Queue for processes waiting on buffer object status change. |
| 111 | * @lock: spinlock protecting mostly synchronization members. | 134 | * @lock: spinlock protecting mostly synchronization members. |
| 112 | * @proposed_placement: Proposed placement for the buffer. Changed only by the | ||
| 113 | * creator prior to validation as opposed to bo->mem.proposed_flags which is | ||
| 114 | * changed by the implementation prior to a buffer move if it wants to outsmart | ||
| 115 | * the buffer creator / user. This latter happens, for example, at eviction. | ||
| 116 | * @mem: structure describing current placement. | 135 | * @mem: structure describing current placement. |
| 117 | * @persistant_swap_storage: Usually the swap storage is deleted for buffers | 136 | * @persistant_swap_storage: Usually the swap storage is deleted for buffers |
| 118 | * pinned in physical memory. If this behaviour is not desired, this member | 137 | * pinned in physical memory. If this behaviour is not desired, this member |
| @@ -177,7 +196,6 @@ struct ttm_buffer_object { | |||
| 177 | * Members protected by the bo::reserved lock. | 196 | * Members protected by the bo::reserved lock. |
| 178 | */ | 197 | */ |
| 179 | 198 | ||
| 180 | uint32_t proposed_placement; | ||
| 181 | struct ttm_mem_reg mem; | 199 | struct ttm_mem_reg mem; |
| 182 | struct file *persistant_swap_storage; | 200 | struct file *persistant_swap_storage; |
| 183 | struct ttm_tt *ttm; | 201 | struct ttm_tt *ttm; |
| @@ -285,7 +303,7 @@ ttm_bo_reference(struct ttm_buffer_object *bo) | |||
| 285 | * Note: It might be necessary to block validations before the | 303 | * Note: It might be necessary to block validations before the |
| 286 | * wait by reserving the buffer. | 304 | * wait by reserving the buffer. |
| 287 | * Returns -EBUSY if no_wait is true and the buffer is busy. | 305 | * Returns -EBUSY if no_wait is true and the buffer is busy. |
| 288 | * Returns -ERESTART if interrupted by a signal. | 306 | * Returns -ERESTARTSYS if interrupted by a signal. |
| 289 | */ | 307 | */ |
| 290 | extern int ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, | 308 | extern int ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, |
| 291 | bool interruptible, bool no_wait); | 309 | bool interruptible, bool no_wait); |
| @@ -293,21 +311,22 @@ extern int ttm_bo_wait(struct ttm_buffer_object *bo, bool lazy, | |||
| 293 | * ttm_buffer_object_validate | 311 | * ttm_buffer_object_validate |
| 294 | * | 312 | * |
| 295 | * @bo: The buffer object. | 313 | * @bo: The buffer object. |
| 296 | * @proposed_placement: Proposed_placement for the buffer object. | 314 | * @placement: Proposed placement for the buffer object. |
| 297 | * @interruptible: Sleep interruptible if sleeping. | 315 | * @interruptible: Sleep interruptible if sleeping. |
| 298 | * @no_wait: Return immediately if the buffer is busy. | 316 | * @no_wait: Return immediately if the buffer is busy. |
| 299 | * | 317 | * |
| 300 | * Changes placement and caching policy of the buffer object | 318 | * Changes placement and caching policy of the buffer object |
| 301 | * according to bo::proposed_flags. | 319 | * according proposed placement. |
| 302 | * Returns | 320 | * Returns |
| 303 | * -EINVAL on invalid proposed_flags. | 321 | * -EINVAL on invalid proposed placement. |
| 304 | * -ENOMEM on out-of-memory condition. | 322 | * -ENOMEM on out-of-memory condition. |
| 305 | * -EBUSY if no_wait is true and buffer busy. | 323 | * -EBUSY if no_wait is true and buffer busy. |
| 306 | * -ERESTART if interrupted by a signal. | 324 | * -ERESTARTSYS if interrupted by a signal. |
| 307 | */ | 325 | */ |
| 308 | extern int ttm_buffer_object_validate(struct ttm_buffer_object *bo, | 326 | extern int ttm_buffer_object_validate(struct ttm_buffer_object *bo, |
| 309 | uint32_t proposed_placement, | 327 | struct ttm_placement *placement, |
| 310 | bool interruptible, bool no_wait); | 328 | bool interruptible, bool no_wait); |
| 329 | |||
| 311 | /** | 330 | /** |
| 312 | * ttm_bo_unref | 331 | * ttm_bo_unref |
| 313 | * | 332 | * |
| @@ -328,7 +347,7 @@ extern void ttm_bo_unref(struct ttm_buffer_object **bo); | |||
| 328 | * waiting for buffer idle. This lock is recursive. | 347 | * waiting for buffer idle. This lock is recursive. |
| 329 | * Returns | 348 | * Returns |
| 330 | * -EBUSY if the buffer is busy and no_wait is true. | 349 | * -EBUSY if the buffer is busy and no_wait is true. |
| 331 | * -ERESTART if interrupted by a signal. | 350 | * -ERESTARTSYS if interrupted by a signal. |
| 332 | */ | 351 | */ |
| 333 | 352 | ||
| 334 | extern int | 353 | extern int |
| @@ -371,7 +390,7 @@ extern void ttm_bo_synccpu_write_release(struct ttm_buffer_object *bo); | |||
| 371 | * Returns | 390 | * Returns |
| 372 | * -ENOMEM: Out of memory. | 391 | * -ENOMEM: Out of memory. |
| 373 | * -EINVAL: Invalid placement flags. | 392 | * -EINVAL: Invalid placement flags. |
| 374 | * -ERESTART: Interrupted by signal while sleeping waiting for resources. | 393 | * -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources. |
| 375 | */ | 394 | */ |
| 376 | 395 | ||
| 377 | extern int ttm_buffer_object_init(struct ttm_bo_device *bdev, | 396 | extern int ttm_buffer_object_init(struct ttm_bo_device *bdev, |
| @@ -411,7 +430,7 @@ extern int ttm_buffer_object_init(struct ttm_bo_device *bdev, | |||
| 411 | * Returns | 430 | * Returns |
| 412 | * -ENOMEM: Out of memory. | 431 | * -ENOMEM: Out of memory. |
| 413 | * -EINVAL: Invalid placement flags. | 432 | * -EINVAL: Invalid placement flags. |
| 414 | * -ERESTART: Interrupted by signal while waiting for resources. | 433 | * -ERESTARTSYS: Interrupted by signal while waiting for resources. |
| 415 | */ | 434 | */ |
| 416 | 435 | ||
| 417 | extern int ttm_buffer_object_create(struct ttm_bo_device *bdev, | 436 | extern int ttm_buffer_object_create(struct ttm_bo_device *bdev, |
| @@ -445,7 +464,6 @@ extern int ttm_bo_check_placement(struct ttm_buffer_object *bo, | |||
| 445 | * | 464 | * |
| 446 | * @bdev: Pointer to a ttm_bo_device struct. | 465 | * @bdev: Pointer to a ttm_bo_device struct. |
| 447 | * @mem_type: The memory type. | 466 | * @mem_type: The memory type. |
| 448 | * @p_offset: offset for managed area in pages. | ||
| 449 | * @p_size: size managed area in pages. | 467 | * @p_size: size managed area in pages. |
| 450 | * | 468 | * |
| 451 | * Initialize a manager for a given memory type. | 469 | * Initialize a manager for a given memory type. |
| @@ -458,7 +476,7 @@ extern int ttm_bo_check_placement(struct ttm_buffer_object *bo, | |||
| 458 | */ | 476 | */ |
| 459 | 477 | ||
| 460 | extern int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, | 478 | extern int ttm_bo_init_mm(struct ttm_bo_device *bdev, unsigned type, |
| 461 | unsigned long p_offset, unsigned long p_size); | 479 | unsigned long p_size); |
| 462 | /** | 480 | /** |
| 463 | * ttm_bo_clean_mm | 481 | * ttm_bo_clean_mm |
| 464 | * | 482 | * |
| @@ -503,7 +521,7 @@ extern int ttm_bo_clean_mm(struct ttm_bo_device *bdev, unsigned mem_type); | |||
| 503 | * | 521 | * |
| 504 | * Returns: | 522 | * Returns: |
| 505 | * -EINVAL: Invalid or uninitialized memory type. | 523 | * -EINVAL: Invalid or uninitialized memory type. |
| 506 | * -ERESTART: The call was interrupted by a signal while waiting to | 524 | * -ERESTARTSYS: The call was interrupted by a signal while waiting to |
| 507 | * evict a buffer. | 525 | * evict a buffer. |
| 508 | */ | 526 | */ |
| 509 | 527 | ||
| @@ -606,7 +624,7 @@ extern int ttm_bo_mmap(struct file *filp, struct vm_area_struct *vma, | |||
| 606 | * be called from the fops::read and fops::write method. | 624 | * be called from the fops::read and fops::write method. |
| 607 | * Returns: | 625 | * Returns: |
| 608 | * See man (2) write, man(2) read. In particular, | 626 | * See man (2) write, man(2) read. In particular, |
| 609 | * the function may return -EINTR if | 627 | * the function may return -ERESTARTSYS if |
| 610 | * interrupted by a signal. | 628 | * interrupted by a signal. |
| 611 | */ | 629 | */ |
| 612 | 630 | ||
diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h index e8cd6d20aed2..ff7664e0c3cd 100644 --- a/include/drm/ttm/ttm_bo_driver.h +++ b/include/drm/ttm/ttm_bo_driver.h | |||
| @@ -242,12 +242,6 @@ struct ttm_mem_type_manager { | |||
| 242 | /** | 242 | /** |
| 243 | * struct ttm_bo_driver | 243 | * struct ttm_bo_driver |
| 244 | * | 244 | * |
| 245 | * @mem_type_prio: Priority array of memory types to place a buffer object in | ||
| 246 | * if it fits without evicting buffers from any of these memory types. | ||
| 247 | * @mem_busy_prio: Priority array of memory types to place a buffer object in | ||
| 248 | * if it needs to evict buffers to make room. | ||
| 249 | * @num_mem_type_prio: Number of elements in the @mem_type_prio array. | ||
| 250 | * @num_mem_busy_prio: Number of elements in the @num_mem_busy_prio array. | ||
| 251 | * @create_ttm_backend_entry: Callback to create a struct ttm_backend. | 245 | * @create_ttm_backend_entry: Callback to create a struct ttm_backend. |
| 252 | * @invalidate_caches: Callback to invalidate read caches when a buffer object | 246 | * @invalidate_caches: Callback to invalidate read caches when a buffer object |
| 253 | * has been evicted. | 247 | * has been evicted. |
| @@ -265,11 +259,6 @@ struct ttm_mem_type_manager { | |||
| 265 | */ | 259 | */ |
| 266 | 260 | ||
| 267 | struct ttm_bo_driver { | 261 | struct ttm_bo_driver { |
| 268 | const uint32_t *mem_type_prio; | ||
| 269 | const uint32_t *mem_busy_prio; | ||
| 270 | uint32_t num_mem_type_prio; | ||
| 271 | uint32_t num_mem_busy_prio; | ||
| 272 | |||
| 273 | /** | 262 | /** |
| 274 | * struct ttm_bo_driver member create_ttm_backend_entry | 263 | * struct ttm_bo_driver member create_ttm_backend_entry |
| 275 | * | 264 | * |
| @@ -306,7 +295,8 @@ struct ttm_bo_driver { | |||
| 306 | * finished, they'll end up in bo->mem.flags | 295 | * finished, they'll end up in bo->mem.flags |
| 307 | */ | 296 | */ |
| 308 | 297 | ||
| 309 | uint32_t(*evict_flags) (struct ttm_buffer_object *bo); | 298 | void(*evict_flags) (struct ttm_buffer_object *bo, |
| 299 | struct ttm_placement *placement); | ||
| 310 | /** | 300 | /** |
| 311 | * struct ttm_bo_driver member move: | 301 | * struct ttm_bo_driver member move: |
| 312 | * | 302 | * |
| @@ -545,6 +535,15 @@ extern int ttm_tt_set_user(struct ttm_tt *ttm, | |||
| 545 | extern int ttm_tt_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem); | 535 | extern int ttm_tt_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem); |
| 546 | 536 | ||
| 547 | /** | 537 | /** |
| 538 | * ttm_tt_populate: | ||
| 539 | * | ||
| 540 | * @ttm: The struct ttm_tt to contain the backing pages. | ||
| 541 | * | ||
| 542 | * Add backing pages to all of @ttm | ||
| 543 | */ | ||
| 544 | extern int ttm_tt_populate(struct ttm_tt *ttm); | ||
| 545 | |||
| 546 | /** | ||
| 548 | * ttm_ttm_destroy: | 547 | * ttm_ttm_destroy: |
| 549 | * | 548 | * |
| 550 | * @ttm: The struct ttm_tt. | 549 | * @ttm: The struct ttm_tt. |
| @@ -639,12 +638,12 @@ extern bool ttm_mem_reg_is_pci(struct ttm_bo_device *bdev, | |||
| 639 | * -EBUSY: No space available (only if no_wait == 1). | 638 | * -EBUSY: No space available (only if no_wait == 1). |
| 640 | * -ENOMEM: Could not allocate memory for the buffer object, either due to | 639 | * -ENOMEM: Could not allocate memory for the buffer object, either due to |
| 641 | * fragmentation or concurrent allocators. | 640 | * fragmentation or concurrent allocators. |
| 642 | * -ERESTART: An interruptible sleep was interrupted by a signal. | 641 | * -ERESTARTSYS: An interruptible sleep was interrupted by a signal. |
| 643 | */ | 642 | */ |
| 644 | extern int ttm_bo_mem_space(struct ttm_buffer_object *bo, | 643 | extern int ttm_bo_mem_space(struct ttm_buffer_object *bo, |
| 645 | uint32_t proposed_placement, | 644 | struct ttm_placement *placement, |
| 646 | struct ttm_mem_reg *mem, | 645 | struct ttm_mem_reg *mem, |
| 647 | bool interruptible, bool no_wait); | 646 | bool interruptible, bool no_wait); |
| 648 | /** | 647 | /** |
| 649 | * ttm_bo_wait_for_cpu | 648 | * ttm_bo_wait_for_cpu |
| 650 | * | 649 | * |
| @@ -654,7 +653,7 @@ extern int ttm_bo_mem_space(struct ttm_buffer_object *bo, | |||
| 654 | * Wait until a buffer object is no longer sync'ed for CPU access. | 653 | * Wait until a buffer object is no longer sync'ed for CPU access. |
| 655 | * Returns: | 654 | * Returns: |
| 656 | * -EBUSY: Buffer object was sync'ed for CPU access. (only if no_wait == 1). | 655 | * -EBUSY: Buffer object was sync'ed for CPU access. (only if no_wait == 1). |
| 657 | * -ERESTART: An interruptible sleep was interrupted by a signal. | 656 | * -ERESTARTSYS: An interruptible sleep was interrupted by a signal. |
| 658 | */ | 657 | */ |
| 659 | 658 | ||
| 660 | extern int ttm_bo_wait_cpu(struct ttm_buffer_object *bo, bool no_wait); | 659 | extern int ttm_bo_wait_cpu(struct ttm_buffer_object *bo, bool no_wait); |
| @@ -758,7 +757,7 @@ extern void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo); | |||
| 758 | * -EAGAIN: The reservation may cause a deadlock. | 757 | * -EAGAIN: The reservation may cause a deadlock. |
| 759 | * Release all buffer reservations, wait for @bo to become unreserved and | 758 | * Release all buffer reservations, wait for @bo to become unreserved and |
| 760 | * try again. (only if use_sequence == 1). | 759 | * try again. (only if use_sequence == 1). |
| 761 | * -ERESTART: A wait for the buffer to become unreserved was interrupted by | 760 | * -ERESTARTSYS: A wait for the buffer to become unreserved was interrupted by |
| 762 | * a signal. Release all buffer reservations and return to user-space. | 761 | * a signal. Release all buffer reservations and return to user-space. |
| 763 | */ | 762 | */ |
| 764 | extern int ttm_bo_reserve(struct ttm_buffer_object *bo, | 763 | extern int ttm_bo_reserve(struct ttm_buffer_object *bo, |
| @@ -799,7 +798,7 @@ extern int ttm_bo_wait_unreserved(struct ttm_buffer_object *bo, | |||
| 799 | * | 798 | * |
| 800 | * Returns: | 799 | * Returns: |
| 801 | * -EBUSY: If no_wait == 1 and the buffer is already reserved. | 800 | * -EBUSY: If no_wait == 1 and the buffer is already reserved. |
| 802 | * -ERESTART: If interruptible == 1 and the process received a signal | 801 | * -ERESTARTSYS: If interruptible == 1 and the process received a signal |
| 803 | * while sleeping. | 802 | * while sleeping. |
| 804 | */ | 803 | */ |
| 805 | extern int ttm_bo_block_reservation(struct ttm_buffer_object *bo, | 804 | extern int ttm_bo_block_reservation(struct ttm_buffer_object *bo, |
diff --git a/include/drm/ttm/ttm_execbuf_util.h b/include/drm/ttm/ttm_execbuf_util.h new file mode 100644 index 000000000000..cd2c475da9ea --- /dev/null +++ b/include/drm/ttm/ttm_execbuf_util.h | |||
| @@ -0,0 +1,107 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2006-2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | /* | ||
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | ||
| 29 | */ | ||
| 30 | |||
| 31 | #ifndef _TTM_EXECBUF_UTIL_H_ | ||
| 32 | #define _TTM_EXECBUF_UTIL_H_ | ||
| 33 | |||
| 34 | #include "ttm/ttm_bo_api.h" | ||
| 35 | #include <linux/list.h> | ||
| 36 | |||
| 37 | /** | ||
| 38 | * struct ttm_validate_buffer | ||
| 39 | * | ||
| 40 | * @head: list head for thread-private list. | ||
| 41 | * @bo: refcounted buffer object pointer. | ||
| 42 | * @new_sync_obj_arg: New sync_obj_arg for @bo, to be used once | ||
| 43 | * adding a new sync object. | ||
| 44 | * @reservied: Indicates whether @bo has been reserved for validation. | ||
| 45 | */ | ||
| 46 | |||
| 47 | struct ttm_validate_buffer { | ||
| 48 | struct list_head head; | ||
| 49 | struct ttm_buffer_object *bo; | ||
| 50 | void *new_sync_obj_arg; | ||
| 51 | bool reserved; | ||
| 52 | }; | ||
| 53 | |||
| 54 | /** | ||
| 55 | * function ttm_eu_backoff_reservation | ||
| 56 | * | ||
| 57 | * @list: thread private list of ttm_validate_buffer structs. | ||
| 58 | * | ||
| 59 | * Undoes all buffer validation reservations for bos pointed to by | ||
| 60 | * the list entries. | ||
| 61 | */ | ||
| 62 | |||
| 63 | extern void ttm_eu_backoff_reservation(struct list_head *list); | ||
| 64 | |||
| 65 | /** | ||
| 66 | * function ttm_eu_reserve_buffers | ||
| 67 | * | ||
| 68 | * @list: thread private list of ttm_validate_buffer structs. | ||
| 69 | * @val_seq: A unique sequence number. | ||
| 70 | * | ||
| 71 | * Tries to reserve bos pointed to by the list entries for validation. | ||
| 72 | * If the function returns 0, all buffers are marked as "unfenced", | ||
| 73 | * taken off the lru lists and are not synced for write CPU usage. | ||
| 74 | * | ||
| 75 | * If the function detects a deadlock due to multiple threads trying to | ||
| 76 | * reserve the same buffers in reverse order, all threads except one will | ||
| 77 | * back off and retry. This function may sleep while waiting for | ||
| 78 | * CPU write reservations to be cleared, and for other threads to | ||
| 79 | * unreserve their buffers. | ||
| 80 | * | ||
| 81 | * This function may return -ERESTART or -EAGAIN if the calling process | ||
| 82 | * receives a signal while waiting. In that case, no buffers on the list | ||
| 83 | * will be reserved upon return. | ||
| 84 | * | ||
| 85 | * Buffers reserved by this function should be unreserved by | ||
| 86 | * a call to either ttm_eu_backoff_reservation() or | ||
| 87 | * ttm_eu_fence_buffer_objects() when command submission is complete or | ||
| 88 | * has failed. | ||
| 89 | */ | ||
| 90 | |||
| 91 | extern int ttm_eu_reserve_buffers(struct list_head *list, uint32_t val_seq); | ||
| 92 | |||
| 93 | /** | ||
| 94 | * function ttm_eu_fence_buffer_objects. | ||
| 95 | * | ||
| 96 | * @list: thread private list of ttm_validate_buffer structs. | ||
| 97 | * @sync_obj: The new sync object for the buffers. | ||
| 98 | * | ||
| 99 | * This function should be called when command submission is complete, and | ||
| 100 | * it will add a new sync object to bos pointed to by entries on @list. | ||
| 101 | * It also unreserves all buffers, putting them on lru lists. | ||
| 102 | * | ||
| 103 | */ | ||
| 104 | |||
| 105 | extern void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj); | ||
| 106 | |||
| 107 | #endif | ||
diff --git a/include/drm/ttm/ttm_lock.h b/include/drm/ttm/ttm_lock.h new file mode 100644 index 000000000000..81ba0b0b891a --- /dev/null +++ b/include/drm/ttm/ttm_lock.h | |||
| @@ -0,0 +1,247 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2007-2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | /* | ||
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | ||
| 29 | */ | ||
| 30 | |||
| 31 | /** @file ttm_lock.h | ||
| 32 | * This file implements a simple replacement for the buffer manager use | ||
| 33 | * of the DRM heavyweight hardware lock. | ||
| 34 | * The lock is a read-write lock. Taking it in read mode and write mode | ||
| 35 | * is relatively fast, and intended for in-kernel use only. | ||
| 36 | * | ||
| 37 | * The vt mode is used only when there is a need to block all | ||
| 38 | * user-space processes from validating buffers. | ||
| 39 | * It's allowed to leave kernel space with the vt lock held. | ||
| 40 | * If a user-space process dies while having the vt-lock, | ||
| 41 | * it will be released during the file descriptor release. The vt lock | ||
| 42 | * excludes write lock and read lock. | ||
| 43 | * | ||
| 44 | * The suspend mode is used to lock out all TTM users when preparing for | ||
| 45 | * and executing suspend operations. | ||
| 46 | * | ||
| 47 | */ | ||
| 48 | |||
| 49 | #ifndef _TTM_LOCK_H_ | ||
| 50 | #define _TTM_LOCK_H_ | ||
| 51 | |||
| 52 | #include "ttm/ttm_object.h" | ||
| 53 | #include <linux/wait.h> | ||
| 54 | #include <asm/atomic.h> | ||
| 55 | |||
| 56 | /** | ||
| 57 | * struct ttm_lock | ||
| 58 | * | ||
| 59 | * @base: ttm base object used solely to release the lock if the client | ||
| 60 | * holding the lock dies. | ||
| 61 | * @queue: Queue for processes waiting for lock change-of-status. | ||
| 62 | * @lock: Spinlock protecting some lock members. | ||
| 63 | * @rw: Read-write lock counter. Protected by @lock. | ||
| 64 | * @flags: Lock state. Protected by @lock. | ||
| 65 | * @kill_takers: Boolean whether to kill takers of the lock. | ||
| 66 | * @signal: Signal to send when kill_takers is true. | ||
| 67 | */ | ||
| 68 | |||
| 69 | struct ttm_lock { | ||
| 70 | struct ttm_base_object base; | ||
| 71 | wait_queue_head_t queue; | ||
| 72 | spinlock_t lock; | ||
| 73 | int32_t rw; | ||
| 74 | uint32_t flags; | ||
| 75 | bool kill_takers; | ||
| 76 | int signal; | ||
| 77 | struct ttm_object_file *vt_holder; | ||
| 78 | }; | ||
| 79 | |||
| 80 | |||
| 81 | /** | ||
| 82 | * ttm_lock_init | ||
| 83 | * | ||
| 84 | * @lock: Pointer to a struct ttm_lock | ||
| 85 | * Initializes the lock. | ||
| 86 | */ | ||
| 87 | extern void ttm_lock_init(struct ttm_lock *lock); | ||
| 88 | |||
| 89 | /** | ||
| 90 | * ttm_read_unlock | ||
| 91 | * | ||
| 92 | * @lock: Pointer to a struct ttm_lock | ||
| 93 | * | ||
| 94 | * Releases a read lock. | ||
| 95 | */ | ||
| 96 | extern void ttm_read_unlock(struct ttm_lock *lock); | ||
| 97 | |||
| 98 | /** | ||
| 99 | * ttm_read_lock | ||
| 100 | * | ||
| 101 | * @lock: Pointer to a struct ttm_lock | ||
| 102 | * @interruptible: Interruptible sleeping while waiting for a lock. | ||
| 103 | * | ||
| 104 | * Takes the lock in read mode. | ||
| 105 | * Returns: | ||
| 106 | * -ERESTARTSYS If interrupted by a signal and interruptible is true. | ||
| 107 | */ | ||
| 108 | extern int ttm_read_lock(struct ttm_lock *lock, bool interruptible); | ||
| 109 | |||
| 110 | /** | ||
| 111 | * ttm_read_trylock | ||
| 112 | * | ||
| 113 | * @lock: Pointer to a struct ttm_lock | ||
| 114 | * @interruptible: Interruptible sleeping while waiting for a lock. | ||
| 115 | * | ||
| 116 | * Tries to take the lock in read mode. If the lock is already held | ||
| 117 | * in write mode, the function will return -EBUSY. If the lock is held | ||
| 118 | * in vt or suspend mode, the function will sleep until these modes | ||
| 119 | * are unlocked. | ||
| 120 | * | ||
| 121 | * Returns: | ||
| 122 | * -EBUSY The lock was already held in write mode. | ||
| 123 | * -ERESTARTSYS If interrupted by a signal and interruptible is true. | ||
| 124 | */ | ||
| 125 | extern int ttm_read_trylock(struct ttm_lock *lock, bool interruptible); | ||
| 126 | |||
| 127 | /** | ||
| 128 | * ttm_write_unlock | ||
| 129 | * | ||
| 130 | * @lock: Pointer to a struct ttm_lock | ||
| 131 | * | ||
| 132 | * Releases a write lock. | ||
| 133 | */ | ||
| 134 | extern void ttm_write_unlock(struct ttm_lock *lock); | ||
| 135 | |||
| 136 | /** | ||
| 137 | * ttm_write_lock | ||
| 138 | * | ||
| 139 | * @lock: Pointer to a struct ttm_lock | ||
| 140 | * @interruptible: Interruptible sleeping while waiting for a lock. | ||
| 141 | * | ||
| 142 | * Takes the lock in write mode. | ||
| 143 | * Returns: | ||
| 144 | * -ERESTARTSYS If interrupted by a signal and interruptible is true. | ||
| 145 | */ | ||
| 146 | extern int ttm_write_lock(struct ttm_lock *lock, bool interruptible); | ||
| 147 | |||
| 148 | /** | ||
| 149 | * ttm_lock_downgrade | ||
| 150 | * | ||
| 151 | * @lock: Pointer to a struct ttm_lock | ||
| 152 | * | ||
| 153 | * Downgrades a write lock to a read lock. | ||
| 154 | */ | ||
| 155 | extern void ttm_lock_downgrade(struct ttm_lock *lock); | ||
| 156 | |||
| 157 | /** | ||
| 158 | * ttm_suspend_lock | ||
| 159 | * | ||
| 160 | * @lock: Pointer to a struct ttm_lock | ||
| 161 | * | ||
| 162 | * Takes the lock in suspend mode. Excludes read and write mode. | ||
| 163 | */ | ||
| 164 | extern void ttm_suspend_lock(struct ttm_lock *lock); | ||
| 165 | |||
| 166 | /** | ||
| 167 | * ttm_suspend_unlock | ||
| 168 | * | ||
| 169 | * @lock: Pointer to a struct ttm_lock | ||
| 170 | * | ||
| 171 | * Releases a suspend lock | ||
| 172 | */ | ||
| 173 | extern void ttm_suspend_unlock(struct ttm_lock *lock); | ||
| 174 | |||
| 175 | /** | ||
| 176 | * ttm_vt_lock | ||
| 177 | * | ||
| 178 | * @lock: Pointer to a struct ttm_lock | ||
| 179 | * @interruptible: Interruptible sleeping while waiting for a lock. | ||
| 180 | * @tfile: Pointer to a struct ttm_object_file to register the lock with. | ||
| 181 | * | ||
| 182 | * Takes the lock in vt mode. | ||
| 183 | * Returns: | ||
| 184 | * -ERESTARTSYS If interrupted by a signal and interruptible is true. | ||
| 185 | * -ENOMEM: Out of memory when locking. | ||
| 186 | */ | ||
| 187 | extern int ttm_vt_lock(struct ttm_lock *lock, bool interruptible, | ||
| 188 | struct ttm_object_file *tfile); | ||
| 189 | |||
| 190 | /** | ||
| 191 | * ttm_vt_unlock | ||
| 192 | * | ||
| 193 | * @lock: Pointer to a struct ttm_lock | ||
| 194 | * | ||
| 195 | * Releases a vt lock. | ||
| 196 | * Returns: | ||
| 197 | * -EINVAL If the lock was not held. | ||
| 198 | */ | ||
| 199 | extern int ttm_vt_unlock(struct ttm_lock *lock); | ||
| 200 | |||
| 201 | /** | ||
| 202 | * ttm_write_unlock | ||
| 203 | * | ||
| 204 | * @lock: Pointer to a struct ttm_lock | ||
| 205 | * | ||
| 206 | * Releases a write lock. | ||
| 207 | */ | ||
| 208 | extern void ttm_write_unlock(struct ttm_lock *lock); | ||
| 209 | |||
| 210 | /** | ||
| 211 | * ttm_write_lock | ||
| 212 | * | ||
| 213 | * @lock: Pointer to a struct ttm_lock | ||
| 214 | * @interruptible: Interruptible sleeping while waiting for a lock. | ||
| 215 | * | ||
| 216 | * Takes the lock in write mode. | ||
| 217 | * Returns: | ||
| 218 | * -ERESTARTSYS If interrupted by a signal and interruptible is true. | ||
| 219 | */ | ||
| 220 | extern int ttm_write_lock(struct ttm_lock *lock, bool interruptible); | ||
| 221 | |||
| 222 | /** | ||
| 223 | * ttm_lock_set_kill | ||
| 224 | * | ||
| 225 | * @lock: Pointer to a struct ttm_lock | ||
| 226 | * @val: Boolean whether to kill processes taking the lock. | ||
| 227 | * @signal: Signal to send to the process taking the lock. | ||
| 228 | * | ||
| 229 | * The kill-when-taking-lock functionality is used to kill processes that keep | ||
| 230 | * on using the TTM functionality when its resources has been taken down, for | ||
| 231 | * example when the X server exits. A typical sequence would look like this: | ||
| 232 | * - X server takes lock in write mode. | ||
| 233 | * - ttm_lock_set_kill() is called with @val set to true. | ||
| 234 | * - As part of X server exit, TTM resources are taken down. | ||
| 235 | * - X server releases the lock on file release. | ||
| 236 | * - Another dri client wants to render, takes the lock and is killed. | ||
| 237 | * | ||
| 238 | */ | ||
| 239 | static inline void ttm_lock_set_kill(struct ttm_lock *lock, bool val, | ||
| 240 | int signal) | ||
| 241 | { | ||
| 242 | lock->kill_takers = val; | ||
| 243 | if (val) | ||
| 244 | lock->signal = signal; | ||
| 245 | } | ||
| 246 | |||
| 247 | #endif | ||
diff --git a/include/drm/ttm/ttm_memory.h b/include/drm/ttm/ttm_memory.h index 6983a7cf4da4..b199170b3c2c 100644 --- a/include/drm/ttm/ttm_memory.h +++ b/include/drm/ttm/ttm_memory.h | |||
| @@ -33,6 +33,7 @@ | |||
| 33 | #include <linux/wait.h> | 33 | #include <linux/wait.h> |
| 34 | #include <linux/errno.h> | 34 | #include <linux/errno.h> |
| 35 | #include <linux/kobject.h> | 35 | #include <linux/kobject.h> |
| 36 | #include <linux/mm.h> | ||
| 36 | 37 | ||
| 37 | /** | 38 | /** |
| 38 | * struct ttm_mem_shrink - callback to shrink TTM memory usage. | 39 | * struct ttm_mem_shrink - callback to shrink TTM memory usage. |
diff --git a/include/drm/ttm/ttm_object.h b/include/drm/ttm/ttm_object.h new file mode 100644 index 000000000000..703ca4db0a29 --- /dev/null +++ b/include/drm/ttm/ttm_object.h | |||
| @@ -0,0 +1,267 @@ | |||
| 1 | /************************************************************************** | ||
| 2 | * | ||
| 3 | * Copyright (c) 2006-2009 VMware, Inc., Palo Alto, CA., USA | ||
| 4 | * All Rights Reserved. | ||
| 5 | * | ||
| 6 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
| 7 | * copy of this software and associated documentation files (the | ||
| 8 | * "Software"), to deal in the Software without restriction, including | ||
| 9 | * without limitation the rights to use, copy, modify, merge, publish, | ||
| 10 | * distribute, sub license, and/or sell copies of the Software, and to | ||
| 11 | * permit persons to whom the Software is furnished to do so, subject to | ||
| 12 | * the following conditions: | ||
| 13 | * | ||
| 14 | * The above copyright notice and this permission notice (including the | ||
| 15 | * next paragraph) shall be included in all copies or substantial portions | ||
| 16 | * of the Software. | ||
| 17 | * | ||
| 18 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 19 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 20 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
| 21 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
| 22 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
| 23 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
| 24 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
| 25 | * | ||
| 26 | **************************************************************************/ | ||
| 27 | /* | ||
| 28 | * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com> | ||
| 29 | */ | ||
| 30 | /** @file ttm_object.h | ||
| 31 | * | ||
| 32 | * Base- and reference object implementation for the various | ||
| 33 | * ttm objects. Implements reference counting, minimal security checks | ||
| 34 | * and release on file close. | ||
| 35 | */ | ||
| 36 | |||
| 37 | #ifndef _TTM_OBJECT_H_ | ||
| 38 | #define _TTM_OBJECT_H_ | ||
| 39 | |||
| 40 | #include <linux/list.h> | ||
| 41 | #include "drm_hashtab.h" | ||
| 42 | #include <linux/kref.h> | ||
| 43 | #include <ttm/ttm_memory.h> | ||
| 44 | |||
| 45 | /** | ||
| 46 | * enum ttm_ref_type | ||
| 47 | * | ||
| 48 | * Describes what type of reference a ref object holds. | ||
| 49 | * | ||
| 50 | * TTM_REF_USAGE is a simple refcount on a base object. | ||
| 51 | * | ||
| 52 | * TTM_REF_SYNCCPU_READ is a SYNCCPU_READ reference on a | ||
| 53 | * buffer object. | ||
| 54 | * | ||
| 55 | * TTM_REF_SYNCCPU_WRITE is a SYNCCPU_WRITE reference on a | ||
| 56 | * buffer object. | ||
| 57 | * | ||
| 58 | */ | ||
| 59 | |||
| 60 | enum ttm_ref_type { | ||
| 61 | TTM_REF_USAGE, | ||
| 62 | TTM_REF_SYNCCPU_READ, | ||
| 63 | TTM_REF_SYNCCPU_WRITE, | ||
| 64 | TTM_REF_NUM | ||
| 65 | }; | ||
| 66 | |||
| 67 | /** | ||
| 68 | * enum ttm_object_type | ||
| 69 | * | ||
| 70 | * One entry per ttm object type. | ||
| 71 | * Device-specific types should use the | ||
| 72 | * ttm_driver_typex types. | ||
| 73 | */ | ||
| 74 | |||
| 75 | enum ttm_object_type { | ||
| 76 | ttm_fence_type, | ||
| 77 | ttm_buffer_type, | ||
| 78 | ttm_lock_type, | ||
| 79 | ttm_driver_type0 = 256, | ||
| 80 | ttm_driver_type1 | ||
| 81 | }; | ||
| 82 | |||
| 83 | struct ttm_object_file; | ||
| 84 | struct ttm_object_device; | ||
| 85 | |||
| 86 | /** | ||
| 87 | * struct ttm_base_object | ||
| 88 | * | ||
| 89 | * @hash: hash entry for the per-device object hash. | ||
| 90 | * @type: derived type this object is base class for. | ||
| 91 | * @shareable: Other ttm_object_files can access this object. | ||
| 92 | * | ||
| 93 | * @tfile: Pointer to ttm_object_file of the creator. | ||
| 94 | * NULL if the object was not created by a user request. | ||
| 95 | * (kernel object). | ||
| 96 | * | ||
| 97 | * @refcount: Number of references to this object, not | ||
| 98 | * including the hash entry. A reference to a base object can | ||
| 99 | * only be held by a ref object. | ||
| 100 | * | ||
| 101 | * @refcount_release: A function to be called when there are | ||
| 102 | * no more references to this object. This function should | ||
| 103 | * destroy the object (or make sure destruction eventually happens), | ||
| 104 | * and when it is called, the object has | ||
| 105 | * already been taken out of the per-device hash. The parameter | ||
| 106 | * "base" should be set to NULL by the function. | ||
| 107 | * | ||
| 108 | * @ref_obj_release: A function to be called when a reference object | ||
| 109 | * with another ttm_ref_type than TTM_REF_USAGE is deleted. | ||
| 110 | * this function may, for example, release a lock held by a user-space | ||
| 111 | * process. | ||
| 112 | * | ||
| 113 | * This struct is intended to be used as a base struct for objects that | ||
| 114 | * are visible to user-space. It provides a global name, race-safe | ||
| 115 | * access and refcounting, minimal access contol and hooks for unref actions. | ||
| 116 | */ | ||
| 117 | |||
| 118 | struct ttm_base_object { | ||
| 119 | struct drm_hash_item hash; | ||
| 120 | enum ttm_object_type object_type; | ||
| 121 | bool shareable; | ||
| 122 | struct ttm_object_file *tfile; | ||
| 123 | struct kref refcount; | ||
| 124 | void (*refcount_release) (struct ttm_base_object **base); | ||
| 125 | void (*ref_obj_release) (struct ttm_base_object *base, | ||
| 126 | enum ttm_ref_type ref_type); | ||
| 127 | }; | ||
| 128 | |||
| 129 | /** | ||
| 130 | * ttm_base_object_init | ||
| 131 | * | ||
| 132 | * @tfile: Pointer to a struct ttm_object_file. | ||
| 133 | * @base: The struct ttm_base_object to initialize. | ||
| 134 | * @shareable: This object is shareable with other applcations. | ||
| 135 | * (different @tfile pointers.) | ||
| 136 | * @type: The object type. | ||
| 137 | * @refcount_release: See the struct ttm_base_object description. | ||
| 138 | * @ref_obj_release: See the struct ttm_base_object description. | ||
| 139 | * | ||
| 140 | * Initializes a struct ttm_base_object. | ||
| 141 | */ | ||
| 142 | |||
| 143 | extern int ttm_base_object_init(struct ttm_object_file *tfile, | ||
| 144 | struct ttm_base_object *base, | ||
| 145 | bool shareable, | ||
| 146 | enum ttm_object_type type, | ||
| 147 | void (*refcount_release) (struct ttm_base_object | ||
| 148 | **), | ||
| 149 | void (*ref_obj_release) (struct ttm_base_object | ||
| 150 | *, | ||
| 151 | enum ttm_ref_type | ||
| 152 | ref_type)); | ||
| 153 | |||
| 154 | /** | ||
| 155 | * ttm_base_object_lookup | ||
| 156 | * | ||
| 157 | * @tfile: Pointer to a struct ttm_object_file. | ||
| 158 | * @key: Hash key | ||
| 159 | * | ||
| 160 | * Looks up a struct ttm_base_object with the key @key. | ||
| 161 | * Also verifies that the object is visible to the application, by | ||
| 162 | * comparing the @tfile argument and checking the object shareable flag. | ||
| 163 | */ | ||
| 164 | |||
| 165 | extern struct ttm_base_object *ttm_base_object_lookup(struct ttm_object_file | ||
| 166 | *tfile, uint32_t key); | ||
| 167 | |||
| 168 | /** | ||
| 169 | * ttm_base_object_unref | ||
| 170 | * | ||
| 171 | * @p_base: Pointer to a pointer referncing a struct ttm_base_object. | ||
| 172 | * | ||
| 173 | * Decrements the base object refcount and clears the pointer pointed to by | ||
| 174 | * p_base. | ||
| 175 | */ | ||
| 176 | |||
| 177 | extern void ttm_base_object_unref(struct ttm_base_object **p_base); | ||
| 178 | |||
| 179 | /** | ||
| 180 | * ttm_ref_object_add. | ||
| 181 | * | ||
| 182 | * @tfile: A struct ttm_object_file representing the application owning the | ||
| 183 | * ref_object. | ||
| 184 | * @base: The base object to reference. | ||
| 185 | * @ref_type: The type of reference. | ||
| 186 | * @existed: Upon completion, indicates that an identical reference object | ||
| 187 | * already existed, and the refcount was upped on that object instead. | ||
| 188 | * | ||
| 189 | * Adding a ref object to a base object is basically like referencing the | ||
| 190 | * base object, but a user-space application holds the reference. When the | ||
| 191 | * file corresponding to @tfile is closed, all its reference objects are | ||
| 192 | * deleted. A reference object can have different types depending on what | ||
| 193 | * it's intended for. It can be refcounting to prevent object destruction, | ||
| 194 | * When user-space takes a lock, it can add a ref object to that lock to | ||
| 195 | * make sure the lock is released if the application dies. A ref object | ||
| 196 | * will hold a single reference on a base object. | ||
| 197 | */ | ||
| 198 | extern int ttm_ref_object_add(struct ttm_object_file *tfile, | ||
| 199 | struct ttm_base_object *base, | ||
| 200 | enum ttm_ref_type ref_type, bool *existed); | ||
| 201 | /** | ||
| 202 | * ttm_ref_object_base_unref | ||
| 203 | * | ||
| 204 | * @key: Key representing the base object. | ||
| 205 | * @ref_type: Ref type of the ref object to be dereferenced. | ||
| 206 | * | ||
| 207 | * Unreference a ref object with type @ref_type | ||
| 208 | * on the base object identified by @key. If there are no duplicate | ||
| 209 | * references, the ref object will be destroyed and the base object | ||
| 210 | * will be unreferenced. | ||
| 211 | */ | ||
| 212 | extern int ttm_ref_object_base_unref(struct ttm_object_file *tfile, | ||
| 213 | unsigned long key, | ||
| 214 | enum ttm_ref_type ref_type); | ||
| 215 | |||
| 216 | /** | ||
| 217 | * ttm_object_file_init - initialize a struct ttm_object file | ||
| 218 | * | ||
| 219 | * @tdev: A struct ttm_object device this file is initialized on. | ||
| 220 | * @hash_order: Order of the hash table used to hold the reference objects. | ||
| 221 | * | ||
| 222 | * This is typically called by the file_ops::open function. | ||
| 223 | */ | ||
| 224 | |||
| 225 | extern struct ttm_object_file *ttm_object_file_init(struct ttm_object_device | ||
| 226 | *tdev, | ||
| 227 | unsigned int hash_order); | ||
| 228 | |||
| 229 | /** | ||
| 230 | * ttm_object_file_release - release data held by a ttm_object_file | ||
| 231 | * | ||
| 232 | * @p_tfile: Pointer to pointer to the ttm_object_file object to release. | ||
| 233 | * *p_tfile will be set to NULL by this function. | ||
| 234 | * | ||
| 235 | * Releases all data associated by a ttm_object_file. | ||
| 236 | * Typically called from file_ops::release. The caller must | ||
| 237 | * ensure that there are no concurrent users of tfile. | ||
| 238 | */ | ||
| 239 | |||
| 240 | extern void ttm_object_file_release(struct ttm_object_file **p_tfile); | ||
| 241 | |||
| 242 | /** | ||
| 243 | * ttm_object device init - initialize a struct ttm_object_device | ||
| 244 | * | ||
| 245 | * @hash_order: Order of hash table used to hash the base objects. | ||
| 246 | * | ||
| 247 | * This function is typically called on device initialization to prepare | ||
| 248 | * data structures needed for ttm base and ref objects. | ||
| 249 | */ | ||
| 250 | |||
| 251 | extern struct ttm_object_device *ttm_object_device_init | ||
| 252 | (struct ttm_mem_global *mem_glob, unsigned int hash_order); | ||
| 253 | |||
| 254 | /** | ||
| 255 | * ttm_object_device_release - release data held by a ttm_object_device | ||
| 256 | * | ||
| 257 | * @p_tdev: Pointer to pointer to the ttm_object_device object to release. | ||
| 258 | * *p_tdev will be set to NULL by this function. | ||
| 259 | * | ||
| 260 | * Releases all data associated by a ttm_object_device. | ||
| 261 | * Typically called from driver::unload before the destruction of the | ||
| 262 | * device private data structure. | ||
| 263 | */ | ||
| 264 | |||
| 265 | extern void ttm_object_device_release(struct ttm_object_device **p_tdev); | ||
| 266 | |||
| 267 | #endif | ||
diff --git a/include/drm/via_drm.h b/include/drm/via_drm.h index 170786e5c2ff..fd11a5bd892d 100644 --- a/include/drm/via_drm.h +++ b/include/drm/via_drm.h | |||
| @@ -24,7 +24,7 @@ | |||
| 24 | #ifndef _VIA_DRM_H_ | 24 | #ifndef _VIA_DRM_H_ |
| 25 | #define _VIA_DRM_H_ | 25 | #define _VIA_DRM_H_ |
| 26 | 26 | ||
| 27 | #include <linux/types.h> | 27 | #include "drm.h" |
| 28 | 28 | ||
| 29 | /* WARNING: These defines must be the same as what the Xserver uses. | 29 | /* WARNING: These defines must be the same as what the Xserver uses. |
| 30 | * if you change them, you must change the defines in the Xserver. | 30 | * if you change them, you must change the defines in the Xserver. |
