diff options
| -rw-r--r-- | Documentation/devicetree/bindings/ata/ahci-platform.txt | 1 | ||||
| -rw-r--r-- | Documentation/devicetree/bindings/dma/arm-pl330.txt | 3 | ||||
| -rw-r--r-- | Documentation/devicetree/bindings/net/calxeda-xgmac.txt | 3 | ||||
| -rw-r--r-- | arch/arm/boot/dts/highbank.dts | 1 | ||||
| -rw-r--r-- | arch/arm/include/asm/barrier.h | 7 | ||||
| -rw-r--r-- | arch/arm/include/asm/dma-mapping.h | 1 | ||||
| -rw-r--r-- | arch/arm/include/asm/memory.h | 8 | ||||
| -rw-r--r-- | arch/arm/mach-highbank/highbank.c | 52 | ||||
| -rw-r--r-- | arch/arm/mm/dma-mapping.c | 264 | ||||
| -rw-r--r-- | arch/arm/mm/mmu.c | 11 | ||||
| -rw-r--r-- | drivers/base/dma-contiguous.c | 18 |
11 files changed, 278 insertions, 91 deletions
diff --git a/Documentation/devicetree/bindings/ata/ahci-platform.txt b/Documentation/devicetree/bindings/ata/ahci-platform.txt index 147c1f6653fe..b519f9b699c3 100644 --- a/Documentation/devicetree/bindings/ata/ahci-platform.txt +++ b/Documentation/devicetree/bindings/ata/ahci-platform.txt | |||
| @@ -12,6 +12,7 @@ Optional properties: | |||
| 12 | - calxeda,port-phys: phandle-combophy and lane assignment, which maps each | 12 | - calxeda,port-phys: phandle-combophy and lane assignment, which maps each |
| 13 | SATA port to a combophy and a lane within that | 13 | SATA port to a combophy and a lane within that |
| 14 | combophy | 14 | combophy |
| 15 | - dma-coherent : Present if dma operations are coherent | ||
| 15 | 16 | ||
| 16 | Example: | 17 | Example: |
| 17 | sata@ffe08000 { | 18 | sata@ffe08000 { |
diff --git a/Documentation/devicetree/bindings/dma/arm-pl330.txt b/Documentation/devicetree/bindings/dma/arm-pl330.txt index a4cd273b2a67..36e27d54260b 100644 --- a/Documentation/devicetree/bindings/dma/arm-pl330.txt +++ b/Documentation/devicetree/bindings/dma/arm-pl330.txt | |||
| @@ -9,6 +9,9 @@ Required properties: | |||
| 9 | region. | 9 | region. |
| 10 | - interrupts: interrupt number to the cpu. | 10 | - interrupts: interrupt number to the cpu. |
| 11 | 11 | ||
| 12 | Optional properties: | ||
| 13 | - dma-coherent : Present if dma operations are coherent | ||
| 14 | |||
| 12 | Example: | 15 | Example: |
| 13 | 16 | ||
| 14 | pdma0: pdma@12680000 { | 17 | pdma0: pdma@12680000 { |
diff --git a/Documentation/devicetree/bindings/net/calxeda-xgmac.txt b/Documentation/devicetree/bindings/net/calxeda-xgmac.txt index 411727a3f82d..c8ae996bd8f2 100644 --- a/Documentation/devicetree/bindings/net/calxeda-xgmac.txt +++ b/Documentation/devicetree/bindings/net/calxeda-xgmac.txt | |||
| @@ -6,6 +6,9 @@ Required properties: | |||
| 6 | - interrupts : Should contain 3 xgmac interrupts. The 1st is main interrupt. | 6 | - interrupts : Should contain 3 xgmac interrupts. The 1st is main interrupt. |
| 7 | The 2nd is pwr mgt interrupt. The 3rd is low power state interrupt. | 7 | The 2nd is pwr mgt interrupt. The 3rd is low power state interrupt. |
| 8 | 8 | ||
| 9 | Optional properties: | ||
| 10 | - dma-coherent : Present if dma operations are coherent | ||
| 11 | |||
| 9 | Example: | 12 | Example: |
| 10 | 13 | ||
| 11 | ethernet@fff50000 { | 14 | ethernet@fff50000 { |
diff --git a/arch/arm/boot/dts/highbank.dts b/arch/arm/boot/dts/highbank.dts index 5204cf73c2d8..0c6fc34821f9 100644 --- a/arch/arm/boot/dts/highbank.dts +++ b/arch/arm/boot/dts/highbank.dts | |||
| @@ -124,6 +124,7 @@ | |||
| 124 | calxeda,port-phys = <&combophy5 0 &combophy0 0 | 124 | calxeda,port-phys = <&combophy5 0 &combophy0 0 |
| 125 | &combophy0 1 &combophy0 2 | 125 | &combophy0 1 &combophy0 2 |
| 126 | &combophy0 3>; | 126 | &combophy0 3>; |
| 127 | dma-coherent; | ||
| 127 | }; | 128 | }; |
| 128 | 129 | ||
| 129 | sdhci@ffe0e000 { | 130 | sdhci@ffe0e000 { |
diff --git a/arch/arm/include/asm/barrier.h b/arch/arm/include/asm/barrier.h index 05112380dc53..8dcd9c702d90 100644 --- a/arch/arm/include/asm/barrier.h +++ b/arch/arm/include/asm/barrier.h | |||
| @@ -44,10 +44,9 @@ | |||
| 44 | #define rmb() dsb() | 44 | #define rmb() dsb() |
| 45 | #define wmb() mb() | 45 | #define wmb() mb() |
| 46 | #else | 46 | #else |
| 47 | #include <asm/memory.h> | 47 | #define mb() barrier() |
| 48 | #define mb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | 48 | #define rmb() barrier() |
| 49 | #define rmb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | 49 | #define wmb() barrier() |
| 50 | #define wmb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | ||
| 51 | #endif | 50 | #endif |
| 52 | 51 | ||
| 53 | #ifndef CONFIG_SMP | 52 | #ifndef CONFIG_SMP |
diff --git a/arch/arm/include/asm/dma-mapping.h b/arch/arm/include/asm/dma-mapping.h index 5c44dcb0987b..23004847bb05 100644 --- a/arch/arm/include/asm/dma-mapping.h +++ b/arch/arm/include/asm/dma-mapping.h | |||
| @@ -13,6 +13,7 @@ | |||
| 13 | 13 | ||
| 14 | #define DMA_ERROR_CODE (~0) | 14 | #define DMA_ERROR_CODE (~0) |
| 15 | extern struct dma_map_ops arm_dma_ops; | 15 | extern struct dma_map_ops arm_dma_ops; |
| 16 | extern struct dma_map_ops arm_coherent_dma_ops; | ||
| 16 | 17 | ||
| 17 | static inline struct dma_map_ops *get_dma_ops(struct device *dev) | 18 | static inline struct dma_map_ops *get_dma_ops(struct device *dev) |
| 18 | { | 19 | { |
diff --git a/arch/arm/include/asm/memory.h b/arch/arm/include/asm/memory.h index 5f6ddcc56452..73cf03aa981e 100644 --- a/arch/arm/include/asm/memory.h +++ b/arch/arm/include/asm/memory.h | |||
| @@ -275,14 +275,6 @@ static inline __deprecated void *bus_to_virt(unsigned long x) | |||
| 275 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) | 275 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) |
| 276 | #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory) | 276 | #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory) |
| 277 | 277 | ||
| 278 | /* | ||
| 279 | * Optional coherency support. Currently used only by selected | ||
| 280 | * Intel XSC3-based systems. | ||
| 281 | */ | ||
| 282 | #ifndef arch_is_coherent | ||
| 283 | #define arch_is_coherent() 0 | ||
| 284 | #endif | ||
| 285 | |||
| 286 | #endif | 278 | #endif |
| 287 | 279 | ||
| 288 | #include <asm-generic/memory_model.h> | 280 | #include <asm-generic/memory_model.h> |
diff --git a/arch/arm/mach-highbank/highbank.c b/arch/arm/mach-highbank/highbank.c index af1da34ccf9d..40e36a50304c 100644 --- a/arch/arm/mach-highbank/highbank.c +++ b/arch/arm/mach-highbank/highbank.c | |||
| @@ -15,6 +15,7 @@ | |||
| 15 | */ | 15 | */ |
| 16 | #include <linux/clk.h> | 16 | #include <linux/clk.h> |
| 17 | #include <linux/clkdev.h> | 17 | #include <linux/clkdev.h> |
| 18 | #include <linux/dma-mapping.h> | ||
| 18 | #include <linux/io.h> | 19 | #include <linux/io.h> |
| 19 | #include <linux/irq.h> | 20 | #include <linux/irq.h> |
| 20 | #include <linux/irqdomain.h> | 21 | #include <linux/irqdomain.h> |
| @@ -23,6 +24,7 @@ | |||
| 23 | #include <linux/of_platform.h> | 24 | #include <linux/of_platform.h> |
| 24 | #include <linux/of_address.h> | 25 | #include <linux/of_address.h> |
| 25 | #include <linux/smp.h> | 26 | #include <linux/smp.h> |
| 27 | #include <linux/amba/bus.h> | ||
| 26 | 28 | ||
| 27 | #include <asm/cacheflush.h> | 29 | #include <asm/cacheflush.h> |
| 28 | #include <asm/smp_plat.h> | 30 | #include <asm/smp_plat.h> |
| @@ -149,11 +151,61 @@ static void highbank_power_off(void) | |||
| 149 | cpu_do_idle(); | 151 | cpu_do_idle(); |
| 150 | } | 152 | } |
| 151 | 153 | ||
| 154 | static int highbank_platform_notifier(struct notifier_block *nb, | ||
| 155 | unsigned long event, void *__dev) | ||
| 156 | { | ||
| 157 | struct resource *res; | ||
| 158 | int reg = -1; | ||
| 159 | struct device *dev = __dev; | ||
| 160 | |||
| 161 | if (event != BUS_NOTIFY_ADD_DEVICE) | ||
| 162 | return NOTIFY_DONE; | ||
| 163 | |||
| 164 | if (of_device_is_compatible(dev->of_node, "calxeda,hb-ahci")) | ||
| 165 | reg = 0xc; | ||
| 166 | else if (of_device_is_compatible(dev->of_node, "calxeda,hb-sdhci")) | ||
| 167 | reg = 0x18; | ||
| 168 | else if (of_device_is_compatible(dev->of_node, "arm,pl330")) | ||
| 169 | reg = 0x20; | ||
| 170 | else if (of_device_is_compatible(dev->of_node, "calxeda,hb-xgmac")) { | ||
| 171 | res = platform_get_resource(to_platform_device(dev), | ||
| 172 | IORESOURCE_MEM, 0); | ||
| 173 | if (res) { | ||
| 174 | if (res->start == 0xfff50000) | ||
| 175 | reg = 0; | ||
| 176 | else if (res->start == 0xfff51000) | ||
| 177 | reg = 4; | ||
| 178 | } | ||
| 179 | } | ||
| 180 | |||
| 181 | if (reg < 0) | ||
| 182 | return NOTIFY_DONE; | ||
| 183 | |||
| 184 | if (of_property_read_bool(dev->of_node, "dma-coherent")) { | ||
| 185 | writel(0xff31, sregs_base + reg); | ||
| 186 | set_dma_ops(dev, &arm_coherent_dma_ops); | ||
| 187 | } else | ||
| 188 | writel(0, sregs_base + reg); | ||
| 189 | |||
| 190 | return NOTIFY_OK; | ||
| 191 | } | ||
| 192 | |||
| 193 | static struct notifier_block highbank_amba_nb = { | ||
| 194 | .notifier_call = highbank_platform_notifier, | ||
| 195 | }; | ||
| 196 | |||
| 197 | static struct notifier_block highbank_platform_nb = { | ||
| 198 | .notifier_call = highbank_platform_notifier, | ||
| 199 | }; | ||
| 200 | |||
| 152 | static void __init highbank_init(void) | 201 | static void __init highbank_init(void) |
| 153 | { | 202 | { |
| 154 | pm_power_off = highbank_power_off; | 203 | pm_power_off = highbank_power_off; |
| 155 | highbank_pm_init(); | 204 | highbank_pm_init(); |
| 156 | 205 | ||
| 206 | bus_register_notifier(&platform_bus_type, &highbank_platform_nb); | ||
| 207 | bus_register_notifier(&amba_bustype, &highbank_amba_nb); | ||
| 208 | |||
| 157 | of_platform_populate(NULL, of_default_bus_match_table, NULL, NULL); | 209 | of_platform_populate(NULL, of_default_bus_match_table, NULL, NULL); |
| 158 | } | 210 | } |
| 159 | 211 | ||
diff --git a/arch/arm/mm/dma-mapping.c b/arch/arm/mm/dma-mapping.c index 13f555d62491..477a2d23ddf1 100644 --- a/arch/arm/mm/dma-mapping.c +++ b/arch/arm/mm/dma-mapping.c | |||
| @@ -73,11 +73,18 @@ static dma_addr_t arm_dma_map_page(struct device *dev, struct page *page, | |||
| 73 | unsigned long offset, size_t size, enum dma_data_direction dir, | 73 | unsigned long offset, size_t size, enum dma_data_direction dir, |
| 74 | struct dma_attrs *attrs) | 74 | struct dma_attrs *attrs) |
| 75 | { | 75 | { |
| 76 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 76 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
| 77 | __dma_page_cpu_to_dev(page, offset, size, dir); | 77 | __dma_page_cpu_to_dev(page, offset, size, dir); |
| 78 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; | 78 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; |
| 79 | } | 79 | } |
| 80 | 80 | ||
| 81 | static dma_addr_t arm_coherent_dma_map_page(struct device *dev, struct page *page, | ||
| 82 | unsigned long offset, size_t size, enum dma_data_direction dir, | ||
| 83 | struct dma_attrs *attrs) | ||
| 84 | { | ||
| 85 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; | ||
| 86 | } | ||
| 87 | |||
| 81 | /** | 88 | /** |
| 82 | * arm_dma_unmap_page - unmap a buffer previously mapped through dma_map_page() | 89 | * arm_dma_unmap_page - unmap a buffer previously mapped through dma_map_page() |
| 83 | * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices | 90 | * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices |
| @@ -96,7 +103,7 @@ static void arm_dma_unmap_page(struct device *dev, dma_addr_t handle, | |||
| 96 | size_t size, enum dma_data_direction dir, | 103 | size_t size, enum dma_data_direction dir, |
| 97 | struct dma_attrs *attrs) | 104 | struct dma_attrs *attrs) |
| 98 | { | 105 | { |
| 99 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 106 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
| 100 | __dma_page_dev_to_cpu(pfn_to_page(dma_to_pfn(dev, handle)), | 107 | __dma_page_dev_to_cpu(pfn_to_page(dma_to_pfn(dev, handle)), |
| 101 | handle & ~PAGE_MASK, size, dir); | 108 | handle & ~PAGE_MASK, size, dir); |
| 102 | } | 109 | } |
| @@ -106,8 +113,7 @@ static void arm_dma_sync_single_for_cpu(struct device *dev, | |||
| 106 | { | 113 | { |
| 107 | unsigned int offset = handle & (PAGE_SIZE - 1); | 114 | unsigned int offset = handle & (PAGE_SIZE - 1); |
| 108 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); | 115 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); |
| 109 | if (!arch_is_coherent()) | 116 | __dma_page_dev_to_cpu(page, offset, size, dir); |
| 110 | __dma_page_dev_to_cpu(page, offset, size, dir); | ||
| 111 | } | 117 | } |
| 112 | 118 | ||
| 113 | static void arm_dma_sync_single_for_device(struct device *dev, | 119 | static void arm_dma_sync_single_for_device(struct device *dev, |
| @@ -115,8 +121,7 @@ static void arm_dma_sync_single_for_device(struct device *dev, | |||
| 115 | { | 121 | { |
| 116 | unsigned int offset = handle & (PAGE_SIZE - 1); | 122 | unsigned int offset = handle & (PAGE_SIZE - 1); |
| 117 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); | 123 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); |
| 118 | if (!arch_is_coherent()) | 124 | __dma_page_cpu_to_dev(page, offset, size, dir); |
| 119 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
| 120 | } | 125 | } |
| 121 | 126 | ||
| 122 | static int arm_dma_set_mask(struct device *dev, u64 dma_mask); | 127 | static int arm_dma_set_mask(struct device *dev, u64 dma_mask); |
| @@ -138,6 +143,22 @@ struct dma_map_ops arm_dma_ops = { | |||
| 138 | }; | 143 | }; |
| 139 | EXPORT_SYMBOL(arm_dma_ops); | 144 | EXPORT_SYMBOL(arm_dma_ops); |
| 140 | 145 | ||
| 146 | static void *arm_coherent_dma_alloc(struct device *dev, size_t size, | ||
| 147 | dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs); | ||
| 148 | static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
| 149 | dma_addr_t handle, struct dma_attrs *attrs); | ||
| 150 | |||
| 151 | struct dma_map_ops arm_coherent_dma_ops = { | ||
| 152 | .alloc = arm_coherent_dma_alloc, | ||
| 153 | .free = arm_coherent_dma_free, | ||
| 154 | .mmap = arm_dma_mmap, | ||
| 155 | .get_sgtable = arm_dma_get_sgtable, | ||
| 156 | .map_page = arm_coherent_dma_map_page, | ||
| 157 | .map_sg = arm_dma_map_sg, | ||
| 158 | .set_dma_mask = arm_dma_set_mask, | ||
| 159 | }; | ||
| 160 | EXPORT_SYMBOL(arm_coherent_dma_ops); | ||
| 161 | |||
| 141 | static u64 get_coherent_dma_mask(struct device *dev) | 162 | static u64 get_coherent_dma_mask(struct device *dev) |
| 142 | { | 163 | { |
| 143 | u64 mask = (u64)arm_dma_limit; | 164 | u64 mask = (u64)arm_dma_limit; |
| @@ -586,7 +607,7 @@ static void *__alloc_simple_buffer(struct device *dev, size_t size, gfp_t gfp, | |||
| 586 | 607 | ||
| 587 | 608 | ||
| 588 | static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | 609 | static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, |
| 589 | gfp_t gfp, pgprot_t prot, const void *caller) | 610 | gfp_t gfp, pgprot_t prot, bool is_coherent, const void *caller) |
| 590 | { | 611 | { |
| 591 | u64 mask = get_coherent_dma_mask(dev); | 612 | u64 mask = get_coherent_dma_mask(dev); |
| 592 | struct page *page; | 613 | struct page *page; |
| @@ -619,7 +640,7 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | |||
| 619 | *handle = DMA_ERROR_CODE; | 640 | *handle = DMA_ERROR_CODE; |
| 620 | size = PAGE_ALIGN(size); | 641 | size = PAGE_ALIGN(size); |
| 621 | 642 | ||
| 622 | if (arch_is_coherent() || nommu()) | 643 | if (is_coherent || nommu()) |
| 623 | addr = __alloc_simple_buffer(dev, size, gfp, &page); | 644 | addr = __alloc_simple_buffer(dev, size, gfp, &page); |
| 624 | else if (gfp & GFP_ATOMIC) | 645 | else if (gfp & GFP_ATOMIC) |
| 625 | addr = __alloc_from_pool(size, &page); | 646 | addr = __alloc_from_pool(size, &page); |
| @@ -647,7 +668,20 @@ void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | |||
| 647 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) | 668 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) |
| 648 | return memory; | 669 | return memory; |
| 649 | 670 | ||
| 650 | return __dma_alloc(dev, size, handle, gfp, prot, | 671 | return __dma_alloc(dev, size, handle, gfp, prot, false, |
| 672 | __builtin_return_address(0)); | ||
| 673 | } | ||
| 674 | |||
| 675 | static void *arm_coherent_dma_alloc(struct device *dev, size_t size, | ||
| 676 | dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs) | ||
| 677 | { | ||
| 678 | pgprot_t prot = __get_dma_pgprot(attrs, pgprot_kernel); | ||
| 679 | void *memory; | ||
| 680 | |||
| 681 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) | ||
| 682 | return memory; | ||
| 683 | |||
| 684 | return __dma_alloc(dev, size, handle, gfp, prot, true, | ||
| 651 | __builtin_return_address(0)); | 685 | __builtin_return_address(0)); |
| 652 | } | 686 | } |
| 653 | 687 | ||
| @@ -684,8 +718,9 @@ int arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, | |||
| 684 | /* | 718 | /* |
| 685 | * Free a buffer as defined by the above mapping. | 719 | * Free a buffer as defined by the above mapping. |
| 686 | */ | 720 | */ |
| 687 | void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | 721 | static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, |
| 688 | dma_addr_t handle, struct dma_attrs *attrs) | 722 | dma_addr_t handle, struct dma_attrs *attrs, |
| 723 | bool is_coherent) | ||
| 689 | { | 724 | { |
| 690 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); | 725 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); |
| 691 | 726 | ||
| @@ -694,7 +729,7 @@ void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | |||
| 694 | 729 | ||
| 695 | size = PAGE_ALIGN(size); | 730 | size = PAGE_ALIGN(size); |
| 696 | 731 | ||
| 697 | if (arch_is_coherent() || nommu()) { | 732 | if (is_coherent || nommu()) { |
| 698 | __dma_free_buffer(page, size); | 733 | __dma_free_buffer(page, size); |
| 699 | } else if (__free_from_pool(cpu_addr, size)) { | 734 | } else if (__free_from_pool(cpu_addr, size)) { |
| 700 | return; | 735 | return; |
| @@ -710,6 +745,18 @@ void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | |||
| 710 | } | 745 | } |
| 711 | } | 746 | } |
| 712 | 747 | ||
| 748 | void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
| 749 | dma_addr_t handle, struct dma_attrs *attrs) | ||
| 750 | { | ||
| 751 | __arm_dma_free(dev, size, cpu_addr, handle, attrs, false); | ||
| 752 | } | ||
| 753 | |||
| 754 | static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
| 755 | dma_addr_t handle, struct dma_attrs *attrs) | ||
| 756 | { | ||
| 757 | __arm_dma_free(dev, size, cpu_addr, handle, attrs, true); | ||
| 758 | } | ||
| 759 | |||
| 713 | int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, | 760 | int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, |
| 714 | void *cpu_addr, dma_addr_t handle, size_t size, | 761 | void *cpu_addr, dma_addr_t handle, size_t size, |
| 715 | struct dma_attrs *attrs) | 762 | struct dma_attrs *attrs) |
| @@ -1012,11 +1059,12 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size, gfp_t | |||
| 1012 | if (!pages[i]) | 1059 | if (!pages[i]) |
| 1013 | goto error; | 1060 | goto error; |
| 1014 | 1061 | ||
| 1015 | if (order) | 1062 | if (order) { |
| 1016 | split_page(pages[i], order); | 1063 | split_page(pages[i], order); |
| 1017 | j = 1 << order; | 1064 | j = 1 << order; |
| 1018 | while (--j) | 1065 | while (--j) |
| 1019 | pages[i + j] = pages[i] + j; | 1066 | pages[i + j] = pages[i] + j; |
| 1067 | } | ||
| 1020 | 1068 | ||
| 1021 | __dma_clear_buffer(pages[i], PAGE_SIZE << order); | 1069 | __dma_clear_buffer(pages[i], PAGE_SIZE << order); |
| 1022 | i += 1 << order; | 1070 | i += 1 << order; |
| @@ -1303,7 +1351,8 @@ static int arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, | |||
| 1303 | */ | 1351 | */ |
| 1304 | static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, | 1352 | static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, |
| 1305 | size_t size, dma_addr_t *handle, | 1353 | size_t size, dma_addr_t *handle, |
| 1306 | enum dma_data_direction dir, struct dma_attrs *attrs) | 1354 | enum dma_data_direction dir, struct dma_attrs *attrs, |
| 1355 | bool is_coherent) | ||
| 1307 | { | 1356 | { |
| 1308 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; | 1357 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; |
| 1309 | dma_addr_t iova, iova_base; | 1358 | dma_addr_t iova, iova_base; |
| @@ -1322,8 +1371,8 @@ static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, | |||
| 1322 | phys_addr_t phys = page_to_phys(sg_page(s)); | 1371 | phys_addr_t phys = page_to_phys(sg_page(s)); |
| 1323 | unsigned int len = PAGE_ALIGN(s->offset + s->length); | 1372 | unsigned int len = PAGE_ALIGN(s->offset + s->length); |
| 1324 | 1373 | ||
| 1325 | if (!arch_is_coherent() && | 1374 | if (!is_coherent && |
| 1326 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1375 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
| 1327 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); | 1376 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); |
| 1328 | 1377 | ||
| 1329 | ret = iommu_map(mapping->domain, iova, phys, len, 0); | 1378 | ret = iommu_map(mapping->domain, iova, phys, len, 0); |
| @@ -1341,20 +1390,9 @@ fail: | |||
| 1341 | return ret; | 1390 | return ret; |
| 1342 | } | 1391 | } |
| 1343 | 1392 | ||
| 1344 | /** | 1393 | static int __iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, |
| 1345 | * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA | 1394 | enum dma_data_direction dir, struct dma_attrs *attrs, |
| 1346 | * @dev: valid struct device pointer | 1395 | bool is_coherent) |
| 1347 | * @sg: list of buffers | ||
| 1348 | * @nents: number of buffers to map | ||
| 1349 | * @dir: DMA transfer direction | ||
| 1350 | * | ||
| 1351 | * Map a set of buffers described by scatterlist in streaming mode for DMA. | ||
| 1352 | * The scatter gather list elements are merged together (if possible) and | ||
| 1353 | * tagged with the appropriate dma address and length. They are obtained via | ||
| 1354 | * sg_dma_{address,length}. | ||
| 1355 | */ | ||
| 1356 | int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | ||
| 1357 | enum dma_data_direction dir, struct dma_attrs *attrs) | ||
| 1358 | { | 1396 | { |
| 1359 | struct scatterlist *s = sg, *dma = sg, *start = sg; | 1397 | struct scatterlist *s = sg, *dma = sg, *start = sg; |
| 1360 | int i, count = 0; | 1398 | int i, count = 0; |
| @@ -1370,7 +1408,7 @@ int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
| 1370 | 1408 | ||
| 1371 | if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { | 1409 | if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { |
| 1372 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, | 1410 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, |
| 1373 | dir, attrs) < 0) | 1411 | dir, attrs, is_coherent) < 0) |
| 1374 | goto bad_mapping; | 1412 | goto bad_mapping; |
| 1375 | 1413 | ||
| 1376 | dma->dma_address += offset; | 1414 | dma->dma_address += offset; |
| @@ -1383,7 +1421,8 @@ int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
| 1383 | } | 1421 | } |
| 1384 | size += s->length; | 1422 | size += s->length; |
| 1385 | } | 1423 | } |
| 1386 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs) < 0) | 1424 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs, |
| 1425 | is_coherent) < 0) | ||
| 1387 | goto bad_mapping; | 1426 | goto bad_mapping; |
| 1388 | 1427 | ||
| 1389 | dma->dma_address += offset; | 1428 | dma->dma_address += offset; |
| @@ -1398,17 +1437,44 @@ bad_mapping: | |||
| 1398 | } | 1437 | } |
| 1399 | 1438 | ||
| 1400 | /** | 1439 | /** |
| 1401 | * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | 1440 | * arm_coherent_iommu_map_sg - map a set of SG buffers for streaming mode DMA |
| 1402 | * @dev: valid struct device pointer | 1441 | * @dev: valid struct device pointer |
| 1403 | * @sg: list of buffers | 1442 | * @sg: list of buffers |
| 1404 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | 1443 | * @nents: number of buffers to map |
| 1405 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | 1444 | * @dir: DMA transfer direction |
| 1406 | * | 1445 | * |
| 1407 | * Unmap a set of streaming mode DMA translations. Again, CPU access | 1446 | * Map a set of i/o coherent buffers described by scatterlist in streaming |
| 1408 | * rules concerning calls here are the same as for dma_unmap_single(). | 1447 | * mode for DMA. The scatter gather list elements are merged together (if |
| 1448 | * possible) and tagged with the appropriate dma address and length. They are | ||
| 1449 | * obtained via sg_dma_{address,length}. | ||
| 1409 | */ | 1450 | */ |
| 1410 | void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | 1451 | int arm_coherent_iommu_map_sg(struct device *dev, struct scatterlist *sg, |
| 1411 | enum dma_data_direction dir, struct dma_attrs *attrs) | 1452 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) |
| 1453 | { | ||
| 1454 | return __iommu_map_sg(dev, sg, nents, dir, attrs, true); | ||
| 1455 | } | ||
| 1456 | |||
| 1457 | /** | ||
| 1458 | * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA | ||
| 1459 | * @dev: valid struct device pointer | ||
| 1460 | * @sg: list of buffers | ||
| 1461 | * @nents: number of buffers to map | ||
| 1462 | * @dir: DMA transfer direction | ||
| 1463 | * | ||
| 1464 | * Map a set of buffers described by scatterlist in streaming mode for DMA. | ||
| 1465 | * The scatter gather list elements are merged together (if possible) and | ||
| 1466 | * tagged with the appropriate dma address and length. They are obtained via | ||
| 1467 | * sg_dma_{address,length}. | ||
| 1468 | */ | ||
| 1469 | int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, | ||
| 1470 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) | ||
| 1471 | { | ||
| 1472 | return __iommu_map_sg(dev, sg, nents, dir, attrs, false); | ||
| 1473 | } | ||
| 1474 | |||
| 1475 | static void __iommu_unmap_sg(struct device *dev, struct scatterlist *sg, | ||
| 1476 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs, | ||
| 1477 | bool is_coherent) | ||
| 1412 | { | 1478 | { |
| 1413 | struct scatterlist *s; | 1479 | struct scatterlist *s; |
| 1414 | int i; | 1480 | int i; |
| @@ -1417,7 +1483,7 @@ void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
| 1417 | if (sg_dma_len(s)) | 1483 | if (sg_dma_len(s)) |
| 1418 | __iommu_remove_mapping(dev, sg_dma_address(s), | 1484 | __iommu_remove_mapping(dev, sg_dma_address(s), |
| 1419 | sg_dma_len(s)); | 1485 | sg_dma_len(s)); |
| 1420 | if (!arch_is_coherent() && | 1486 | if (!is_coherent && |
| 1421 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1487 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
| 1422 | __dma_page_dev_to_cpu(sg_page(s), s->offset, | 1488 | __dma_page_dev_to_cpu(sg_page(s), s->offset, |
| 1423 | s->length, dir); | 1489 | s->length, dir); |
| @@ -1425,6 +1491,38 @@ void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
| 1425 | } | 1491 | } |
| 1426 | 1492 | ||
| 1427 | /** | 1493 | /** |
| 1494 | * arm_coherent_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | ||
| 1495 | * @dev: valid struct device pointer | ||
| 1496 | * @sg: list of buffers | ||
| 1497 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | ||
| 1498 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | ||
| 1499 | * | ||
| 1500 | * Unmap a set of streaming mode DMA translations. Again, CPU access | ||
| 1501 | * rules concerning calls here are the same as for dma_unmap_single(). | ||
| 1502 | */ | ||
| 1503 | void arm_coherent_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, | ||
| 1504 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) | ||
| 1505 | { | ||
| 1506 | __iommu_unmap_sg(dev, sg, nents, dir, attrs, true); | ||
| 1507 | } | ||
| 1508 | |||
| 1509 | /** | ||
| 1510 | * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | ||
| 1511 | * @dev: valid struct device pointer | ||
| 1512 | * @sg: list of buffers | ||
| 1513 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | ||
| 1514 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | ||
| 1515 | * | ||
| 1516 | * Unmap a set of streaming mode DMA translations. Again, CPU access | ||
| 1517 | * rules concerning calls here are the same as for dma_unmap_single(). | ||
| 1518 | */ | ||
| 1519 | void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | ||
| 1520 | enum dma_data_direction dir, struct dma_attrs *attrs) | ||
| 1521 | { | ||
| 1522 | __iommu_unmap_sg(dev, sg, nents, dir, attrs, false); | ||
| 1523 | } | ||
| 1524 | |||
| 1525 | /** | ||
| 1428 | * arm_iommu_sync_sg_for_cpu | 1526 | * arm_iommu_sync_sg_for_cpu |
| 1429 | * @dev: valid struct device pointer | 1527 | * @dev: valid struct device pointer |
| 1430 | * @sg: list of buffers | 1528 | * @sg: list of buffers |
| @@ -1438,8 +1536,7 @@ void arm_iommu_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, | |||
| 1438 | int i; | 1536 | int i; |
| 1439 | 1537 | ||
| 1440 | for_each_sg(sg, s, nents, i) | 1538 | for_each_sg(sg, s, nents, i) |
| 1441 | if (!arch_is_coherent()) | 1539 | __dma_page_dev_to_cpu(sg_page(s), s->offset, s->length, dir); |
| 1442 | __dma_page_dev_to_cpu(sg_page(s), s->offset, s->length, dir); | ||
| 1443 | 1540 | ||
| 1444 | } | 1541 | } |
| 1445 | 1542 | ||
| @@ -1457,22 +1554,21 @@ void arm_iommu_sync_sg_for_device(struct device *dev, struct scatterlist *sg, | |||
| 1457 | int i; | 1554 | int i; |
| 1458 | 1555 | ||
| 1459 | for_each_sg(sg, s, nents, i) | 1556 | for_each_sg(sg, s, nents, i) |
| 1460 | if (!arch_is_coherent()) | 1557 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); |
| 1461 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); | ||
| 1462 | } | 1558 | } |
| 1463 | 1559 | ||
| 1464 | 1560 | ||
| 1465 | /** | 1561 | /** |
| 1466 | * arm_iommu_map_page | 1562 | * arm_coherent_iommu_map_page |
| 1467 | * @dev: valid struct device pointer | 1563 | * @dev: valid struct device pointer |
| 1468 | * @page: page that buffer resides in | 1564 | * @page: page that buffer resides in |
| 1469 | * @offset: offset into page for start of buffer | 1565 | * @offset: offset into page for start of buffer |
| 1470 | * @size: size of buffer to map | 1566 | * @size: size of buffer to map |
| 1471 | * @dir: DMA transfer direction | 1567 | * @dir: DMA transfer direction |
| 1472 | * | 1568 | * |
| 1473 | * IOMMU aware version of arm_dma_map_page() | 1569 | * Coherent IOMMU aware version of arm_dma_map_page() |
| 1474 | */ | 1570 | */ |
| 1475 | static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | 1571 | static dma_addr_t arm_coherent_iommu_map_page(struct device *dev, struct page *page, |
| 1476 | unsigned long offset, size_t size, enum dma_data_direction dir, | 1572 | unsigned long offset, size_t size, enum dma_data_direction dir, |
| 1477 | struct dma_attrs *attrs) | 1573 | struct dma_attrs *attrs) |
| 1478 | { | 1574 | { |
| @@ -1480,9 +1576,6 @@ static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | |||
| 1480 | dma_addr_t dma_addr; | 1576 | dma_addr_t dma_addr; |
| 1481 | int ret, len = PAGE_ALIGN(size + offset); | 1577 | int ret, len = PAGE_ALIGN(size + offset); |
| 1482 | 1578 | ||
| 1483 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | ||
| 1484 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
| 1485 | |||
| 1486 | dma_addr = __alloc_iova(mapping, len); | 1579 | dma_addr = __alloc_iova(mapping, len); |
| 1487 | if (dma_addr == DMA_ERROR_CODE) | 1580 | if (dma_addr == DMA_ERROR_CODE) |
| 1488 | return dma_addr; | 1581 | return dma_addr; |
| @@ -1498,6 +1591,51 @@ fail: | |||
| 1498 | } | 1591 | } |
| 1499 | 1592 | ||
| 1500 | /** | 1593 | /** |
| 1594 | * arm_iommu_map_page | ||
| 1595 | * @dev: valid struct device pointer | ||
| 1596 | * @page: page that buffer resides in | ||
| 1597 | * @offset: offset into page for start of buffer | ||
| 1598 | * @size: size of buffer to map | ||
| 1599 | * @dir: DMA transfer direction | ||
| 1600 | * | ||
| 1601 | * IOMMU aware version of arm_dma_map_page() | ||
| 1602 | */ | ||
| 1603 | static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | ||
| 1604 | unsigned long offset, size_t size, enum dma_data_direction dir, | ||
| 1605 | struct dma_attrs *attrs) | ||
| 1606 | { | ||
| 1607 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | ||
| 1608 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
| 1609 | |||
| 1610 | return arm_coherent_iommu_map_page(dev, page, offset, size, dir, attrs); | ||
| 1611 | } | ||
| 1612 | |||
| 1613 | /** | ||
| 1614 | * arm_coherent_iommu_unmap_page | ||
| 1615 | * @dev: valid struct device pointer | ||
| 1616 | * @handle: DMA address of buffer | ||
| 1617 | * @size: size of buffer (same as passed to dma_map_page) | ||
| 1618 | * @dir: DMA transfer direction (same as passed to dma_map_page) | ||
| 1619 | * | ||
| 1620 | * Coherent IOMMU aware version of arm_dma_unmap_page() | ||
| 1621 | */ | ||
| 1622 | static void arm_coherent_iommu_unmap_page(struct device *dev, dma_addr_t handle, | ||
| 1623 | size_t size, enum dma_data_direction dir, | ||
| 1624 | struct dma_attrs *attrs) | ||
| 1625 | { | ||
| 1626 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; | ||
| 1627 | dma_addr_t iova = handle & PAGE_MASK; | ||
| 1628 | int offset = handle & ~PAGE_MASK; | ||
| 1629 | int len = PAGE_ALIGN(size + offset); | ||
| 1630 | |||
| 1631 | if (!iova) | ||
| 1632 | return; | ||
| 1633 | |||
| 1634 | iommu_unmap(mapping->domain, iova, len); | ||
| 1635 | __free_iova(mapping, iova, len); | ||
| 1636 | } | ||
| 1637 | |||
| 1638 | /** | ||
| 1501 | * arm_iommu_unmap_page | 1639 | * arm_iommu_unmap_page |
| 1502 | * @dev: valid struct device pointer | 1640 | * @dev: valid struct device pointer |
| 1503 | * @handle: DMA address of buffer | 1641 | * @handle: DMA address of buffer |
| @@ -1519,7 +1657,7 @@ static void arm_iommu_unmap_page(struct device *dev, dma_addr_t handle, | |||
| 1519 | if (!iova) | 1657 | if (!iova) |
| 1520 | return; | 1658 | return; |
| 1521 | 1659 | ||
| 1522 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1660 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
| 1523 | __dma_page_dev_to_cpu(page, offset, size, dir); | 1661 | __dma_page_dev_to_cpu(page, offset, size, dir); |
| 1524 | 1662 | ||
| 1525 | iommu_unmap(mapping->domain, iova, len); | 1663 | iommu_unmap(mapping->domain, iova, len); |
| @@ -1537,8 +1675,7 @@ static void arm_iommu_sync_single_for_cpu(struct device *dev, | |||
| 1537 | if (!iova) | 1675 | if (!iova) |
| 1538 | return; | 1676 | return; |
| 1539 | 1677 | ||
| 1540 | if (!arch_is_coherent()) | 1678 | __dma_page_dev_to_cpu(page, offset, size, dir); |
| 1541 | __dma_page_dev_to_cpu(page, offset, size, dir); | ||
| 1542 | } | 1679 | } |
| 1543 | 1680 | ||
| 1544 | static void arm_iommu_sync_single_for_device(struct device *dev, | 1681 | static void arm_iommu_sync_single_for_device(struct device *dev, |
| @@ -1572,6 +1709,19 @@ struct dma_map_ops iommu_ops = { | |||
| 1572 | .sync_sg_for_device = arm_iommu_sync_sg_for_device, | 1709 | .sync_sg_for_device = arm_iommu_sync_sg_for_device, |
| 1573 | }; | 1710 | }; |
| 1574 | 1711 | ||
| 1712 | struct dma_map_ops iommu_coherent_ops = { | ||
| 1713 | .alloc = arm_iommu_alloc_attrs, | ||
| 1714 | .free = arm_iommu_free_attrs, | ||
| 1715 | .mmap = arm_iommu_mmap_attrs, | ||
| 1716 | .get_sgtable = arm_iommu_get_sgtable, | ||
| 1717 | |||
| 1718 | .map_page = arm_coherent_iommu_map_page, | ||
| 1719 | .unmap_page = arm_coherent_iommu_unmap_page, | ||
| 1720 | |||
| 1721 | .map_sg = arm_coherent_iommu_map_sg, | ||
| 1722 | .unmap_sg = arm_coherent_iommu_unmap_sg, | ||
| 1723 | }; | ||
| 1724 | |||
| 1575 | /** | 1725 | /** |
| 1576 | * arm_iommu_create_mapping | 1726 | * arm_iommu_create_mapping |
| 1577 | * @bus: pointer to the bus holding the client device (for IOMMU calls) | 1727 | * @bus: pointer to the bus holding the client device (for IOMMU calls) |
| @@ -1665,7 +1815,7 @@ int arm_iommu_attach_device(struct device *dev, | |||
| 1665 | dev->archdata.mapping = mapping; | 1815 | dev->archdata.mapping = mapping; |
| 1666 | set_dma_ops(dev, &iommu_ops); | 1816 | set_dma_ops(dev, &iommu_ops); |
| 1667 | 1817 | ||
| 1668 | pr_info("Attached IOMMU controller to %s device.\n", dev_name(dev)); | 1818 | pr_debug("Attached IOMMU controller to %s device.\n", dev_name(dev)); |
| 1669 | return 0; | 1819 | return 0; |
| 1670 | } | 1820 | } |
| 1671 | 1821 | ||
diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c index 18144e6a3115..941dfb9e9a78 100644 --- a/arch/arm/mm/mmu.c +++ b/arch/arm/mm/mmu.c | |||
| @@ -423,17 +423,6 @@ static void __init build_mem_type_table(void) | |||
| 423 | vecs_pgprot = kern_pgprot = user_pgprot = cp->pte; | 423 | vecs_pgprot = kern_pgprot = user_pgprot = cp->pte; |
| 424 | 424 | ||
| 425 | /* | 425 | /* |
| 426 | * Enable CPU-specific coherency if supported. | ||
| 427 | * (Only available on XSC3 at the moment.) | ||
| 428 | */ | ||
| 429 | if (arch_is_coherent() && cpu_is_xsc3()) { | ||
| 430 | mem_types[MT_MEMORY].prot_sect |= PMD_SECT_S; | ||
| 431 | mem_types[MT_MEMORY].prot_pte |= L_PTE_SHARED; | ||
| 432 | mem_types[MT_MEMORY_DMA_READY].prot_pte |= L_PTE_SHARED; | ||
| 433 | mem_types[MT_MEMORY_NONCACHED].prot_sect |= PMD_SECT_S; | ||
| 434 | mem_types[MT_MEMORY_NONCACHED].prot_pte |= L_PTE_SHARED; | ||
| 435 | } | ||
| 436 | /* | ||
| 437 | * ARMv6 and above have extended page tables. | 426 | * ARMv6 and above have extended page tables. |
| 438 | */ | 427 | */ |
| 439 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { | 428 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { |
diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c index 34d94c762a1e..9a1469474f55 100644 --- a/drivers/base/dma-contiguous.c +++ b/drivers/base/dma-contiguous.c | |||
| @@ -315,6 +315,7 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
| 315 | { | 315 | { |
| 316 | unsigned long mask, pfn, pageno, start = 0; | 316 | unsigned long mask, pfn, pageno, start = 0; |
| 317 | struct cma *cma = dev_get_cma_area(dev); | 317 | struct cma *cma = dev_get_cma_area(dev); |
| 318 | struct page *page = NULL; | ||
| 318 | int ret; | 319 | int ret; |
| 319 | 320 | ||
| 320 | if (!cma || !cma->count) | 321 | if (!cma || !cma->count) |
| @@ -336,18 +337,17 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
| 336 | for (;;) { | 337 | for (;;) { |
| 337 | pageno = bitmap_find_next_zero_area(cma->bitmap, cma->count, | 338 | pageno = bitmap_find_next_zero_area(cma->bitmap, cma->count, |
| 338 | start, count, mask); | 339 | start, count, mask); |
| 339 | if (pageno >= cma->count) { | 340 | if (pageno >= cma->count) |
| 340 | ret = -ENOMEM; | 341 | break; |
| 341 | goto error; | ||
| 342 | } | ||
| 343 | 342 | ||
| 344 | pfn = cma->base_pfn + pageno; | 343 | pfn = cma->base_pfn + pageno; |
| 345 | ret = alloc_contig_range(pfn, pfn + count, MIGRATE_CMA); | 344 | ret = alloc_contig_range(pfn, pfn + count, MIGRATE_CMA); |
| 346 | if (ret == 0) { | 345 | if (ret == 0) { |
| 347 | bitmap_set(cma->bitmap, pageno, count); | 346 | bitmap_set(cma->bitmap, pageno, count); |
| 347 | page = pfn_to_page(pfn); | ||
| 348 | break; | 348 | break; |
| 349 | } else if (ret != -EBUSY) { | 349 | } else if (ret != -EBUSY) { |
| 350 | goto error; | 350 | break; |
| 351 | } | 351 | } |
| 352 | pr_debug("%s(): memory range at %p is busy, retrying\n", | 352 | pr_debug("%s(): memory range at %p is busy, retrying\n", |
| 353 | __func__, pfn_to_page(pfn)); | 353 | __func__, pfn_to_page(pfn)); |
| @@ -356,12 +356,8 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
| 356 | } | 356 | } |
| 357 | 357 | ||
| 358 | mutex_unlock(&cma_mutex); | 358 | mutex_unlock(&cma_mutex); |
| 359 | 359 | pr_debug("%s(): returned %p\n", __func__, page); | |
| 360 | pr_debug("%s(): returned %p\n", __func__, pfn_to_page(pfn)); | 360 | return page; |
| 361 | return pfn_to_page(pfn); | ||
| 362 | error: | ||
| 363 | mutex_unlock(&cma_mutex); | ||
| 364 | return NULL; | ||
| 365 | } | 361 | } |
| 366 | 362 | ||
| 367 | /** | 363 | /** |
