diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2012-10-02 22:13:12 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-10-02 22:13:12 -0400 |
commit | ca41cc96b2813221b05af57d0355157924de5a07 (patch) | |
tree | 093d4ca55e6dc4f77c7c8486ec47e9e625e84f17 | |
parent | 3151367f8778a1789d6f6e6f6c642681b6cd6d64 (diff) | |
parent | 461b6f0d3d7d4e556035463b531136b034b7433e (diff) |
Merge branch 'for-v3.7' of git://git.linaro.org/people/mszyprowski/linux-dma-mapping
Pull CMA and DMA-mapping updates from Marek Szyprowski:
"This time the pull request is rather small, because the further
redesign patches were not ready on time.
This pull request consists of the patches which extend ARM DMA-mapping
subsystem with support for CPU coherent (ACP) DMA busses. The first
client of the new version is HighBank SATA driver. The second part of
the pull request includes various cleanup for both CMA common code and
ARM DMA-mapping subsystem."
Fix up trivial add-add conflict due to the "dma-coherent" DT property
being added next to the "calxeda,port-phys" property for the Calxeda
AHCI controller.
* 'for-v3.7' of git://git.linaro.org/people/mszyprowski/linux-dma-mapping:
ARM: dma-mapping: Remove unsed var at arm_coherent_iommu_unmap_page
ARM: highbank: add coherent DMA setup
ARM: kill off arch_is_coherent
ARM: add coherent iommu dma ops
ARM: add coherent dma ops
ARM: dma-mapping: Refrain noisy console message
ARM: dma-mapping: Small logical clean up
drivers: dma-contiguous: refactor dma_alloc_from_contiguous()
-rw-r--r-- | Documentation/devicetree/bindings/ata/ahci-platform.txt | 1 | ||||
-rw-r--r-- | Documentation/devicetree/bindings/dma/arm-pl330.txt | 3 | ||||
-rw-r--r-- | Documentation/devicetree/bindings/net/calxeda-xgmac.txt | 3 | ||||
-rw-r--r-- | arch/arm/boot/dts/highbank.dts | 1 | ||||
-rw-r--r-- | arch/arm/include/asm/barrier.h | 7 | ||||
-rw-r--r-- | arch/arm/include/asm/dma-mapping.h | 1 | ||||
-rw-r--r-- | arch/arm/include/asm/memory.h | 8 | ||||
-rw-r--r-- | arch/arm/mach-highbank/highbank.c | 52 | ||||
-rw-r--r-- | arch/arm/mm/dma-mapping.c | 264 | ||||
-rw-r--r-- | arch/arm/mm/mmu.c | 11 | ||||
-rw-r--r-- | drivers/base/dma-contiguous.c | 18 |
11 files changed, 278 insertions, 91 deletions
diff --git a/Documentation/devicetree/bindings/ata/ahci-platform.txt b/Documentation/devicetree/bindings/ata/ahci-platform.txt index 147c1f6653fe..b519f9b699c3 100644 --- a/Documentation/devicetree/bindings/ata/ahci-platform.txt +++ b/Documentation/devicetree/bindings/ata/ahci-platform.txt | |||
@@ -12,6 +12,7 @@ Optional properties: | |||
12 | - calxeda,port-phys: phandle-combophy and lane assignment, which maps each | 12 | - calxeda,port-phys: phandle-combophy and lane assignment, which maps each |
13 | SATA port to a combophy and a lane within that | 13 | SATA port to a combophy and a lane within that |
14 | combophy | 14 | combophy |
15 | - dma-coherent : Present if dma operations are coherent | ||
15 | 16 | ||
16 | Example: | 17 | Example: |
17 | sata@ffe08000 { | 18 | sata@ffe08000 { |
diff --git a/Documentation/devicetree/bindings/dma/arm-pl330.txt b/Documentation/devicetree/bindings/dma/arm-pl330.txt index a4cd273b2a67..36e27d54260b 100644 --- a/Documentation/devicetree/bindings/dma/arm-pl330.txt +++ b/Documentation/devicetree/bindings/dma/arm-pl330.txt | |||
@@ -9,6 +9,9 @@ Required properties: | |||
9 | region. | 9 | region. |
10 | - interrupts: interrupt number to the cpu. | 10 | - interrupts: interrupt number to the cpu. |
11 | 11 | ||
12 | Optional properties: | ||
13 | - dma-coherent : Present if dma operations are coherent | ||
14 | |||
12 | Example: | 15 | Example: |
13 | 16 | ||
14 | pdma0: pdma@12680000 { | 17 | pdma0: pdma@12680000 { |
diff --git a/Documentation/devicetree/bindings/net/calxeda-xgmac.txt b/Documentation/devicetree/bindings/net/calxeda-xgmac.txt index 411727a3f82d..c8ae996bd8f2 100644 --- a/Documentation/devicetree/bindings/net/calxeda-xgmac.txt +++ b/Documentation/devicetree/bindings/net/calxeda-xgmac.txt | |||
@@ -6,6 +6,9 @@ Required properties: | |||
6 | - interrupts : Should contain 3 xgmac interrupts. The 1st is main interrupt. | 6 | - interrupts : Should contain 3 xgmac interrupts. The 1st is main interrupt. |
7 | The 2nd is pwr mgt interrupt. The 3rd is low power state interrupt. | 7 | The 2nd is pwr mgt interrupt. The 3rd is low power state interrupt. |
8 | 8 | ||
9 | Optional properties: | ||
10 | - dma-coherent : Present if dma operations are coherent | ||
11 | |||
9 | Example: | 12 | Example: |
10 | 13 | ||
11 | ethernet@fff50000 { | 14 | ethernet@fff50000 { |
diff --git a/arch/arm/boot/dts/highbank.dts b/arch/arm/boot/dts/highbank.dts index 5204cf73c2d8..0c6fc34821f9 100644 --- a/arch/arm/boot/dts/highbank.dts +++ b/arch/arm/boot/dts/highbank.dts | |||
@@ -124,6 +124,7 @@ | |||
124 | calxeda,port-phys = <&combophy5 0 &combophy0 0 | 124 | calxeda,port-phys = <&combophy5 0 &combophy0 0 |
125 | &combophy0 1 &combophy0 2 | 125 | &combophy0 1 &combophy0 2 |
126 | &combophy0 3>; | 126 | &combophy0 3>; |
127 | dma-coherent; | ||
127 | }; | 128 | }; |
128 | 129 | ||
129 | sdhci@ffe0e000 { | 130 | sdhci@ffe0e000 { |
diff --git a/arch/arm/include/asm/barrier.h b/arch/arm/include/asm/barrier.h index 05112380dc53..8dcd9c702d90 100644 --- a/arch/arm/include/asm/barrier.h +++ b/arch/arm/include/asm/barrier.h | |||
@@ -44,10 +44,9 @@ | |||
44 | #define rmb() dsb() | 44 | #define rmb() dsb() |
45 | #define wmb() mb() | 45 | #define wmb() mb() |
46 | #else | 46 | #else |
47 | #include <asm/memory.h> | 47 | #define mb() barrier() |
48 | #define mb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | 48 | #define rmb() barrier() |
49 | #define rmb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | 49 | #define wmb() barrier() |
50 | #define wmb() do { if (arch_is_coherent()) dmb(); else barrier(); } while (0) | ||
51 | #endif | 50 | #endif |
52 | 51 | ||
53 | #ifndef CONFIG_SMP | 52 | #ifndef CONFIG_SMP |
diff --git a/arch/arm/include/asm/dma-mapping.h b/arch/arm/include/asm/dma-mapping.h index 5c44dcb0987b..23004847bb05 100644 --- a/arch/arm/include/asm/dma-mapping.h +++ b/arch/arm/include/asm/dma-mapping.h | |||
@@ -13,6 +13,7 @@ | |||
13 | 13 | ||
14 | #define DMA_ERROR_CODE (~0) | 14 | #define DMA_ERROR_CODE (~0) |
15 | extern struct dma_map_ops arm_dma_ops; | 15 | extern struct dma_map_ops arm_dma_ops; |
16 | extern struct dma_map_ops arm_coherent_dma_ops; | ||
16 | 17 | ||
17 | static inline struct dma_map_ops *get_dma_ops(struct device *dev) | 18 | static inline struct dma_map_ops *get_dma_ops(struct device *dev) |
18 | { | 19 | { |
diff --git a/arch/arm/include/asm/memory.h b/arch/arm/include/asm/memory.h index 5f6ddcc56452..73cf03aa981e 100644 --- a/arch/arm/include/asm/memory.h +++ b/arch/arm/include/asm/memory.h | |||
@@ -275,14 +275,6 @@ static inline __deprecated void *bus_to_virt(unsigned long x) | |||
275 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) | 275 | #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) |
276 | #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory) | 276 | #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory) |
277 | 277 | ||
278 | /* | ||
279 | * Optional coherency support. Currently used only by selected | ||
280 | * Intel XSC3-based systems. | ||
281 | */ | ||
282 | #ifndef arch_is_coherent | ||
283 | #define arch_is_coherent() 0 | ||
284 | #endif | ||
285 | |||
286 | #endif | 278 | #endif |
287 | 279 | ||
288 | #include <asm-generic/memory_model.h> | 280 | #include <asm-generic/memory_model.h> |
diff --git a/arch/arm/mach-highbank/highbank.c b/arch/arm/mach-highbank/highbank.c index af1da34ccf9d..40e36a50304c 100644 --- a/arch/arm/mach-highbank/highbank.c +++ b/arch/arm/mach-highbank/highbank.c | |||
@@ -15,6 +15,7 @@ | |||
15 | */ | 15 | */ |
16 | #include <linux/clk.h> | 16 | #include <linux/clk.h> |
17 | #include <linux/clkdev.h> | 17 | #include <linux/clkdev.h> |
18 | #include <linux/dma-mapping.h> | ||
18 | #include <linux/io.h> | 19 | #include <linux/io.h> |
19 | #include <linux/irq.h> | 20 | #include <linux/irq.h> |
20 | #include <linux/irqdomain.h> | 21 | #include <linux/irqdomain.h> |
@@ -23,6 +24,7 @@ | |||
23 | #include <linux/of_platform.h> | 24 | #include <linux/of_platform.h> |
24 | #include <linux/of_address.h> | 25 | #include <linux/of_address.h> |
25 | #include <linux/smp.h> | 26 | #include <linux/smp.h> |
27 | #include <linux/amba/bus.h> | ||
26 | 28 | ||
27 | #include <asm/cacheflush.h> | 29 | #include <asm/cacheflush.h> |
28 | #include <asm/smp_plat.h> | 30 | #include <asm/smp_plat.h> |
@@ -149,11 +151,61 @@ static void highbank_power_off(void) | |||
149 | cpu_do_idle(); | 151 | cpu_do_idle(); |
150 | } | 152 | } |
151 | 153 | ||
154 | static int highbank_platform_notifier(struct notifier_block *nb, | ||
155 | unsigned long event, void *__dev) | ||
156 | { | ||
157 | struct resource *res; | ||
158 | int reg = -1; | ||
159 | struct device *dev = __dev; | ||
160 | |||
161 | if (event != BUS_NOTIFY_ADD_DEVICE) | ||
162 | return NOTIFY_DONE; | ||
163 | |||
164 | if (of_device_is_compatible(dev->of_node, "calxeda,hb-ahci")) | ||
165 | reg = 0xc; | ||
166 | else if (of_device_is_compatible(dev->of_node, "calxeda,hb-sdhci")) | ||
167 | reg = 0x18; | ||
168 | else if (of_device_is_compatible(dev->of_node, "arm,pl330")) | ||
169 | reg = 0x20; | ||
170 | else if (of_device_is_compatible(dev->of_node, "calxeda,hb-xgmac")) { | ||
171 | res = platform_get_resource(to_platform_device(dev), | ||
172 | IORESOURCE_MEM, 0); | ||
173 | if (res) { | ||
174 | if (res->start == 0xfff50000) | ||
175 | reg = 0; | ||
176 | else if (res->start == 0xfff51000) | ||
177 | reg = 4; | ||
178 | } | ||
179 | } | ||
180 | |||
181 | if (reg < 0) | ||
182 | return NOTIFY_DONE; | ||
183 | |||
184 | if (of_property_read_bool(dev->of_node, "dma-coherent")) { | ||
185 | writel(0xff31, sregs_base + reg); | ||
186 | set_dma_ops(dev, &arm_coherent_dma_ops); | ||
187 | } else | ||
188 | writel(0, sregs_base + reg); | ||
189 | |||
190 | return NOTIFY_OK; | ||
191 | } | ||
192 | |||
193 | static struct notifier_block highbank_amba_nb = { | ||
194 | .notifier_call = highbank_platform_notifier, | ||
195 | }; | ||
196 | |||
197 | static struct notifier_block highbank_platform_nb = { | ||
198 | .notifier_call = highbank_platform_notifier, | ||
199 | }; | ||
200 | |||
152 | static void __init highbank_init(void) | 201 | static void __init highbank_init(void) |
153 | { | 202 | { |
154 | pm_power_off = highbank_power_off; | 203 | pm_power_off = highbank_power_off; |
155 | highbank_pm_init(); | 204 | highbank_pm_init(); |
156 | 205 | ||
206 | bus_register_notifier(&platform_bus_type, &highbank_platform_nb); | ||
207 | bus_register_notifier(&amba_bustype, &highbank_amba_nb); | ||
208 | |||
157 | of_platform_populate(NULL, of_default_bus_match_table, NULL, NULL); | 209 | of_platform_populate(NULL, of_default_bus_match_table, NULL, NULL); |
158 | } | 210 | } |
159 | 211 | ||
diff --git a/arch/arm/mm/dma-mapping.c b/arch/arm/mm/dma-mapping.c index 13f555d62491..477a2d23ddf1 100644 --- a/arch/arm/mm/dma-mapping.c +++ b/arch/arm/mm/dma-mapping.c | |||
@@ -73,11 +73,18 @@ static dma_addr_t arm_dma_map_page(struct device *dev, struct page *page, | |||
73 | unsigned long offset, size_t size, enum dma_data_direction dir, | 73 | unsigned long offset, size_t size, enum dma_data_direction dir, |
74 | struct dma_attrs *attrs) | 74 | struct dma_attrs *attrs) |
75 | { | 75 | { |
76 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 76 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
77 | __dma_page_cpu_to_dev(page, offset, size, dir); | 77 | __dma_page_cpu_to_dev(page, offset, size, dir); |
78 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; | 78 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; |
79 | } | 79 | } |
80 | 80 | ||
81 | static dma_addr_t arm_coherent_dma_map_page(struct device *dev, struct page *page, | ||
82 | unsigned long offset, size_t size, enum dma_data_direction dir, | ||
83 | struct dma_attrs *attrs) | ||
84 | { | ||
85 | return pfn_to_dma(dev, page_to_pfn(page)) + offset; | ||
86 | } | ||
87 | |||
81 | /** | 88 | /** |
82 | * arm_dma_unmap_page - unmap a buffer previously mapped through dma_map_page() | 89 | * arm_dma_unmap_page - unmap a buffer previously mapped through dma_map_page() |
83 | * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices | 90 | * @dev: valid struct device pointer, or NULL for ISA and EISA-like devices |
@@ -96,7 +103,7 @@ static void arm_dma_unmap_page(struct device *dev, dma_addr_t handle, | |||
96 | size_t size, enum dma_data_direction dir, | 103 | size_t size, enum dma_data_direction dir, |
97 | struct dma_attrs *attrs) | 104 | struct dma_attrs *attrs) |
98 | { | 105 | { |
99 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 106 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
100 | __dma_page_dev_to_cpu(pfn_to_page(dma_to_pfn(dev, handle)), | 107 | __dma_page_dev_to_cpu(pfn_to_page(dma_to_pfn(dev, handle)), |
101 | handle & ~PAGE_MASK, size, dir); | 108 | handle & ~PAGE_MASK, size, dir); |
102 | } | 109 | } |
@@ -106,8 +113,7 @@ static void arm_dma_sync_single_for_cpu(struct device *dev, | |||
106 | { | 113 | { |
107 | unsigned int offset = handle & (PAGE_SIZE - 1); | 114 | unsigned int offset = handle & (PAGE_SIZE - 1); |
108 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); | 115 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); |
109 | if (!arch_is_coherent()) | 116 | __dma_page_dev_to_cpu(page, offset, size, dir); |
110 | __dma_page_dev_to_cpu(page, offset, size, dir); | ||
111 | } | 117 | } |
112 | 118 | ||
113 | static void arm_dma_sync_single_for_device(struct device *dev, | 119 | static void arm_dma_sync_single_for_device(struct device *dev, |
@@ -115,8 +121,7 @@ static void arm_dma_sync_single_for_device(struct device *dev, | |||
115 | { | 121 | { |
116 | unsigned int offset = handle & (PAGE_SIZE - 1); | 122 | unsigned int offset = handle & (PAGE_SIZE - 1); |
117 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); | 123 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle-offset)); |
118 | if (!arch_is_coherent()) | 124 | __dma_page_cpu_to_dev(page, offset, size, dir); |
119 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
120 | } | 125 | } |
121 | 126 | ||
122 | static int arm_dma_set_mask(struct device *dev, u64 dma_mask); | 127 | static int arm_dma_set_mask(struct device *dev, u64 dma_mask); |
@@ -138,6 +143,22 @@ struct dma_map_ops arm_dma_ops = { | |||
138 | }; | 143 | }; |
139 | EXPORT_SYMBOL(arm_dma_ops); | 144 | EXPORT_SYMBOL(arm_dma_ops); |
140 | 145 | ||
146 | static void *arm_coherent_dma_alloc(struct device *dev, size_t size, | ||
147 | dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs); | ||
148 | static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
149 | dma_addr_t handle, struct dma_attrs *attrs); | ||
150 | |||
151 | struct dma_map_ops arm_coherent_dma_ops = { | ||
152 | .alloc = arm_coherent_dma_alloc, | ||
153 | .free = arm_coherent_dma_free, | ||
154 | .mmap = arm_dma_mmap, | ||
155 | .get_sgtable = arm_dma_get_sgtable, | ||
156 | .map_page = arm_coherent_dma_map_page, | ||
157 | .map_sg = arm_dma_map_sg, | ||
158 | .set_dma_mask = arm_dma_set_mask, | ||
159 | }; | ||
160 | EXPORT_SYMBOL(arm_coherent_dma_ops); | ||
161 | |||
141 | static u64 get_coherent_dma_mask(struct device *dev) | 162 | static u64 get_coherent_dma_mask(struct device *dev) |
142 | { | 163 | { |
143 | u64 mask = (u64)arm_dma_limit; | 164 | u64 mask = (u64)arm_dma_limit; |
@@ -586,7 +607,7 @@ static void *__alloc_simple_buffer(struct device *dev, size_t size, gfp_t gfp, | |||
586 | 607 | ||
587 | 608 | ||
588 | static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | 609 | static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, |
589 | gfp_t gfp, pgprot_t prot, const void *caller) | 610 | gfp_t gfp, pgprot_t prot, bool is_coherent, const void *caller) |
590 | { | 611 | { |
591 | u64 mask = get_coherent_dma_mask(dev); | 612 | u64 mask = get_coherent_dma_mask(dev); |
592 | struct page *page; | 613 | struct page *page; |
@@ -619,7 +640,7 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | |||
619 | *handle = DMA_ERROR_CODE; | 640 | *handle = DMA_ERROR_CODE; |
620 | size = PAGE_ALIGN(size); | 641 | size = PAGE_ALIGN(size); |
621 | 642 | ||
622 | if (arch_is_coherent() || nommu()) | 643 | if (is_coherent || nommu()) |
623 | addr = __alloc_simple_buffer(dev, size, gfp, &page); | 644 | addr = __alloc_simple_buffer(dev, size, gfp, &page); |
624 | else if (gfp & GFP_ATOMIC) | 645 | else if (gfp & GFP_ATOMIC) |
625 | addr = __alloc_from_pool(size, &page); | 646 | addr = __alloc_from_pool(size, &page); |
@@ -647,7 +668,20 @@ void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, | |||
647 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) | 668 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) |
648 | return memory; | 669 | return memory; |
649 | 670 | ||
650 | return __dma_alloc(dev, size, handle, gfp, prot, | 671 | return __dma_alloc(dev, size, handle, gfp, prot, false, |
672 | __builtin_return_address(0)); | ||
673 | } | ||
674 | |||
675 | static void *arm_coherent_dma_alloc(struct device *dev, size_t size, | ||
676 | dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs) | ||
677 | { | ||
678 | pgprot_t prot = __get_dma_pgprot(attrs, pgprot_kernel); | ||
679 | void *memory; | ||
680 | |||
681 | if (dma_alloc_from_coherent(dev, size, handle, &memory)) | ||
682 | return memory; | ||
683 | |||
684 | return __dma_alloc(dev, size, handle, gfp, prot, true, | ||
651 | __builtin_return_address(0)); | 685 | __builtin_return_address(0)); |
652 | } | 686 | } |
653 | 687 | ||
@@ -684,8 +718,9 @@ int arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, | |||
684 | /* | 718 | /* |
685 | * Free a buffer as defined by the above mapping. | 719 | * Free a buffer as defined by the above mapping. |
686 | */ | 720 | */ |
687 | void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | 721 | static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, |
688 | dma_addr_t handle, struct dma_attrs *attrs) | 722 | dma_addr_t handle, struct dma_attrs *attrs, |
723 | bool is_coherent) | ||
689 | { | 724 | { |
690 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); | 725 | struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); |
691 | 726 | ||
@@ -694,7 +729,7 @@ void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | |||
694 | 729 | ||
695 | size = PAGE_ALIGN(size); | 730 | size = PAGE_ALIGN(size); |
696 | 731 | ||
697 | if (arch_is_coherent() || nommu()) { | 732 | if (is_coherent || nommu()) { |
698 | __dma_free_buffer(page, size); | 733 | __dma_free_buffer(page, size); |
699 | } else if (__free_from_pool(cpu_addr, size)) { | 734 | } else if (__free_from_pool(cpu_addr, size)) { |
700 | return; | 735 | return; |
@@ -710,6 +745,18 @@ void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | |||
710 | } | 745 | } |
711 | } | 746 | } |
712 | 747 | ||
748 | void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
749 | dma_addr_t handle, struct dma_attrs *attrs) | ||
750 | { | ||
751 | __arm_dma_free(dev, size, cpu_addr, handle, attrs, false); | ||
752 | } | ||
753 | |||
754 | static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, | ||
755 | dma_addr_t handle, struct dma_attrs *attrs) | ||
756 | { | ||
757 | __arm_dma_free(dev, size, cpu_addr, handle, attrs, true); | ||
758 | } | ||
759 | |||
713 | int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, | 760 | int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, |
714 | void *cpu_addr, dma_addr_t handle, size_t size, | 761 | void *cpu_addr, dma_addr_t handle, size_t size, |
715 | struct dma_attrs *attrs) | 762 | struct dma_attrs *attrs) |
@@ -1012,11 +1059,12 @@ static struct page **__iommu_alloc_buffer(struct device *dev, size_t size, gfp_t | |||
1012 | if (!pages[i]) | 1059 | if (!pages[i]) |
1013 | goto error; | 1060 | goto error; |
1014 | 1061 | ||
1015 | if (order) | 1062 | if (order) { |
1016 | split_page(pages[i], order); | 1063 | split_page(pages[i], order); |
1017 | j = 1 << order; | 1064 | j = 1 << order; |
1018 | while (--j) | 1065 | while (--j) |
1019 | pages[i + j] = pages[i] + j; | 1066 | pages[i + j] = pages[i] + j; |
1067 | } | ||
1020 | 1068 | ||
1021 | __dma_clear_buffer(pages[i], PAGE_SIZE << order); | 1069 | __dma_clear_buffer(pages[i], PAGE_SIZE << order); |
1022 | i += 1 << order; | 1070 | i += 1 << order; |
@@ -1303,7 +1351,8 @@ static int arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, | |||
1303 | */ | 1351 | */ |
1304 | static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, | 1352 | static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, |
1305 | size_t size, dma_addr_t *handle, | 1353 | size_t size, dma_addr_t *handle, |
1306 | enum dma_data_direction dir, struct dma_attrs *attrs) | 1354 | enum dma_data_direction dir, struct dma_attrs *attrs, |
1355 | bool is_coherent) | ||
1307 | { | 1356 | { |
1308 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; | 1357 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; |
1309 | dma_addr_t iova, iova_base; | 1358 | dma_addr_t iova, iova_base; |
@@ -1322,8 +1371,8 @@ static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, | |||
1322 | phys_addr_t phys = page_to_phys(sg_page(s)); | 1371 | phys_addr_t phys = page_to_phys(sg_page(s)); |
1323 | unsigned int len = PAGE_ALIGN(s->offset + s->length); | 1372 | unsigned int len = PAGE_ALIGN(s->offset + s->length); |
1324 | 1373 | ||
1325 | if (!arch_is_coherent() && | 1374 | if (!is_coherent && |
1326 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1375 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
1327 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); | 1376 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); |
1328 | 1377 | ||
1329 | ret = iommu_map(mapping->domain, iova, phys, len, 0); | 1378 | ret = iommu_map(mapping->domain, iova, phys, len, 0); |
@@ -1341,20 +1390,9 @@ fail: | |||
1341 | return ret; | 1390 | return ret; |
1342 | } | 1391 | } |
1343 | 1392 | ||
1344 | /** | 1393 | static int __iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, |
1345 | * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA | 1394 | enum dma_data_direction dir, struct dma_attrs *attrs, |
1346 | * @dev: valid struct device pointer | 1395 | bool is_coherent) |
1347 | * @sg: list of buffers | ||
1348 | * @nents: number of buffers to map | ||
1349 | * @dir: DMA transfer direction | ||
1350 | * | ||
1351 | * Map a set of buffers described by scatterlist in streaming mode for DMA. | ||
1352 | * The scatter gather list elements are merged together (if possible) and | ||
1353 | * tagged with the appropriate dma address and length. They are obtained via | ||
1354 | * sg_dma_{address,length}. | ||
1355 | */ | ||
1356 | int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | ||
1357 | enum dma_data_direction dir, struct dma_attrs *attrs) | ||
1358 | { | 1396 | { |
1359 | struct scatterlist *s = sg, *dma = sg, *start = sg; | 1397 | struct scatterlist *s = sg, *dma = sg, *start = sg; |
1360 | int i, count = 0; | 1398 | int i, count = 0; |
@@ -1370,7 +1408,7 @@ int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
1370 | 1408 | ||
1371 | if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { | 1409 | if (s->offset || (size & ~PAGE_MASK) || size + s->length > max) { |
1372 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, | 1410 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, |
1373 | dir, attrs) < 0) | 1411 | dir, attrs, is_coherent) < 0) |
1374 | goto bad_mapping; | 1412 | goto bad_mapping; |
1375 | 1413 | ||
1376 | dma->dma_address += offset; | 1414 | dma->dma_address += offset; |
@@ -1383,7 +1421,8 @@ int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
1383 | } | 1421 | } |
1384 | size += s->length; | 1422 | size += s->length; |
1385 | } | 1423 | } |
1386 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs) < 0) | 1424 | if (__map_sg_chunk(dev, start, size, &dma->dma_address, dir, attrs, |
1425 | is_coherent) < 0) | ||
1387 | goto bad_mapping; | 1426 | goto bad_mapping; |
1388 | 1427 | ||
1389 | dma->dma_address += offset; | 1428 | dma->dma_address += offset; |
@@ -1398,17 +1437,44 @@ bad_mapping: | |||
1398 | } | 1437 | } |
1399 | 1438 | ||
1400 | /** | 1439 | /** |
1401 | * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | 1440 | * arm_coherent_iommu_map_sg - map a set of SG buffers for streaming mode DMA |
1402 | * @dev: valid struct device pointer | 1441 | * @dev: valid struct device pointer |
1403 | * @sg: list of buffers | 1442 | * @sg: list of buffers |
1404 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | 1443 | * @nents: number of buffers to map |
1405 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | 1444 | * @dir: DMA transfer direction |
1406 | * | 1445 | * |
1407 | * Unmap a set of streaming mode DMA translations. Again, CPU access | 1446 | * Map a set of i/o coherent buffers described by scatterlist in streaming |
1408 | * rules concerning calls here are the same as for dma_unmap_single(). | 1447 | * mode for DMA. The scatter gather list elements are merged together (if |
1448 | * possible) and tagged with the appropriate dma address and length. They are | ||
1449 | * obtained via sg_dma_{address,length}. | ||
1409 | */ | 1450 | */ |
1410 | void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | 1451 | int arm_coherent_iommu_map_sg(struct device *dev, struct scatterlist *sg, |
1411 | enum dma_data_direction dir, struct dma_attrs *attrs) | 1452 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) |
1453 | { | ||
1454 | return __iommu_map_sg(dev, sg, nents, dir, attrs, true); | ||
1455 | } | ||
1456 | |||
1457 | /** | ||
1458 | * arm_iommu_map_sg - map a set of SG buffers for streaming mode DMA | ||
1459 | * @dev: valid struct device pointer | ||
1460 | * @sg: list of buffers | ||
1461 | * @nents: number of buffers to map | ||
1462 | * @dir: DMA transfer direction | ||
1463 | * | ||
1464 | * Map a set of buffers described by scatterlist in streaming mode for DMA. | ||
1465 | * The scatter gather list elements are merged together (if possible) and | ||
1466 | * tagged with the appropriate dma address and length. They are obtained via | ||
1467 | * sg_dma_{address,length}. | ||
1468 | */ | ||
1469 | int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, | ||
1470 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) | ||
1471 | { | ||
1472 | return __iommu_map_sg(dev, sg, nents, dir, attrs, false); | ||
1473 | } | ||
1474 | |||
1475 | static void __iommu_unmap_sg(struct device *dev, struct scatterlist *sg, | ||
1476 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs, | ||
1477 | bool is_coherent) | ||
1412 | { | 1478 | { |
1413 | struct scatterlist *s; | 1479 | struct scatterlist *s; |
1414 | int i; | 1480 | int i; |
@@ -1417,7 +1483,7 @@ void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
1417 | if (sg_dma_len(s)) | 1483 | if (sg_dma_len(s)) |
1418 | __iommu_remove_mapping(dev, sg_dma_address(s), | 1484 | __iommu_remove_mapping(dev, sg_dma_address(s), |
1419 | sg_dma_len(s)); | 1485 | sg_dma_len(s)); |
1420 | if (!arch_is_coherent() && | 1486 | if (!is_coherent && |
1421 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1487 | !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
1422 | __dma_page_dev_to_cpu(sg_page(s), s->offset, | 1488 | __dma_page_dev_to_cpu(sg_page(s), s->offset, |
1423 | s->length, dir); | 1489 | s->length, dir); |
@@ -1425,6 +1491,38 @@ void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | |||
1425 | } | 1491 | } |
1426 | 1492 | ||
1427 | /** | 1493 | /** |
1494 | * arm_coherent_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | ||
1495 | * @dev: valid struct device pointer | ||
1496 | * @sg: list of buffers | ||
1497 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | ||
1498 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | ||
1499 | * | ||
1500 | * Unmap a set of streaming mode DMA translations. Again, CPU access | ||
1501 | * rules concerning calls here are the same as for dma_unmap_single(). | ||
1502 | */ | ||
1503 | void arm_coherent_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, | ||
1504 | int nents, enum dma_data_direction dir, struct dma_attrs *attrs) | ||
1505 | { | ||
1506 | __iommu_unmap_sg(dev, sg, nents, dir, attrs, true); | ||
1507 | } | ||
1508 | |||
1509 | /** | ||
1510 | * arm_iommu_unmap_sg - unmap a set of SG buffers mapped by dma_map_sg | ||
1511 | * @dev: valid struct device pointer | ||
1512 | * @sg: list of buffers | ||
1513 | * @nents: number of buffers to unmap (same as was passed to dma_map_sg) | ||
1514 | * @dir: DMA transfer direction (same as was passed to dma_map_sg) | ||
1515 | * | ||
1516 | * Unmap a set of streaming mode DMA translations. Again, CPU access | ||
1517 | * rules concerning calls here are the same as for dma_unmap_single(). | ||
1518 | */ | ||
1519 | void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, | ||
1520 | enum dma_data_direction dir, struct dma_attrs *attrs) | ||
1521 | { | ||
1522 | __iommu_unmap_sg(dev, sg, nents, dir, attrs, false); | ||
1523 | } | ||
1524 | |||
1525 | /** | ||
1428 | * arm_iommu_sync_sg_for_cpu | 1526 | * arm_iommu_sync_sg_for_cpu |
1429 | * @dev: valid struct device pointer | 1527 | * @dev: valid struct device pointer |
1430 | * @sg: list of buffers | 1528 | * @sg: list of buffers |
@@ -1438,8 +1536,7 @@ void arm_iommu_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, | |||
1438 | int i; | 1536 | int i; |
1439 | 1537 | ||
1440 | for_each_sg(sg, s, nents, i) | 1538 | for_each_sg(sg, s, nents, i) |
1441 | if (!arch_is_coherent()) | 1539 | __dma_page_dev_to_cpu(sg_page(s), s->offset, s->length, dir); |
1442 | __dma_page_dev_to_cpu(sg_page(s), s->offset, s->length, dir); | ||
1443 | 1540 | ||
1444 | } | 1541 | } |
1445 | 1542 | ||
@@ -1457,22 +1554,21 @@ void arm_iommu_sync_sg_for_device(struct device *dev, struct scatterlist *sg, | |||
1457 | int i; | 1554 | int i; |
1458 | 1555 | ||
1459 | for_each_sg(sg, s, nents, i) | 1556 | for_each_sg(sg, s, nents, i) |
1460 | if (!arch_is_coherent()) | 1557 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); |
1461 | __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); | ||
1462 | } | 1558 | } |
1463 | 1559 | ||
1464 | 1560 | ||
1465 | /** | 1561 | /** |
1466 | * arm_iommu_map_page | 1562 | * arm_coherent_iommu_map_page |
1467 | * @dev: valid struct device pointer | 1563 | * @dev: valid struct device pointer |
1468 | * @page: page that buffer resides in | 1564 | * @page: page that buffer resides in |
1469 | * @offset: offset into page for start of buffer | 1565 | * @offset: offset into page for start of buffer |
1470 | * @size: size of buffer to map | 1566 | * @size: size of buffer to map |
1471 | * @dir: DMA transfer direction | 1567 | * @dir: DMA transfer direction |
1472 | * | 1568 | * |
1473 | * IOMMU aware version of arm_dma_map_page() | 1569 | * Coherent IOMMU aware version of arm_dma_map_page() |
1474 | */ | 1570 | */ |
1475 | static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | 1571 | static dma_addr_t arm_coherent_iommu_map_page(struct device *dev, struct page *page, |
1476 | unsigned long offset, size_t size, enum dma_data_direction dir, | 1572 | unsigned long offset, size_t size, enum dma_data_direction dir, |
1477 | struct dma_attrs *attrs) | 1573 | struct dma_attrs *attrs) |
1478 | { | 1574 | { |
@@ -1480,9 +1576,6 @@ static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | |||
1480 | dma_addr_t dma_addr; | 1576 | dma_addr_t dma_addr; |
1481 | int ret, len = PAGE_ALIGN(size + offset); | 1577 | int ret, len = PAGE_ALIGN(size + offset); |
1482 | 1578 | ||
1483 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | ||
1484 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
1485 | |||
1486 | dma_addr = __alloc_iova(mapping, len); | 1579 | dma_addr = __alloc_iova(mapping, len); |
1487 | if (dma_addr == DMA_ERROR_CODE) | 1580 | if (dma_addr == DMA_ERROR_CODE) |
1488 | return dma_addr; | 1581 | return dma_addr; |
@@ -1498,6 +1591,51 @@ fail: | |||
1498 | } | 1591 | } |
1499 | 1592 | ||
1500 | /** | 1593 | /** |
1594 | * arm_iommu_map_page | ||
1595 | * @dev: valid struct device pointer | ||
1596 | * @page: page that buffer resides in | ||
1597 | * @offset: offset into page for start of buffer | ||
1598 | * @size: size of buffer to map | ||
1599 | * @dir: DMA transfer direction | ||
1600 | * | ||
1601 | * IOMMU aware version of arm_dma_map_page() | ||
1602 | */ | ||
1603 | static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, | ||
1604 | unsigned long offset, size_t size, enum dma_data_direction dir, | ||
1605 | struct dma_attrs *attrs) | ||
1606 | { | ||
1607 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | ||
1608 | __dma_page_cpu_to_dev(page, offset, size, dir); | ||
1609 | |||
1610 | return arm_coherent_iommu_map_page(dev, page, offset, size, dir, attrs); | ||
1611 | } | ||
1612 | |||
1613 | /** | ||
1614 | * arm_coherent_iommu_unmap_page | ||
1615 | * @dev: valid struct device pointer | ||
1616 | * @handle: DMA address of buffer | ||
1617 | * @size: size of buffer (same as passed to dma_map_page) | ||
1618 | * @dir: DMA transfer direction (same as passed to dma_map_page) | ||
1619 | * | ||
1620 | * Coherent IOMMU aware version of arm_dma_unmap_page() | ||
1621 | */ | ||
1622 | static void arm_coherent_iommu_unmap_page(struct device *dev, dma_addr_t handle, | ||
1623 | size_t size, enum dma_data_direction dir, | ||
1624 | struct dma_attrs *attrs) | ||
1625 | { | ||
1626 | struct dma_iommu_mapping *mapping = dev->archdata.mapping; | ||
1627 | dma_addr_t iova = handle & PAGE_MASK; | ||
1628 | int offset = handle & ~PAGE_MASK; | ||
1629 | int len = PAGE_ALIGN(size + offset); | ||
1630 | |||
1631 | if (!iova) | ||
1632 | return; | ||
1633 | |||
1634 | iommu_unmap(mapping->domain, iova, len); | ||
1635 | __free_iova(mapping, iova, len); | ||
1636 | } | ||
1637 | |||
1638 | /** | ||
1501 | * arm_iommu_unmap_page | 1639 | * arm_iommu_unmap_page |
1502 | * @dev: valid struct device pointer | 1640 | * @dev: valid struct device pointer |
1503 | * @handle: DMA address of buffer | 1641 | * @handle: DMA address of buffer |
@@ -1519,7 +1657,7 @@ static void arm_iommu_unmap_page(struct device *dev, dma_addr_t handle, | |||
1519 | if (!iova) | 1657 | if (!iova) |
1520 | return; | 1658 | return; |
1521 | 1659 | ||
1522 | if (!arch_is_coherent() && !dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) | 1660 | if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) |
1523 | __dma_page_dev_to_cpu(page, offset, size, dir); | 1661 | __dma_page_dev_to_cpu(page, offset, size, dir); |
1524 | 1662 | ||
1525 | iommu_unmap(mapping->domain, iova, len); | 1663 | iommu_unmap(mapping->domain, iova, len); |
@@ -1537,8 +1675,7 @@ static void arm_iommu_sync_single_for_cpu(struct device *dev, | |||
1537 | if (!iova) | 1675 | if (!iova) |
1538 | return; | 1676 | return; |
1539 | 1677 | ||
1540 | if (!arch_is_coherent()) | 1678 | __dma_page_dev_to_cpu(page, offset, size, dir); |
1541 | __dma_page_dev_to_cpu(page, offset, size, dir); | ||
1542 | } | 1679 | } |
1543 | 1680 | ||
1544 | static void arm_iommu_sync_single_for_device(struct device *dev, | 1681 | static void arm_iommu_sync_single_for_device(struct device *dev, |
@@ -1572,6 +1709,19 @@ struct dma_map_ops iommu_ops = { | |||
1572 | .sync_sg_for_device = arm_iommu_sync_sg_for_device, | 1709 | .sync_sg_for_device = arm_iommu_sync_sg_for_device, |
1573 | }; | 1710 | }; |
1574 | 1711 | ||
1712 | struct dma_map_ops iommu_coherent_ops = { | ||
1713 | .alloc = arm_iommu_alloc_attrs, | ||
1714 | .free = arm_iommu_free_attrs, | ||
1715 | .mmap = arm_iommu_mmap_attrs, | ||
1716 | .get_sgtable = arm_iommu_get_sgtable, | ||
1717 | |||
1718 | .map_page = arm_coherent_iommu_map_page, | ||
1719 | .unmap_page = arm_coherent_iommu_unmap_page, | ||
1720 | |||
1721 | .map_sg = arm_coherent_iommu_map_sg, | ||
1722 | .unmap_sg = arm_coherent_iommu_unmap_sg, | ||
1723 | }; | ||
1724 | |||
1575 | /** | 1725 | /** |
1576 | * arm_iommu_create_mapping | 1726 | * arm_iommu_create_mapping |
1577 | * @bus: pointer to the bus holding the client device (for IOMMU calls) | 1727 | * @bus: pointer to the bus holding the client device (for IOMMU calls) |
@@ -1665,7 +1815,7 @@ int arm_iommu_attach_device(struct device *dev, | |||
1665 | dev->archdata.mapping = mapping; | 1815 | dev->archdata.mapping = mapping; |
1666 | set_dma_ops(dev, &iommu_ops); | 1816 | set_dma_ops(dev, &iommu_ops); |
1667 | 1817 | ||
1668 | pr_info("Attached IOMMU controller to %s device.\n", dev_name(dev)); | 1818 | pr_debug("Attached IOMMU controller to %s device.\n", dev_name(dev)); |
1669 | return 0; | 1819 | return 0; |
1670 | } | 1820 | } |
1671 | 1821 | ||
diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c index 18144e6a3115..941dfb9e9a78 100644 --- a/arch/arm/mm/mmu.c +++ b/arch/arm/mm/mmu.c | |||
@@ -423,17 +423,6 @@ static void __init build_mem_type_table(void) | |||
423 | vecs_pgprot = kern_pgprot = user_pgprot = cp->pte; | 423 | vecs_pgprot = kern_pgprot = user_pgprot = cp->pte; |
424 | 424 | ||
425 | /* | 425 | /* |
426 | * Enable CPU-specific coherency if supported. | ||
427 | * (Only available on XSC3 at the moment.) | ||
428 | */ | ||
429 | if (arch_is_coherent() && cpu_is_xsc3()) { | ||
430 | mem_types[MT_MEMORY].prot_sect |= PMD_SECT_S; | ||
431 | mem_types[MT_MEMORY].prot_pte |= L_PTE_SHARED; | ||
432 | mem_types[MT_MEMORY_DMA_READY].prot_pte |= L_PTE_SHARED; | ||
433 | mem_types[MT_MEMORY_NONCACHED].prot_sect |= PMD_SECT_S; | ||
434 | mem_types[MT_MEMORY_NONCACHED].prot_pte |= L_PTE_SHARED; | ||
435 | } | ||
436 | /* | ||
437 | * ARMv6 and above have extended page tables. | 426 | * ARMv6 and above have extended page tables. |
438 | */ | 427 | */ |
439 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { | 428 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { |
diff --git a/drivers/base/dma-contiguous.c b/drivers/base/dma-contiguous.c index 34d94c762a1e..9a1469474f55 100644 --- a/drivers/base/dma-contiguous.c +++ b/drivers/base/dma-contiguous.c | |||
@@ -315,6 +315,7 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
315 | { | 315 | { |
316 | unsigned long mask, pfn, pageno, start = 0; | 316 | unsigned long mask, pfn, pageno, start = 0; |
317 | struct cma *cma = dev_get_cma_area(dev); | 317 | struct cma *cma = dev_get_cma_area(dev); |
318 | struct page *page = NULL; | ||
318 | int ret; | 319 | int ret; |
319 | 320 | ||
320 | if (!cma || !cma->count) | 321 | if (!cma || !cma->count) |
@@ -336,18 +337,17 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
336 | for (;;) { | 337 | for (;;) { |
337 | pageno = bitmap_find_next_zero_area(cma->bitmap, cma->count, | 338 | pageno = bitmap_find_next_zero_area(cma->bitmap, cma->count, |
338 | start, count, mask); | 339 | start, count, mask); |
339 | if (pageno >= cma->count) { | 340 | if (pageno >= cma->count) |
340 | ret = -ENOMEM; | 341 | break; |
341 | goto error; | ||
342 | } | ||
343 | 342 | ||
344 | pfn = cma->base_pfn + pageno; | 343 | pfn = cma->base_pfn + pageno; |
345 | ret = alloc_contig_range(pfn, pfn + count, MIGRATE_CMA); | 344 | ret = alloc_contig_range(pfn, pfn + count, MIGRATE_CMA); |
346 | if (ret == 0) { | 345 | if (ret == 0) { |
347 | bitmap_set(cma->bitmap, pageno, count); | 346 | bitmap_set(cma->bitmap, pageno, count); |
347 | page = pfn_to_page(pfn); | ||
348 | break; | 348 | break; |
349 | } else if (ret != -EBUSY) { | 349 | } else if (ret != -EBUSY) { |
350 | goto error; | 350 | break; |
351 | } | 351 | } |
352 | pr_debug("%s(): memory range at %p is busy, retrying\n", | 352 | pr_debug("%s(): memory range at %p is busy, retrying\n", |
353 | __func__, pfn_to_page(pfn)); | 353 | __func__, pfn_to_page(pfn)); |
@@ -356,12 +356,8 @@ struct page *dma_alloc_from_contiguous(struct device *dev, int count, | |||
356 | } | 356 | } |
357 | 357 | ||
358 | mutex_unlock(&cma_mutex); | 358 | mutex_unlock(&cma_mutex); |
359 | 359 | pr_debug("%s(): returned %p\n", __func__, page); | |
360 | pr_debug("%s(): returned %p\n", __func__, pfn_to_page(pfn)); | 360 | return page; |
361 | return pfn_to_page(pfn); | ||
362 | error: | ||
363 | mutex_unlock(&cma_mutex); | ||
364 | return NULL; | ||
365 | } | 361 | } |
366 | 362 | ||
367 | /** | 363 | /** |