diff options
| author | Linus Torvalds <torvalds@linux-foundation.org> | 2010-03-07 18:47:19 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@linux-foundation.org> | 2010-03-07 18:47:19 -0500 |
| commit | 4a31c08d2fecc74a630653828f5388fbb037f8c2 (patch) | |
| tree | c3baf80157bab2cf6bdf3d26772001e43233aad6 | |
| parent | 2ddb3b15f1b46836c61cfac5b00d8f08a24236e6 (diff) | |
| parent | 0272282f7cffb469cd2676dcb6e58bc942fcf8a8 (diff) | |
Merge git://git.kernel.org/pub/scm/linux/kernel/git/lethal/sh-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/lethal/sh-2.6: (26 commits)
sh: Convert sh to use read/update_persistent_clock
sh: Move PMB debugfs entry initialization to later stage
sh: Fix up flush_cache_vmap() on SMP.
sh: fix up MMU reset with variable PMB mapping sizes.
sh: establish PMB mappings for NUMA nodes.
sh: check for existing mappings for bolted PMB entries.
sh: fixed virt/phys mapping helpers for PMB.
sh: make pmb iomapping configurable.
sh: reworked dynamic PMB mapping.
sh: Fix up cpumask_of_pcibus() for the NUMA build.
serial: sh-sci: Tidy up build warnings.
sh: Fix up ctrl_read/write stragglers in migor setup.
serial: sh-sci: Add DMA support.
dmaengine: shdma: extend .device_terminate_all() to record partial transfer
sh: merge sh7722 and sh7724 DMA register definitions
sh: activate runtime PM for dmaengine on sh7722 and sh7724
dmaengine: shdma: add runtime PM support.
dmaengine: shdma: separate DMA headers.
dmaengine: shdma: convert to platform device resources
dmaengine: shdma: fix DMA error handling.
...
35 files changed, 2201 insertions, 728 deletions
diff --git a/arch/sh/boards/mach-migor/setup.c b/arch/sh/boards/mach-migor/setup.c index be300aaca6fe..7da0fc94a01e 100644 --- a/arch/sh/boards/mach-migor/setup.c +++ b/arch/sh/boards/mach-migor/setup.c | |||
| @@ -419,6 +419,9 @@ static struct i2c_board_info migor_i2c_devices[] = { | |||
| 419 | I2C_BOARD_INFO("migor_ts", 0x51), | 419 | I2C_BOARD_INFO("migor_ts", 0x51), |
| 420 | .irq = 38, /* IRQ6 */ | 420 | .irq = 38, /* IRQ6 */ |
| 421 | }, | 421 | }, |
| 422 | { | ||
| 423 | I2C_BOARD_INFO("wm8978", 0x1a), | ||
| 424 | }, | ||
| 422 | }; | 425 | }; |
| 423 | 426 | ||
| 424 | static struct i2c_board_info migor_i2c_camera[] = { | 427 | static struct i2c_board_info migor_i2c_camera[] = { |
| @@ -619,6 +622,19 @@ static int __init migor_devices_setup(void) | |||
| 619 | 622 | ||
| 620 | platform_resource_setup_memory(&migor_ceu_device, "ceu", 4 << 20); | 623 | platform_resource_setup_memory(&migor_ceu_device, "ceu", 4 << 20); |
| 621 | 624 | ||
| 625 | /* SIU: Port B */ | ||
| 626 | gpio_request(GPIO_FN_SIUBOLR, NULL); | ||
| 627 | gpio_request(GPIO_FN_SIUBOBT, NULL); | ||
| 628 | gpio_request(GPIO_FN_SIUBISLD, NULL); | ||
| 629 | gpio_request(GPIO_FN_SIUBOSLD, NULL); | ||
| 630 | gpio_request(GPIO_FN_SIUMCKB, NULL); | ||
| 631 | |||
| 632 | /* | ||
| 633 | * The original driver sets SIUB OLR/OBT, ILR/IBT, and SIUA OLR/OBT to | ||
| 634 | * output. Need only SIUB, set to output for master mode (table 34.2) | ||
| 635 | */ | ||
| 636 | __raw_writew(__raw_readw(PORT_MSELCRA) | 1, PORT_MSELCRA); | ||
| 637 | |||
| 622 | i2c_register_board_info(0, migor_i2c_devices, | 638 | i2c_register_board_info(0, migor_i2c_devices, |
| 623 | ARRAY_SIZE(migor_i2c_devices)); | 639 | ARRAY_SIZE(migor_i2c_devices)); |
| 624 | 640 | ||
diff --git a/arch/sh/boot/compressed/cache.c b/arch/sh/boot/compressed/cache.c index e27fc74f228c..d0b77b68a4d0 100644 --- a/arch/sh/boot/compressed/cache.c +++ b/arch/sh/boot/compressed/cache.c | |||
| @@ -5,7 +5,7 @@ int cache_control(unsigned int command) | |||
| 5 | 5 | ||
| 6 | for (i = 0; i < (32 * 1024); i += 32) { | 6 | for (i = 0; i < (32 * 1024); i += 32) { |
| 7 | (void)*p; | 7 | (void)*p; |
| 8 | p += (32 / sizeof (int)); | 8 | p += (32 / sizeof(int)); |
| 9 | } | 9 | } |
| 10 | 10 | ||
| 11 | return 0; | 11 | return 0; |
diff --git a/arch/sh/include/asm/cacheflush.h b/arch/sh/include/asm/cacheflush.h index da3ebec921a7..1f4e562c5e8c 100644 --- a/arch/sh/include/asm/cacheflush.h +++ b/arch/sh/include/asm/cacheflush.h | |||
| @@ -86,8 +86,8 @@ extern void copy_from_user_page(struct vm_area_struct *vma, | |||
| 86 | struct page *page, unsigned long vaddr, void *dst, const void *src, | 86 | struct page *page, unsigned long vaddr, void *dst, const void *src, |
| 87 | unsigned long len); | 87 | unsigned long len); |
| 88 | 88 | ||
| 89 | #define flush_cache_vmap(start, end) flush_cache_all() | 89 | #define flush_cache_vmap(start, end) local_flush_cache_all(NULL) |
| 90 | #define flush_cache_vunmap(start, end) flush_cache_all() | 90 | #define flush_cache_vunmap(start, end) local_flush_cache_all(NULL) |
| 91 | 91 | ||
| 92 | #define flush_dcache_mmap_lock(mapping) do { } while (0) | 92 | #define flush_dcache_mmap_lock(mapping) do { } while (0) |
| 93 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) | 93 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) |
diff --git a/arch/sh/include/asm/dma-register.h b/arch/sh/include/asm/dma-register.h new file mode 100644 index 000000000000..51cd78feacff --- /dev/null +++ b/arch/sh/include/asm/dma-register.h | |||
| @@ -0,0 +1,51 @@ | |||
| 1 | /* | ||
| 2 | * Common header for the legacy SH DMA driver and the new dmaengine driver | ||
| 3 | * | ||
| 4 | * extracted from arch/sh/include/asm/dma-sh.h: | ||
| 5 | * | ||
| 6 | * Copyright (C) 2000 Takashi YOSHII | ||
| 7 | * Copyright (C) 2003 Paul Mundt | ||
| 8 | * | ||
| 9 | * This file is subject to the terms and conditions of the GNU General Public | ||
| 10 | * License. See the file "COPYING" in the main directory of this archive | ||
| 11 | * for more details. | ||
| 12 | */ | ||
| 13 | #ifndef DMA_REGISTER_H | ||
| 14 | #define DMA_REGISTER_H | ||
| 15 | |||
| 16 | /* DMA register */ | ||
| 17 | #define SAR 0x00 | ||
| 18 | #define DAR 0x04 | ||
| 19 | #define TCR 0x08 | ||
| 20 | #define CHCR 0x0C | ||
| 21 | #define DMAOR 0x40 | ||
| 22 | |||
| 23 | /* DMAOR definitions */ | ||
| 24 | #define DMAOR_AE 0x00000004 | ||
| 25 | #define DMAOR_NMIF 0x00000002 | ||
| 26 | #define DMAOR_DME 0x00000001 | ||
| 27 | |||
| 28 | /* Definitions for the SuperH DMAC */ | ||
| 29 | #define REQ_L 0x00000000 | ||
| 30 | #define REQ_E 0x00080000 | ||
| 31 | #define RACK_H 0x00000000 | ||
| 32 | #define RACK_L 0x00040000 | ||
| 33 | #define ACK_R 0x00000000 | ||
| 34 | #define ACK_W 0x00020000 | ||
| 35 | #define ACK_H 0x00000000 | ||
| 36 | #define ACK_L 0x00010000 | ||
| 37 | #define DM_INC 0x00004000 | ||
| 38 | #define DM_DEC 0x00008000 | ||
| 39 | #define DM_FIX 0x0000c000 | ||
| 40 | #define SM_INC 0x00001000 | ||
| 41 | #define SM_DEC 0x00002000 | ||
| 42 | #define SM_FIX 0x00003000 | ||
| 43 | #define RS_IN 0x00000200 | ||
| 44 | #define RS_OUT 0x00000300 | ||
| 45 | #define TS_BLK 0x00000040 | ||
| 46 | #define TM_BUR 0x00000020 | ||
| 47 | #define CHCR_DE 0x00000001 | ||
| 48 | #define CHCR_TE 0x00000002 | ||
| 49 | #define CHCR_IE 0x00000004 | ||
| 50 | |||
| 51 | #endif | ||
diff --git a/arch/sh/include/asm/dma-sh.h b/arch/sh/include/asm/dma-sh.h index e934a2e66651..f3acb8e34c6b 100644 --- a/arch/sh/include/asm/dma-sh.h +++ b/arch/sh/include/asm/dma-sh.h | |||
| @@ -11,7 +11,8 @@ | |||
| 11 | #ifndef __DMA_SH_H | 11 | #ifndef __DMA_SH_H |
| 12 | #define __DMA_SH_H | 12 | #define __DMA_SH_H |
| 13 | 13 | ||
| 14 | #include <asm/dma.h> | 14 | #include <asm/dma-register.h> |
| 15 | #include <cpu/dma-register.h> | ||
| 15 | #include <cpu/dma.h> | 16 | #include <cpu/dma.h> |
| 16 | 17 | ||
| 17 | /* DMAOR contorl: The DMAOR access size is different by CPU.*/ | 18 | /* DMAOR contorl: The DMAOR access size is different by CPU.*/ |
| @@ -53,34 +54,6 @@ static int dmte_irq_map[] __maybe_unused = { | |||
| 53 | #endif | 54 | #endif |
| 54 | }; | 55 | }; |
| 55 | 56 | ||
| 56 | /* Definitions for the SuperH DMAC */ | ||
| 57 | #define REQ_L 0x00000000 | ||
| 58 | #define REQ_E 0x00080000 | ||
| 59 | #define RACK_H 0x00000000 | ||
| 60 | #define RACK_L 0x00040000 | ||
| 61 | #define ACK_R 0x00000000 | ||
| 62 | #define ACK_W 0x00020000 | ||
| 63 | #define ACK_H 0x00000000 | ||
| 64 | #define ACK_L 0x00010000 | ||
| 65 | #define DM_INC 0x00004000 | ||
| 66 | #define DM_DEC 0x00008000 | ||
| 67 | #define DM_FIX 0x0000c000 | ||
| 68 | #define SM_INC 0x00001000 | ||
| 69 | #define SM_DEC 0x00002000 | ||
| 70 | #define SM_FIX 0x00003000 | ||
| 71 | #define RS_IN 0x00000200 | ||
| 72 | #define RS_OUT 0x00000300 | ||
| 73 | #define TS_BLK 0x00000040 | ||
| 74 | #define TM_BUR 0x00000020 | ||
| 75 | #define CHCR_DE 0x00000001 | ||
| 76 | #define CHCR_TE 0x00000002 | ||
| 77 | #define CHCR_IE 0x00000004 | ||
| 78 | |||
| 79 | /* DMAOR definitions */ | ||
| 80 | #define DMAOR_AE 0x00000004 | ||
| 81 | #define DMAOR_NMIF 0x00000002 | ||
| 82 | #define DMAOR_DME 0x00000001 | ||
| 83 | |||
| 84 | /* | 57 | /* |
| 85 | * Define the default configuration for dual address memory-memory transfer. | 58 | * Define the default configuration for dual address memory-memory transfer. |
| 86 | * The 0x400 value represents auto-request, external->external. | 59 | * The 0x400 value represents auto-request, external->external. |
| @@ -111,61 +84,4 @@ static u32 dma_base_addr[] __maybe_unused = { | |||
| 111 | #endif | 84 | #endif |
| 112 | }; | 85 | }; |
| 113 | 86 | ||
| 114 | /* DMA register */ | ||
| 115 | #define SAR 0x00 | ||
| 116 | #define DAR 0x04 | ||
| 117 | #define TCR 0x08 | ||
| 118 | #define CHCR 0x0C | ||
| 119 | #define DMAOR 0x40 | ||
| 120 | |||
| 121 | /* | ||
| 122 | * for dma engine | ||
| 123 | * | ||
| 124 | * SuperH DMA mode | ||
| 125 | */ | ||
| 126 | #define SHDMA_MIX_IRQ (1 << 1) | ||
| 127 | #define SHDMA_DMAOR1 (1 << 2) | ||
| 128 | #define SHDMA_DMAE1 (1 << 3) | ||
| 129 | |||
| 130 | enum sh_dmae_slave_chan_id { | ||
| 131 | SHDMA_SLAVE_SCIF0_TX, | ||
| 132 | SHDMA_SLAVE_SCIF0_RX, | ||
| 133 | SHDMA_SLAVE_SCIF1_TX, | ||
| 134 | SHDMA_SLAVE_SCIF1_RX, | ||
| 135 | SHDMA_SLAVE_SCIF2_TX, | ||
| 136 | SHDMA_SLAVE_SCIF2_RX, | ||
| 137 | SHDMA_SLAVE_SCIF3_TX, | ||
| 138 | SHDMA_SLAVE_SCIF3_RX, | ||
| 139 | SHDMA_SLAVE_SCIF4_TX, | ||
| 140 | SHDMA_SLAVE_SCIF4_RX, | ||
| 141 | SHDMA_SLAVE_SCIF5_TX, | ||
| 142 | SHDMA_SLAVE_SCIF5_RX, | ||
| 143 | SHDMA_SLAVE_SIUA_TX, | ||
| 144 | SHDMA_SLAVE_SIUA_RX, | ||
| 145 | SHDMA_SLAVE_SIUB_TX, | ||
| 146 | SHDMA_SLAVE_SIUB_RX, | ||
| 147 | SHDMA_SLAVE_NUMBER, /* Must stay last */ | ||
| 148 | }; | ||
| 149 | |||
| 150 | struct sh_dmae_slave_config { | ||
| 151 | enum sh_dmae_slave_chan_id slave_id; | ||
| 152 | dma_addr_t addr; | ||
| 153 | u32 chcr; | ||
| 154 | char mid_rid; | ||
| 155 | }; | ||
| 156 | |||
| 157 | struct sh_dmae_pdata { | ||
| 158 | unsigned int mode; | ||
| 159 | struct sh_dmae_slave_config *config; | ||
| 160 | int config_num; | ||
| 161 | }; | ||
| 162 | |||
| 163 | struct device; | ||
| 164 | |||
| 165 | struct sh_dmae_slave { | ||
| 166 | enum sh_dmae_slave_chan_id slave_id; /* Set by the platform */ | ||
| 167 | struct device *dma_dev; /* Set by the platform */ | ||
| 168 | struct sh_dmae_slave_config *config; /* Set by the driver */ | ||
| 169 | }; | ||
| 170 | |||
| 171 | #endif /* __DMA_SH_H */ | 87 | #endif /* __DMA_SH_H */ |
diff --git a/arch/sh/include/asm/dmaengine.h b/arch/sh/include/asm/dmaengine.h new file mode 100644 index 000000000000..bf2f30cf0a27 --- /dev/null +++ b/arch/sh/include/asm/dmaengine.h | |||
| @@ -0,0 +1,93 @@ | |||
| 1 | /* | ||
| 2 | * Header for the new SH dmaengine driver | ||
| 3 | * | ||
| 4 | * Copyright (C) 2010 Guennadi Liakhovetski <g.liakhovetski@gmx.de> | ||
| 5 | * | ||
| 6 | * This program is free software; you can redistribute it and/or modify | ||
| 7 | * it under the terms of the GNU General Public License version 2 as | ||
| 8 | * published by the Free Software Foundation. | ||
| 9 | */ | ||
| 10 | #ifndef ASM_DMAENGINE_H | ||
| 11 | #define ASM_DMAENGINE_H | ||
| 12 | |||
| 13 | #include <linux/dmaengine.h> | ||
| 14 | #include <linux/list.h> | ||
| 15 | |||
| 16 | #include <asm/dma-register.h> | ||
| 17 | |||
| 18 | #define SH_DMAC_MAX_CHANNELS 6 | ||
| 19 | |||
| 20 | enum sh_dmae_slave_chan_id { | ||
| 21 | SHDMA_SLAVE_SCIF0_TX, | ||
| 22 | SHDMA_SLAVE_SCIF0_RX, | ||
| 23 | SHDMA_SLAVE_SCIF1_TX, | ||
| 24 | SHDMA_SLAVE_SCIF1_RX, | ||
| 25 | SHDMA_SLAVE_SCIF2_TX, | ||
| 26 | SHDMA_SLAVE_SCIF2_RX, | ||
| 27 | SHDMA_SLAVE_SCIF3_TX, | ||
| 28 | SHDMA_SLAVE_SCIF3_RX, | ||
| 29 | SHDMA_SLAVE_SCIF4_TX, | ||
| 30 | SHDMA_SLAVE_SCIF4_RX, | ||
| 31 | SHDMA_SLAVE_SCIF5_TX, | ||
| 32 | SHDMA_SLAVE_SCIF5_RX, | ||
| 33 | SHDMA_SLAVE_SIUA_TX, | ||
| 34 | SHDMA_SLAVE_SIUA_RX, | ||
| 35 | SHDMA_SLAVE_SIUB_TX, | ||
| 36 | SHDMA_SLAVE_SIUB_RX, | ||
| 37 | SHDMA_SLAVE_NUMBER, /* Must stay last */ | ||
| 38 | }; | ||
| 39 | |||
| 40 | struct sh_dmae_slave_config { | ||
| 41 | enum sh_dmae_slave_chan_id slave_id; | ||
| 42 | dma_addr_t addr; | ||
| 43 | u32 chcr; | ||
| 44 | char mid_rid; | ||
| 45 | }; | ||
| 46 | |||
| 47 | struct sh_dmae_channel { | ||
| 48 | unsigned int offset; | ||
| 49 | unsigned int dmars; | ||
| 50 | unsigned int dmars_bit; | ||
| 51 | }; | ||
| 52 | |||
| 53 | struct sh_dmae_pdata { | ||
| 54 | struct sh_dmae_slave_config *slave; | ||
| 55 | int slave_num; | ||
| 56 | struct sh_dmae_channel *channel; | ||
| 57 | int channel_num; | ||
| 58 | unsigned int ts_low_shift; | ||
| 59 | unsigned int ts_low_mask; | ||
| 60 | unsigned int ts_high_shift; | ||
| 61 | unsigned int ts_high_mask; | ||
| 62 | unsigned int *ts_shift; | ||
| 63 | int ts_shift_num; | ||
| 64 | u16 dmaor_init; | ||
| 65 | }; | ||
| 66 | |||
| 67 | struct device; | ||
| 68 | |||
| 69 | /* Used by slave DMA clients to request DMA to/from a specific peripheral */ | ||
| 70 | struct sh_dmae_slave { | ||
| 71 | enum sh_dmae_slave_chan_id slave_id; /* Set by the platform */ | ||
| 72 | struct device *dma_dev; /* Set by the platform */ | ||
| 73 | struct sh_dmae_slave_config *config; /* Set by the driver */ | ||
| 74 | }; | ||
| 75 | |||
| 76 | struct sh_dmae_regs { | ||
| 77 | u32 sar; /* SAR / source address */ | ||
| 78 | u32 dar; /* DAR / destination address */ | ||
| 79 | u32 tcr; /* TCR / transfer count */ | ||
| 80 | }; | ||
| 81 | |||
| 82 | struct sh_desc { | ||
| 83 | struct sh_dmae_regs hw; | ||
| 84 | struct list_head node; | ||
| 85 | struct dma_async_tx_descriptor async_tx; | ||
| 86 | enum dma_data_direction direction; | ||
| 87 | dma_cookie_t cookie; | ||
| 88 | size_t partial; | ||
| 89 | int chunks; | ||
| 90 | int mark; | ||
| 91 | }; | ||
| 92 | |||
| 93 | #endif | ||
diff --git a/arch/sh/include/asm/io.h b/arch/sh/include/asm/io.h index 7dab7b23a5ec..f689554e17c1 100644 --- a/arch/sh/include/asm/io.h +++ b/arch/sh/include/asm/io.h | |||
| @@ -291,21 +291,21 @@ unsigned long long poke_real_address_q(unsigned long long addr, | |||
| 291 | * doesn't exist, so everything must go through page tables. | 291 | * doesn't exist, so everything must go through page tables. |
| 292 | */ | 292 | */ |
| 293 | #ifdef CONFIG_MMU | 293 | #ifdef CONFIG_MMU |
| 294 | void __iomem *__ioremap_caller(unsigned long offset, unsigned long size, | 294 | void __iomem *__ioremap_caller(phys_addr_t offset, unsigned long size, |
| 295 | pgprot_t prot, void *caller); | 295 | pgprot_t prot, void *caller); |
| 296 | void __iounmap(void __iomem *addr); | 296 | void __iounmap(void __iomem *addr); |
| 297 | 297 | ||
| 298 | static inline void __iomem * | 298 | static inline void __iomem * |
| 299 | __ioremap(unsigned long offset, unsigned long size, pgprot_t prot) | 299 | __ioremap(phys_addr_t offset, unsigned long size, pgprot_t prot) |
| 300 | { | 300 | { |
| 301 | return __ioremap_caller(offset, size, prot, __builtin_return_address(0)); | 301 | return __ioremap_caller(offset, size, prot, __builtin_return_address(0)); |
| 302 | } | 302 | } |
| 303 | 303 | ||
| 304 | static inline void __iomem * | 304 | static inline void __iomem * |
| 305 | __ioremap_29bit(unsigned long offset, unsigned long size, pgprot_t prot) | 305 | __ioremap_29bit(phys_addr_t offset, unsigned long size, pgprot_t prot) |
| 306 | { | 306 | { |
| 307 | #ifdef CONFIG_29BIT | 307 | #ifdef CONFIG_29BIT |
| 308 | unsigned long last_addr = offset + size - 1; | 308 | phys_addr_t last_addr = offset + size - 1; |
| 309 | 309 | ||
| 310 | /* | 310 | /* |
| 311 | * For P1 and P2 space this is trivial, as everything is already | 311 | * For P1 and P2 space this is trivial, as everything is already |
| @@ -329,7 +329,7 @@ __ioremap_29bit(unsigned long offset, unsigned long size, pgprot_t prot) | |||
| 329 | } | 329 | } |
| 330 | 330 | ||
| 331 | static inline void __iomem * | 331 | static inline void __iomem * |
| 332 | __ioremap_mode(unsigned long offset, unsigned long size, pgprot_t prot) | 332 | __ioremap_mode(phys_addr_t offset, unsigned long size, pgprot_t prot) |
| 333 | { | 333 | { |
| 334 | void __iomem *ret; | 334 | void __iomem *ret; |
| 335 | 335 | ||
| @@ -349,35 +349,32 @@ __ioremap_mode(unsigned long offset, unsigned long size, pgprot_t prot) | |||
| 349 | #define __iounmap(addr) do { } while (0) | 349 | #define __iounmap(addr) do { } while (0) |
| 350 | #endif /* CONFIG_MMU */ | 350 | #endif /* CONFIG_MMU */ |
| 351 | 351 | ||
| 352 | static inline void __iomem * | 352 | static inline void __iomem *ioremap(phys_addr_t offset, unsigned long size) |
| 353 | ioremap(unsigned long offset, unsigned long size) | ||
| 354 | { | 353 | { |
| 355 | return __ioremap_mode(offset, size, PAGE_KERNEL_NOCACHE); | 354 | return __ioremap_mode(offset, size, PAGE_KERNEL_NOCACHE); |
| 356 | } | 355 | } |
| 357 | 356 | ||
| 358 | static inline void __iomem * | 357 | static inline void __iomem * |
| 359 | ioremap_cache(unsigned long offset, unsigned long size) | 358 | ioremap_cache(phys_addr_t offset, unsigned long size) |
| 360 | { | 359 | { |
| 361 | return __ioremap_mode(offset, size, PAGE_KERNEL); | 360 | return __ioremap_mode(offset, size, PAGE_KERNEL); |
| 362 | } | 361 | } |
| 363 | 362 | ||
| 364 | #ifdef CONFIG_HAVE_IOREMAP_PROT | 363 | #ifdef CONFIG_HAVE_IOREMAP_PROT |
| 365 | static inline void __iomem * | 364 | static inline void __iomem * |
| 366 | ioremap_prot(resource_size_t offset, unsigned long size, unsigned long flags) | 365 | ioremap_prot(phys_addr_t offset, unsigned long size, unsigned long flags) |
| 367 | { | 366 | { |
| 368 | return __ioremap_mode(offset, size, __pgprot(flags)); | 367 | return __ioremap_mode(offset, size, __pgprot(flags)); |
| 369 | } | 368 | } |
| 370 | #endif | 369 | #endif |
| 371 | 370 | ||
| 372 | #ifdef CONFIG_IOREMAP_FIXED | 371 | #ifdef CONFIG_IOREMAP_FIXED |
| 373 | extern void __iomem *ioremap_fixed(resource_size_t, unsigned long, | 372 | extern void __iomem *ioremap_fixed(phys_addr_t, unsigned long, pgprot_t); |
| 374 | unsigned long, pgprot_t); | ||
| 375 | extern int iounmap_fixed(void __iomem *); | 373 | extern int iounmap_fixed(void __iomem *); |
| 376 | extern void ioremap_fixed_init(void); | 374 | extern void ioremap_fixed_init(void); |
| 377 | #else | 375 | #else |
| 378 | static inline void __iomem * | 376 | static inline void __iomem * |
| 379 | ioremap_fixed(resource_size_t phys_addr, unsigned long offset, | 377 | ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) |
| 380 | unsigned long size, pgprot_t prot) | ||
| 381 | { | 378 | { |
| 382 | BUG(); | 379 | BUG(); |
| 383 | return NULL; | 380 | return NULL; |
diff --git a/arch/sh/include/asm/mmu.h b/arch/sh/include/asm/mmu.h index 15a05b615ba7..19fe84550b49 100644 --- a/arch/sh/include/asm/mmu.h +++ b/arch/sh/include/asm/mmu.h | |||
| @@ -55,19 +55,29 @@ typedef struct { | |||
| 55 | 55 | ||
| 56 | #ifdef CONFIG_PMB | 56 | #ifdef CONFIG_PMB |
| 57 | /* arch/sh/mm/pmb.c */ | 57 | /* arch/sh/mm/pmb.c */ |
| 58 | long pmb_remap(unsigned long virt, unsigned long phys, | ||
| 59 | unsigned long size, pgprot_t prot); | ||
| 60 | void pmb_unmap(unsigned long addr); | ||
| 61 | void pmb_init(void); | ||
| 62 | bool __in_29bit_mode(void); | 58 | bool __in_29bit_mode(void); |
| 59 | |||
| 60 | void pmb_init(void); | ||
| 61 | int pmb_bolt_mapping(unsigned long virt, phys_addr_t phys, | ||
| 62 | unsigned long size, pgprot_t prot); | ||
| 63 | void __iomem *pmb_remap_caller(phys_addr_t phys, unsigned long size, | ||
| 64 | pgprot_t prot, void *caller); | ||
| 65 | int pmb_unmap(void __iomem *addr); | ||
| 66 | |||
| 63 | #else | 67 | #else |
| 64 | static inline long pmb_remap(unsigned long virt, unsigned long phys, | 68 | |
| 65 | unsigned long size, pgprot_t prot) | 69 | static inline void __iomem * |
| 70 | pmb_remap_caller(phys_addr_t phys, unsigned long size, | ||
| 71 | pgprot_t prot, void *caller) | ||
| 72 | { | ||
| 73 | return NULL; | ||
| 74 | } | ||
| 75 | |||
| 76 | static inline int pmb_unmap(void __iomem *addr) | ||
| 66 | { | 77 | { |
| 67 | return -EINVAL; | 78 | return -EINVAL; |
| 68 | } | 79 | } |
| 69 | 80 | ||
| 70 | #define pmb_unmap(addr) do { } while (0) | ||
| 71 | #define pmb_init(addr) do { } while (0) | 81 | #define pmb_init(addr) do { } while (0) |
| 72 | 82 | ||
| 73 | #ifdef CONFIG_29BIT | 83 | #ifdef CONFIG_29BIT |
| @@ -77,6 +87,13 @@ static inline long pmb_remap(unsigned long virt, unsigned long phys, | |||
| 77 | #endif | 87 | #endif |
| 78 | 88 | ||
| 79 | #endif /* CONFIG_PMB */ | 89 | #endif /* CONFIG_PMB */ |
| 90 | |||
| 91 | static inline void __iomem * | ||
| 92 | pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot) | ||
| 93 | { | ||
| 94 | return pmb_remap_caller(phys, size, prot, __builtin_return_address(0)); | ||
| 95 | } | ||
| 96 | |||
| 80 | #endif /* __ASSEMBLY__ */ | 97 | #endif /* __ASSEMBLY__ */ |
| 81 | 98 | ||
| 82 | #endif /* __MMU_H */ | 99 | #endif /* __MMU_H */ |
diff --git a/arch/sh/include/asm/siu.h b/arch/sh/include/asm/siu.h index 57565a3b551f..f1b1e6944a5f 100644 --- a/arch/sh/include/asm/siu.h +++ b/arch/sh/include/asm/siu.h | |||
| @@ -11,7 +11,7 @@ | |||
| 11 | #ifndef ASM_SIU_H | 11 | #ifndef ASM_SIU_H |
| 12 | #define ASM_SIU_H | 12 | #define ASM_SIU_H |
| 13 | 13 | ||
| 14 | #include <asm/dma-sh.h> | 14 | #include <asm/dmaengine.h> |
| 15 | 15 | ||
| 16 | struct device; | 16 | struct device; |
| 17 | 17 | ||
diff --git a/arch/sh/include/asm/topology.h b/arch/sh/include/asm/topology.h index 37cdadd975ac..88e734069fa6 100644 --- a/arch/sh/include/asm/topology.h +++ b/arch/sh/include/asm/topology.h | |||
| @@ -35,7 +35,7 @@ | |||
| 35 | 35 | ||
| 36 | #define pcibus_to_node(bus) ((void)(bus), -1) | 36 | #define pcibus_to_node(bus) ((void)(bus), -1) |
| 37 | #define cpumask_of_pcibus(bus) (pcibus_to_node(bus) == -1 ? \ | 37 | #define cpumask_of_pcibus(bus) (pcibus_to_node(bus) == -1 ? \ |
| 38 | CPU_MASK_ALL_PTR : \ | 38 | cpu_all_mask : \ |
| 39 | cpumask_of_node(pcibus_to_node(bus))) | 39 | cpumask_of_node(pcibus_to_node(bus))) |
| 40 | 40 | ||
| 41 | #endif | 41 | #endif |
diff --git a/arch/sh/include/cpu-sh3/cpu/dma-register.h b/arch/sh/include/cpu-sh3/cpu/dma-register.h new file mode 100644 index 000000000000..2349e488c9a6 --- /dev/null +++ b/arch/sh/include/cpu-sh3/cpu/dma-register.h | |||
| @@ -0,0 +1,41 @@ | |||
| 1 | /* | ||
| 2 | * SH3 CPU-specific DMA definitions, used by both DMA drivers | ||
| 3 | * | ||
| 4 | * Copyright (C) 2010 Guennadi Liakhovetski <g.liakhovetski@gmx.de> | ||
| 5 | * | ||
| 6 | * This program is free software; you can redistribute it and/or modify | ||
| 7 | * it under the terms of the GNU General Public License version 2 as | ||
| 8 | * published by the Free Software Foundation. | ||
| 9 | */ | ||
| 10 | #ifndef CPU_DMA_REGISTER_H | ||
| 11 | #define CPU_DMA_REGISTER_H | ||
| 12 | |||
| 13 | #define CHCR_TS_LOW_MASK 0x18 | ||
| 14 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 15 | #define CHCR_TS_HIGH_MASK 0 | ||
| 16 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 17 | |||
| 18 | #define DMAOR_INIT DMAOR_DME | ||
| 19 | |||
| 20 | /* | ||
| 21 | * The SuperH DMAC supports a number of transmit sizes, we list them here, | ||
| 22 | * with their respective values as they appear in the CHCR registers. | ||
| 23 | */ | ||
| 24 | enum { | ||
| 25 | XMIT_SZ_8BIT, | ||
| 26 | XMIT_SZ_16BIT, | ||
| 27 | XMIT_SZ_32BIT, | ||
| 28 | XMIT_SZ_128BIT, | ||
| 29 | }; | ||
| 30 | |||
| 31 | /* log2(size / 8) - used to calculate number of transfers */ | ||
| 32 | #define TS_SHIFT { \ | ||
| 33 | [XMIT_SZ_8BIT] = 0, \ | ||
| 34 | [XMIT_SZ_16BIT] = 1, \ | ||
| 35 | [XMIT_SZ_32BIT] = 2, \ | ||
| 36 | [XMIT_SZ_128BIT] = 4, \ | ||
| 37 | } | ||
| 38 | |||
| 39 | #define TS_INDEX2VAL(i) (((i) & 3) << CHCR_TS_LOW_SHIFT) | ||
| 40 | |||
| 41 | #endif | ||
diff --git a/arch/sh/include/cpu-sh3/cpu/dma.h b/arch/sh/include/cpu-sh3/cpu/dma.h index 207811a7a650..24e28b91c9d5 100644 --- a/arch/sh/include/cpu-sh3/cpu/dma.h +++ b/arch/sh/include/cpu-sh3/cpu/dma.h | |||
| @@ -20,31 +20,4 @@ | |||
| 20 | #define TS_32 0x00000010 | 20 | #define TS_32 0x00000010 |
| 21 | #define TS_128 0x00000018 | 21 | #define TS_128 0x00000018 |
| 22 | 22 | ||
| 23 | #define CHCR_TS_LOW_MASK 0x18 | ||
| 24 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 25 | #define CHCR_TS_HIGH_MASK 0 | ||
| 26 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 27 | |||
| 28 | #define DMAOR_INIT DMAOR_DME | ||
| 29 | |||
| 30 | /* | ||
| 31 | * The SuperH DMAC supports a number of transmit sizes, we list them here, | ||
| 32 | * with their respective values as they appear in the CHCR registers. | ||
| 33 | */ | ||
| 34 | enum { | ||
| 35 | XMIT_SZ_8BIT, | ||
| 36 | XMIT_SZ_16BIT, | ||
| 37 | XMIT_SZ_32BIT, | ||
| 38 | XMIT_SZ_128BIT, | ||
| 39 | }; | ||
| 40 | |||
| 41 | #define TS_SHIFT { \ | ||
| 42 | [XMIT_SZ_8BIT] = 0, \ | ||
| 43 | [XMIT_SZ_16BIT] = 1, \ | ||
| 44 | [XMIT_SZ_32BIT] = 2, \ | ||
| 45 | [XMIT_SZ_128BIT] = 4, \ | ||
| 46 | } | ||
| 47 | |||
| 48 | #define TS_INDEX2VAL(i) (((i) & 3) << CHCR_TS_LOW_SHIFT) | ||
| 49 | |||
| 50 | #endif /* __ASM_CPU_SH3_DMA_H */ | 23 | #endif /* __ASM_CPU_SH3_DMA_H */ |
diff --git a/arch/sh/include/cpu-sh4/cpu/dma-register.h b/arch/sh/include/cpu-sh4/cpu/dma-register.h new file mode 100644 index 000000000000..55f9fec082d4 --- /dev/null +++ b/arch/sh/include/cpu-sh4/cpu/dma-register.h | |||
| @@ -0,0 +1,112 @@ | |||
| 1 | /* | ||
| 2 | * SH4 CPU-specific DMA definitions, used by both DMA drivers | ||
| 3 | * | ||
| 4 | * Copyright (C) 2010 Guennadi Liakhovetski <g.liakhovetski@gmx.de> | ||
| 5 | * | ||
| 6 | * This program is free software; you can redistribute it and/or modify | ||
| 7 | * it under the terms of the GNU General Public License version 2 as | ||
| 8 | * published by the Free Software Foundation. | ||
| 9 | */ | ||
| 10 | #ifndef CPU_DMA_REGISTER_H | ||
| 11 | #define CPU_DMA_REGISTER_H | ||
| 12 | |||
| 13 | /* SH7751/7760/7780 DMA IRQ sources */ | ||
| 14 | |||
| 15 | #ifdef CONFIG_CPU_SH4A | ||
| 16 | |||
| 17 | #define DMAOR_INIT DMAOR_DME | ||
| 18 | |||
| 19 | #if defined(CONFIG_CPU_SUBTYPE_SH7343) || \ | ||
| 20 | defined(CONFIG_CPU_SUBTYPE_SH7730) | ||
| 21 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 22 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 23 | #define CHCR_TS_HIGH_MASK 0 | ||
| 24 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 25 | #elif defined(CONFIG_CPU_SUBTYPE_SH7722) || \ | ||
| 26 | defined(CONFIG_CPU_SUBTYPE_SH7724) | ||
| 27 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 28 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 29 | #define CHCR_TS_HIGH_MASK 0x00300000 | ||
| 30 | #define CHCR_TS_HIGH_SHIFT (20 - 2) /* 2 bits for shifted low TS */ | ||
| 31 | #elif defined(CONFIG_CPU_SUBTYPE_SH7763) || \ | ||
| 32 | defined(CONFIG_CPU_SUBTYPE_SH7764) | ||
| 33 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 34 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 35 | #define CHCR_TS_HIGH_MASK 0 | ||
| 36 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 37 | #elif defined(CONFIG_CPU_SUBTYPE_SH7723) | ||
| 38 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 39 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 40 | #define CHCR_TS_HIGH_MASK 0 | ||
| 41 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 42 | #elif defined(CONFIG_CPU_SUBTYPE_SH7780) | ||
| 43 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 44 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 45 | #define CHCR_TS_HIGH_MASK 0 | ||
| 46 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 47 | #else /* SH7785 */ | ||
| 48 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 49 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 50 | #define CHCR_TS_HIGH_MASK 0 | ||
| 51 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 52 | #endif | ||
| 53 | |||
| 54 | /* Transmit sizes and respective CHCR register values */ | ||
| 55 | enum { | ||
| 56 | XMIT_SZ_8BIT = 0, | ||
| 57 | XMIT_SZ_16BIT = 1, | ||
| 58 | XMIT_SZ_32BIT = 2, | ||
| 59 | XMIT_SZ_64BIT = 7, | ||
| 60 | XMIT_SZ_128BIT = 3, | ||
| 61 | XMIT_SZ_256BIT = 4, | ||
| 62 | XMIT_SZ_128BIT_BLK = 0xb, | ||
| 63 | XMIT_SZ_256BIT_BLK = 0xc, | ||
| 64 | }; | ||
| 65 | |||
| 66 | /* log2(size / 8) - used to calculate number of transfers */ | ||
| 67 | #define TS_SHIFT { \ | ||
| 68 | [XMIT_SZ_8BIT] = 0, \ | ||
| 69 | [XMIT_SZ_16BIT] = 1, \ | ||
| 70 | [XMIT_SZ_32BIT] = 2, \ | ||
| 71 | [XMIT_SZ_64BIT] = 3, \ | ||
| 72 | [XMIT_SZ_128BIT] = 4, \ | ||
| 73 | [XMIT_SZ_256BIT] = 5, \ | ||
| 74 | [XMIT_SZ_128BIT_BLK] = 4, \ | ||
| 75 | [XMIT_SZ_256BIT_BLK] = 5, \ | ||
| 76 | } | ||
| 77 | |||
| 78 | #define TS_INDEX2VAL(i) ((((i) & 3) << CHCR_TS_LOW_SHIFT) | \ | ||
| 79 | ((((i) >> 2) & 3) << CHCR_TS_HIGH_SHIFT)) | ||
| 80 | |||
| 81 | #else /* CONFIG_CPU_SH4A */ | ||
| 82 | |||
| 83 | #define DMAOR_INIT (0x8000 | DMAOR_DME) | ||
| 84 | |||
| 85 | #define CHCR_TS_LOW_MASK 0x70 | ||
| 86 | #define CHCR_TS_LOW_SHIFT 4 | ||
| 87 | #define CHCR_TS_HIGH_MASK 0 | ||
| 88 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 89 | |||
| 90 | /* Transmit sizes and respective CHCR register values */ | ||
| 91 | enum { | ||
| 92 | XMIT_SZ_8BIT = 1, | ||
| 93 | XMIT_SZ_16BIT = 2, | ||
| 94 | XMIT_SZ_32BIT = 3, | ||
| 95 | XMIT_SZ_64BIT = 0, | ||
| 96 | XMIT_SZ_256BIT = 4, | ||
| 97 | }; | ||
| 98 | |||
| 99 | /* log2(size / 8) - used to calculate number of transfers */ | ||
| 100 | #define TS_SHIFT { \ | ||
| 101 | [XMIT_SZ_8BIT] = 0, \ | ||
| 102 | [XMIT_SZ_16BIT] = 1, \ | ||
| 103 | [XMIT_SZ_32BIT] = 2, \ | ||
| 104 | [XMIT_SZ_64BIT] = 3, \ | ||
| 105 | [XMIT_SZ_256BIT] = 5, \ | ||
| 106 | } | ||
| 107 | |||
| 108 | #define TS_INDEX2VAL(i) (((i) & 7) << CHCR_TS_LOW_SHIFT) | ||
| 109 | |||
| 110 | #endif /* CONFIG_CPU_SH4A */ | ||
| 111 | |||
| 112 | #endif | ||
diff --git a/arch/sh/include/cpu-sh4/cpu/dma-sh4a.h b/arch/sh/include/cpu-sh4/cpu/dma-sh4a.h index e734ea47d8a0..9647e681fd27 100644 --- a/arch/sh/include/cpu-sh4/cpu/dma-sh4a.h +++ b/arch/sh/include/cpu-sh4/cpu/dma-sh4a.h | |||
| @@ -8,20 +8,12 @@ | |||
| 8 | #define DMAE0_IRQ 78 /* DMA Error IRQ*/ | 8 | #define DMAE0_IRQ 78 /* DMA Error IRQ*/ |
| 9 | #define SH_DMAC_BASE0 0xFE008020 | 9 | #define SH_DMAC_BASE0 0xFE008020 |
| 10 | #define SH_DMARS_BASE0 0xFE009000 | 10 | #define SH_DMARS_BASE0 0xFE009000 |
| 11 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 12 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 13 | #define CHCR_TS_HIGH_MASK 0 | ||
| 14 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 15 | #elif defined(CONFIG_CPU_SUBTYPE_SH7722) | 11 | #elif defined(CONFIG_CPU_SUBTYPE_SH7722) |
| 16 | #define DMTE0_IRQ 48 | 12 | #define DMTE0_IRQ 48 |
| 17 | #define DMTE4_IRQ 76 | 13 | #define DMTE4_IRQ 76 |
| 18 | #define DMAE0_IRQ 78 /* DMA Error IRQ*/ | 14 | #define DMAE0_IRQ 78 /* DMA Error IRQ*/ |
| 19 | #define SH_DMAC_BASE0 0xFE008020 | 15 | #define SH_DMAC_BASE0 0xFE008020 |
| 20 | #define SH_DMARS_BASE0 0xFE009000 | 16 | #define SH_DMARS_BASE0 0xFE009000 |
| 21 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 22 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 23 | #define CHCR_TS_HIGH_MASK 0x00300000 | ||
| 24 | #define CHCR_TS_HIGH_SHIFT 20 | ||
| 25 | #elif defined(CONFIG_CPU_SUBTYPE_SH7763) || \ | 17 | #elif defined(CONFIG_CPU_SUBTYPE_SH7763) || \ |
| 26 | defined(CONFIG_CPU_SUBTYPE_SH7764) | 18 | defined(CONFIG_CPU_SUBTYPE_SH7764) |
| 27 | #define DMTE0_IRQ 34 | 19 | #define DMTE0_IRQ 34 |
| @@ -29,10 +21,6 @@ | |||
| 29 | #define DMAE0_IRQ 38 | 21 | #define DMAE0_IRQ 38 |
| 30 | #define SH_DMAC_BASE0 0xFF608020 | 22 | #define SH_DMAC_BASE0 0xFF608020 |
| 31 | #define SH_DMARS_BASE0 0xFF609000 | 23 | #define SH_DMARS_BASE0 0xFF609000 |
| 32 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 33 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 34 | #define CHCR_TS_HIGH_MASK 0 | ||
| 35 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 36 | #elif defined(CONFIG_CPU_SUBTYPE_SH7723) | 24 | #elif defined(CONFIG_CPU_SUBTYPE_SH7723) |
| 37 | #define DMTE0_IRQ 48 /* DMAC0A*/ | 25 | #define DMTE0_IRQ 48 /* DMAC0A*/ |
| 38 | #define DMTE4_IRQ 76 /* DMAC0B */ | 26 | #define DMTE4_IRQ 76 /* DMAC0B */ |
| @@ -46,10 +34,6 @@ | |||
| 46 | #define SH_DMAC_BASE0 0xFE008020 | 34 | #define SH_DMAC_BASE0 0xFE008020 |
| 47 | #define SH_DMAC_BASE1 0xFDC08020 | 35 | #define SH_DMAC_BASE1 0xFDC08020 |
| 48 | #define SH_DMARS_BASE0 0xFDC09000 | 36 | #define SH_DMARS_BASE0 0xFDC09000 |
| 49 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 50 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 51 | #define CHCR_TS_HIGH_MASK 0 | ||
| 52 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 53 | #elif defined(CONFIG_CPU_SUBTYPE_SH7724) | 37 | #elif defined(CONFIG_CPU_SUBTYPE_SH7724) |
| 54 | #define DMTE0_IRQ 48 /* DMAC0A*/ | 38 | #define DMTE0_IRQ 48 /* DMAC0A*/ |
| 55 | #define DMTE4_IRQ 76 /* DMAC0B */ | 39 | #define DMTE4_IRQ 76 /* DMAC0B */ |
| @@ -64,10 +48,6 @@ | |||
| 64 | #define SH_DMAC_BASE1 0xFDC08020 | 48 | #define SH_DMAC_BASE1 0xFDC08020 |
| 65 | #define SH_DMARS_BASE0 0xFE009000 | 49 | #define SH_DMARS_BASE0 0xFE009000 |
| 66 | #define SH_DMARS_BASE1 0xFDC09000 | 50 | #define SH_DMARS_BASE1 0xFDC09000 |
| 67 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 68 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 69 | #define CHCR_TS_HIGH_MASK 0x00600000 | ||
| 70 | #define CHCR_TS_HIGH_SHIFT 21 | ||
| 71 | #elif defined(CONFIG_CPU_SUBTYPE_SH7780) | 51 | #elif defined(CONFIG_CPU_SUBTYPE_SH7780) |
| 72 | #define DMTE0_IRQ 34 | 52 | #define DMTE0_IRQ 34 |
| 73 | #define DMTE4_IRQ 44 | 53 | #define DMTE4_IRQ 44 |
| @@ -80,10 +60,6 @@ | |||
| 80 | #define SH_DMAC_BASE0 0xFC808020 | 60 | #define SH_DMAC_BASE0 0xFC808020 |
| 81 | #define SH_DMAC_BASE1 0xFC818020 | 61 | #define SH_DMAC_BASE1 0xFC818020 |
| 82 | #define SH_DMARS_BASE0 0xFC809000 | 62 | #define SH_DMARS_BASE0 0xFC809000 |
| 83 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 84 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 85 | #define CHCR_TS_HIGH_MASK 0 | ||
| 86 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 87 | #else /* SH7785 */ | 63 | #else /* SH7785 */ |
| 88 | #define DMTE0_IRQ 33 | 64 | #define DMTE0_IRQ 33 |
| 89 | #define DMTE4_IRQ 37 | 65 | #define DMTE4_IRQ 37 |
| @@ -97,10 +73,6 @@ | |||
| 97 | #define SH_DMAC_BASE0 0xFC808020 | 73 | #define SH_DMAC_BASE0 0xFC808020 |
| 98 | #define SH_DMAC_BASE1 0xFCC08020 | 74 | #define SH_DMAC_BASE1 0xFCC08020 |
| 99 | #define SH_DMARS_BASE0 0xFC809000 | 75 | #define SH_DMARS_BASE0 0xFC809000 |
| 100 | #define CHCR_TS_LOW_MASK 0x00000018 | ||
| 101 | #define CHCR_TS_LOW_SHIFT 3 | ||
| 102 | #define CHCR_TS_HIGH_MASK 0 | ||
| 103 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 104 | #endif | 76 | #endif |
| 105 | 77 | ||
| 106 | #define REQ_HE 0x000000C0 | 78 | #define REQ_HE 0x000000C0 |
| @@ -108,38 +80,4 @@ | |||
| 108 | #define REQ_LE 0x00000040 | 80 | #define REQ_LE 0x00000040 |
| 109 | #define TM_BURST 0x00000020 | 81 | #define TM_BURST 0x00000020 |
| 110 | 82 | ||
| 111 | /* | ||
| 112 | * The SuperH DMAC supports a number of transmit sizes, we list them here, | ||
| 113 | * with their respective values as they appear in the CHCR registers. | ||
| 114 | * | ||
| 115 | * Defaults to a 64-bit transfer size. | ||
| 116 | */ | ||
| 117 | enum { | ||
| 118 | XMIT_SZ_8BIT = 0, | ||
| 119 | XMIT_SZ_16BIT = 1, | ||
| 120 | XMIT_SZ_32BIT = 2, | ||
| 121 | XMIT_SZ_64BIT = 7, | ||
| 122 | XMIT_SZ_128BIT = 3, | ||
| 123 | XMIT_SZ_256BIT = 4, | ||
| 124 | XMIT_SZ_128BIT_BLK = 0xb, | ||
| 125 | XMIT_SZ_256BIT_BLK = 0xc, | ||
| 126 | }; | ||
| 127 | |||
| 128 | /* | ||
| 129 | * The DMA count is defined as the number of bytes to transfer. | ||
| 130 | */ | ||
| 131 | #define TS_SHIFT { \ | ||
| 132 | [XMIT_SZ_8BIT] = 0, \ | ||
| 133 | [XMIT_SZ_16BIT] = 1, \ | ||
| 134 | [XMIT_SZ_32BIT] = 2, \ | ||
| 135 | [XMIT_SZ_64BIT] = 3, \ | ||
| 136 | [XMIT_SZ_128BIT] = 4, \ | ||
| 137 | [XMIT_SZ_256BIT] = 5, \ | ||
| 138 | [XMIT_SZ_128BIT_BLK] = 4, \ | ||
| 139 | [XMIT_SZ_256BIT_BLK] = 5, \ | ||
| 140 | } | ||
| 141 | |||
| 142 | #define TS_INDEX2VAL(i) ((((i) & 3) << CHCR_TS_LOW_SHIFT) | \ | ||
| 143 | ((((i) >> 2) & 3) << CHCR_TS_HIGH_SHIFT)) | ||
| 144 | |||
| 145 | #endif /* __ASM_SH_CPU_SH4_DMA_SH7780_H */ | 83 | #endif /* __ASM_SH_CPU_SH4_DMA_SH7780_H */ |
diff --git a/arch/sh/include/cpu-sh4/cpu/dma.h b/arch/sh/include/cpu-sh4/cpu/dma.h index 114a369705bc..ca747e93c2ed 100644 --- a/arch/sh/include/cpu-sh4/cpu/dma.h +++ b/arch/sh/include/cpu-sh4/cpu/dma.h | |||
| @@ -5,9 +5,8 @@ | |||
| 5 | 5 | ||
| 6 | #ifdef CONFIG_CPU_SH4A | 6 | #ifdef CONFIG_CPU_SH4A |
| 7 | 7 | ||
| 8 | #define DMAOR_INIT (DMAOR_DME) | ||
| 9 | |||
| 10 | #include <cpu/dma-sh4a.h> | 8 | #include <cpu/dma-sh4a.h> |
| 9 | |||
| 11 | #else /* CONFIG_CPU_SH4A */ | 10 | #else /* CONFIG_CPU_SH4A */ |
| 12 | /* | 11 | /* |
| 13 | * SH7750/SH7751/SH7760 | 12 | * SH7750/SH7751/SH7760 |
| @@ -17,7 +16,6 @@ | |||
| 17 | #define DMTE6_IRQ 46 | 16 | #define DMTE6_IRQ 46 |
| 18 | #define DMAE0_IRQ 38 | 17 | #define DMAE0_IRQ 38 |
| 19 | 18 | ||
| 20 | #define DMAOR_INIT (0x8000|DMAOR_DME) | ||
| 21 | #define SH_DMAC_BASE0 0xffa00000 | 19 | #define SH_DMAC_BASE0 0xffa00000 |
| 22 | #define SH_DMAC_BASE1 0xffa00070 | 20 | #define SH_DMAC_BASE1 0xffa00070 |
| 23 | /* Definitions for the SuperH DMAC */ | 21 | /* Definitions for the SuperH DMAC */ |
| @@ -27,40 +25,8 @@ | |||
| 27 | #define TS_32 0x00000030 | 25 | #define TS_32 0x00000030 |
| 28 | #define TS_64 0x00000000 | 26 | #define TS_64 0x00000000 |
| 29 | 27 | ||
| 30 | #define CHCR_TS_LOW_MASK 0x70 | ||
| 31 | #define CHCR_TS_LOW_SHIFT 4 | ||
| 32 | #define CHCR_TS_HIGH_MASK 0 | ||
| 33 | #define CHCR_TS_HIGH_SHIFT 0 | ||
| 34 | |||
| 35 | #define DMAOR_COD 0x00000008 | 28 | #define DMAOR_COD 0x00000008 |
| 36 | 29 | ||
| 37 | /* | ||
| 38 | * The SuperH DMAC supports a number of transmit sizes, we list them here, | ||
| 39 | * with their respective values as they appear in the CHCR registers. | ||
| 40 | * | ||
| 41 | * Defaults to a 64-bit transfer size. | ||
| 42 | */ | ||
| 43 | enum { | ||
| 44 | XMIT_SZ_8BIT = 1, | ||
| 45 | XMIT_SZ_16BIT = 2, | ||
| 46 | XMIT_SZ_32BIT = 3, | ||
| 47 | XMIT_SZ_64BIT = 0, | ||
| 48 | XMIT_SZ_256BIT = 4, | ||
| 49 | }; | ||
| 50 | |||
| 51 | /* | ||
| 52 | * The DMA count is defined as the number of bytes to transfer. | ||
| 53 | */ | ||
| 54 | #define TS_SHIFT { \ | ||
| 55 | [XMIT_SZ_8BIT] = 0, \ | ||
| 56 | [XMIT_SZ_16BIT] = 1, \ | ||
| 57 | [XMIT_SZ_32BIT] = 2, \ | ||
| 58 | [XMIT_SZ_64BIT] = 3, \ | ||
| 59 | [XMIT_SZ_256BIT] = 5, \ | ||
| 60 | } | ||
| 61 | |||
| 62 | #define TS_INDEX2VAL(i) (((i) & 7) << CHCR_TS_LOW_SHIFT) | ||
| 63 | |||
| 64 | #endif | 30 | #endif |
| 65 | 31 | ||
| 66 | #endif /* __ASM_CPU_SH4_DMA_H */ | 32 | #endif /* __ASM_CPU_SH4_DMA_H */ |
diff --git a/arch/sh/include/mach-migor/mach/migor.h b/arch/sh/include/mach-migor/mach/migor.h index cee6cb88e020..42fccf93412e 100644 --- a/arch/sh/include/mach-migor/mach/migor.h +++ b/arch/sh/include/mach-migor/mach/migor.h | |||
| @@ -1,6 +1,7 @@ | |||
| 1 | #ifndef __ASM_SH_MIGOR_H | 1 | #ifndef __ASM_SH_MIGOR_H |
| 2 | #define __ASM_SH_MIGOR_H | 2 | #define __ASM_SH_MIGOR_H |
| 3 | 3 | ||
| 4 | #define PORT_MSELCRA 0xa4050180 | ||
| 4 | #define PORT_MSELCRB 0xa4050182 | 5 | #define PORT_MSELCRB 0xa4050182 |
| 5 | #define BSC_CS4BCR 0xfec10010 | 6 | #define BSC_CS4BCR 0xfec10010 |
| 6 | #define BSC_CS6ABCR 0xfec1001c | 7 | #define BSC_CS6ABCR 0xfec1001c |
diff --git a/arch/sh/kernel/cpu/sh4a/setup-sh7722.c b/arch/sh/kernel/cpu/sh4a/setup-sh7722.c index ef3f97827808..fd7e3639e845 100644 --- a/arch/sh/kernel/cpu/sh4a/setup-sh7722.c +++ b/arch/sh/kernel/cpu/sh4a/setup-sh7722.c | |||
| @@ -7,19 +7,167 @@ | |||
| 7 | * License. See the file "COPYING" in the main directory of this archive | 7 | * License. See the file "COPYING" in the main directory of this archive |
| 8 | * for more details. | 8 | * for more details. |
| 9 | */ | 9 | */ |
| 10 | #include <linux/platform_device.h> | ||
| 11 | #include <linux/init.h> | 10 | #include <linux/init.h> |
| 11 | #include <linux/mm.h> | ||
| 12 | #include <linux/platform_device.h> | ||
| 12 | #include <linux/serial.h> | 13 | #include <linux/serial.h> |
| 13 | #include <linux/serial_sci.h> | 14 | #include <linux/serial_sci.h> |
| 14 | #include <linux/mm.h> | 15 | #include <linux/sh_timer.h> |
| 15 | #include <linux/uio_driver.h> | 16 | #include <linux/uio_driver.h> |
| 16 | #include <linux/usb/m66592.h> | 17 | #include <linux/usb/m66592.h> |
| 17 | #include <linux/sh_timer.h> | 18 | |
| 18 | #include <asm/clock.h> | 19 | #include <asm/clock.h> |
| 20 | #include <asm/dmaengine.h> | ||
| 19 | #include <asm/mmzone.h> | 21 | #include <asm/mmzone.h> |
| 20 | #include <asm/dma-sh.h> | 22 | #include <asm/siu.h> |
| 23 | |||
| 24 | #include <cpu/dma-register.h> | ||
| 21 | #include <cpu/sh7722.h> | 25 | #include <cpu/sh7722.h> |
| 22 | 26 | ||
| 27 | static struct sh_dmae_slave_config sh7722_dmae_slaves[] = { | ||
| 28 | { | ||
| 29 | .slave_id = SHDMA_SLAVE_SCIF0_TX, | ||
| 30 | .addr = 0xffe0000c, | ||
| 31 | .chcr = DM_FIX | SM_INC | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 32 | .mid_rid = 0x21, | ||
| 33 | }, { | ||
| 34 | .slave_id = SHDMA_SLAVE_SCIF0_RX, | ||
| 35 | .addr = 0xffe00014, | ||
| 36 | .chcr = DM_INC | SM_FIX | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 37 | .mid_rid = 0x22, | ||
| 38 | }, { | ||
| 39 | .slave_id = SHDMA_SLAVE_SCIF1_TX, | ||
| 40 | .addr = 0xffe1000c, | ||
| 41 | .chcr = DM_FIX | SM_INC | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 42 | .mid_rid = 0x25, | ||
| 43 | }, { | ||
| 44 | .slave_id = SHDMA_SLAVE_SCIF1_RX, | ||
| 45 | .addr = 0xffe10014, | ||
| 46 | .chcr = DM_INC | SM_FIX | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 47 | .mid_rid = 0x26, | ||
| 48 | }, { | ||
| 49 | .slave_id = SHDMA_SLAVE_SCIF2_TX, | ||
| 50 | .addr = 0xffe2000c, | ||
| 51 | .chcr = DM_FIX | SM_INC | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 52 | .mid_rid = 0x29, | ||
| 53 | }, { | ||
| 54 | .slave_id = SHDMA_SLAVE_SCIF2_RX, | ||
| 55 | .addr = 0xffe20014, | ||
| 56 | .chcr = DM_INC | SM_FIX | 0x800 | TS_INDEX2VAL(XMIT_SZ_8BIT), | ||
| 57 | .mid_rid = 0x2a, | ||
| 58 | }, { | ||
| 59 | .slave_id = SHDMA_SLAVE_SIUA_TX, | ||
| 60 | .addr = 0xa454c098, | ||
| 61 | .chcr = DM_FIX | SM_INC | 0x800 | TS_INDEX2VAL(XMIT_SZ_32BIT), | ||
| 62 | .mid_rid = 0xb1, | ||
| 63 | }, { | ||
| 64 | .slave_id = SHDMA_SLAVE_SIUA_RX, | ||
| 65 | .addr = 0xa454c090, | ||
| 66 | .chcr = DM_INC | SM_FIX | 0x800 | TS_INDEX2VAL(XMIT_SZ_32BIT), | ||
| 67 | .mid_rid = 0xb2, | ||
| 68 | }, { | ||
| 69 | .slave_id = SHDMA_SLAVE_SIUB_TX, | ||
| 70 | .addr = 0xa454c09c, | ||
| 71 | .chcr = DM_FIX | SM_INC | 0x800 | TS_INDEX2VAL(XMIT_SZ_32BIT), | ||
| 72 | .mid_rid = 0xb5, | ||
| 73 | }, { | ||
| 74 | .slave_id = SHDMA_SLAVE_SIUB_RX, | ||
| 75 | .addr = 0xa454c094, | ||
| 76 | .chcr = DM_INC | SM_FIX | 0x800 | TS_INDEX2VAL(XMIT_SZ_32BIT), | ||
| 77 | .mid_rid = 0xb6, | ||
| 78 | }, | ||
| 79 | }; | ||
| 80 | |||
| 81 | static struct sh_dmae_channel sh7722_dmae_channels[] = { | ||
| 82 | { | ||
| 83 | .offset = 0, | ||
| 84 | .dmars = 0, | ||
| 85 | .dmars_bit = 0, | ||
| 86 | }, { | ||
| 87 | .offset = 0x10, | ||
| 88 | .dmars = 0, | ||
| 89 | .dmars_bit = 8, | ||
| 90 | }, { | ||
| 91 | .offset = 0x20, | ||
| 92 | .dmars = 4, | ||
| 93 | .dmars_bit = 0, | ||
| 94 | }, { | ||
| 95 | .offset = 0x30, | ||
| 96 | .dmars = 4, | ||
| 97 | .dmars_bit = 8, | ||
| 98 | }, { | ||
| 99 | .offset = 0x50, | ||
| 100 | .dmars = 8, | ||
| 101 | .dmars_bit = 0, | ||
| 102 | }, { | ||
| 103 | .offset = 0x60, | ||
| 104 | .dmars = 8, | ||
| 105 | .dmars_bit = 8, | ||
| 106 | } | ||
| 107 | }; | ||
| 108 | |||
| 109 | static unsigned int ts_shift[] = TS_SHIFT; | ||
| 110 | |||
| 111 | static struct sh_dmae_pdata dma_platform_data = { | ||
| 112 | .slave = sh7722_dmae_slaves, | ||
| 113 | .slave_num = ARRAY_SIZE(sh7722_dmae_slaves), | ||
| 114 | .channel = sh7722_dmae_channels, | ||
| 115 | .channel_num = ARRAY_SIZE(sh7722_dmae_channels), | ||
| 116 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 117 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 118 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 119 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 120 | .ts_shift = ts_shift, | ||
| 121 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 122 | .dmaor_init = DMAOR_INIT, | ||
| 123 | }; | ||
| 124 | |||
| 125 | static struct resource sh7722_dmae_resources[] = { | ||
| 126 | [0] = { | ||
| 127 | /* Channel registers and DMAOR */ | ||
| 128 | .start = 0xfe008020, | ||
| 129 | .end = 0xfe00808f, | ||
| 130 | .flags = IORESOURCE_MEM, | ||
| 131 | }, | ||
| 132 | [1] = { | ||
| 133 | /* DMARSx */ | ||
| 134 | .start = 0xfe009000, | ||
| 135 | .end = 0xfe00900b, | ||
| 136 | .flags = IORESOURCE_MEM, | ||
| 137 | }, | ||
| 138 | { | ||
| 139 | /* DMA error IRQ */ | ||
| 140 | .start = 78, | ||
| 141 | .end = 78, | ||
| 142 | .flags = IORESOURCE_IRQ, | ||
| 143 | }, | ||
| 144 | { | ||
| 145 | /* IRQ for channels 0-3 */ | ||
| 146 | .start = 48, | ||
| 147 | .end = 51, | ||
| 148 | .flags = IORESOURCE_IRQ, | ||
| 149 | }, | ||
| 150 | { | ||
| 151 | /* IRQ for channels 4-5 */ | ||
| 152 | .start = 76, | ||
| 153 | .end = 77, | ||
| 154 | .flags = IORESOURCE_IRQ, | ||
| 155 | }, | ||
| 156 | }; | ||
| 157 | |||
| 158 | struct platform_device dma_device = { | ||
| 159 | .name = "sh-dma-engine", | ||
| 160 | .id = -1, | ||
| 161 | .resource = sh7722_dmae_resources, | ||
| 162 | .num_resources = ARRAY_SIZE(sh7722_dmae_resources), | ||
| 163 | .dev = { | ||
| 164 | .platform_data = &dma_platform_data, | ||
| 165 | }, | ||
| 166 | .archdata = { | ||
| 167 | .hwblk_id = HWBLK_DMAC, | ||
| 168 | }, | ||
| 169 | }; | ||
| 170 | |||
| 23 | /* Serial */ | 171 | /* Serial */ |
| 24 | static struct plat_sci_port scif0_platform_data = { | 172 | static struct plat_sci_port scif0_platform_data = { |
| 25 | .mapbase = 0xffe00000, | 173 | .mapbase = 0xffe00000, |
| @@ -388,15 +536,36 @@ static struct platform_device tmu2_device = { | |||
| 388 | }, | 536 | }, |
| 389 | }; | 537 | }; |
| 390 | 538 | ||
| 391 | static struct sh_dmae_pdata dma_platform_data = { | 539 | static struct siu_platform siu_platform_data = { |
| 392 | .mode = 0, | 540 | .dma_dev = &dma_device.dev, |
| 541 | .dma_slave_tx_a = SHDMA_SLAVE_SIUA_TX, | ||
| 542 | .dma_slave_rx_a = SHDMA_SLAVE_SIUA_RX, | ||
| 543 | .dma_slave_tx_b = SHDMA_SLAVE_SIUB_TX, | ||
| 544 | .dma_slave_rx_b = SHDMA_SLAVE_SIUB_RX, | ||
| 393 | }; | 545 | }; |
| 394 | 546 | ||
| 395 | static struct platform_device dma_device = { | 547 | static struct resource siu_resources[] = { |
| 396 | .name = "sh-dma-engine", | 548 | [0] = { |
| 549 | .start = 0xa4540000, | ||
| 550 | .end = 0xa454c10f, | ||
| 551 | .flags = IORESOURCE_MEM, | ||
| 552 | }, | ||
| 553 | [1] = { | ||
| 554 | .start = 108, | ||
| 555 | .flags = IORESOURCE_IRQ, | ||
| 556 | }, | ||
| 557 | }; | ||
| 558 | |||
| 559 | static struct platform_device siu_device = { | ||
| 560 | .name = "sh_siu", | ||
| 397 | .id = -1, | 561 | .id = -1, |
| 398 | .dev = { | 562 | .dev = { |
| 399 | .platform_data = &dma_platform_data, | 563 | .platform_data = &siu_platform_data, |
| 564 | }, | ||
| 565 | .resource = siu_resources, | ||
| 566 | .num_resources = ARRAY_SIZE(siu_resources), | ||
| 567 | .archdata = { | ||
| 568 | .hwblk_id = HWBLK_SIU, | ||
| 400 | }, | 569 | }, |
| 401 | }; | 570 | }; |
| 402 | 571 | ||
| @@ -414,6 +583,7 @@ static struct platform_device *sh7722_devices[] __initdata = { | |||
| 414 | &vpu_device, | 583 | &vpu_device, |
| 415 | &veu_device, | 584 | &veu_device, |
| 416 | &jpu_device, | 585 | &jpu_device, |
| 586 | &siu_device, | ||
| 417 | &dma_device, | 587 | &dma_device, |
| 418 | }; | 588 | }; |
| 419 | 589 | ||
diff --git a/arch/sh/kernel/cpu/sh4a/setup-sh7724.c b/arch/sh/kernel/cpu/sh4a/setup-sh7724.c index 31e3451f7e3d..e7fa2a92fc1f 100644 --- a/arch/sh/kernel/cpu/sh4a/setup-sh7724.c +++ b/arch/sh/kernel/cpu/sh4a/setup-sh7724.c | |||
| @@ -21,22 +21,189 @@ | |||
| 21 | #include <linux/sh_timer.h> | 21 | #include <linux/sh_timer.h> |
| 22 | #include <linux/io.h> | 22 | #include <linux/io.h> |
| 23 | #include <linux/notifier.h> | 23 | #include <linux/notifier.h> |
| 24 | |||
| 24 | #include <asm/suspend.h> | 25 | #include <asm/suspend.h> |
| 25 | #include <asm/clock.h> | 26 | #include <asm/clock.h> |
| 26 | #include <asm/dma-sh.h> | 27 | #include <asm/dmaengine.h> |
| 27 | #include <asm/mmzone.h> | 28 | #include <asm/mmzone.h> |
| 29 | |||
| 30 | #include <cpu/dma-register.h> | ||
| 28 | #include <cpu/sh7724.h> | 31 | #include <cpu/sh7724.h> |
| 29 | 32 | ||
| 30 | /* DMA */ | 33 | /* DMA */ |
| 31 | static struct sh_dmae_pdata dma_platform_data = { | 34 | static struct sh_dmae_channel sh7724_dmae0_channels[] = { |
| 32 | .mode = SHDMA_DMAOR1, | 35 | { |
| 36 | .offset = 0, | ||
| 37 | .dmars = 0, | ||
| 38 | .dmars_bit = 0, | ||
| 39 | }, { | ||
| 40 | .offset = 0x10, | ||
| 41 | .dmars = 0, | ||
| 42 | .dmars_bit = 8, | ||
| 43 | }, { | ||
| 44 | .offset = 0x20, | ||
| 45 | .dmars = 4, | ||
| 46 | .dmars_bit = 0, | ||
| 47 | }, { | ||
| 48 | .offset = 0x30, | ||
| 49 | .dmars = 4, | ||
| 50 | .dmars_bit = 8, | ||
| 51 | }, { | ||
| 52 | .offset = 0x50, | ||
| 53 | .dmars = 8, | ||
| 54 | .dmars_bit = 0, | ||
| 55 | }, { | ||
| 56 | .offset = 0x60, | ||
| 57 | .dmars = 8, | ||
| 58 | .dmars_bit = 8, | ||
| 59 | } | ||
| 60 | }; | ||
| 61 | |||
| 62 | static struct sh_dmae_channel sh7724_dmae1_channels[] = { | ||
| 63 | { | ||
| 64 | .offset = 0, | ||
| 65 | .dmars = 0, | ||
| 66 | .dmars_bit = 0, | ||
| 67 | }, { | ||
| 68 | .offset = 0x10, | ||
| 69 | .dmars = 0, | ||
| 70 | .dmars_bit = 8, | ||
| 71 | }, { | ||
| 72 | .offset = 0x20, | ||
| 73 | .dmars = 4, | ||
| 74 | .dmars_bit = 0, | ||
| 75 | }, { | ||
| 76 | .offset = 0x30, | ||
| 77 | .dmars = 4, | ||
| 78 | .dmars_bit = 8, | ||
| 79 | }, { | ||
| 80 | .offset = 0x50, | ||
| 81 | .dmars = 8, | ||
| 82 | .dmars_bit = 0, | ||
| 83 | }, { | ||
| 84 | .offset = 0x60, | ||
| 85 | .dmars = 8, | ||
| 86 | .dmars_bit = 8, | ||
| 87 | } | ||
| 88 | }; | ||
| 89 | |||
| 90 | static unsigned int ts_shift[] = TS_SHIFT; | ||
| 91 | |||
| 92 | static struct sh_dmae_pdata dma0_platform_data = { | ||
| 93 | .channel = sh7724_dmae0_channels, | ||
| 94 | .channel_num = ARRAY_SIZE(sh7724_dmae0_channels), | ||
| 95 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 96 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 97 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 98 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 99 | .ts_shift = ts_shift, | ||
| 100 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 101 | .dmaor_init = DMAOR_INIT, | ||
| 102 | }; | ||
| 103 | |||
| 104 | static struct sh_dmae_pdata dma1_platform_data = { | ||
| 105 | .channel = sh7724_dmae1_channels, | ||
| 106 | .channel_num = ARRAY_SIZE(sh7724_dmae1_channels), | ||
| 107 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 108 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 109 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 110 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 111 | .ts_shift = ts_shift, | ||
| 112 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 113 | .dmaor_init = DMAOR_INIT, | ||
| 114 | }; | ||
| 115 | |||
| 116 | /* Resource order important! */ | ||
| 117 | static struct resource sh7724_dmae0_resources[] = { | ||
| 118 | { | ||
| 119 | /* Channel registers and DMAOR */ | ||
| 120 | .start = 0xfe008020, | ||
| 121 | .end = 0xfe00808f, | ||
| 122 | .flags = IORESOURCE_MEM, | ||
| 123 | }, | ||
| 124 | { | ||
| 125 | /* DMARSx */ | ||
| 126 | .start = 0xfe009000, | ||
| 127 | .end = 0xfe00900b, | ||
| 128 | .flags = IORESOURCE_MEM, | ||
| 129 | }, | ||
| 130 | { | ||
| 131 | /* DMA error IRQ */ | ||
| 132 | .start = 78, | ||
| 133 | .end = 78, | ||
| 134 | .flags = IORESOURCE_IRQ, | ||
| 135 | }, | ||
| 136 | { | ||
| 137 | /* IRQ for channels 0-3 */ | ||
| 138 | .start = 48, | ||
| 139 | .end = 51, | ||
| 140 | .flags = IORESOURCE_IRQ, | ||
| 141 | }, | ||
| 142 | { | ||
| 143 | /* IRQ for channels 4-5 */ | ||
| 144 | .start = 76, | ||
| 145 | .end = 77, | ||
| 146 | .flags = IORESOURCE_IRQ, | ||
| 147 | }, | ||
| 33 | }; | 148 | }; |
| 34 | 149 | ||
| 35 | static struct platform_device dma_device = { | 150 | /* Resource order important! */ |
| 36 | .name = "sh-dma-engine", | 151 | static struct resource sh7724_dmae1_resources[] = { |
| 37 | .id = -1, | 152 | { |
| 38 | .dev = { | 153 | /* Channel registers and DMAOR */ |
| 39 | .platform_data = &dma_platform_data, | 154 | .start = 0xfdc08020, |
| 155 | .end = 0xfdc0808f, | ||
| 156 | .flags = IORESOURCE_MEM, | ||
| 157 | }, | ||
| 158 | { | ||
| 159 | /* DMARSx */ | ||
| 160 | .start = 0xfdc09000, | ||
| 161 | .end = 0xfdc0900b, | ||
| 162 | .flags = IORESOURCE_MEM, | ||
| 163 | }, | ||
| 164 | { | ||
| 165 | /* DMA error IRQ */ | ||
| 166 | .start = 74, | ||
| 167 | .end = 74, | ||
| 168 | .flags = IORESOURCE_IRQ, | ||
| 169 | }, | ||
| 170 | { | ||
| 171 | /* IRQ for channels 0-3 */ | ||
| 172 | .start = 40, | ||
| 173 | .end = 43, | ||
| 174 | .flags = IORESOURCE_IRQ, | ||
| 175 | }, | ||
| 176 | { | ||
| 177 | /* IRQ for channels 4-5 */ | ||
| 178 | .start = 72, | ||
| 179 | .end = 73, | ||
| 180 | .flags = IORESOURCE_IRQ, | ||
| 181 | }, | ||
| 182 | }; | ||
| 183 | |||
| 184 | static struct platform_device dma0_device = { | ||
| 185 | .name = "sh-dma-engine", | ||
| 186 | .id = 0, | ||
| 187 | .resource = sh7724_dmae0_resources, | ||
| 188 | .num_resources = ARRAY_SIZE(sh7724_dmae0_resources), | ||
| 189 | .dev = { | ||
| 190 | .platform_data = &dma0_platform_data, | ||
| 191 | }, | ||
| 192 | .archdata = { | ||
| 193 | .hwblk_id = HWBLK_DMAC0, | ||
| 194 | }, | ||
| 195 | }; | ||
| 196 | |||
| 197 | static struct platform_device dma1_device = { | ||
| 198 | .name = "sh-dma-engine", | ||
| 199 | .id = 1, | ||
| 200 | .resource = sh7724_dmae1_resources, | ||
| 201 | .num_resources = ARRAY_SIZE(sh7724_dmae1_resources), | ||
| 202 | .dev = { | ||
| 203 | .platform_data = &dma1_platform_data, | ||
| 204 | }, | ||
| 205 | .archdata = { | ||
| 206 | .hwblk_id = HWBLK_DMAC1, | ||
| 40 | }, | 207 | }, |
| 41 | }; | 208 | }; |
| 42 | 209 | ||
| @@ -663,7 +830,8 @@ static struct platform_device *sh7724_devices[] __initdata = { | |||
| 663 | &tmu3_device, | 830 | &tmu3_device, |
| 664 | &tmu4_device, | 831 | &tmu4_device, |
| 665 | &tmu5_device, | 832 | &tmu5_device, |
| 666 | &dma_device, | 833 | &dma0_device, |
| 834 | &dma1_device, | ||
| 667 | &rtc_device, | 835 | &rtc_device, |
| 668 | &iic0_device, | 836 | &iic0_device, |
| 669 | &iic1_device, | 837 | &iic1_device, |
diff --git a/arch/sh/kernel/cpu/sh4a/setup-sh7780.c b/arch/sh/kernel/cpu/sh4a/setup-sh7780.c index f8f21618d785..02e792c90de6 100644 --- a/arch/sh/kernel/cpu/sh4a/setup-sh7780.c +++ b/arch/sh/kernel/cpu/sh4a/setup-sh7780.c | |||
| @@ -13,7 +13,10 @@ | |||
| 13 | #include <linux/io.h> | 13 | #include <linux/io.h> |
| 14 | #include <linux/serial_sci.h> | 14 | #include <linux/serial_sci.h> |
| 15 | #include <linux/sh_timer.h> | 15 | #include <linux/sh_timer.h> |
| 16 | #include <asm/dma-sh.h> | 16 | |
| 17 | #include <asm/dmaengine.h> | ||
| 18 | |||
| 19 | #include <cpu/dma-register.h> | ||
| 17 | 20 | ||
| 18 | static struct plat_sci_port scif0_platform_data = { | 21 | static struct plat_sci_port scif0_platform_data = { |
| 19 | .mapbase = 0xffe00000, | 22 | .mapbase = 0xffe00000, |
| @@ -247,15 +250,131 @@ static struct platform_device rtc_device = { | |||
| 247 | .resource = rtc_resources, | 250 | .resource = rtc_resources, |
| 248 | }; | 251 | }; |
| 249 | 252 | ||
| 250 | static struct sh_dmae_pdata dma_platform_data = { | 253 | /* DMA */ |
| 251 | .mode = (SHDMA_MIX_IRQ | SHDMA_DMAOR1), | 254 | static struct sh_dmae_channel sh7780_dmae0_channels[] = { |
| 255 | { | ||
| 256 | .offset = 0, | ||
| 257 | .dmars = 0, | ||
| 258 | .dmars_bit = 0, | ||
| 259 | }, { | ||
| 260 | .offset = 0x10, | ||
| 261 | .dmars = 0, | ||
| 262 | .dmars_bit = 8, | ||
| 263 | }, { | ||
| 264 | .offset = 0x20, | ||
| 265 | .dmars = 4, | ||
| 266 | .dmars_bit = 0, | ||
| 267 | }, { | ||
| 268 | .offset = 0x30, | ||
| 269 | .dmars = 4, | ||
| 270 | .dmars_bit = 8, | ||
| 271 | }, { | ||
| 272 | .offset = 0x50, | ||
| 273 | .dmars = 8, | ||
| 274 | .dmars_bit = 0, | ||
| 275 | }, { | ||
| 276 | .offset = 0x60, | ||
| 277 | .dmars = 8, | ||
| 278 | .dmars_bit = 8, | ||
| 279 | } | ||
| 280 | }; | ||
| 281 | |||
| 282 | static struct sh_dmae_channel sh7780_dmae1_channels[] = { | ||
| 283 | { | ||
| 284 | .offset = 0, | ||
| 285 | }, { | ||
| 286 | .offset = 0x10, | ||
| 287 | }, { | ||
| 288 | .offset = 0x20, | ||
| 289 | }, { | ||
| 290 | .offset = 0x30, | ||
| 291 | }, { | ||
| 292 | .offset = 0x50, | ||
| 293 | }, { | ||
| 294 | .offset = 0x60, | ||
| 295 | } | ||
| 296 | }; | ||
| 297 | |||
| 298 | static unsigned int ts_shift[] = TS_SHIFT; | ||
| 299 | |||
| 300 | static struct sh_dmae_pdata dma0_platform_data = { | ||
| 301 | .channel = sh7780_dmae0_channels, | ||
| 302 | .channel_num = ARRAY_SIZE(sh7780_dmae0_channels), | ||
| 303 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 304 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 305 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 306 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 307 | .ts_shift = ts_shift, | ||
| 308 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 309 | .dmaor_init = DMAOR_INIT, | ||
| 310 | }; | ||
| 311 | |||
| 312 | static struct sh_dmae_pdata dma1_platform_data = { | ||
| 313 | .channel = sh7780_dmae1_channels, | ||
| 314 | .channel_num = ARRAY_SIZE(sh7780_dmae1_channels), | ||
| 315 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 316 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 317 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 318 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 319 | .ts_shift = ts_shift, | ||
| 320 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 321 | .dmaor_init = DMAOR_INIT, | ||
| 252 | }; | 322 | }; |
| 253 | 323 | ||
| 254 | static struct platform_device dma_device = { | 324 | static struct resource sh7780_dmae0_resources[] = { |
| 325 | [0] = { | ||
| 326 | /* Channel registers and DMAOR */ | ||
| 327 | .start = 0xfc808020, | ||
| 328 | .end = 0xfc80808f, | ||
| 329 | .flags = IORESOURCE_MEM, | ||
| 330 | }, | ||
| 331 | [1] = { | ||
| 332 | /* DMARSx */ | ||
| 333 | .start = 0xfc809000, | ||
| 334 | .end = 0xfc80900b, | ||
| 335 | .flags = IORESOURCE_MEM, | ||
| 336 | }, | ||
| 337 | { | ||
| 338 | /* Real DMA error IRQ is 38, and channel IRQs are 34-37, 44-45 */ | ||
| 339 | .start = 34, | ||
| 340 | .end = 34, | ||
| 341 | .flags = IORESOURCE_IRQ | IORESOURCE_IRQ_SHAREABLE, | ||
| 342 | }, | ||
| 343 | }; | ||
| 344 | |||
| 345 | static struct resource sh7780_dmae1_resources[] = { | ||
| 346 | [0] = { | ||
| 347 | /* Channel registers and DMAOR */ | ||
| 348 | .start = 0xfc818020, | ||
| 349 | .end = 0xfc81808f, | ||
| 350 | .flags = IORESOURCE_MEM, | ||
| 351 | }, | ||
| 352 | /* DMAC1 has no DMARS */ | ||
| 353 | { | ||
| 354 | /* Real DMA error IRQ is 38, and channel IRQs are 46-47, 92-95 */ | ||
| 355 | .start = 46, | ||
| 356 | .end = 46, | ||
| 357 | .flags = IORESOURCE_IRQ | IORESOURCE_IRQ_SHAREABLE, | ||
| 358 | }, | ||
| 359 | }; | ||
| 360 | |||
| 361 | static struct platform_device dma0_device = { | ||
| 255 | .name = "sh-dma-engine", | 362 | .name = "sh-dma-engine", |
| 256 | .id = -1, | 363 | .id = 0, |
| 364 | .resource = sh7780_dmae0_resources, | ||
| 365 | .num_resources = ARRAY_SIZE(sh7780_dmae0_resources), | ||
| 257 | .dev = { | 366 | .dev = { |
| 258 | .platform_data = &dma_platform_data, | 367 | .platform_data = &dma0_platform_data, |
| 368 | }, | ||
| 369 | }; | ||
| 370 | |||
| 371 | static struct platform_device dma1_device = { | ||
| 372 | .name = "sh-dma-engine", | ||
| 373 | .id = 1, | ||
| 374 | .resource = sh7780_dmae1_resources, | ||
| 375 | .num_resources = ARRAY_SIZE(sh7780_dmae1_resources), | ||
| 376 | .dev = { | ||
| 377 | .platform_data = &dma1_platform_data, | ||
| 259 | }, | 378 | }, |
| 260 | }; | 379 | }; |
| 261 | 380 | ||
| @@ -269,7 +388,8 @@ static struct platform_device *sh7780_devices[] __initdata = { | |||
| 269 | &tmu4_device, | 388 | &tmu4_device, |
| 270 | &tmu5_device, | 389 | &tmu5_device, |
| 271 | &rtc_device, | 390 | &rtc_device, |
| 272 | &dma_device, | 391 | &dma0_device, |
| 392 | &dma1_device, | ||
| 273 | }; | 393 | }; |
| 274 | 394 | ||
| 275 | static int __init sh7780_devices_setup(void) | 395 | static int __init sh7780_devices_setup(void) |
diff --git a/arch/sh/kernel/cpu/sh4a/setup-sh7785.c b/arch/sh/kernel/cpu/sh4a/setup-sh7785.c index 23448d8c6711..1fcd88b1671e 100644 --- a/arch/sh/kernel/cpu/sh4a/setup-sh7785.c +++ b/arch/sh/kernel/cpu/sh4a/setup-sh7785.c | |||
| @@ -14,9 +14,12 @@ | |||
| 14 | #include <linux/io.h> | 14 | #include <linux/io.h> |
| 15 | #include <linux/mm.h> | 15 | #include <linux/mm.h> |
| 16 | #include <linux/sh_timer.h> | 16 | #include <linux/sh_timer.h> |
| 17 | #include <asm/dma-sh.h> | 17 | |
| 18 | #include <asm/dmaengine.h> | ||
| 18 | #include <asm/mmzone.h> | 19 | #include <asm/mmzone.h> |
| 19 | 20 | ||
| 21 | #include <cpu/dma-register.h> | ||
| 22 | |||
| 20 | static struct plat_sci_port scif0_platform_data = { | 23 | static struct plat_sci_port scif0_platform_data = { |
| 21 | .mapbase = 0xffea0000, | 24 | .mapbase = 0xffea0000, |
| 22 | .flags = UPF_BOOT_AUTOCONF, | 25 | .flags = UPF_BOOT_AUTOCONF, |
| @@ -295,15 +298,131 @@ static struct platform_device tmu5_device = { | |||
| 295 | .num_resources = ARRAY_SIZE(tmu5_resources), | 298 | .num_resources = ARRAY_SIZE(tmu5_resources), |
| 296 | }; | 299 | }; |
| 297 | 300 | ||
| 298 | static struct sh_dmae_pdata dma_platform_data = { | 301 | /* DMA */ |
| 299 | .mode = (SHDMA_MIX_IRQ | SHDMA_DMAOR1), | 302 | static struct sh_dmae_channel sh7785_dmae0_channels[] = { |
| 303 | { | ||
| 304 | .offset = 0, | ||
| 305 | .dmars = 0, | ||
| 306 | .dmars_bit = 0, | ||
| 307 | }, { | ||
| 308 | .offset = 0x10, | ||
| 309 | .dmars = 0, | ||
| 310 | .dmars_bit = 8, | ||
| 311 | }, { | ||
| 312 | .offset = 0x20, | ||
| 313 | .dmars = 4, | ||
| 314 | .dmars_bit = 0, | ||
| 315 | }, { | ||
| 316 | .offset = 0x30, | ||
| 317 | .dmars = 4, | ||
| 318 | .dmars_bit = 8, | ||
| 319 | }, { | ||
| 320 | .offset = 0x50, | ||
| 321 | .dmars = 8, | ||
| 322 | .dmars_bit = 0, | ||
| 323 | }, { | ||
| 324 | .offset = 0x60, | ||
| 325 | .dmars = 8, | ||
| 326 | .dmars_bit = 8, | ||
| 327 | } | ||
| 328 | }; | ||
| 329 | |||
| 330 | static struct sh_dmae_channel sh7785_dmae1_channels[] = { | ||
| 331 | { | ||
| 332 | .offset = 0, | ||
| 333 | }, { | ||
| 334 | .offset = 0x10, | ||
| 335 | }, { | ||
| 336 | .offset = 0x20, | ||
| 337 | }, { | ||
| 338 | .offset = 0x30, | ||
| 339 | }, { | ||
| 340 | .offset = 0x50, | ||
| 341 | }, { | ||
| 342 | .offset = 0x60, | ||
| 343 | } | ||
| 344 | }; | ||
| 345 | |||
| 346 | static unsigned int ts_shift[] = TS_SHIFT; | ||
| 347 | |||
| 348 | static struct sh_dmae_pdata dma0_platform_data = { | ||
| 349 | .channel = sh7785_dmae0_channels, | ||
| 350 | .channel_num = ARRAY_SIZE(sh7785_dmae0_channels), | ||
| 351 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 352 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 353 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 354 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 355 | .ts_shift = ts_shift, | ||
| 356 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 357 | .dmaor_init = DMAOR_INIT, | ||
| 358 | }; | ||
| 359 | |||
| 360 | static struct sh_dmae_pdata dma1_platform_data = { | ||
| 361 | .channel = sh7785_dmae1_channels, | ||
| 362 | .channel_num = ARRAY_SIZE(sh7785_dmae1_channels), | ||
| 363 | .ts_low_shift = CHCR_TS_LOW_SHIFT, | ||
| 364 | .ts_low_mask = CHCR_TS_LOW_MASK, | ||
| 365 | .ts_high_shift = CHCR_TS_HIGH_SHIFT, | ||
| 366 | .ts_high_mask = CHCR_TS_HIGH_MASK, | ||
| 367 | .ts_shift = ts_shift, | ||
| 368 | .ts_shift_num = ARRAY_SIZE(ts_shift), | ||
| 369 | .dmaor_init = DMAOR_INIT, | ||
| 300 | }; | 370 | }; |
| 301 | 371 | ||
| 302 | static struct platform_device dma_device = { | 372 | static struct resource sh7785_dmae0_resources[] = { |
| 373 | [0] = { | ||
| 374 | /* Channel registers and DMAOR */ | ||
| 375 | .start = 0xfc808020, | ||
| 376 | .end = 0xfc80808f, | ||
| 377 | .flags = IORESOURCE_MEM, | ||
| 378 | }, | ||
| 379 | [1] = { | ||
| 380 | /* DMARSx */ | ||
| 381 | .start = 0xfc809000, | ||
| 382 | .end = 0xfc80900b, | ||
| 383 | .flags = IORESOURCE_MEM, | ||
| 384 | }, | ||
| 385 | { | ||
| 386 | /* Real DMA error IRQ is 39, and channel IRQs are 33-38 */ | ||
| 387 | .start = 33, | ||
| 388 | .end = 33, | ||
| 389 | .flags = IORESOURCE_IRQ | IORESOURCE_IRQ_SHAREABLE, | ||
| 390 | }, | ||
| 391 | }; | ||
| 392 | |||
| 393 | static struct resource sh7785_dmae1_resources[] = { | ||
| 394 | [0] = { | ||
| 395 | /* Channel registers and DMAOR */ | ||
| 396 | .start = 0xfcc08020, | ||
| 397 | .end = 0xfcc0808f, | ||
| 398 | .flags = IORESOURCE_MEM, | ||
| 399 | }, | ||
| 400 | /* DMAC1 has no DMARS */ | ||
| 401 | { | ||
| 402 | /* Real DMA error IRQ is 58, and channel IRQs are 52-57 */ | ||
| 403 | .start = 52, | ||
| 404 | .end = 52, | ||
| 405 | .flags = IORESOURCE_IRQ | IORESOURCE_IRQ_SHAREABLE, | ||
| 406 | }, | ||
| 407 | }; | ||
| 408 | |||
| 409 | static struct platform_device dma0_device = { | ||
| 303 | .name = "sh-dma-engine", | 410 | .name = "sh-dma-engine", |
| 304 | .id = -1, | 411 | .id = 0, |
| 412 | .resource = sh7785_dmae0_resources, | ||
| 413 | .num_resources = ARRAY_SIZE(sh7785_dmae0_resources), | ||
| 305 | .dev = { | 414 | .dev = { |
| 306 | .platform_data = &dma_platform_data, | 415 | .platform_data = &dma0_platform_data, |
| 416 | }, | ||
| 417 | }; | ||
| 418 | |||
| 419 | static struct platform_device dma1_device = { | ||
| 420 | .name = "sh-dma-engine", | ||
| 421 | .id = 1, | ||
| 422 | .resource = sh7785_dmae1_resources, | ||
| 423 | .num_resources = ARRAY_SIZE(sh7785_dmae1_resources), | ||
| 424 | .dev = { | ||
| 425 | .platform_data = &dma1_platform_data, | ||
| 307 | }, | 426 | }, |
| 308 | }; | 427 | }; |
| 309 | 428 | ||
| @@ -320,7 +439,8 @@ static struct platform_device *sh7785_devices[] __initdata = { | |||
| 320 | &tmu3_device, | 439 | &tmu3_device, |
| 321 | &tmu4_device, | 440 | &tmu4_device, |
| 322 | &tmu5_device, | 441 | &tmu5_device, |
| 323 | &dma_device, | 442 | &dma0_device, |
| 443 | &dma1_device, | ||
| 324 | }; | 444 | }; |
| 325 | 445 | ||
| 326 | static int __init sh7785_devices_setup(void) | 446 | static int __init sh7785_devices_setup(void) |
diff --git a/arch/sh/kernel/hw_breakpoint.c b/arch/sh/kernel/hw_breakpoint.c index e2f1753d275c..675eea7785d9 100644 --- a/arch/sh/kernel/hw_breakpoint.c +++ b/arch/sh/kernel/hw_breakpoint.c | |||
| @@ -143,26 +143,6 @@ static int arch_check_va_in_kernelspace(unsigned long va, u8 hbp_len) | |||
| 143 | return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE); | 143 | return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE); |
| 144 | } | 144 | } |
| 145 | 145 | ||
| 146 | /* | ||
| 147 | * Store a breakpoint's encoded address, length, and type. | ||
| 148 | */ | ||
| 149 | static int arch_store_info(struct perf_event *bp) | ||
| 150 | { | ||
| 151 | struct arch_hw_breakpoint *info = counter_arch_bp(bp); | ||
| 152 | |||
| 153 | /* | ||
| 154 | * User-space requests will always have the address field populated | ||
| 155 | * For kernel-addresses, either the address or symbol name can be | ||
| 156 | * specified. | ||
| 157 | */ | ||
| 158 | if (info->name) | ||
| 159 | info->address = (unsigned long)kallsyms_lookup_name(info->name); | ||
| 160 | if (info->address) | ||
| 161 | return 0; | ||
| 162 | |||
| 163 | return -EINVAL; | ||
| 164 | } | ||
| 165 | |||
| 166 | int arch_bp_generic_fields(int sh_len, int sh_type, | 146 | int arch_bp_generic_fields(int sh_len, int sh_type, |
| 167 | int *gen_len, int *gen_type) | 147 | int *gen_len, int *gen_type) |
| 168 | { | 148 | { |
| @@ -276,10 +256,12 @@ int arch_validate_hwbkpt_settings(struct perf_event *bp, | |||
| 276 | return ret; | 256 | return ret; |
| 277 | } | 257 | } |
| 278 | 258 | ||
| 279 | ret = arch_store_info(bp); | 259 | /* |
| 280 | 260 | * For kernel-addresses, either the address or symbol name can be | |
| 281 | if (ret < 0) | 261 | * specified. |
| 282 | return ret; | 262 | */ |
| 263 | if (info->name) | ||
| 264 | info->address = (unsigned long)kallsyms_lookup_name(info->name); | ||
| 283 | 265 | ||
| 284 | /* | 266 | /* |
| 285 | * Check that the low-order bits of the address are appropriate | 267 | * Check that the low-order bits of the address are appropriate |
diff --git a/arch/sh/kernel/setup.c b/arch/sh/kernel/setup.c index 3459e70eed72..8870d6ba64bf 100644 --- a/arch/sh/kernel/setup.c +++ b/arch/sh/kernel/setup.c | |||
| @@ -443,7 +443,7 @@ void __init setup_arch(char **cmdline_p) | |||
| 443 | 443 | ||
| 444 | nodes_clear(node_online_map); | 444 | nodes_clear(node_online_map); |
| 445 | 445 | ||
| 446 | /* Setup bootmem with available RAM */ | 446 | pmb_init(); |
| 447 | lmb_init(); | 447 | lmb_init(); |
| 448 | setup_memory(); | 448 | setup_memory(); |
| 449 | sparse_init(); | 449 | sparse_init(); |
| @@ -452,7 +452,6 @@ void __init setup_arch(char **cmdline_p) | |||
| 452 | conswitchp = &dummy_con; | 452 | conswitchp = &dummy_con; |
| 453 | #endif | 453 | #endif |
| 454 | paging_init(); | 454 | paging_init(); |
| 455 | pmb_init(); | ||
| 456 | 455 | ||
| 457 | ioremap_fixed_init(); | 456 | ioremap_fixed_init(); |
| 458 | 457 | ||
diff --git a/arch/sh/kernel/time.c b/arch/sh/kernel/time.c index 953fa1613312..8a0072de2bcc 100644 --- a/arch/sh/kernel/time.c +++ b/arch/sh/kernel/time.c | |||
| @@ -39,12 +39,12 @@ static int null_rtc_set_time(const time_t secs) | |||
| 39 | void (*rtc_sh_get_time)(struct timespec *) = null_rtc_get_time; | 39 | void (*rtc_sh_get_time)(struct timespec *) = null_rtc_get_time; |
| 40 | int (*rtc_sh_set_time)(const time_t) = null_rtc_set_time; | 40 | int (*rtc_sh_set_time)(const time_t) = null_rtc_set_time; |
| 41 | 41 | ||
| 42 | #ifdef CONFIG_GENERIC_CMOS_UPDATE | ||
| 43 | void read_persistent_clock(struct timespec *ts) | 42 | void read_persistent_clock(struct timespec *ts) |
| 44 | { | 43 | { |
| 45 | rtc_sh_get_time(ts); | 44 | rtc_sh_get_time(ts); |
| 46 | } | 45 | } |
| 47 | 46 | ||
| 47 | #ifdef CONFIG_GENERIC_CMOS_UPDATE | ||
| 48 | int update_persistent_clock(struct timespec now) | 48 | int update_persistent_clock(struct timespec now) |
| 49 | { | 49 | { |
| 50 | return rtc_sh_set_time(now.tv_sec); | 50 | return rtc_sh_set_time(now.tv_sec); |
| @@ -113,9 +113,5 @@ void __init time_init(void) | |||
| 113 | hwblk_init(); | 113 | hwblk_init(); |
| 114 | clk_init(); | 114 | clk_init(); |
| 115 | 115 | ||
| 116 | rtc_sh_get_time(&xtime); | ||
| 117 | set_normalized_timespec(&wall_to_monotonic, | ||
| 118 | -xtime.tv_sec, -xtime.tv_nsec); | ||
| 119 | |||
| 120 | late_time_init = sh_late_time_init; | 116 | late_time_init = sh_late_time_init; |
| 121 | } | 117 | } |
diff --git a/arch/sh/lib/libgcc.h b/arch/sh/lib/libgcc.h index 3f19d1c5d942..05909d58e2fe 100644 --- a/arch/sh/lib/libgcc.h +++ b/arch/sh/lib/libgcc.h | |||
| @@ -17,8 +17,7 @@ struct DWstruct { | |||
| 17 | #error I feel sick. | 17 | #error I feel sick. |
| 18 | #endif | 18 | #endif |
| 19 | 19 | ||
| 20 | typedef union | 20 | typedef union { |
| 21 | { | ||
| 22 | struct DWstruct s; | 21 | struct DWstruct s; |
| 23 | long long ll; | 22 | long long ll; |
| 24 | } DWunion; | 23 | } DWunion; |
diff --git a/arch/sh/mm/ioremap.c b/arch/sh/mm/ioremap.c index c68d2d7d00a9..1ab2385ecefe 100644 --- a/arch/sh/mm/ioremap.c +++ b/arch/sh/mm/ioremap.c | |||
| @@ -34,11 +34,12 @@ | |||
| 34 | * caller shouldn't need to know that small detail. | 34 | * caller shouldn't need to know that small detail. |
| 35 | */ | 35 | */ |
| 36 | void __iomem * __init_refok | 36 | void __iomem * __init_refok |
| 37 | __ioremap_caller(unsigned long phys_addr, unsigned long size, | 37 | __ioremap_caller(phys_addr_t phys_addr, unsigned long size, |
| 38 | pgprot_t pgprot, void *caller) | 38 | pgprot_t pgprot, void *caller) |
| 39 | { | 39 | { |
| 40 | struct vm_struct *area; | 40 | struct vm_struct *area; |
| 41 | unsigned long offset, last_addr, addr, orig_addr; | 41 | unsigned long offset, last_addr, addr, orig_addr; |
| 42 | void __iomem *mapped; | ||
| 42 | 43 | ||
| 43 | /* Don't allow wraparound or zero size */ | 44 | /* Don't allow wraparound or zero size */ |
| 44 | last_addr = phys_addr + size - 1; | 45 | last_addr = phys_addr + size - 1; |
| @@ -46,6 +47,20 @@ __ioremap_caller(unsigned long phys_addr, unsigned long size, | |||
| 46 | return NULL; | 47 | return NULL; |
| 47 | 48 | ||
| 48 | /* | 49 | /* |
| 50 | * If we can't yet use the regular approach, go the fixmap route. | ||
| 51 | */ | ||
| 52 | if (!mem_init_done) | ||
| 53 | return ioremap_fixed(phys_addr, size, pgprot); | ||
| 54 | |||
| 55 | /* | ||
| 56 | * First try to remap through the PMB. | ||
| 57 | * PMB entries are all pre-faulted. | ||
| 58 | */ | ||
| 59 | mapped = pmb_remap_caller(phys_addr, size, pgprot, caller); | ||
| 60 | if (mapped && !IS_ERR(mapped)) | ||
| 61 | return mapped; | ||
| 62 | |||
| 63 | /* | ||
| 49 | * Mappings have to be page-aligned | 64 | * Mappings have to be page-aligned |
| 50 | */ | 65 | */ |
| 51 | offset = phys_addr & ~PAGE_MASK; | 66 | offset = phys_addr & ~PAGE_MASK; |
| @@ -53,12 +68,6 @@ __ioremap_caller(unsigned long phys_addr, unsigned long size, | |||
| 53 | size = PAGE_ALIGN(last_addr+1) - phys_addr; | 68 | size = PAGE_ALIGN(last_addr+1) - phys_addr; |
| 54 | 69 | ||
| 55 | /* | 70 | /* |
| 56 | * If we can't yet use the regular approach, go the fixmap route. | ||
| 57 | */ | ||
| 58 | if (!mem_init_done) | ||
| 59 | return ioremap_fixed(phys_addr, offset, size, pgprot); | ||
| 60 | |||
| 61 | /* | ||
| 62 | * Ok, go for it.. | 71 | * Ok, go for it.. |
| 63 | */ | 72 | */ |
| 64 | area = get_vm_area_caller(size, VM_IOREMAP, caller); | 73 | area = get_vm_area_caller(size, VM_IOREMAP, caller); |
| @@ -67,33 +76,10 @@ __ioremap_caller(unsigned long phys_addr, unsigned long size, | |||
| 67 | area->phys_addr = phys_addr; | 76 | area->phys_addr = phys_addr; |
| 68 | orig_addr = addr = (unsigned long)area->addr; | 77 | orig_addr = addr = (unsigned long)area->addr; |
| 69 | 78 | ||
| 70 | #ifdef CONFIG_PMB | 79 | if (ioremap_page_range(addr, addr + size, phys_addr, pgprot)) { |
| 71 | /* | 80 | vunmap((void *)orig_addr); |
| 72 | * First try to remap through the PMB once a valid VMA has been | 81 | return NULL; |
| 73 | * established. Smaller allocations (or the rest of the size | ||
| 74 | * remaining after a PMB mapping due to the size not being | ||
| 75 | * perfectly aligned on a PMB size boundary) are then mapped | ||
| 76 | * through the UTLB using conventional page tables. | ||
| 77 | * | ||
| 78 | * PMB entries are all pre-faulted. | ||
| 79 | */ | ||
| 80 | if (unlikely(phys_addr >= P1SEG)) { | ||
| 81 | unsigned long mapped; | ||
| 82 | |||
| 83 | mapped = pmb_remap(addr, phys_addr, size, pgprot); | ||
| 84 | if (likely(mapped)) { | ||
| 85 | addr += mapped; | ||
| 86 | phys_addr += mapped; | ||
| 87 | size -= mapped; | ||
| 88 | } | ||
| 89 | } | 82 | } |
| 90 | #endif | ||
| 91 | |||
| 92 | if (likely(size)) | ||
| 93 | if (ioremap_page_range(addr, addr + size, phys_addr, pgprot)) { | ||
| 94 | vunmap((void *)orig_addr); | ||
| 95 | return NULL; | ||
| 96 | } | ||
| 97 | 83 | ||
| 98 | return (void __iomem *)(offset + (char *)orig_addr); | 84 | return (void __iomem *)(offset + (char *)orig_addr); |
| 99 | } | 85 | } |
| @@ -133,23 +119,11 @@ void __iounmap(void __iomem *addr) | |||
| 133 | if (iounmap_fixed(addr) == 0) | 119 | if (iounmap_fixed(addr) == 0) |
| 134 | return; | 120 | return; |
| 135 | 121 | ||
| 136 | #ifdef CONFIG_PMB | ||
| 137 | /* | 122 | /* |
| 138 | * Purge any PMB entries that may have been established for this | 123 | * If the PMB handled it, there's nothing else to do. |
| 139 | * mapping, then proceed with conventional VMA teardown. | ||
| 140 | * | ||
| 141 | * XXX: Note that due to the way that remove_vm_area() does | ||
| 142 | * matching of the resultant VMA, we aren't able to fast-forward | ||
| 143 | * the address past the PMB space until the end of the VMA where | ||
| 144 | * the page tables reside. As such, unmap_vm_area() will be | ||
| 145 | * forced to linearly scan over the area until it finds the page | ||
| 146 | * tables where PTEs that need to be unmapped actually reside, | ||
| 147 | * which is far from optimal. Perhaps we need to use a separate | ||
| 148 | * VMA for the PMB mappings? | ||
| 149 | * -- PFM. | ||
| 150 | */ | 124 | */ |
| 151 | pmb_unmap(vaddr); | 125 | if (pmb_unmap(addr) == 0) |
| 152 | #endif | 126 | return; |
| 153 | 127 | ||
| 154 | p = remove_vm_area((void *)(vaddr & PAGE_MASK)); | 128 | p = remove_vm_area((void *)(vaddr & PAGE_MASK)); |
| 155 | if (!p) { | 129 | if (!p) { |
diff --git a/arch/sh/mm/ioremap_fixed.c b/arch/sh/mm/ioremap_fixed.c index 0b78b1e20ef1..7f682e5dafcf 100644 --- a/arch/sh/mm/ioremap_fixed.c +++ b/arch/sh/mm/ioremap_fixed.c | |||
| @@ -45,14 +45,21 @@ void __init ioremap_fixed_init(void) | |||
| 45 | } | 45 | } |
| 46 | 46 | ||
| 47 | void __init __iomem * | 47 | void __init __iomem * |
| 48 | ioremap_fixed(resource_size_t phys_addr, unsigned long offset, | 48 | ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) |
| 49 | unsigned long size, pgprot_t prot) | ||
| 50 | { | 49 | { |
| 51 | enum fixed_addresses idx0, idx; | 50 | enum fixed_addresses idx0, idx; |
| 52 | struct ioremap_map *map; | 51 | struct ioremap_map *map; |
| 53 | unsigned int nrpages; | 52 | unsigned int nrpages; |
| 53 | unsigned long offset; | ||
| 54 | int i, slot; | 54 | int i, slot; |
| 55 | 55 | ||
| 56 | /* | ||
| 57 | * Mappings have to be page-aligned | ||
| 58 | */ | ||
| 59 | offset = phys_addr & ~PAGE_MASK; | ||
| 60 | phys_addr &= PAGE_MASK; | ||
| 61 | size = PAGE_ALIGN(phys_addr + size) - phys_addr; | ||
| 62 | |||
| 56 | slot = -1; | 63 | slot = -1; |
| 57 | for (i = 0; i < FIX_N_IOREMAPS; i++) { | 64 | for (i = 0; i < FIX_N_IOREMAPS; i++) { |
| 58 | map = &ioremap_maps[i]; | 65 | map = &ioremap_maps[i]; |
diff --git a/arch/sh/mm/numa.c b/arch/sh/mm/numa.c index 422e92721878..961b34085e3b 100644 --- a/arch/sh/mm/numa.c +++ b/arch/sh/mm/numa.c | |||
| @@ -74,6 +74,9 @@ void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end) | |||
| 74 | start_pfn = start >> PAGE_SHIFT; | 74 | start_pfn = start >> PAGE_SHIFT; |
| 75 | end_pfn = end >> PAGE_SHIFT; | 75 | end_pfn = end >> PAGE_SHIFT; |
| 76 | 76 | ||
| 77 | pmb_bolt_mapping((unsigned long)__va(start), start, end - start, | ||
| 78 | PAGE_KERNEL); | ||
| 79 | |||
| 77 | lmb_add(start, end - start); | 80 | lmb_add(start, end - start); |
| 78 | 81 | ||
| 79 | __add_active_range(nid, start_pfn, end_pfn); | 82 | __add_active_range(nid, start_pfn, end_pfn); |
diff --git a/arch/sh/mm/pmb.c b/arch/sh/mm/pmb.c index 198bcff5e96f..a4662e2782c3 100644 --- a/arch/sh/mm/pmb.c +++ b/arch/sh/mm/pmb.c | |||
| @@ -23,7 +23,8 @@ | |||
| 23 | #include <linux/err.h> | 23 | #include <linux/err.h> |
| 24 | #include <linux/io.h> | 24 | #include <linux/io.h> |
| 25 | #include <linux/spinlock.h> | 25 | #include <linux/spinlock.h> |
| 26 | #include <linux/rwlock.h> | 26 | #include <linux/vmalloc.h> |
| 27 | #include <asm/cacheflush.h> | ||
| 27 | #include <asm/sizes.h> | 28 | #include <asm/sizes.h> |
| 28 | #include <asm/system.h> | 29 | #include <asm/system.h> |
| 29 | #include <asm/uaccess.h> | 30 | #include <asm/uaccess.h> |
| @@ -52,12 +53,24 @@ struct pmb_entry { | |||
| 52 | struct pmb_entry *link; | 53 | struct pmb_entry *link; |
| 53 | }; | 54 | }; |
| 54 | 55 | ||
| 56 | static struct { | ||
| 57 | unsigned long size; | ||
| 58 | int flag; | ||
| 59 | } pmb_sizes[] = { | ||
| 60 | { .size = SZ_512M, .flag = PMB_SZ_512M, }, | ||
| 61 | { .size = SZ_128M, .flag = PMB_SZ_128M, }, | ||
| 62 | { .size = SZ_64M, .flag = PMB_SZ_64M, }, | ||
| 63 | { .size = SZ_16M, .flag = PMB_SZ_16M, }, | ||
| 64 | }; | ||
| 65 | |||
| 55 | static void pmb_unmap_entry(struct pmb_entry *, int depth); | 66 | static void pmb_unmap_entry(struct pmb_entry *, int depth); |
| 56 | 67 | ||
| 57 | static DEFINE_RWLOCK(pmb_rwlock); | 68 | static DEFINE_RWLOCK(pmb_rwlock); |
| 58 | static struct pmb_entry pmb_entry_list[NR_PMB_ENTRIES]; | 69 | static struct pmb_entry pmb_entry_list[NR_PMB_ENTRIES]; |
| 59 | static DECLARE_BITMAP(pmb_map, NR_PMB_ENTRIES); | 70 | static DECLARE_BITMAP(pmb_map, NR_PMB_ENTRIES); |
| 60 | 71 | ||
| 72 | static unsigned int pmb_iomapping_enabled; | ||
| 73 | |||
| 61 | static __always_inline unsigned long mk_pmb_entry(unsigned int entry) | 74 | static __always_inline unsigned long mk_pmb_entry(unsigned int entry) |
| 62 | { | 75 | { |
| 63 | return (entry & PMB_E_MASK) << PMB_E_SHIFT; | 76 | return (entry & PMB_E_MASK) << PMB_E_SHIFT; |
| @@ -73,6 +86,142 @@ static __always_inline unsigned long mk_pmb_data(unsigned int entry) | |||
| 73 | return mk_pmb_entry(entry) | PMB_DATA; | 86 | return mk_pmb_entry(entry) | PMB_DATA; |
| 74 | } | 87 | } |
| 75 | 88 | ||
| 89 | static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn) | ||
| 90 | { | ||
| 91 | return ppn >= __pa(memory_start) && ppn < __pa(memory_end); | ||
| 92 | } | ||
| 93 | |||
| 94 | /* | ||
| 95 | * Ensure that the PMB entries match our cache configuration. | ||
| 96 | * | ||
| 97 | * When we are in 32-bit address extended mode, CCR.CB becomes | ||
| 98 | * invalid, so care must be taken to manually adjust cacheable | ||
| 99 | * translations. | ||
| 100 | */ | ||
| 101 | static __always_inline unsigned long pmb_cache_flags(void) | ||
| 102 | { | ||
| 103 | unsigned long flags = 0; | ||
| 104 | |||
| 105 | #if defined(CONFIG_CACHE_OFF) | ||
| 106 | flags |= PMB_WT | PMB_UB; | ||
| 107 | #elif defined(CONFIG_CACHE_WRITETHROUGH) | ||
| 108 | flags |= PMB_C | PMB_WT | PMB_UB; | ||
| 109 | #elif defined(CONFIG_CACHE_WRITEBACK) | ||
| 110 | flags |= PMB_C; | ||
| 111 | #endif | ||
| 112 | |||
| 113 | return flags; | ||
| 114 | } | ||
| 115 | |||
| 116 | /* | ||
| 117 | * Convert typical pgprot value to the PMB equivalent | ||
| 118 | */ | ||
| 119 | static inline unsigned long pgprot_to_pmb_flags(pgprot_t prot) | ||
| 120 | { | ||
| 121 | unsigned long pmb_flags = 0; | ||
| 122 | u64 flags = pgprot_val(prot); | ||
| 123 | |||
| 124 | if (flags & _PAGE_CACHABLE) | ||
| 125 | pmb_flags |= PMB_C; | ||
| 126 | if (flags & _PAGE_WT) | ||
| 127 | pmb_flags |= PMB_WT | PMB_UB; | ||
| 128 | |||
| 129 | return pmb_flags; | ||
| 130 | } | ||
| 131 | |||
| 132 | static inline bool pmb_can_merge(struct pmb_entry *a, struct pmb_entry *b) | ||
| 133 | { | ||
| 134 | return (b->vpn == (a->vpn + a->size)) && | ||
| 135 | (b->ppn == (a->ppn + a->size)) && | ||
| 136 | (b->flags == a->flags); | ||
| 137 | } | ||
| 138 | |||
| 139 | static bool pmb_mapping_exists(unsigned long vaddr, phys_addr_t phys, | ||
| 140 | unsigned long size) | ||
| 141 | { | ||
| 142 | int i; | ||
| 143 | |||
| 144 | read_lock(&pmb_rwlock); | ||
| 145 | |||
| 146 | for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) { | ||
| 147 | struct pmb_entry *pmbe, *iter; | ||
| 148 | unsigned long span; | ||
| 149 | |||
| 150 | if (!test_bit(i, pmb_map)) | ||
| 151 | continue; | ||
| 152 | |||
| 153 | pmbe = &pmb_entry_list[i]; | ||
| 154 | |||
| 155 | /* | ||
| 156 | * See if VPN and PPN are bounded by an existing mapping. | ||
| 157 | */ | ||
| 158 | if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) | ||
| 159 | continue; | ||
| 160 | if ((phys < pmbe->ppn) || (phys >= (pmbe->ppn + pmbe->size))) | ||
| 161 | continue; | ||
| 162 | |||
| 163 | /* | ||
| 164 | * Now see if we're in range of a simple mapping. | ||
| 165 | */ | ||
| 166 | if (size <= pmbe->size) { | ||
| 167 | read_unlock(&pmb_rwlock); | ||
| 168 | return true; | ||
| 169 | } | ||
| 170 | |||
| 171 | span = pmbe->size; | ||
| 172 | |||
| 173 | /* | ||
| 174 | * Finally for sizes that involve compound mappings, walk | ||
| 175 | * the chain. | ||
| 176 | */ | ||
| 177 | for (iter = pmbe->link; iter; iter = iter->link) | ||
| 178 | span += iter->size; | ||
| 179 | |||
| 180 | /* | ||
| 181 | * Nothing else to do if the range requirements are met. | ||
| 182 | */ | ||
| 183 | if (size <= span) { | ||
| 184 | read_unlock(&pmb_rwlock); | ||
| 185 | return true; | ||
| 186 | } | ||
| 187 | } | ||
| 188 | |||
| 189 | read_unlock(&pmb_rwlock); | ||
| 190 | return false; | ||
| 191 | } | ||
| 192 | |||
| 193 | static bool pmb_size_valid(unsigned long size) | ||
| 194 | { | ||
| 195 | int i; | ||
| 196 | |||
| 197 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) | ||
| 198 | if (pmb_sizes[i].size == size) | ||
| 199 | return true; | ||
| 200 | |||
| 201 | return false; | ||
| 202 | } | ||
| 203 | |||
| 204 | static inline bool pmb_addr_valid(unsigned long addr, unsigned long size) | ||
| 205 | { | ||
| 206 | return (addr >= P1SEG && (addr + size - 1) < P3SEG); | ||
| 207 | } | ||
| 208 | |||
| 209 | static inline bool pmb_prot_valid(pgprot_t prot) | ||
| 210 | { | ||
| 211 | return (pgprot_val(prot) & _PAGE_USER) == 0; | ||
| 212 | } | ||
| 213 | |||
| 214 | static int pmb_size_to_flags(unsigned long size) | ||
| 215 | { | ||
| 216 | int i; | ||
| 217 | |||
| 218 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) | ||
| 219 | if (pmb_sizes[i].size == size) | ||
| 220 | return pmb_sizes[i].flag; | ||
| 221 | |||
| 222 | return 0; | ||
| 223 | } | ||
| 224 | |||
| 76 | static int pmb_alloc_entry(void) | 225 | static int pmb_alloc_entry(void) |
| 77 | { | 226 | { |
| 78 | int pos; | 227 | int pos; |
| @@ -140,33 +289,22 @@ static void pmb_free(struct pmb_entry *pmbe) | |||
| 140 | } | 289 | } |
| 141 | 290 | ||
| 142 | /* | 291 | /* |
| 143 | * Ensure that the PMB entries match our cache configuration. | 292 | * Must be run uncached. |
| 144 | * | ||
| 145 | * When we are in 32-bit address extended mode, CCR.CB becomes | ||
| 146 | * invalid, so care must be taken to manually adjust cacheable | ||
| 147 | * translations. | ||
| 148 | */ | 293 | */ |
| 149 | static __always_inline unsigned long pmb_cache_flags(void) | 294 | static void __set_pmb_entry(struct pmb_entry *pmbe) |
| 150 | { | 295 | { |
| 151 | unsigned long flags = 0; | 296 | unsigned long addr, data; |
| 152 | 297 | ||
| 153 | #if defined(CONFIG_CACHE_WRITETHROUGH) | 298 | addr = mk_pmb_addr(pmbe->entry); |
| 154 | flags |= PMB_C | PMB_WT | PMB_UB; | 299 | data = mk_pmb_data(pmbe->entry); |
| 155 | #elif defined(CONFIG_CACHE_WRITEBACK) | ||
| 156 | flags |= PMB_C; | ||
| 157 | #endif | ||
| 158 | 300 | ||
| 159 | return flags; | 301 | jump_to_uncached(); |
| 160 | } | ||
| 161 | 302 | ||
| 162 | /* | 303 | /* Set V-bit */ |
| 163 | * Must be run uncached. | 304 | __raw_writel(pmbe->vpn | PMB_V, addr); |
| 164 | */ | 305 | __raw_writel(pmbe->ppn | pmbe->flags | PMB_V, data); |
| 165 | static void __set_pmb_entry(struct pmb_entry *pmbe) | 306 | |
| 166 | { | 307 | back_to_cached(); |
| 167 | writel_uncached(pmbe->vpn | PMB_V, mk_pmb_addr(pmbe->entry)); | ||
| 168 | writel_uncached(pmbe->ppn | pmbe->flags | PMB_V, | ||
| 169 | mk_pmb_data(pmbe->entry)); | ||
| 170 | } | 308 | } |
| 171 | 309 | ||
| 172 | static void __clear_pmb_entry(struct pmb_entry *pmbe) | 310 | static void __clear_pmb_entry(struct pmb_entry *pmbe) |
| @@ -194,144 +332,155 @@ static void set_pmb_entry(struct pmb_entry *pmbe) | |||
| 194 | spin_unlock_irqrestore(&pmbe->lock, flags); | 332 | spin_unlock_irqrestore(&pmbe->lock, flags); |
| 195 | } | 333 | } |
| 196 | 334 | ||
| 197 | static struct { | 335 | int pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, |
| 198 | unsigned long size; | 336 | unsigned long size, pgprot_t prot) |
| 199 | int flag; | ||
| 200 | } pmb_sizes[] = { | ||
| 201 | { .size = SZ_512M, .flag = PMB_SZ_512M, }, | ||
| 202 | { .size = SZ_128M, .flag = PMB_SZ_128M, }, | ||
| 203 | { .size = SZ_64M, .flag = PMB_SZ_64M, }, | ||
| 204 | { .size = SZ_16M, .flag = PMB_SZ_16M, }, | ||
| 205 | }; | ||
| 206 | |||
| 207 | long pmb_remap(unsigned long vaddr, unsigned long phys, | ||
| 208 | unsigned long size, pgprot_t prot) | ||
| 209 | { | 337 | { |
| 210 | struct pmb_entry *pmbp, *pmbe; | 338 | struct pmb_entry *pmbp, *pmbe; |
| 211 | unsigned long wanted; | 339 | unsigned long orig_addr, orig_size; |
| 212 | int pmb_flags, i; | 340 | unsigned long flags, pmb_flags; |
| 213 | long err; | 341 | int i, mapped; |
| 214 | u64 flags; | ||
| 215 | 342 | ||
| 216 | flags = pgprot_val(prot); | 343 | if (!pmb_addr_valid(vaddr, size)) |
| 344 | return -EFAULT; | ||
| 345 | if (pmb_mapping_exists(vaddr, phys, size)) | ||
| 346 | return 0; | ||
| 217 | 347 | ||
| 218 | pmb_flags = PMB_WT | PMB_UB; | 348 | orig_addr = vaddr; |
| 219 | 349 | orig_size = size; | |
| 220 | /* Convert typical pgprot value to the PMB equivalent */ | ||
| 221 | if (flags & _PAGE_CACHABLE) { | ||
| 222 | pmb_flags |= PMB_C; | ||
| 223 | 350 | ||
| 224 | if ((flags & _PAGE_WT) == 0) | 351 | flush_tlb_kernel_range(vaddr, vaddr + size); |
| 225 | pmb_flags &= ~(PMB_WT | PMB_UB); | ||
| 226 | } | ||
| 227 | 352 | ||
| 353 | pmb_flags = pgprot_to_pmb_flags(prot); | ||
| 228 | pmbp = NULL; | 354 | pmbp = NULL; |
| 229 | wanted = size; | ||
| 230 | 355 | ||
| 231 | again: | 356 | do { |
| 232 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) { | 357 | for (i = mapped = 0; i < ARRAY_SIZE(pmb_sizes); i++) { |
| 233 | unsigned long flags; | 358 | if (size < pmb_sizes[i].size) |
| 359 | continue; | ||
| 360 | |||
| 361 | pmbe = pmb_alloc(vaddr, phys, pmb_flags | | ||
| 362 | pmb_sizes[i].flag, PMB_NO_ENTRY); | ||
| 363 | if (IS_ERR(pmbe)) { | ||
| 364 | pmb_unmap_entry(pmbp, mapped); | ||
| 365 | return PTR_ERR(pmbe); | ||
| 366 | } | ||
| 234 | 367 | ||
| 235 | if (size < pmb_sizes[i].size) | 368 | spin_lock_irqsave(&pmbe->lock, flags); |
| 236 | continue; | ||
| 237 | 369 | ||
| 238 | pmbe = pmb_alloc(vaddr, phys, pmb_flags | pmb_sizes[i].flag, | 370 | pmbe->size = pmb_sizes[i].size; |
| 239 | PMB_NO_ENTRY); | ||
| 240 | if (IS_ERR(pmbe)) { | ||
| 241 | err = PTR_ERR(pmbe); | ||
| 242 | goto out; | ||
| 243 | } | ||
| 244 | 371 | ||
| 245 | spin_lock_irqsave(&pmbe->lock, flags); | 372 | __set_pmb_entry(pmbe); |
| 246 | 373 | ||
| 247 | __set_pmb_entry(pmbe); | 374 | phys += pmbe->size; |
| 375 | vaddr += pmbe->size; | ||
| 376 | size -= pmbe->size; | ||
| 248 | 377 | ||
| 249 | phys += pmb_sizes[i].size; | 378 | /* |
| 250 | vaddr += pmb_sizes[i].size; | 379 | * Link adjacent entries that span multiple PMB |
| 251 | size -= pmb_sizes[i].size; | 380 | * entries for easier tear-down. |
| 381 | */ | ||
| 382 | if (likely(pmbp)) { | ||
| 383 | spin_lock(&pmbp->lock); | ||
| 384 | pmbp->link = pmbe; | ||
| 385 | spin_unlock(&pmbp->lock); | ||
| 386 | } | ||
| 252 | 387 | ||
| 253 | pmbe->size = pmb_sizes[i].size; | 388 | pmbp = pmbe; |
| 254 | 389 | ||
| 255 | /* | 390 | /* |
| 256 | * Link adjacent entries that span multiple PMB entries | 391 | * Instead of trying smaller sizes on every |
| 257 | * for easier tear-down. | 392 | * iteration (even if we succeed in allocating |
| 258 | */ | 393 | * space), try using pmb_sizes[i].size again. |
| 259 | if (likely(pmbp)) { | 394 | */ |
| 260 | spin_lock(&pmbp->lock); | 395 | i--; |
| 261 | pmbp->link = pmbe; | 396 | mapped++; |
| 262 | spin_unlock(&pmbp->lock); | 397 | |
| 398 | spin_unlock_irqrestore(&pmbe->lock, flags); | ||
| 263 | } | 399 | } |
| 400 | } while (size >= SZ_16M); | ||
| 264 | 401 | ||
| 265 | pmbp = pmbe; | 402 | flush_cache_vmap(orig_addr, orig_addr + orig_size); |
| 266 | 403 | ||
| 267 | /* | 404 | return 0; |
| 268 | * Instead of trying smaller sizes on every iteration | 405 | } |
| 269 | * (even if we succeed in allocating space), try using | ||
| 270 | * pmb_sizes[i].size again. | ||
| 271 | */ | ||
| 272 | i--; | ||
| 273 | 406 | ||
| 274 | spin_unlock_irqrestore(&pmbe->lock, flags); | 407 | void __iomem *pmb_remap_caller(phys_addr_t phys, unsigned long size, |
| 275 | } | 408 | pgprot_t prot, void *caller) |
| 409 | { | ||
| 410 | unsigned long vaddr; | ||
| 411 | phys_addr_t offset, last_addr; | ||
| 412 | phys_addr_t align_mask; | ||
| 413 | unsigned long aligned; | ||
| 414 | struct vm_struct *area; | ||
| 415 | int i, ret; | ||
| 276 | 416 | ||
| 277 | if (size >= SZ_16M) | 417 | if (!pmb_iomapping_enabled) |
| 278 | goto again; | 418 | return NULL; |
| 279 | 419 | ||
| 280 | return wanted - size; | 420 | /* |
| 421 | * Small mappings need to go through the TLB. | ||
| 422 | */ | ||
| 423 | if (size < SZ_16M) | ||
| 424 | return ERR_PTR(-EINVAL); | ||
| 425 | if (!pmb_prot_valid(prot)) | ||
| 426 | return ERR_PTR(-EINVAL); | ||
| 281 | 427 | ||
| 282 | out: | 428 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) |
| 283 | pmb_unmap_entry(pmbp, NR_PMB_ENTRIES); | 429 | if (size >= pmb_sizes[i].size) |
| 430 | break; | ||
| 431 | |||
| 432 | last_addr = phys + size; | ||
| 433 | align_mask = ~(pmb_sizes[i].size - 1); | ||
| 434 | offset = phys & ~align_mask; | ||
| 435 | phys &= align_mask; | ||
| 436 | aligned = ALIGN(last_addr, pmb_sizes[i].size) - phys; | ||
| 437 | |||
| 438 | /* | ||
| 439 | * XXX: This should really start from uncached_end, but this | ||
| 440 | * causes the MMU to reset, so for now we restrict it to the | ||
| 441 | * 0xb000...0xc000 range. | ||
| 442 | */ | ||
| 443 | area = __get_vm_area_caller(aligned, VM_IOREMAP, 0xb0000000, | ||
| 444 | P3SEG, caller); | ||
| 445 | if (!area) | ||
| 446 | return NULL; | ||
| 447 | |||
| 448 | area->phys_addr = phys; | ||
| 449 | vaddr = (unsigned long)area->addr; | ||
| 450 | |||
| 451 | ret = pmb_bolt_mapping(vaddr, phys, size, prot); | ||
| 452 | if (unlikely(ret != 0)) | ||
| 453 | return ERR_PTR(ret); | ||
| 284 | 454 | ||
| 285 | return err; | 455 | return (void __iomem *)(offset + (char *)vaddr); |
| 286 | } | 456 | } |
| 287 | 457 | ||
| 288 | void pmb_unmap(unsigned long addr) | 458 | int pmb_unmap(void __iomem *addr) |
| 289 | { | 459 | { |
| 290 | struct pmb_entry *pmbe = NULL; | 460 | struct pmb_entry *pmbe = NULL; |
| 291 | int i; | 461 | unsigned long vaddr = (unsigned long __force)addr; |
| 462 | int i, found = 0; | ||
| 292 | 463 | ||
| 293 | read_lock(&pmb_rwlock); | 464 | read_lock(&pmb_rwlock); |
| 294 | 465 | ||
| 295 | for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) { | 466 | for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) { |
| 296 | if (test_bit(i, pmb_map)) { | 467 | if (test_bit(i, pmb_map)) { |
| 297 | pmbe = &pmb_entry_list[i]; | 468 | pmbe = &pmb_entry_list[i]; |
| 298 | if (pmbe->vpn == addr) | 469 | if (pmbe->vpn == vaddr) { |
| 470 | found = 1; | ||
| 299 | break; | 471 | break; |
| 472 | } | ||
| 300 | } | 473 | } |
| 301 | } | 474 | } |
| 302 | 475 | ||
| 303 | read_unlock(&pmb_rwlock); | 476 | read_unlock(&pmb_rwlock); |
| 304 | 477 | ||
| 305 | pmb_unmap_entry(pmbe, NR_PMB_ENTRIES); | 478 | if (found) { |
| 306 | } | 479 | pmb_unmap_entry(pmbe, NR_PMB_ENTRIES); |
| 307 | 480 | return 0; | |
| 308 | static bool pmb_can_merge(struct pmb_entry *a, struct pmb_entry *b) | 481 | } |
| 309 | { | ||
| 310 | return (b->vpn == (a->vpn + a->size)) && | ||
| 311 | (b->ppn == (a->ppn + a->size)) && | ||
| 312 | (b->flags == a->flags); | ||
| 313 | } | ||
| 314 | |||
| 315 | static bool pmb_size_valid(unsigned long size) | ||
| 316 | { | ||
| 317 | int i; | ||
| 318 | |||
| 319 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) | ||
| 320 | if (pmb_sizes[i].size == size) | ||
| 321 | return true; | ||
| 322 | |||
| 323 | return false; | ||
| 324 | } | ||
| 325 | |||
| 326 | static int pmb_size_to_flags(unsigned long size) | ||
| 327 | { | ||
| 328 | int i; | ||
| 329 | |||
| 330 | for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) | ||
| 331 | if (pmb_sizes[i].size == size) | ||
| 332 | return pmb_sizes[i].flag; | ||
| 333 | 482 | ||
| 334 | return 0; | 483 | return -EINVAL; |
| 335 | } | 484 | } |
| 336 | 485 | ||
| 337 | static void __pmb_unmap_entry(struct pmb_entry *pmbe, int depth) | 486 | static void __pmb_unmap_entry(struct pmb_entry *pmbe, int depth) |
| @@ -351,6 +500,8 @@ static void __pmb_unmap_entry(struct pmb_entry *pmbe, int depth) | |||
| 351 | */ | 500 | */ |
| 352 | __clear_pmb_entry(pmbe); | 501 | __clear_pmb_entry(pmbe); |
| 353 | 502 | ||
| 503 | flush_cache_vunmap(pmbe->vpn, pmbe->vpn + pmbe->size); | ||
| 504 | |||
| 354 | pmbe = pmblink->link; | 505 | pmbe = pmblink->link; |
| 355 | 506 | ||
| 356 | pmb_free(pmblink); | 507 | pmb_free(pmblink); |
| @@ -369,11 +520,6 @@ static void pmb_unmap_entry(struct pmb_entry *pmbe, int depth) | |||
| 369 | write_unlock_irqrestore(&pmb_rwlock, flags); | 520 | write_unlock_irqrestore(&pmb_rwlock, flags); |
| 370 | } | 521 | } |
| 371 | 522 | ||
| 372 | static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn) | ||
| 373 | { | ||
| 374 | return ppn >= __pa(memory_start) && ppn < __pa(memory_end); | ||
| 375 | } | ||
| 376 | |||
| 377 | static void __init pmb_notify(void) | 523 | static void __init pmb_notify(void) |
| 378 | { | 524 | { |
| 379 | int i; | 525 | int i; |
| @@ -625,6 +771,18 @@ static void __init pmb_resize(void) | |||
| 625 | } | 771 | } |
| 626 | #endif | 772 | #endif |
| 627 | 773 | ||
| 774 | static int __init early_pmb(char *p) | ||
| 775 | { | ||
| 776 | if (!p) | ||
| 777 | return 0; | ||
| 778 | |||
| 779 | if (strstr(p, "iomap")) | ||
| 780 | pmb_iomapping_enabled = 1; | ||
| 781 | |||
| 782 | return 0; | ||
| 783 | } | ||
| 784 | early_param("pmb", early_pmb); | ||
| 785 | |||
| 628 | void __init pmb_init(void) | 786 | void __init pmb_init(void) |
| 629 | { | 787 | { |
| 630 | /* Synchronize software state */ | 788 | /* Synchronize software state */ |
| @@ -713,7 +871,7 @@ static int __init pmb_debugfs_init(void) | |||
| 713 | 871 | ||
| 714 | return 0; | 872 | return 0; |
| 715 | } | 873 | } |
| 716 | postcore_initcall(pmb_debugfs_init); | 874 | subsys_initcall(pmb_debugfs_init); |
| 717 | 875 | ||
| 718 | #ifdef CONFIG_PM | 876 | #ifdef CONFIG_PM |
| 719 | static int pmb_sysdev_suspend(struct sys_device *dev, pm_message_t state) | 877 | static int pmb_sysdev_suspend(struct sys_device *dev, pm_message_t state) |
diff --git a/drivers/dma/shdma.c b/drivers/dma/shdma.c index b75ce8b84c46..5d17e09cb625 100644 --- a/drivers/dma/shdma.c +++ b/drivers/dma/shdma.c | |||
| @@ -24,8 +24,10 @@ | |||
| 24 | #include <linux/delay.h> | 24 | #include <linux/delay.h> |
| 25 | #include <linux/dma-mapping.h> | 25 | #include <linux/dma-mapping.h> |
| 26 | #include <linux/platform_device.h> | 26 | #include <linux/platform_device.h> |
| 27 | #include <cpu/dma.h> | 27 | #include <linux/pm_runtime.h> |
| 28 | #include <asm/dma-sh.h> | 28 | |
| 29 | #include <asm/dmaengine.h> | ||
| 30 | |||
| 29 | #include "shdma.h" | 31 | #include "shdma.h" |
| 30 | 32 | ||
| 31 | /* DMA descriptor control */ | 33 | /* DMA descriptor control */ |
| @@ -38,30 +40,32 @@ enum sh_dmae_desc_status { | |||
| 38 | }; | 40 | }; |
| 39 | 41 | ||
| 40 | #define NR_DESCS_PER_CHANNEL 32 | 42 | #define NR_DESCS_PER_CHANNEL 32 |
| 41 | /* | 43 | /* Default MEMCPY transfer size = 2^2 = 4 bytes */ |
| 42 | * Define the default configuration for dual address memory-memory transfer. | 44 | #define LOG2_DEFAULT_XFER_SIZE 2 |
| 43 | * The 0x400 value represents auto-request, external->external. | ||
| 44 | * | ||
| 45 | * And this driver set 4byte burst mode. | ||
| 46 | * If you want to change mode, you need to change RS_DEFAULT of value. | ||
| 47 | * (ex 1byte burst mode -> (RS_DUAL & ~TS_32) | ||
| 48 | */ | ||
| 49 | #define RS_DEFAULT (RS_DUAL) | ||
| 50 | 45 | ||
| 51 | /* A bitmask with bits enough for enum sh_dmae_slave_chan_id */ | 46 | /* A bitmask with bits enough for enum sh_dmae_slave_chan_id */ |
| 52 | static unsigned long sh_dmae_slave_used[BITS_TO_LONGS(SHDMA_SLAVE_NUMBER)]; | 47 | static unsigned long sh_dmae_slave_used[BITS_TO_LONGS(SHDMA_SLAVE_NUMBER)]; |
| 53 | 48 | ||
| 54 | static void sh_dmae_chan_ld_cleanup(struct sh_dmae_chan *sh_chan, bool all); | 49 | static void sh_dmae_chan_ld_cleanup(struct sh_dmae_chan *sh_chan, bool all); |
| 55 | 50 | ||
| 56 | #define SH_DMAC_CHAN_BASE(id) (dma_base_addr[id]) | ||
| 57 | static void sh_dmae_writel(struct sh_dmae_chan *sh_dc, u32 data, u32 reg) | 51 | static void sh_dmae_writel(struct sh_dmae_chan *sh_dc, u32 data, u32 reg) |
| 58 | { | 52 | { |
| 59 | ctrl_outl(data, SH_DMAC_CHAN_BASE(sh_dc->id) + reg); | 53 | __raw_writel(data, sh_dc->base + reg / sizeof(u32)); |
| 60 | } | 54 | } |
| 61 | 55 | ||
| 62 | static u32 sh_dmae_readl(struct sh_dmae_chan *sh_dc, u32 reg) | 56 | static u32 sh_dmae_readl(struct sh_dmae_chan *sh_dc, u32 reg) |
| 63 | { | 57 | { |
| 64 | return ctrl_inl(SH_DMAC_CHAN_BASE(sh_dc->id) + reg); | 58 | return __raw_readl(sh_dc->base + reg / sizeof(u32)); |
| 59 | } | ||
| 60 | |||
| 61 | static u16 dmaor_read(struct sh_dmae_device *shdev) | ||
| 62 | { | ||
| 63 | return __raw_readw(shdev->chan_reg + DMAOR / sizeof(u32)); | ||
| 64 | } | ||
| 65 | |||
| 66 | static void dmaor_write(struct sh_dmae_device *shdev, u16 data) | ||
| 67 | { | ||
| 68 | __raw_writew(data, shdev->chan_reg + DMAOR / sizeof(u32)); | ||
| 65 | } | 69 | } |
| 66 | 70 | ||
| 67 | /* | 71 | /* |
| @@ -69,24 +73,23 @@ static u32 sh_dmae_readl(struct sh_dmae_chan *sh_dc, u32 reg) | |||
| 69 | * | 73 | * |
| 70 | * SH7780 has two DMAOR register | 74 | * SH7780 has two DMAOR register |
| 71 | */ | 75 | */ |
| 72 | static void sh_dmae_ctl_stop(int id) | 76 | static void sh_dmae_ctl_stop(struct sh_dmae_device *shdev) |
| 73 | { | 77 | { |
| 74 | unsigned short dmaor = dmaor_read_reg(id); | 78 | unsigned short dmaor = dmaor_read(shdev); |
| 75 | 79 | ||
| 76 | dmaor &= ~(DMAOR_NMIF | DMAOR_AE); | 80 | dmaor_write(shdev, dmaor & ~(DMAOR_NMIF | DMAOR_AE | DMAOR_DME)); |
| 77 | dmaor_write_reg(id, dmaor); | ||
| 78 | } | 81 | } |
| 79 | 82 | ||
| 80 | static int sh_dmae_rst(int id) | 83 | static int sh_dmae_rst(struct sh_dmae_device *shdev) |
| 81 | { | 84 | { |
| 82 | unsigned short dmaor; | 85 | unsigned short dmaor; |
| 83 | 86 | ||
| 84 | sh_dmae_ctl_stop(id); | 87 | sh_dmae_ctl_stop(shdev); |
| 85 | dmaor = dmaor_read_reg(id) | DMAOR_INIT; | 88 | dmaor = dmaor_read(shdev) | shdev->pdata->dmaor_init; |
| 86 | 89 | ||
| 87 | dmaor_write_reg(id, dmaor); | 90 | dmaor_write(shdev, dmaor); |
| 88 | if (dmaor_read_reg(id) & (DMAOR_AE | DMAOR_NMIF)) { | 91 | if (dmaor_read(shdev) & (DMAOR_AE | DMAOR_NMIF)) { |
| 89 | pr_warning(KERN_ERR "dma-sh: Can't initialize DMAOR.\n"); | 92 | pr_warning("dma-sh: Can't initialize DMAOR.\n"); |
| 90 | return -EINVAL; | 93 | return -EINVAL; |
| 91 | } | 94 | } |
| 92 | return 0; | 95 | return 0; |
| @@ -102,13 +105,36 @@ static bool dmae_is_busy(struct sh_dmae_chan *sh_chan) | |||
| 102 | return false; /* waiting */ | 105 | return false; /* waiting */ |
| 103 | } | 106 | } |
| 104 | 107 | ||
| 105 | static unsigned int ts_shift[] = TS_SHIFT; | 108 | static unsigned int calc_xmit_shift(struct sh_dmae_chan *sh_chan, u32 chcr) |
| 106 | static inline unsigned int calc_xmit_shift(u32 chcr) | ||
| 107 | { | 109 | { |
| 108 | int cnt = ((chcr & CHCR_TS_LOW_MASK) >> CHCR_TS_LOW_SHIFT) | | 110 | struct sh_dmae_device *shdev = container_of(sh_chan->common.device, |
| 109 | ((chcr & CHCR_TS_HIGH_MASK) >> CHCR_TS_HIGH_SHIFT); | 111 | struct sh_dmae_device, common); |
| 112 | struct sh_dmae_pdata *pdata = shdev->pdata; | ||
| 113 | int cnt = ((chcr & pdata->ts_low_mask) >> pdata->ts_low_shift) | | ||
| 114 | ((chcr & pdata->ts_high_mask) >> pdata->ts_high_shift); | ||
| 115 | |||
| 116 | if (cnt >= pdata->ts_shift_num) | ||
| 117 | cnt = 0; | ||
| 110 | 118 | ||
| 111 | return ts_shift[cnt]; | 119 | return pdata->ts_shift[cnt]; |
| 120 | } | ||
| 121 | |||
| 122 | static u32 log2size_to_chcr(struct sh_dmae_chan *sh_chan, int l2size) | ||
| 123 | { | ||
| 124 | struct sh_dmae_device *shdev = container_of(sh_chan->common.device, | ||
| 125 | struct sh_dmae_device, common); | ||
| 126 | struct sh_dmae_pdata *pdata = shdev->pdata; | ||
| 127 | int i; | ||
| 128 | |||
| 129 | for (i = 0; i < pdata->ts_shift_num; i++) | ||
| 130 | if (pdata->ts_shift[i] == l2size) | ||
| 131 | break; | ||
| 132 | |||
| 133 | if (i == pdata->ts_shift_num) | ||
| 134 | i = 0; | ||
| 135 | |||
| 136 | return ((i << pdata->ts_low_shift) & pdata->ts_low_mask) | | ||
| 137 | ((i << pdata->ts_high_shift) & pdata->ts_high_mask); | ||
| 112 | } | 138 | } |
| 113 | 139 | ||
| 114 | static void dmae_set_reg(struct sh_dmae_chan *sh_chan, struct sh_dmae_regs *hw) | 140 | static void dmae_set_reg(struct sh_dmae_chan *sh_chan, struct sh_dmae_regs *hw) |
| @@ -136,8 +162,13 @@ static void dmae_halt(struct sh_dmae_chan *sh_chan) | |||
| 136 | 162 | ||
| 137 | static void dmae_init(struct sh_dmae_chan *sh_chan) | 163 | static void dmae_init(struct sh_dmae_chan *sh_chan) |
| 138 | { | 164 | { |
| 139 | u32 chcr = RS_DEFAULT; /* default is DUAL mode */ | 165 | /* |
| 140 | sh_chan->xmit_shift = calc_xmit_shift(chcr); | 166 | * Default configuration for dual address memory-memory transfer. |
| 167 | * 0x400 represents auto-request. | ||
| 168 | */ | ||
| 169 | u32 chcr = DM_INC | SM_INC | 0x400 | log2size_to_chcr(sh_chan, | ||
| 170 | LOG2_DEFAULT_XFER_SIZE); | ||
| 171 | sh_chan->xmit_shift = calc_xmit_shift(sh_chan, chcr); | ||
| 141 | sh_dmae_writel(sh_chan, chcr, CHCR); | 172 | sh_dmae_writel(sh_chan, chcr, CHCR); |
| 142 | } | 173 | } |
| 143 | 174 | ||
| @@ -147,37 +178,26 @@ static int dmae_set_chcr(struct sh_dmae_chan *sh_chan, u32 val) | |||
| 147 | if (dmae_is_busy(sh_chan)) | 178 | if (dmae_is_busy(sh_chan)) |
| 148 | return -EBUSY; | 179 | return -EBUSY; |
| 149 | 180 | ||
| 150 | sh_chan->xmit_shift = calc_xmit_shift(val); | 181 | sh_chan->xmit_shift = calc_xmit_shift(sh_chan, val); |
| 151 | sh_dmae_writel(sh_chan, val, CHCR); | 182 | sh_dmae_writel(sh_chan, val, CHCR); |
| 152 | 183 | ||
| 153 | return 0; | 184 | return 0; |
| 154 | } | 185 | } |
| 155 | 186 | ||
| 156 | #define DMARS_SHIFT 8 | ||
| 157 | #define DMARS_CHAN_MSK 0x01 | ||
| 158 | static int dmae_set_dmars(struct sh_dmae_chan *sh_chan, u16 val) | 187 | static int dmae_set_dmars(struct sh_dmae_chan *sh_chan, u16 val) |
| 159 | { | 188 | { |
| 160 | u32 addr; | 189 | struct sh_dmae_device *shdev = container_of(sh_chan->common.device, |
| 161 | int shift = 0; | 190 | struct sh_dmae_device, common); |
| 191 | struct sh_dmae_pdata *pdata = shdev->pdata; | ||
| 192 | struct sh_dmae_channel *chan_pdata = &pdata->channel[sh_chan->id]; | ||
| 193 | u16 __iomem *addr = shdev->dmars + chan_pdata->dmars / sizeof(u16); | ||
| 194 | int shift = chan_pdata->dmars_bit; | ||
| 162 | 195 | ||
| 163 | if (dmae_is_busy(sh_chan)) | 196 | if (dmae_is_busy(sh_chan)) |
| 164 | return -EBUSY; | 197 | return -EBUSY; |
| 165 | 198 | ||
| 166 | if (sh_chan->id & DMARS_CHAN_MSK) | 199 | __raw_writew((__raw_readw(addr) & (0xff00 >> shift)) | (val << shift), |
| 167 | shift = DMARS_SHIFT; | 200 | addr); |
| 168 | |||
| 169 | if (sh_chan->id < 6) | ||
| 170 | /* DMA0RS0 - DMA0RS2 */ | ||
| 171 | addr = SH_DMARS_BASE0 + (sh_chan->id / 2) * 4; | ||
| 172 | #ifdef SH_DMARS_BASE1 | ||
| 173 | else if (sh_chan->id < 12) | ||
| 174 | /* DMA1RS0 - DMA1RS2 */ | ||
| 175 | addr = SH_DMARS_BASE1 + ((sh_chan->id - 6) / 2) * 4; | ||
| 176 | #endif | ||
| 177 | else | ||
| 178 | return -EINVAL; | ||
| 179 | |||
| 180 | ctrl_outw((val << shift) | (ctrl_inw(addr) & (0xFF00 >> shift)), addr); | ||
| 181 | 201 | ||
| 182 | return 0; | 202 | return 0; |
| 183 | } | 203 | } |
| @@ -251,15 +271,15 @@ static struct sh_dmae_slave_config *sh_dmae_find_slave( | |||
| 251 | struct dma_device *dma_dev = sh_chan->common.device; | 271 | struct dma_device *dma_dev = sh_chan->common.device; |
| 252 | struct sh_dmae_device *shdev = container_of(dma_dev, | 272 | struct sh_dmae_device *shdev = container_of(dma_dev, |
| 253 | struct sh_dmae_device, common); | 273 | struct sh_dmae_device, common); |
| 254 | struct sh_dmae_pdata *pdata = &shdev->pdata; | 274 | struct sh_dmae_pdata *pdata = shdev->pdata; |
| 255 | int i; | 275 | int i; |
| 256 | 276 | ||
| 257 | if ((unsigned)slave_id >= SHDMA_SLAVE_NUMBER) | 277 | if ((unsigned)slave_id >= SHDMA_SLAVE_NUMBER) |
| 258 | return NULL; | 278 | return NULL; |
| 259 | 279 | ||
| 260 | for (i = 0; i < pdata->config_num; i++) | 280 | for (i = 0; i < pdata->slave_num; i++) |
| 261 | if (pdata->config[i].slave_id == slave_id) | 281 | if (pdata->slave[i].slave_id == slave_id) |
| 262 | return pdata->config + i; | 282 | return pdata->slave + i; |
| 263 | 283 | ||
| 264 | return NULL; | 284 | return NULL; |
| 265 | } | 285 | } |
| @@ -270,6 +290,8 @@ static int sh_dmae_alloc_chan_resources(struct dma_chan *chan) | |||
| 270 | struct sh_desc *desc; | 290 | struct sh_desc *desc; |
| 271 | struct sh_dmae_slave *param = chan->private; | 291 | struct sh_dmae_slave *param = chan->private; |
| 272 | 292 | ||
| 293 | pm_runtime_get_sync(sh_chan->dev); | ||
| 294 | |||
| 273 | /* | 295 | /* |
| 274 | * This relies on the guarantee from dmaengine that alloc_chan_resources | 296 | * This relies on the guarantee from dmaengine that alloc_chan_resources |
| 275 | * never runs concurrently with itself or free_chan_resources. | 297 | * never runs concurrently with itself or free_chan_resources. |
| @@ -288,9 +310,8 @@ static int sh_dmae_alloc_chan_resources(struct dma_chan *chan) | |||
| 288 | 310 | ||
| 289 | dmae_set_dmars(sh_chan, cfg->mid_rid); | 311 | dmae_set_dmars(sh_chan, cfg->mid_rid); |
| 290 | dmae_set_chcr(sh_chan, cfg->chcr); | 312 | dmae_set_chcr(sh_chan, cfg->chcr); |
| 291 | } else { | 313 | } else if ((sh_dmae_readl(sh_chan, CHCR) & 0xf00) != 0x400) { |
| 292 | if ((sh_dmae_readl(sh_chan, CHCR) & 0x700) != 0x400) | 314 | dmae_init(sh_chan); |
| 293 | dmae_set_chcr(sh_chan, RS_DEFAULT); | ||
| 294 | } | 315 | } |
| 295 | 316 | ||
| 296 | spin_lock_bh(&sh_chan->desc_lock); | 317 | spin_lock_bh(&sh_chan->desc_lock); |
| @@ -312,6 +333,9 @@ static int sh_dmae_alloc_chan_resources(struct dma_chan *chan) | |||
| 312 | } | 333 | } |
| 313 | spin_unlock_bh(&sh_chan->desc_lock); | 334 | spin_unlock_bh(&sh_chan->desc_lock); |
| 314 | 335 | ||
| 336 | if (!sh_chan->descs_allocated) | ||
| 337 | pm_runtime_put(sh_chan->dev); | ||
| 338 | |||
| 315 | return sh_chan->descs_allocated; | 339 | return sh_chan->descs_allocated; |
| 316 | } | 340 | } |
| 317 | 341 | ||
| @@ -323,6 +347,7 @@ static void sh_dmae_free_chan_resources(struct dma_chan *chan) | |||
| 323 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); | 347 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); |
| 324 | struct sh_desc *desc, *_desc; | 348 | struct sh_desc *desc, *_desc; |
| 325 | LIST_HEAD(list); | 349 | LIST_HEAD(list); |
| 350 | int descs = sh_chan->descs_allocated; | ||
| 326 | 351 | ||
| 327 | dmae_halt(sh_chan); | 352 | dmae_halt(sh_chan); |
| 328 | 353 | ||
| @@ -343,6 +368,9 @@ static void sh_dmae_free_chan_resources(struct dma_chan *chan) | |||
| 343 | 368 | ||
| 344 | spin_unlock_bh(&sh_chan->desc_lock); | 369 | spin_unlock_bh(&sh_chan->desc_lock); |
| 345 | 370 | ||
| 371 | if (descs > 0) | ||
| 372 | pm_runtime_put(sh_chan->dev); | ||
| 373 | |||
| 346 | list_for_each_entry_safe(desc, _desc, &list, node) | 374 | list_for_each_entry_safe(desc, _desc, &list, node) |
| 347 | kfree(desc); | 375 | kfree(desc); |
| 348 | } | 376 | } |
| @@ -559,6 +587,19 @@ static void sh_dmae_terminate_all(struct dma_chan *chan) | |||
| 559 | if (!chan) | 587 | if (!chan) |
| 560 | return; | 588 | return; |
| 561 | 589 | ||
| 590 | dmae_halt(sh_chan); | ||
| 591 | |||
| 592 | spin_lock_bh(&sh_chan->desc_lock); | ||
| 593 | if (!list_empty(&sh_chan->ld_queue)) { | ||
| 594 | /* Record partial transfer */ | ||
| 595 | struct sh_desc *desc = list_entry(sh_chan->ld_queue.next, | ||
| 596 | struct sh_desc, node); | ||
| 597 | desc->partial = (desc->hw.tcr - sh_dmae_readl(sh_chan, TCR)) << | ||
| 598 | sh_chan->xmit_shift; | ||
| 599 | |||
| 600 | } | ||
| 601 | spin_unlock_bh(&sh_chan->desc_lock); | ||
| 602 | |||
| 562 | sh_dmae_chan_ld_cleanup(sh_chan, true); | 603 | sh_dmae_chan_ld_cleanup(sh_chan, true); |
| 563 | } | 604 | } |
| 564 | 605 | ||
| @@ -661,7 +702,7 @@ static void sh_dmae_chan_ld_cleanup(struct sh_dmae_chan *sh_chan, bool all) | |||
| 661 | 702 | ||
| 662 | static void sh_chan_xfer_ld_queue(struct sh_dmae_chan *sh_chan) | 703 | static void sh_chan_xfer_ld_queue(struct sh_dmae_chan *sh_chan) |
| 663 | { | 704 | { |
| 664 | struct sh_desc *sd; | 705 | struct sh_desc *desc; |
| 665 | 706 | ||
| 666 | spin_lock_bh(&sh_chan->desc_lock); | 707 | spin_lock_bh(&sh_chan->desc_lock); |
| 667 | /* DMA work check */ | 708 | /* DMA work check */ |
| @@ -671,10 +712,13 @@ static void sh_chan_xfer_ld_queue(struct sh_dmae_chan *sh_chan) | |||
| 671 | } | 712 | } |
| 672 | 713 | ||
| 673 | /* Find the first not transferred desciptor */ | 714 | /* Find the first not transferred desciptor */ |
| 674 | list_for_each_entry(sd, &sh_chan->ld_queue, node) | 715 | list_for_each_entry(desc, &sh_chan->ld_queue, node) |
| 675 | if (sd->mark == DESC_SUBMITTED) { | 716 | if (desc->mark == DESC_SUBMITTED) { |
| 717 | dev_dbg(sh_chan->dev, "Queue #%d to %d: %u@%x -> %x\n", | ||
| 718 | desc->async_tx.cookie, sh_chan->id, | ||
| 719 | desc->hw.tcr, desc->hw.sar, desc->hw.dar); | ||
| 676 | /* Get the ld start address from ld_queue */ | 720 | /* Get the ld start address from ld_queue */ |
| 677 | dmae_set_reg(sh_chan, &sd->hw); | 721 | dmae_set_reg(sh_chan, &desc->hw); |
| 678 | dmae_start(sh_chan); | 722 | dmae_start(sh_chan); |
| 679 | break; | 723 | break; |
| 680 | } | 724 | } |
| @@ -696,6 +740,7 @@ static enum dma_status sh_dmae_is_complete(struct dma_chan *chan, | |||
| 696 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); | 740 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); |
| 697 | dma_cookie_t last_used; | 741 | dma_cookie_t last_used; |
| 698 | dma_cookie_t last_complete; | 742 | dma_cookie_t last_complete; |
| 743 | enum dma_status status; | ||
| 699 | 744 | ||
| 700 | sh_dmae_chan_ld_cleanup(sh_chan, false); | 745 | sh_dmae_chan_ld_cleanup(sh_chan, false); |
| 701 | 746 | ||
| @@ -709,7 +754,27 @@ static enum dma_status sh_dmae_is_complete(struct dma_chan *chan, | |||
| 709 | if (used) | 754 | if (used) |
| 710 | *used = last_used; | 755 | *used = last_used; |
| 711 | 756 | ||
| 712 | return dma_async_is_complete(cookie, last_complete, last_used); | 757 | spin_lock_bh(&sh_chan->desc_lock); |
| 758 | |||
| 759 | status = dma_async_is_complete(cookie, last_complete, last_used); | ||
| 760 | |||
| 761 | /* | ||
| 762 | * If we don't find cookie on the queue, it has been aborted and we have | ||
| 763 | * to report error | ||
| 764 | */ | ||
| 765 | if (status != DMA_SUCCESS) { | ||
| 766 | struct sh_desc *desc; | ||
| 767 | status = DMA_ERROR; | ||
| 768 | list_for_each_entry(desc, &sh_chan->ld_queue, node) | ||
| 769 | if (desc->cookie == cookie) { | ||
| 770 | status = DMA_IN_PROGRESS; | ||
| 771 | break; | ||
| 772 | } | ||
| 773 | } | ||
| 774 | |||
| 775 | spin_unlock_bh(&sh_chan->desc_lock); | ||
| 776 | |||
| 777 | return status; | ||
| 713 | } | 778 | } |
| 714 | 779 | ||
| 715 | static irqreturn_t sh_dmae_interrupt(int irq, void *data) | 780 | static irqreturn_t sh_dmae_interrupt(int irq, void *data) |
| @@ -732,40 +797,32 @@ static irqreturn_t sh_dmae_interrupt(int irq, void *data) | |||
| 732 | #if defined(CONFIG_CPU_SH4) | 797 | #if defined(CONFIG_CPU_SH4) |
| 733 | static irqreturn_t sh_dmae_err(int irq, void *data) | 798 | static irqreturn_t sh_dmae_err(int irq, void *data) |
| 734 | { | 799 | { |
| 735 | int err = 0; | ||
| 736 | struct sh_dmae_device *shdev = (struct sh_dmae_device *)data; | 800 | struct sh_dmae_device *shdev = (struct sh_dmae_device *)data; |
| 801 | int i; | ||
| 737 | 802 | ||
| 738 | /* IRQ Multi */ | 803 | /* halt the dma controller */ |
| 739 | if (shdev->pdata.mode & SHDMA_MIX_IRQ) { | 804 | sh_dmae_ctl_stop(shdev); |
| 740 | int __maybe_unused cnt = 0; | 805 | |
| 741 | switch (irq) { | 806 | /* We cannot detect, which channel caused the error, have to reset all */ |
| 742 | #if defined(DMTE6_IRQ) && defined(DMAE1_IRQ) | 807 | for (i = 0; i < SH_DMAC_MAX_CHANNELS; i++) { |
| 743 | case DMTE6_IRQ: | 808 | struct sh_dmae_chan *sh_chan = shdev->chan[i]; |
| 744 | cnt++; | 809 | if (sh_chan) { |
| 745 | #endif | 810 | struct sh_desc *desc; |
| 746 | case DMTE0_IRQ: | 811 | /* Stop the channel */ |
| 747 | if (dmaor_read_reg(cnt) & (DMAOR_NMIF | DMAOR_AE)) { | 812 | dmae_halt(sh_chan); |
| 748 | disable_irq(irq); | 813 | /* Complete all */ |
| 749 | return IRQ_HANDLED; | 814 | list_for_each_entry(desc, &sh_chan->ld_queue, node) { |
| 815 | struct dma_async_tx_descriptor *tx = &desc->async_tx; | ||
| 816 | desc->mark = DESC_IDLE; | ||
| 817 | if (tx->callback) | ||
| 818 | tx->callback(tx->callback_param); | ||
| 750 | } | 819 | } |
| 751 | default: | 820 | list_splice_init(&sh_chan->ld_queue, &sh_chan->ld_free); |
| 752 | return IRQ_NONE; | ||
| 753 | } | 821 | } |
| 754 | } else { | ||
| 755 | /* reset dma controller */ | ||
| 756 | err = sh_dmae_rst(0); | ||
| 757 | if (err) | ||
| 758 | return err; | ||
| 759 | #ifdef SH_DMAC_BASE1 | ||
| 760 | if (shdev->pdata.mode & SHDMA_DMAOR1) { | ||
| 761 | err = sh_dmae_rst(1); | ||
| 762 | if (err) | ||
| 763 | return err; | ||
| 764 | } | ||
| 765 | #endif | ||
| 766 | disable_irq(irq); | ||
| 767 | return IRQ_HANDLED; | ||
| 768 | } | 822 | } |
| 823 | sh_dmae_rst(shdev); | ||
| 824 | |||
| 825 | return IRQ_HANDLED; | ||
| 769 | } | 826 | } |
| 770 | #endif | 827 | #endif |
| 771 | 828 | ||
| @@ -796,19 +853,12 @@ static void dmae_do_tasklet(unsigned long data) | |||
| 796 | sh_dmae_chan_ld_cleanup(sh_chan, false); | 853 | sh_dmae_chan_ld_cleanup(sh_chan, false); |
| 797 | } | 854 | } |
| 798 | 855 | ||
| 799 | static unsigned int get_dmae_irq(unsigned int id) | 856 | static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id, |
| 800 | { | 857 | int irq, unsigned long flags) |
| 801 | unsigned int irq = 0; | ||
| 802 | if (id < ARRAY_SIZE(dmte_irq_map)) | ||
| 803 | irq = dmte_irq_map[id]; | ||
| 804 | return irq; | ||
| 805 | } | ||
| 806 | |||
| 807 | static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id) | ||
| 808 | { | 858 | { |
| 809 | int err; | 859 | int err; |
| 810 | unsigned int irq = get_dmae_irq(id); | 860 | struct sh_dmae_channel *chan_pdata = &shdev->pdata->channel[id]; |
| 811 | unsigned long irqflags = IRQF_DISABLED; | 861 | struct platform_device *pdev = to_platform_device(shdev->common.dev); |
| 812 | struct sh_dmae_chan *new_sh_chan; | 862 | struct sh_dmae_chan *new_sh_chan; |
| 813 | 863 | ||
| 814 | /* alloc channel */ | 864 | /* alloc channel */ |
| @@ -819,8 +869,13 @@ static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id) | |||
| 819 | return -ENOMEM; | 869 | return -ENOMEM; |
| 820 | } | 870 | } |
| 821 | 871 | ||
| 872 | /* copy struct dma_device */ | ||
| 873 | new_sh_chan->common.device = &shdev->common; | ||
| 874 | |||
| 822 | new_sh_chan->dev = shdev->common.dev; | 875 | new_sh_chan->dev = shdev->common.dev; |
| 823 | new_sh_chan->id = id; | 876 | new_sh_chan->id = id; |
| 877 | new_sh_chan->irq = irq; | ||
| 878 | new_sh_chan->base = shdev->chan_reg + chan_pdata->offset / sizeof(u32); | ||
| 824 | 879 | ||
| 825 | /* Init DMA tasklet */ | 880 | /* Init DMA tasklet */ |
| 826 | tasklet_init(&new_sh_chan->tasklet, dmae_do_tasklet, | 881 | tasklet_init(&new_sh_chan->tasklet, dmae_do_tasklet, |
| @@ -835,29 +890,20 @@ static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id) | |||
| 835 | INIT_LIST_HEAD(&new_sh_chan->ld_queue); | 890 | INIT_LIST_HEAD(&new_sh_chan->ld_queue); |
| 836 | INIT_LIST_HEAD(&new_sh_chan->ld_free); | 891 | INIT_LIST_HEAD(&new_sh_chan->ld_free); |
| 837 | 892 | ||
| 838 | /* copy struct dma_device */ | ||
| 839 | new_sh_chan->common.device = &shdev->common; | ||
| 840 | |||
| 841 | /* Add the channel to DMA device channel list */ | 893 | /* Add the channel to DMA device channel list */ |
| 842 | list_add_tail(&new_sh_chan->common.device_node, | 894 | list_add_tail(&new_sh_chan->common.device_node, |
| 843 | &shdev->common.channels); | 895 | &shdev->common.channels); |
| 844 | shdev->common.chancnt++; | 896 | shdev->common.chancnt++; |
| 845 | 897 | ||
| 846 | if (shdev->pdata.mode & SHDMA_MIX_IRQ) { | 898 | if (pdev->id >= 0) |
| 847 | irqflags = IRQF_SHARED; | 899 | snprintf(new_sh_chan->dev_id, sizeof(new_sh_chan->dev_id), |
| 848 | #if defined(DMTE6_IRQ) | 900 | "sh-dmae%d.%d", pdev->id, new_sh_chan->id); |
| 849 | if (irq >= DMTE6_IRQ) | 901 | else |
| 850 | irq = DMTE6_IRQ; | 902 | snprintf(new_sh_chan->dev_id, sizeof(new_sh_chan->dev_id), |
| 851 | else | 903 | "sh-dma%d", new_sh_chan->id); |
| 852 | #endif | ||
| 853 | irq = DMTE0_IRQ; | ||
| 854 | } | ||
| 855 | |||
| 856 | snprintf(new_sh_chan->dev_id, sizeof(new_sh_chan->dev_id), | ||
| 857 | "sh-dmae%d", new_sh_chan->id); | ||
| 858 | 904 | ||
| 859 | /* set up channel irq */ | 905 | /* set up channel irq */ |
| 860 | err = request_irq(irq, &sh_dmae_interrupt, irqflags, | 906 | err = request_irq(irq, &sh_dmae_interrupt, flags, |
| 861 | new_sh_chan->dev_id, new_sh_chan); | 907 | new_sh_chan->dev_id, new_sh_chan); |
| 862 | if (err) { | 908 | if (err) { |
| 863 | dev_err(shdev->common.dev, "DMA channel %d request_irq error " | 909 | dev_err(shdev->common.dev, "DMA channel %d request_irq error " |
| @@ -881,12 +927,12 @@ static void sh_dmae_chan_remove(struct sh_dmae_device *shdev) | |||
| 881 | 927 | ||
| 882 | for (i = shdev->common.chancnt - 1 ; i >= 0 ; i--) { | 928 | for (i = shdev->common.chancnt - 1 ; i >= 0 ; i--) { |
| 883 | if (shdev->chan[i]) { | 929 | if (shdev->chan[i]) { |
| 884 | struct sh_dmae_chan *shchan = shdev->chan[i]; | 930 | struct sh_dmae_chan *sh_chan = shdev->chan[i]; |
| 885 | if (!(shdev->pdata.mode & SHDMA_MIX_IRQ)) | ||
| 886 | free_irq(dmte_irq_map[i], shchan); | ||
| 887 | 931 | ||
| 888 | list_del(&shchan->common.device_node); | 932 | free_irq(sh_chan->irq, sh_chan); |
| 889 | kfree(shchan); | 933 | |
| 934 | list_del(&sh_chan->common.device_node); | ||
| 935 | kfree(sh_chan); | ||
| 890 | shdev->chan[i] = NULL; | 936 | shdev->chan[i] = NULL; |
| 891 | } | 937 | } |
| 892 | } | 938 | } |
| @@ -895,47 +941,84 @@ static void sh_dmae_chan_remove(struct sh_dmae_device *shdev) | |||
| 895 | 941 | ||
| 896 | static int __init sh_dmae_probe(struct platform_device *pdev) | 942 | static int __init sh_dmae_probe(struct platform_device *pdev) |
| 897 | { | 943 | { |
| 898 | int err = 0, cnt, ecnt; | 944 | struct sh_dmae_pdata *pdata = pdev->dev.platform_data; |
| 899 | unsigned long irqflags = IRQF_DISABLED; | 945 | unsigned long irqflags = IRQF_DISABLED, |
| 900 | #if defined(CONFIG_CPU_SH4) | 946 | chan_flag[SH_DMAC_MAX_CHANNELS] = {}; |
| 901 | int eirq[] = { DMAE0_IRQ, | 947 | int errirq, chan_irq[SH_DMAC_MAX_CHANNELS]; |
| 902 | #if defined(DMAE1_IRQ) | 948 | int err, i, irq_cnt = 0, irqres = 0; |
| 903 | DMAE1_IRQ | ||
| 904 | #endif | ||
| 905 | }; | ||
| 906 | #endif | ||
| 907 | struct sh_dmae_device *shdev; | 949 | struct sh_dmae_device *shdev; |
| 950 | struct resource *chan, *dmars, *errirq_res, *chanirq_res; | ||
| 908 | 951 | ||
| 909 | /* get platform data */ | 952 | /* get platform data */ |
| 910 | if (!pdev->dev.platform_data) | 953 | if (!pdata || !pdata->channel_num) |
| 911 | return -ENODEV; | 954 | return -ENODEV; |
| 912 | 955 | ||
| 956 | chan = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
| 957 | /* DMARS area is optional, if absent, this controller cannot do slave DMA */ | ||
| 958 | dmars = platform_get_resource(pdev, IORESOURCE_MEM, 1); | ||
| 959 | /* | ||
| 960 | * IRQ resources: | ||
| 961 | * 1. there always must be at least one IRQ IO-resource. On SH4 it is | ||
| 962 | * the error IRQ, in which case it is the only IRQ in this resource: | ||
| 963 | * start == end. If it is the only IRQ resource, all channels also | ||
| 964 | * use the same IRQ. | ||
| 965 | * 2. DMA channel IRQ resources can be specified one per resource or in | ||
| 966 | * ranges (start != end) | ||
| 967 | * 3. iff all events (channels and, optionally, error) on this | ||
| 968 | * controller use the same IRQ, only one IRQ resource can be | ||
| 969 | * specified, otherwise there must be one IRQ per channel, even if | ||
| 970 | * some of them are equal | ||
| 971 | * 4. if all IRQs on this controller are equal or if some specific IRQs | ||
| 972 | * specify IORESOURCE_IRQ_SHAREABLE in their resources, they will be | ||
| 973 | * requested with the IRQF_SHARED flag | ||
| 974 | */ | ||
| 975 | errirq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 0); | ||
| 976 | if (!chan || !errirq_res) | ||
| 977 | return -ENODEV; | ||
| 978 | |||
| 979 | if (!request_mem_region(chan->start, resource_size(chan), pdev->name)) { | ||
| 980 | dev_err(&pdev->dev, "DMAC register region already claimed\n"); | ||
| 981 | return -EBUSY; | ||
| 982 | } | ||
| 983 | |||
| 984 | if (dmars && !request_mem_region(dmars->start, resource_size(dmars), pdev->name)) { | ||
| 985 | dev_err(&pdev->dev, "DMAC DMARS region already claimed\n"); | ||
| 986 | err = -EBUSY; | ||
| 987 | goto ermrdmars; | ||
| 988 | } | ||
| 989 | |||
| 990 | err = -ENOMEM; | ||
| 913 | shdev = kzalloc(sizeof(struct sh_dmae_device), GFP_KERNEL); | 991 | shdev = kzalloc(sizeof(struct sh_dmae_device), GFP_KERNEL); |
| 914 | if (!shdev) { | 992 | if (!shdev) { |
| 915 | dev_err(&pdev->dev, "No enough memory\n"); | 993 | dev_err(&pdev->dev, "Not enough memory\n"); |
| 916 | return -ENOMEM; | 994 | goto ealloc; |
| 995 | } | ||
| 996 | |||
| 997 | shdev->chan_reg = ioremap(chan->start, resource_size(chan)); | ||
| 998 | if (!shdev->chan_reg) | ||
| 999 | goto emapchan; | ||
| 1000 | if (dmars) { | ||
| 1001 | shdev->dmars = ioremap(dmars->start, resource_size(dmars)); | ||
| 1002 | if (!shdev->dmars) | ||
| 1003 | goto emapdmars; | ||
| 917 | } | 1004 | } |
| 918 | 1005 | ||
| 919 | /* platform data */ | 1006 | /* platform data */ |
| 920 | memcpy(&shdev->pdata, pdev->dev.platform_data, | 1007 | shdev->pdata = pdata; |
| 921 | sizeof(struct sh_dmae_pdata)); | 1008 | |
| 1009 | pm_runtime_enable(&pdev->dev); | ||
| 1010 | pm_runtime_get_sync(&pdev->dev); | ||
| 922 | 1011 | ||
| 923 | /* reset dma controller */ | 1012 | /* reset dma controller */ |
| 924 | err = sh_dmae_rst(0); | 1013 | err = sh_dmae_rst(shdev); |
| 925 | if (err) | 1014 | if (err) |
| 926 | goto rst_err; | 1015 | goto rst_err; |
| 927 | 1016 | ||
| 928 | /* SH7780/85/23 has DMAOR1 */ | ||
| 929 | if (shdev->pdata.mode & SHDMA_DMAOR1) { | ||
| 930 | err = sh_dmae_rst(1); | ||
| 931 | if (err) | ||
| 932 | goto rst_err; | ||
| 933 | } | ||
| 934 | |||
| 935 | INIT_LIST_HEAD(&shdev->common.channels); | 1017 | INIT_LIST_HEAD(&shdev->common.channels); |
| 936 | 1018 | ||
| 937 | dma_cap_set(DMA_MEMCPY, shdev->common.cap_mask); | 1019 | dma_cap_set(DMA_MEMCPY, shdev->common.cap_mask); |
| 938 | dma_cap_set(DMA_SLAVE, shdev->common.cap_mask); | 1020 | if (dmars) |
| 1021 | dma_cap_set(DMA_SLAVE, shdev->common.cap_mask); | ||
| 939 | 1022 | ||
| 940 | shdev->common.device_alloc_chan_resources | 1023 | shdev->common.device_alloc_chan_resources |
| 941 | = sh_dmae_alloc_chan_resources; | 1024 | = sh_dmae_alloc_chan_resources; |
| @@ -950,37 +1033,72 @@ static int __init sh_dmae_probe(struct platform_device *pdev) | |||
| 950 | 1033 | ||
| 951 | shdev->common.dev = &pdev->dev; | 1034 | shdev->common.dev = &pdev->dev; |
| 952 | /* Default transfer size of 32 bytes requires 32-byte alignment */ | 1035 | /* Default transfer size of 32 bytes requires 32-byte alignment */ |
| 953 | shdev->common.copy_align = 5; | 1036 | shdev->common.copy_align = LOG2_DEFAULT_XFER_SIZE; |
| 954 | 1037 | ||
| 955 | #if defined(CONFIG_CPU_SH4) | 1038 | #if defined(CONFIG_CPU_SH4) |
| 956 | /* Non Mix IRQ mode SH7722/SH7730 etc... */ | 1039 | chanirq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 1); |
| 957 | if (shdev->pdata.mode & SHDMA_MIX_IRQ) { | 1040 | |
| 1041 | if (!chanirq_res) | ||
| 1042 | chanirq_res = errirq_res; | ||
| 1043 | else | ||
| 1044 | irqres++; | ||
| 1045 | |||
| 1046 | if (chanirq_res == errirq_res || | ||
| 1047 | (errirq_res->flags & IORESOURCE_BITS) == IORESOURCE_IRQ_SHAREABLE) | ||
| 958 | irqflags = IRQF_SHARED; | 1048 | irqflags = IRQF_SHARED; |
| 959 | eirq[0] = DMTE0_IRQ; | 1049 | |
| 960 | #if defined(DMTE6_IRQ) && defined(DMAE1_IRQ) | 1050 | errirq = errirq_res->start; |
| 961 | eirq[1] = DMTE6_IRQ; | 1051 | |
| 962 | #endif | 1052 | err = request_irq(errirq, sh_dmae_err, irqflags, |
| 1053 | "DMAC Address Error", shdev); | ||
| 1054 | if (err) { | ||
| 1055 | dev_err(&pdev->dev, | ||
| 1056 | "DMA failed requesting irq #%d, error %d\n", | ||
| 1057 | errirq, err); | ||
| 1058 | goto eirq_err; | ||
| 963 | } | 1059 | } |
| 964 | 1060 | ||
| 965 | for (ecnt = 0 ; ecnt < ARRAY_SIZE(eirq); ecnt++) { | 1061 | #else |
| 966 | err = request_irq(eirq[ecnt], sh_dmae_err, irqflags, | 1062 | chanirq_res = errirq_res; |
| 967 | "DMAC Address Error", shdev); | 1063 | #endif /* CONFIG_CPU_SH4 */ |
| 968 | if (err) { | 1064 | |
| 969 | dev_err(&pdev->dev, "DMA device request_irq" | 1065 | if (chanirq_res->start == chanirq_res->end && |
| 970 | "error (irq %d) with return %d\n", | 1066 | !platform_get_resource(pdev, IORESOURCE_IRQ, 1)) { |
| 971 | eirq[ecnt], err); | 1067 | /* Special case - all multiplexed */ |
| 972 | goto eirq_err; | 1068 | for (; irq_cnt < pdata->channel_num; irq_cnt++) { |
| 1069 | chan_irq[irq_cnt] = chanirq_res->start; | ||
| 1070 | chan_flag[irq_cnt] = IRQF_SHARED; | ||
| 973 | } | 1071 | } |
| 1072 | } else { | ||
| 1073 | do { | ||
| 1074 | for (i = chanirq_res->start; i <= chanirq_res->end; i++) { | ||
| 1075 | if ((errirq_res->flags & IORESOURCE_BITS) == | ||
| 1076 | IORESOURCE_IRQ_SHAREABLE) | ||
| 1077 | chan_flag[irq_cnt] = IRQF_SHARED; | ||
| 1078 | else | ||
| 1079 | chan_flag[irq_cnt] = IRQF_DISABLED; | ||
| 1080 | dev_dbg(&pdev->dev, | ||
| 1081 | "Found IRQ %d for channel %d\n", | ||
| 1082 | i, irq_cnt); | ||
| 1083 | chan_irq[irq_cnt++] = i; | ||
| 1084 | } | ||
| 1085 | chanirq_res = platform_get_resource(pdev, | ||
| 1086 | IORESOURCE_IRQ, ++irqres); | ||
| 1087 | } while (irq_cnt < pdata->channel_num && chanirq_res); | ||
| 974 | } | 1088 | } |
| 975 | #endif /* CONFIG_CPU_SH4 */ | 1089 | |
| 1090 | if (irq_cnt < pdata->channel_num) | ||
| 1091 | goto eirqres; | ||
| 976 | 1092 | ||
| 977 | /* Create DMA Channel */ | 1093 | /* Create DMA Channel */ |
| 978 | for (cnt = 0 ; cnt < MAX_DMA_CHANNELS ; cnt++) { | 1094 | for (i = 0; i < pdata->channel_num; i++) { |
| 979 | err = sh_dmae_chan_probe(shdev, cnt); | 1095 | err = sh_dmae_chan_probe(shdev, i, chan_irq[i], chan_flag[i]); |
| 980 | if (err) | 1096 | if (err) |
| 981 | goto chan_probe_err; | 1097 | goto chan_probe_err; |
| 982 | } | 1098 | } |
| 983 | 1099 | ||
| 1100 | pm_runtime_put(&pdev->dev); | ||
| 1101 | |||
| 984 | platform_set_drvdata(pdev, shdev); | 1102 | platform_set_drvdata(pdev, shdev); |
| 985 | dma_async_device_register(&shdev->common); | 1103 | dma_async_device_register(&shdev->common); |
| 986 | 1104 | ||
| @@ -988,13 +1106,24 @@ static int __init sh_dmae_probe(struct platform_device *pdev) | |||
| 988 | 1106 | ||
| 989 | chan_probe_err: | 1107 | chan_probe_err: |
| 990 | sh_dmae_chan_remove(shdev); | 1108 | sh_dmae_chan_remove(shdev); |
| 991 | 1109 | eirqres: | |
| 1110 | #if defined(CONFIG_CPU_SH4) | ||
| 1111 | free_irq(errirq, shdev); | ||
| 992 | eirq_err: | 1112 | eirq_err: |
| 993 | for (ecnt-- ; ecnt >= 0; ecnt--) | 1113 | #endif |
| 994 | free_irq(eirq[ecnt], shdev); | ||
| 995 | |||
| 996 | rst_err: | 1114 | rst_err: |
| 1115 | pm_runtime_put(&pdev->dev); | ||
| 1116 | if (dmars) | ||
| 1117 | iounmap(shdev->dmars); | ||
| 1118 | emapdmars: | ||
| 1119 | iounmap(shdev->chan_reg); | ||
| 1120 | emapchan: | ||
| 997 | kfree(shdev); | 1121 | kfree(shdev); |
| 1122 | ealloc: | ||
| 1123 | if (dmars) | ||
| 1124 | release_mem_region(dmars->start, resource_size(dmars)); | ||
| 1125 | ermrdmars: | ||
| 1126 | release_mem_region(chan->start, resource_size(chan)); | ||
| 998 | 1127 | ||
| 999 | return err; | 1128 | return err; |
| 1000 | } | 1129 | } |
| @@ -1002,36 +1131,39 @@ rst_err: | |||
| 1002 | static int __exit sh_dmae_remove(struct platform_device *pdev) | 1131 | static int __exit sh_dmae_remove(struct platform_device *pdev) |
| 1003 | { | 1132 | { |
| 1004 | struct sh_dmae_device *shdev = platform_get_drvdata(pdev); | 1133 | struct sh_dmae_device *shdev = platform_get_drvdata(pdev); |
| 1134 | struct resource *res; | ||
| 1135 | int errirq = platform_get_irq(pdev, 0); | ||
| 1005 | 1136 | ||
| 1006 | dma_async_device_unregister(&shdev->common); | 1137 | dma_async_device_unregister(&shdev->common); |
| 1007 | 1138 | ||
| 1008 | if (shdev->pdata.mode & SHDMA_MIX_IRQ) { | 1139 | if (errirq > 0) |
| 1009 | free_irq(DMTE0_IRQ, shdev); | 1140 | free_irq(errirq, shdev); |
| 1010 | #if defined(DMTE6_IRQ) | ||
| 1011 | free_irq(DMTE6_IRQ, shdev); | ||
| 1012 | #endif | ||
| 1013 | } | ||
| 1014 | 1141 | ||
| 1015 | /* channel data remove */ | 1142 | /* channel data remove */ |
| 1016 | sh_dmae_chan_remove(shdev); | 1143 | sh_dmae_chan_remove(shdev); |
| 1017 | 1144 | ||
| 1018 | if (!(shdev->pdata.mode & SHDMA_MIX_IRQ)) { | 1145 | pm_runtime_disable(&pdev->dev); |
| 1019 | free_irq(DMAE0_IRQ, shdev); | 1146 | |
| 1020 | #if defined(DMAE1_IRQ) | 1147 | if (shdev->dmars) |
| 1021 | free_irq(DMAE1_IRQ, shdev); | 1148 | iounmap(shdev->dmars); |
| 1022 | #endif | 1149 | iounmap(shdev->chan_reg); |
| 1023 | } | 1150 | |
| 1024 | kfree(shdev); | 1151 | kfree(shdev); |
| 1025 | 1152 | ||
| 1153 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
| 1154 | if (res) | ||
| 1155 | release_mem_region(res->start, resource_size(res)); | ||
| 1156 | res = platform_get_resource(pdev, IORESOURCE_MEM, 1); | ||
| 1157 | if (res) | ||
| 1158 | release_mem_region(res->start, resource_size(res)); | ||
| 1159 | |||
| 1026 | return 0; | 1160 | return 0; |
| 1027 | } | 1161 | } |
| 1028 | 1162 | ||
| 1029 | static void sh_dmae_shutdown(struct platform_device *pdev) | 1163 | static void sh_dmae_shutdown(struct platform_device *pdev) |
| 1030 | { | 1164 | { |
| 1031 | struct sh_dmae_device *shdev = platform_get_drvdata(pdev); | 1165 | struct sh_dmae_device *shdev = platform_get_drvdata(pdev); |
| 1032 | sh_dmae_ctl_stop(0); | 1166 | sh_dmae_ctl_stop(shdev); |
| 1033 | if (shdev->pdata.mode & SHDMA_DMAOR1) | ||
| 1034 | sh_dmae_ctl_stop(1); | ||
| 1035 | } | 1167 | } |
| 1036 | 1168 | ||
| 1037 | static struct platform_driver sh_dmae_driver = { | 1169 | static struct platform_driver sh_dmae_driver = { |
diff --git a/drivers/dma/shdma.h b/drivers/dma/shdma.h index 7e227f3c87c4..153609a1e96c 100644 --- a/drivers/dma/shdma.h +++ b/drivers/dma/shdma.h | |||
| @@ -17,23 +17,9 @@ | |||
| 17 | #include <linux/interrupt.h> | 17 | #include <linux/interrupt.h> |
| 18 | #include <linux/list.h> | 18 | #include <linux/list.h> |
| 19 | 19 | ||
| 20 | #define SH_DMA_TCR_MAX 0x00FFFFFF /* 16MB */ | 20 | #include <asm/dmaengine.h> |
| 21 | |||
| 22 | struct sh_dmae_regs { | ||
| 23 | u32 sar; /* SAR / source address */ | ||
| 24 | u32 dar; /* DAR / destination address */ | ||
| 25 | u32 tcr; /* TCR / transfer count */ | ||
| 26 | }; | ||
| 27 | 21 | ||
| 28 | struct sh_desc { | 22 | #define SH_DMA_TCR_MAX 0x00FFFFFF /* 16MB */ |
| 29 | struct sh_dmae_regs hw; | ||
| 30 | struct list_head node; | ||
| 31 | struct dma_async_tx_descriptor async_tx; | ||
| 32 | enum dma_data_direction direction; | ||
| 33 | dma_cookie_t cookie; | ||
| 34 | int chunks; | ||
| 35 | int mark; | ||
| 36 | }; | ||
| 37 | 23 | ||
| 38 | struct device; | 24 | struct device; |
| 39 | 25 | ||
| @@ -47,14 +33,18 @@ struct sh_dmae_chan { | |||
| 47 | struct tasklet_struct tasklet; /* Tasklet */ | 33 | struct tasklet_struct tasklet; /* Tasklet */ |
| 48 | int descs_allocated; /* desc count */ | 34 | int descs_allocated; /* desc count */ |
| 49 | int xmit_shift; /* log_2(bytes_per_xfer) */ | 35 | int xmit_shift; /* log_2(bytes_per_xfer) */ |
| 36 | int irq; | ||
| 50 | int id; /* Raw id of this channel */ | 37 | int id; /* Raw id of this channel */ |
| 38 | u32 __iomem *base; | ||
| 51 | char dev_id[16]; /* unique name per DMAC of channel */ | 39 | char dev_id[16]; /* unique name per DMAC of channel */ |
| 52 | }; | 40 | }; |
| 53 | 41 | ||
| 54 | struct sh_dmae_device { | 42 | struct sh_dmae_device { |
| 55 | struct dma_device common; | 43 | struct dma_device common; |
| 56 | struct sh_dmae_chan *chan[MAX_DMA_CHANNELS]; | 44 | struct sh_dmae_chan *chan[SH_DMAC_MAX_CHANNELS]; |
| 57 | struct sh_dmae_pdata pdata; | 45 | struct sh_dmae_pdata *pdata; |
| 46 | u32 __iomem *chan_reg; | ||
| 47 | u16 __iomem *dmars; | ||
| 58 | }; | 48 | }; |
| 59 | 49 | ||
| 60 | #define to_sh_chan(chan) container_of(chan, struct sh_dmae_chan, common) | 50 | #define to_sh_chan(chan) container_of(chan, struct sh_dmae_chan, common) |
diff --git a/drivers/serial/Kconfig b/drivers/serial/Kconfig index 746e07033dce..d6ff73395623 100644 --- a/drivers/serial/Kconfig +++ b/drivers/serial/Kconfig | |||
| @@ -1009,6 +1009,10 @@ config SERIAL_SH_SCI_CONSOLE | |||
| 1009 | depends on SERIAL_SH_SCI=y | 1009 | depends on SERIAL_SH_SCI=y |
| 1010 | select SERIAL_CORE_CONSOLE | 1010 | select SERIAL_CORE_CONSOLE |
| 1011 | 1011 | ||
| 1012 | config SERIAL_SH_SCI_DMA | ||
| 1013 | bool "DMA support" | ||
| 1014 | depends on SERIAL_SH_SCI && SH_DMAE && EXPERIMENTAL | ||
| 1015 | |||
| 1012 | config SERIAL_PNX8XXX | 1016 | config SERIAL_PNX8XXX |
| 1013 | bool "Enable PNX8XXX SoCs' UART Support" | 1017 | bool "Enable PNX8XXX SoCs' UART Support" |
| 1014 | depends on MIPS && (SOC_PNX8550 || SOC_PNX833X) | 1018 | depends on MIPS && (SOC_PNX8550 || SOC_PNX833X) |
diff --git a/drivers/serial/sh-sci.c b/drivers/serial/sh-sci.c index 42f3333c4ad0..980f39449ee5 100644 --- a/drivers/serial/sh-sci.c +++ b/drivers/serial/sh-sci.c | |||
| @@ -48,6 +48,9 @@ | |||
| 48 | #include <linux/ctype.h> | 48 | #include <linux/ctype.h> |
| 49 | #include <linux/err.h> | 49 | #include <linux/err.h> |
| 50 | #include <linux/list.h> | 50 | #include <linux/list.h> |
| 51 | #include <linux/dmaengine.h> | ||
| 52 | #include <linux/scatterlist.h> | ||
| 53 | #include <linux/timer.h> | ||
| 51 | 54 | ||
| 52 | #ifdef CONFIG_SUPERH | 55 | #ifdef CONFIG_SUPERH |
| 53 | #include <asm/sh_bios.h> | 56 | #include <asm/sh_bios.h> |
| @@ -84,6 +87,27 @@ struct sci_port { | |||
| 84 | struct clk *dclk; | 87 | struct clk *dclk; |
| 85 | 88 | ||
| 86 | struct list_head node; | 89 | struct list_head node; |
| 90 | struct dma_chan *chan_tx; | ||
| 91 | struct dma_chan *chan_rx; | ||
| 92 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 93 | struct device *dma_dev; | ||
| 94 | enum sh_dmae_slave_chan_id slave_tx; | ||
| 95 | enum sh_dmae_slave_chan_id slave_rx; | ||
| 96 | struct dma_async_tx_descriptor *desc_tx; | ||
| 97 | struct dma_async_tx_descriptor *desc_rx[2]; | ||
| 98 | dma_cookie_t cookie_tx; | ||
| 99 | dma_cookie_t cookie_rx[2]; | ||
| 100 | dma_cookie_t active_rx; | ||
| 101 | struct scatterlist sg_tx; | ||
| 102 | unsigned int sg_len_tx; | ||
| 103 | struct scatterlist sg_rx[2]; | ||
| 104 | size_t buf_len_rx; | ||
| 105 | struct sh_dmae_slave param_tx; | ||
| 106 | struct sh_dmae_slave param_rx; | ||
| 107 | struct work_struct work_tx; | ||
| 108 | struct work_struct work_rx; | ||
| 109 | struct timer_list rx_timer; | ||
| 110 | #endif | ||
| 87 | }; | 111 | }; |
| 88 | 112 | ||
| 89 | struct sh_sci_priv { | 113 | struct sh_sci_priv { |
| @@ -269,29 +293,44 @@ static inline void sci_init_pins(struct uart_port *port, unsigned int cflag) | |||
| 269 | defined(CONFIG_CPU_SUBTYPE_SH7780) || \ | 293 | defined(CONFIG_CPU_SUBTYPE_SH7780) || \ |
| 270 | defined(CONFIG_CPU_SUBTYPE_SH7785) || \ | 294 | defined(CONFIG_CPU_SUBTYPE_SH7785) || \ |
| 271 | defined(CONFIG_CPU_SUBTYPE_SH7786) | 295 | defined(CONFIG_CPU_SUBTYPE_SH7786) |
| 272 | static inline int scif_txroom(struct uart_port *port) | 296 | static int scif_txfill(struct uart_port *port) |
| 273 | { | 297 | { |
| 274 | return SCIF_TXROOM_MAX - (sci_in(port, SCTFDR) & 0xff); | 298 | return sci_in(port, SCTFDR) & 0xff; |
| 275 | } | 299 | } |
| 276 | 300 | ||
| 277 | static inline int scif_rxroom(struct uart_port *port) | 301 | static int scif_txroom(struct uart_port *port) |
| 302 | { | ||
| 303 | return SCIF_TXROOM_MAX - scif_txfill(port); | ||
| 304 | } | ||
| 305 | |||
| 306 | static int scif_rxfill(struct uart_port *port) | ||
| 278 | { | 307 | { |
| 279 | return sci_in(port, SCRFDR) & 0xff; | 308 | return sci_in(port, SCRFDR) & 0xff; |
| 280 | } | 309 | } |
| 281 | #elif defined(CONFIG_CPU_SUBTYPE_SH7763) | 310 | #elif defined(CONFIG_CPU_SUBTYPE_SH7763) |
| 282 | static inline int scif_txroom(struct uart_port *port) | 311 | static int scif_txfill(struct uart_port *port) |
| 283 | { | 312 | { |
| 284 | if ((port->mapbase == 0xffe00000) || | 313 | if (port->mapbase == 0xffe00000 || |
| 285 | (port->mapbase == 0xffe08000)) { | 314 | port->mapbase == 0xffe08000) |
| 286 | /* SCIF0/1*/ | 315 | /* SCIF0/1*/ |
| 287 | return SCIF_TXROOM_MAX - (sci_in(port, SCTFDR) & 0xff); | 316 | return sci_in(port, SCTFDR) & 0xff; |
| 288 | } else { | 317 | else |
| 289 | /* SCIF2 */ | 318 | /* SCIF2 */ |
| 290 | return SCIF2_TXROOM_MAX - (sci_in(port, SCFDR) >> 8); | 319 | return sci_in(port, SCFDR) >> 8; |
| 291 | } | ||
| 292 | } | 320 | } |
| 293 | 321 | ||
| 294 | static inline int scif_rxroom(struct uart_port *port) | 322 | static int scif_txroom(struct uart_port *port) |
| 323 | { | ||
| 324 | if (port->mapbase == 0xffe00000 || | ||
| 325 | port->mapbase == 0xffe08000) | ||
| 326 | /* SCIF0/1*/ | ||
| 327 | return SCIF_TXROOM_MAX - scif_txfill(port); | ||
| 328 | else | ||
| 329 | /* SCIF2 */ | ||
| 330 | return SCIF2_TXROOM_MAX - scif_txfill(port); | ||
| 331 | } | ||
| 332 | |||
| 333 | static int scif_rxfill(struct uart_port *port) | ||
| 295 | { | 334 | { |
| 296 | if ((port->mapbase == 0xffe00000) || | 335 | if ((port->mapbase == 0xffe00000) || |
| 297 | (port->mapbase == 0xffe08000)) { | 336 | (port->mapbase == 0xffe08000)) { |
| @@ -303,23 +342,33 @@ static inline int scif_rxroom(struct uart_port *port) | |||
| 303 | } | 342 | } |
| 304 | } | 343 | } |
| 305 | #else | 344 | #else |
| 306 | static inline int scif_txroom(struct uart_port *port) | 345 | static int scif_txfill(struct uart_port *port) |
| 346 | { | ||
| 347 | return sci_in(port, SCFDR) >> 8; | ||
| 348 | } | ||
| 349 | |||
| 350 | static int scif_txroom(struct uart_port *port) | ||
| 307 | { | 351 | { |
| 308 | return SCIF_TXROOM_MAX - (sci_in(port, SCFDR) >> 8); | 352 | return SCIF_TXROOM_MAX - scif_txfill(port); |
| 309 | } | 353 | } |
| 310 | 354 | ||
| 311 | static inline int scif_rxroom(struct uart_port *port) | 355 | static int scif_rxfill(struct uart_port *port) |
| 312 | { | 356 | { |
| 313 | return sci_in(port, SCFDR) & SCIF_RFDC_MASK; | 357 | return sci_in(port, SCFDR) & SCIF_RFDC_MASK; |
| 314 | } | 358 | } |
| 315 | #endif | 359 | #endif |
| 316 | 360 | ||
| 317 | static inline int sci_txroom(struct uart_port *port) | 361 | static int sci_txfill(struct uart_port *port) |
| 318 | { | 362 | { |
| 319 | return (sci_in(port, SCxSR) & SCI_TDRE) != 0; | 363 | return !(sci_in(port, SCxSR) & SCI_TDRE); |
| 320 | } | 364 | } |
| 321 | 365 | ||
| 322 | static inline int sci_rxroom(struct uart_port *port) | 366 | static int sci_txroom(struct uart_port *port) |
| 367 | { | ||
| 368 | return !sci_txfill(port); | ||
| 369 | } | ||
| 370 | |||
| 371 | static int sci_rxfill(struct uart_port *port) | ||
| 323 | { | 372 | { |
| 324 | return (sci_in(port, SCxSR) & SCxSR_RDxF(port)) != 0; | 373 | return (sci_in(port, SCxSR) & SCxSR_RDxF(port)) != 0; |
| 325 | } | 374 | } |
| @@ -406,9 +455,9 @@ static inline void sci_receive_chars(struct uart_port *port) | |||
| 406 | 455 | ||
| 407 | while (1) { | 456 | while (1) { |
| 408 | if (port->type == PORT_SCI) | 457 | if (port->type == PORT_SCI) |
| 409 | count = sci_rxroom(port); | 458 | count = sci_rxfill(port); |
| 410 | else | 459 | else |
| 411 | count = scif_rxroom(port); | 460 | count = scif_rxfill(port); |
| 412 | 461 | ||
| 413 | /* Don't copy more bytes than there is room for in the buffer */ | 462 | /* Don't copy more bytes than there is room for in the buffer */ |
| 414 | count = tty_buffer_request_room(tty, count); | 463 | count = tty_buffer_request_room(tty, count); |
| @@ -453,10 +502,10 @@ static inline void sci_receive_chars(struct uart_port *port) | |||
| 453 | } | 502 | } |
| 454 | 503 | ||
| 455 | /* Store data and status */ | 504 | /* Store data and status */ |
| 456 | if (status&SCxSR_FER(port)) { | 505 | if (status & SCxSR_FER(port)) { |
| 457 | flag = TTY_FRAME; | 506 | flag = TTY_FRAME; |
| 458 | dev_notice(port->dev, "frame error\n"); | 507 | dev_notice(port->dev, "frame error\n"); |
| 459 | } else if (status&SCxSR_PER(port)) { | 508 | } else if (status & SCxSR_PER(port)) { |
| 460 | flag = TTY_PARITY; | 509 | flag = TTY_PARITY; |
| 461 | dev_notice(port->dev, "parity error\n"); | 510 | dev_notice(port->dev, "parity error\n"); |
| 462 | } else | 511 | } else |
| @@ -618,13 +667,39 @@ static inline int sci_handle_breaks(struct uart_port *port) | |||
| 618 | return copied; | 667 | return copied; |
| 619 | } | 668 | } |
| 620 | 669 | ||
| 621 | static irqreturn_t sci_rx_interrupt(int irq, void *port) | 670 | static irqreturn_t sci_rx_interrupt(int irq, void *ptr) |
| 622 | { | 671 | { |
| 672 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 673 | struct uart_port *port = ptr; | ||
| 674 | struct sci_port *s = to_sci_port(port); | ||
| 675 | |||
| 676 | if (s->chan_rx) { | ||
| 677 | unsigned long tout; | ||
| 678 | u16 scr = sci_in(port, SCSCR); | ||
| 679 | u16 ssr = sci_in(port, SCxSR); | ||
| 680 | |||
| 681 | /* Disable future Rx interrupts */ | ||
| 682 | sci_out(port, SCSCR, scr & ~SCI_CTRL_FLAGS_RIE); | ||
| 683 | /* Clear current interrupt */ | ||
| 684 | sci_out(port, SCxSR, ssr & ~(1 | SCxSR_RDxF(port))); | ||
| 685 | /* Calculate delay for 1.5 DMA buffers */ | ||
| 686 | tout = (port->timeout - HZ / 50) * s->buf_len_rx * 3 / | ||
| 687 | port->fifosize / 2; | ||
| 688 | dev_dbg(port->dev, "Rx IRQ: setup timeout in %lu ms\n", | ||
| 689 | tout * 1000 / HZ); | ||
| 690 | if (tout < 2) | ||
| 691 | tout = 2; | ||
| 692 | mod_timer(&s->rx_timer, jiffies + tout); | ||
| 693 | |||
| 694 | return IRQ_HANDLED; | ||
| 695 | } | ||
| 696 | #endif | ||
| 697 | |||
| 623 | /* I think sci_receive_chars has to be called irrespective | 698 | /* I think sci_receive_chars has to be called irrespective |
| 624 | * of whether the I_IXOFF is set, otherwise, how is the interrupt | 699 | * of whether the I_IXOFF is set, otherwise, how is the interrupt |
| 625 | * to be disabled? | 700 | * to be disabled? |
| 626 | */ | 701 | */ |
| 627 | sci_receive_chars(port); | 702 | sci_receive_chars(ptr); |
| 628 | 703 | ||
| 629 | return IRQ_HANDLED; | 704 | return IRQ_HANDLED; |
| 630 | } | 705 | } |
| @@ -680,6 +755,7 @@ static irqreturn_t sci_mpxed_interrupt(int irq, void *ptr) | |||
| 680 | { | 755 | { |
| 681 | unsigned short ssr_status, scr_status, err_enabled; | 756 | unsigned short ssr_status, scr_status, err_enabled; |
| 682 | struct uart_port *port = ptr; | 757 | struct uart_port *port = ptr; |
| 758 | struct sci_port *s = to_sci_port(port); | ||
| 683 | irqreturn_t ret = IRQ_NONE; | 759 | irqreturn_t ret = IRQ_NONE; |
| 684 | 760 | ||
| 685 | ssr_status = sci_in(port, SCxSR); | 761 | ssr_status = sci_in(port, SCxSR); |
| @@ -687,10 +763,15 @@ static irqreturn_t sci_mpxed_interrupt(int irq, void *ptr) | |||
| 687 | err_enabled = scr_status & (SCI_CTRL_FLAGS_REIE | SCI_CTRL_FLAGS_RIE); | 763 | err_enabled = scr_status & (SCI_CTRL_FLAGS_REIE | SCI_CTRL_FLAGS_RIE); |
| 688 | 764 | ||
| 689 | /* Tx Interrupt */ | 765 | /* Tx Interrupt */ |
| 690 | if ((ssr_status & SCxSR_TDxE(port)) && (scr_status & SCI_CTRL_FLAGS_TIE)) | 766 | if ((ssr_status & SCxSR_TDxE(port)) && (scr_status & SCI_CTRL_FLAGS_TIE) && |
| 767 | !s->chan_tx) | ||
| 691 | ret = sci_tx_interrupt(irq, ptr); | 768 | ret = sci_tx_interrupt(irq, ptr); |
| 692 | /* Rx Interrupt */ | 769 | /* |
| 693 | if ((ssr_status & SCxSR_RDxF(port)) && (scr_status & SCI_CTRL_FLAGS_RIE)) | 770 | * Rx Interrupt: if we're using DMA, the DMA controller clears RDF / |
| 771 | * DR flags | ||
| 772 | */ | ||
| 773 | if (((ssr_status & SCxSR_RDxF(port)) || s->chan_rx) && | ||
| 774 | (scr_status & SCI_CTRL_FLAGS_RIE)) | ||
| 694 | ret = sci_rx_interrupt(irq, ptr); | 775 | ret = sci_rx_interrupt(irq, ptr); |
| 695 | /* Error Interrupt */ | 776 | /* Error Interrupt */ |
| 696 | if ((ssr_status & SCxSR_ERRORS(port)) && err_enabled) | 777 | if ((ssr_status & SCxSR_ERRORS(port)) && err_enabled) |
| @@ -699,6 +780,10 @@ static irqreturn_t sci_mpxed_interrupt(int irq, void *ptr) | |||
| 699 | if ((ssr_status & SCxSR_BRK(port)) && err_enabled) | 780 | if ((ssr_status & SCxSR_BRK(port)) && err_enabled) |
| 700 | ret = sci_br_interrupt(irq, ptr); | 781 | ret = sci_br_interrupt(irq, ptr); |
| 701 | 782 | ||
| 783 | WARN_ONCE(ret == IRQ_NONE, | ||
| 784 | "%s: %d IRQ %d, status %x, control %x\n", __func__, | ||
| 785 | irq, port->line, ssr_status, scr_status); | ||
| 786 | |||
| 702 | return ret; | 787 | return ret; |
| 703 | } | 788 | } |
| 704 | 789 | ||
| @@ -800,7 +885,9 @@ static void sci_free_irq(struct sci_port *port) | |||
| 800 | static unsigned int sci_tx_empty(struct uart_port *port) | 885 | static unsigned int sci_tx_empty(struct uart_port *port) |
| 801 | { | 886 | { |
| 802 | unsigned short status = sci_in(port, SCxSR); | 887 | unsigned short status = sci_in(port, SCxSR); |
| 803 | return status & SCxSR_TEND(port) ? TIOCSER_TEMT : 0; | 888 | unsigned short in_tx_fifo = scif_txfill(port); |
| 889 | |||
| 890 | return (status & SCxSR_TEND(port)) && !in_tx_fifo ? TIOCSER_TEMT : 0; | ||
| 804 | } | 891 | } |
| 805 | 892 | ||
| 806 | static void sci_set_mctrl(struct uart_port *port, unsigned int mctrl) | 893 | static void sci_set_mctrl(struct uart_port *port, unsigned int mctrl) |
| @@ -812,16 +899,297 @@ static void sci_set_mctrl(struct uart_port *port, unsigned int mctrl) | |||
| 812 | 899 | ||
| 813 | static unsigned int sci_get_mctrl(struct uart_port *port) | 900 | static unsigned int sci_get_mctrl(struct uart_port *port) |
| 814 | { | 901 | { |
| 815 | /* This routine is used for geting signals of: DTR, DCD, DSR, RI, | 902 | /* This routine is used for getting signals of: DTR, DCD, DSR, RI, |
| 816 | and CTS/RTS */ | 903 | and CTS/RTS */ |
| 817 | 904 | ||
| 818 | return TIOCM_DTR | TIOCM_RTS | TIOCM_DSR; | 905 | return TIOCM_DTR | TIOCM_RTS | TIOCM_DSR; |
| 819 | } | 906 | } |
| 820 | 907 | ||
| 908 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 909 | static void sci_dma_tx_complete(void *arg) | ||
| 910 | { | ||
| 911 | struct sci_port *s = arg; | ||
| 912 | struct uart_port *port = &s->port; | ||
| 913 | struct circ_buf *xmit = &port->state->xmit; | ||
| 914 | unsigned long flags; | ||
| 915 | |||
| 916 | dev_dbg(port->dev, "%s(%d)\n", __func__, port->line); | ||
| 917 | |||
| 918 | spin_lock_irqsave(&port->lock, flags); | ||
| 919 | |||
| 920 | xmit->tail += s->sg_tx.length; | ||
| 921 | xmit->tail &= UART_XMIT_SIZE - 1; | ||
| 922 | |||
| 923 | port->icount.tx += s->sg_tx.length; | ||
| 924 | |||
| 925 | async_tx_ack(s->desc_tx); | ||
| 926 | s->cookie_tx = -EINVAL; | ||
| 927 | s->desc_tx = NULL; | ||
| 928 | |||
| 929 | spin_unlock_irqrestore(&port->lock, flags); | ||
| 930 | |||
| 931 | if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS) | ||
| 932 | uart_write_wakeup(port); | ||
| 933 | |||
| 934 | if (uart_circ_chars_pending(xmit)) | ||
| 935 | schedule_work(&s->work_tx); | ||
| 936 | } | ||
| 937 | |||
| 938 | /* Locking: called with port lock held */ | ||
| 939 | static int sci_dma_rx_push(struct sci_port *s, struct tty_struct *tty, | ||
| 940 | size_t count) | ||
| 941 | { | ||
| 942 | struct uart_port *port = &s->port; | ||
| 943 | int i, active, room; | ||
| 944 | |||
| 945 | room = tty_buffer_request_room(tty, count); | ||
| 946 | |||
| 947 | if (s->active_rx == s->cookie_rx[0]) { | ||
| 948 | active = 0; | ||
| 949 | } else if (s->active_rx == s->cookie_rx[1]) { | ||
| 950 | active = 1; | ||
| 951 | } else { | ||
| 952 | dev_err(port->dev, "cookie %d not found!\n", s->active_rx); | ||
| 953 | return 0; | ||
| 954 | } | ||
| 955 | |||
| 956 | if (room < count) | ||
| 957 | dev_warn(port->dev, "Rx overrun: dropping %u bytes\n", | ||
| 958 | count - room); | ||
| 959 | if (!room) | ||
| 960 | return room; | ||
| 961 | |||
| 962 | for (i = 0; i < room; i++) | ||
| 963 | tty_insert_flip_char(tty, ((u8 *)sg_virt(&s->sg_rx[active]))[i], | ||
| 964 | TTY_NORMAL); | ||
| 965 | |||
| 966 | port->icount.rx += room; | ||
| 967 | |||
| 968 | return room; | ||
| 969 | } | ||
| 970 | |||
| 971 | static void sci_dma_rx_complete(void *arg) | ||
| 972 | { | ||
| 973 | struct sci_port *s = arg; | ||
| 974 | struct uart_port *port = &s->port; | ||
| 975 | struct tty_struct *tty = port->state->port.tty; | ||
| 976 | unsigned long flags; | ||
| 977 | int count; | ||
| 978 | |||
| 979 | dev_dbg(port->dev, "%s(%d)\n", __func__, port->line); | ||
| 980 | |||
| 981 | spin_lock_irqsave(&port->lock, flags); | ||
| 982 | |||
| 983 | count = sci_dma_rx_push(s, tty, s->buf_len_rx); | ||
| 984 | |||
| 985 | mod_timer(&s->rx_timer, jiffies + msecs_to_jiffies(5)); | ||
| 986 | |||
| 987 | spin_unlock_irqrestore(&port->lock, flags); | ||
| 988 | |||
| 989 | if (count) | ||
| 990 | tty_flip_buffer_push(tty); | ||
| 991 | |||
| 992 | schedule_work(&s->work_rx); | ||
| 993 | } | ||
| 994 | |||
| 995 | static void sci_start_rx(struct uart_port *port); | ||
| 996 | static void sci_start_tx(struct uart_port *port); | ||
| 997 | |||
| 998 | static void sci_rx_dma_release(struct sci_port *s, bool enable_pio) | ||
| 999 | { | ||
| 1000 | struct dma_chan *chan = s->chan_rx; | ||
| 1001 | struct uart_port *port = &s->port; | ||
| 1002 | |||
| 1003 | s->chan_rx = NULL; | ||
| 1004 | s->cookie_rx[0] = s->cookie_rx[1] = -EINVAL; | ||
| 1005 | dma_release_channel(chan); | ||
| 1006 | dma_free_coherent(port->dev, s->buf_len_rx * 2, | ||
| 1007 | sg_virt(&s->sg_rx[0]), sg_dma_address(&s->sg_rx[0])); | ||
| 1008 | if (enable_pio) | ||
| 1009 | sci_start_rx(port); | ||
| 1010 | } | ||
| 1011 | |||
| 1012 | static void sci_tx_dma_release(struct sci_port *s, bool enable_pio) | ||
| 1013 | { | ||
| 1014 | struct dma_chan *chan = s->chan_tx; | ||
| 1015 | struct uart_port *port = &s->port; | ||
| 1016 | |||
| 1017 | s->chan_tx = NULL; | ||
| 1018 | s->cookie_tx = -EINVAL; | ||
| 1019 | dma_release_channel(chan); | ||
| 1020 | if (enable_pio) | ||
| 1021 | sci_start_tx(port); | ||
| 1022 | } | ||
| 1023 | |||
| 1024 | static void sci_submit_rx(struct sci_port *s) | ||
| 1025 | { | ||
| 1026 | struct dma_chan *chan = s->chan_rx; | ||
| 1027 | int i; | ||
| 1028 | |||
| 1029 | for (i = 0; i < 2; i++) { | ||
| 1030 | struct scatterlist *sg = &s->sg_rx[i]; | ||
| 1031 | struct dma_async_tx_descriptor *desc; | ||
| 1032 | |||
| 1033 | desc = chan->device->device_prep_slave_sg(chan, | ||
| 1034 | sg, 1, DMA_FROM_DEVICE, DMA_PREP_INTERRUPT); | ||
| 1035 | |||
| 1036 | if (desc) { | ||
| 1037 | s->desc_rx[i] = desc; | ||
| 1038 | desc->callback = sci_dma_rx_complete; | ||
| 1039 | desc->callback_param = s; | ||
| 1040 | s->cookie_rx[i] = desc->tx_submit(desc); | ||
| 1041 | } | ||
| 1042 | |||
| 1043 | if (!desc || s->cookie_rx[i] < 0) { | ||
| 1044 | if (i) { | ||
| 1045 | async_tx_ack(s->desc_rx[0]); | ||
| 1046 | s->cookie_rx[0] = -EINVAL; | ||
| 1047 | } | ||
| 1048 | if (desc) { | ||
| 1049 | async_tx_ack(desc); | ||
| 1050 | s->cookie_rx[i] = -EINVAL; | ||
| 1051 | } | ||
| 1052 | dev_warn(s->port.dev, | ||
| 1053 | "failed to re-start DMA, using PIO\n"); | ||
| 1054 | sci_rx_dma_release(s, true); | ||
| 1055 | return; | ||
| 1056 | } | ||
| 1057 | } | ||
| 1058 | |||
| 1059 | s->active_rx = s->cookie_rx[0]; | ||
| 1060 | |||
| 1061 | dma_async_issue_pending(chan); | ||
| 1062 | } | ||
| 1063 | |||
| 1064 | static void work_fn_rx(struct work_struct *work) | ||
| 1065 | { | ||
| 1066 | struct sci_port *s = container_of(work, struct sci_port, work_rx); | ||
| 1067 | struct uart_port *port = &s->port; | ||
| 1068 | struct dma_async_tx_descriptor *desc; | ||
| 1069 | int new; | ||
| 1070 | |||
| 1071 | if (s->active_rx == s->cookie_rx[0]) { | ||
| 1072 | new = 0; | ||
| 1073 | } else if (s->active_rx == s->cookie_rx[1]) { | ||
| 1074 | new = 1; | ||
| 1075 | } else { | ||
| 1076 | dev_err(port->dev, "cookie %d not found!\n", s->active_rx); | ||
| 1077 | return; | ||
| 1078 | } | ||
| 1079 | desc = s->desc_rx[new]; | ||
| 1080 | |||
| 1081 | if (dma_async_is_tx_complete(s->chan_rx, s->active_rx, NULL, NULL) != | ||
| 1082 | DMA_SUCCESS) { | ||
| 1083 | /* Handle incomplete DMA receive */ | ||
| 1084 | struct tty_struct *tty = port->state->port.tty; | ||
| 1085 | struct dma_chan *chan = s->chan_rx; | ||
| 1086 | struct sh_desc *sh_desc = container_of(desc, struct sh_desc, | ||
| 1087 | async_tx); | ||
| 1088 | unsigned long flags; | ||
| 1089 | int count; | ||
| 1090 | |||
| 1091 | chan->device->device_terminate_all(chan); | ||
| 1092 | dev_dbg(port->dev, "Read %u bytes with cookie %d\n", | ||
| 1093 | sh_desc->partial, sh_desc->cookie); | ||
| 1094 | |||
| 1095 | spin_lock_irqsave(&port->lock, flags); | ||
| 1096 | count = sci_dma_rx_push(s, tty, sh_desc->partial); | ||
| 1097 | spin_unlock_irqrestore(&port->lock, flags); | ||
| 1098 | |||
| 1099 | if (count) | ||
| 1100 | tty_flip_buffer_push(tty); | ||
| 1101 | |||
| 1102 | sci_submit_rx(s); | ||
| 1103 | |||
| 1104 | return; | ||
| 1105 | } | ||
| 1106 | |||
| 1107 | s->cookie_rx[new] = desc->tx_submit(desc); | ||
| 1108 | if (s->cookie_rx[new] < 0) { | ||
| 1109 | dev_warn(port->dev, "Failed submitting Rx DMA descriptor\n"); | ||
| 1110 | sci_rx_dma_release(s, true); | ||
| 1111 | return; | ||
| 1112 | } | ||
| 1113 | |||
| 1114 | dev_dbg(port->dev, "%s: cookie %d #%d\n", __func__, | ||
| 1115 | s->cookie_rx[new], new); | ||
| 1116 | |||
| 1117 | s->active_rx = s->cookie_rx[!new]; | ||
| 1118 | } | ||
| 1119 | |||
| 1120 | static void work_fn_tx(struct work_struct *work) | ||
| 1121 | { | ||
| 1122 | struct sci_port *s = container_of(work, struct sci_port, work_tx); | ||
| 1123 | struct dma_async_tx_descriptor *desc; | ||
| 1124 | struct dma_chan *chan = s->chan_tx; | ||
| 1125 | struct uart_port *port = &s->port; | ||
| 1126 | struct circ_buf *xmit = &port->state->xmit; | ||
| 1127 | struct scatterlist *sg = &s->sg_tx; | ||
| 1128 | |||
| 1129 | /* | ||
| 1130 | * DMA is idle now. | ||
| 1131 | * Port xmit buffer is already mapped, and it is one page... Just adjust | ||
| 1132 | * offsets and lengths. Since it is a circular buffer, we have to | ||
| 1133 | * transmit till the end, and then the rest. Take the port lock to get a | ||
| 1134 | * consistent xmit buffer state. | ||
| 1135 | */ | ||
| 1136 | spin_lock_irq(&port->lock); | ||
| 1137 | sg->offset = xmit->tail & (UART_XMIT_SIZE - 1); | ||
| 1138 | sg->dma_address = (sg_dma_address(sg) & ~(UART_XMIT_SIZE - 1)) + | ||
| 1139 | sg->offset; | ||
| 1140 | sg->length = min((int)CIRC_CNT(xmit->head, xmit->tail, UART_XMIT_SIZE), | ||
| 1141 | CIRC_CNT_TO_END(xmit->head, xmit->tail, UART_XMIT_SIZE)); | ||
| 1142 | sg->dma_length = sg->length; | ||
| 1143 | spin_unlock_irq(&port->lock); | ||
| 1144 | |||
| 1145 | BUG_ON(!sg->length); | ||
| 1146 | |||
| 1147 | desc = chan->device->device_prep_slave_sg(chan, | ||
| 1148 | sg, s->sg_len_tx, DMA_TO_DEVICE, | ||
| 1149 | DMA_PREP_INTERRUPT | DMA_CTRL_ACK); | ||
| 1150 | if (!desc) { | ||
| 1151 | /* switch to PIO */ | ||
| 1152 | sci_tx_dma_release(s, true); | ||
| 1153 | return; | ||
| 1154 | } | ||
| 1155 | |||
| 1156 | dma_sync_sg_for_device(port->dev, sg, 1, DMA_TO_DEVICE); | ||
| 1157 | |||
| 1158 | spin_lock_irq(&port->lock); | ||
| 1159 | s->desc_tx = desc; | ||
| 1160 | desc->callback = sci_dma_tx_complete; | ||
| 1161 | desc->callback_param = s; | ||
| 1162 | spin_unlock_irq(&port->lock); | ||
| 1163 | s->cookie_tx = desc->tx_submit(desc); | ||
| 1164 | if (s->cookie_tx < 0) { | ||
| 1165 | dev_warn(port->dev, "Failed submitting Tx DMA descriptor\n"); | ||
| 1166 | /* switch to PIO */ | ||
| 1167 | sci_tx_dma_release(s, true); | ||
| 1168 | return; | ||
| 1169 | } | ||
| 1170 | |||
| 1171 | dev_dbg(port->dev, "%s: %p: %d...%d, cookie %d\n", __func__, | ||
| 1172 | xmit->buf, xmit->tail, xmit->head, s->cookie_tx); | ||
| 1173 | |||
| 1174 | dma_async_issue_pending(chan); | ||
| 1175 | } | ||
| 1176 | #endif | ||
| 1177 | |||
| 821 | static void sci_start_tx(struct uart_port *port) | 1178 | static void sci_start_tx(struct uart_port *port) |
| 822 | { | 1179 | { |
| 823 | unsigned short ctrl; | 1180 | unsigned short ctrl; |
| 824 | 1181 | ||
| 1182 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 1183 | struct sci_port *s = to_sci_port(port); | ||
| 1184 | |||
| 1185 | if (s->chan_tx) { | ||
| 1186 | if (!uart_circ_empty(&s->port.state->xmit) && s->cookie_tx < 0) | ||
| 1187 | schedule_work(&s->work_tx); | ||
| 1188 | |||
| 1189 | return; | ||
| 1190 | } | ||
| 1191 | #endif | ||
| 1192 | |||
| 825 | /* Set TIE (Transmit Interrupt Enable) bit in SCSCR */ | 1193 | /* Set TIE (Transmit Interrupt Enable) bit in SCSCR */ |
| 826 | ctrl = sci_in(port, SCSCR); | 1194 | ctrl = sci_in(port, SCSCR); |
| 827 | ctrl |= SCI_CTRL_FLAGS_TIE; | 1195 | ctrl |= SCI_CTRL_FLAGS_TIE; |
| @@ -838,13 +1206,12 @@ static void sci_stop_tx(struct uart_port *port) | |||
| 838 | sci_out(port, SCSCR, ctrl); | 1206 | sci_out(port, SCSCR, ctrl); |
| 839 | } | 1207 | } |
| 840 | 1208 | ||
| 841 | static void sci_start_rx(struct uart_port *port, unsigned int tty_start) | 1209 | static void sci_start_rx(struct uart_port *port) |
| 842 | { | 1210 | { |
| 843 | unsigned short ctrl; | 1211 | unsigned short ctrl = SCI_CTRL_FLAGS_RIE | SCI_CTRL_FLAGS_REIE; |
| 844 | 1212 | ||
| 845 | /* Set RIE (Receive Interrupt Enable) bit in SCSCR */ | 1213 | /* Set RIE (Receive Interrupt Enable) bit in SCSCR */ |
| 846 | ctrl = sci_in(port, SCSCR); | 1214 | ctrl |= sci_in(port, SCSCR); |
| 847 | ctrl |= SCI_CTRL_FLAGS_RIE | SCI_CTRL_FLAGS_REIE; | ||
| 848 | sci_out(port, SCSCR, ctrl); | 1215 | sci_out(port, SCSCR, ctrl); |
| 849 | } | 1216 | } |
| 850 | 1217 | ||
| @@ -868,16 +1235,154 @@ static void sci_break_ctl(struct uart_port *port, int break_state) | |||
| 868 | /* Nothing here yet .. */ | 1235 | /* Nothing here yet .. */ |
| 869 | } | 1236 | } |
| 870 | 1237 | ||
| 1238 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 1239 | static bool filter(struct dma_chan *chan, void *slave) | ||
| 1240 | { | ||
| 1241 | struct sh_dmae_slave *param = slave; | ||
| 1242 | |||
| 1243 | dev_dbg(chan->device->dev, "%s: slave ID %d\n", __func__, | ||
| 1244 | param->slave_id); | ||
| 1245 | |||
| 1246 | if (param->dma_dev == chan->device->dev) { | ||
| 1247 | chan->private = param; | ||
| 1248 | return true; | ||
| 1249 | } else { | ||
| 1250 | return false; | ||
| 1251 | } | ||
| 1252 | } | ||
| 1253 | |||
| 1254 | static void rx_timer_fn(unsigned long arg) | ||
| 1255 | { | ||
| 1256 | struct sci_port *s = (struct sci_port *)arg; | ||
| 1257 | struct uart_port *port = &s->port; | ||
| 1258 | |||
| 1259 | u16 scr = sci_in(port, SCSCR); | ||
| 1260 | sci_out(port, SCSCR, scr | SCI_CTRL_FLAGS_RIE); | ||
| 1261 | dev_dbg(port->dev, "DMA Rx timed out\n"); | ||
| 1262 | schedule_work(&s->work_rx); | ||
| 1263 | } | ||
| 1264 | |||
| 1265 | static void sci_request_dma(struct uart_port *port) | ||
| 1266 | { | ||
| 1267 | struct sci_port *s = to_sci_port(port); | ||
| 1268 | struct sh_dmae_slave *param; | ||
| 1269 | struct dma_chan *chan; | ||
| 1270 | dma_cap_mask_t mask; | ||
| 1271 | int nent; | ||
| 1272 | |||
| 1273 | dev_dbg(port->dev, "%s: port %d DMA %p\n", __func__, | ||
| 1274 | port->line, s->dma_dev); | ||
| 1275 | |||
| 1276 | if (!s->dma_dev) | ||
| 1277 | return; | ||
| 1278 | |||
| 1279 | dma_cap_zero(mask); | ||
| 1280 | dma_cap_set(DMA_SLAVE, mask); | ||
| 1281 | |||
| 1282 | param = &s->param_tx; | ||
| 1283 | |||
| 1284 | /* Slave ID, e.g., SHDMA_SLAVE_SCIF0_TX */ | ||
| 1285 | param->slave_id = s->slave_tx; | ||
| 1286 | param->dma_dev = s->dma_dev; | ||
| 1287 | |||
| 1288 | s->cookie_tx = -EINVAL; | ||
| 1289 | chan = dma_request_channel(mask, filter, param); | ||
| 1290 | dev_dbg(port->dev, "%s: TX: got channel %p\n", __func__, chan); | ||
| 1291 | if (chan) { | ||
| 1292 | s->chan_tx = chan; | ||
| 1293 | sg_init_table(&s->sg_tx, 1); | ||
| 1294 | /* UART circular tx buffer is an aligned page. */ | ||
| 1295 | BUG_ON((int)port->state->xmit.buf & ~PAGE_MASK); | ||
| 1296 | sg_set_page(&s->sg_tx, virt_to_page(port->state->xmit.buf), | ||
| 1297 | UART_XMIT_SIZE, (int)port->state->xmit.buf & ~PAGE_MASK); | ||
| 1298 | nent = dma_map_sg(port->dev, &s->sg_tx, 1, DMA_TO_DEVICE); | ||
| 1299 | if (!nent) | ||
| 1300 | sci_tx_dma_release(s, false); | ||
| 1301 | else | ||
| 1302 | dev_dbg(port->dev, "%s: mapped %d@%p to %x\n", __func__, | ||
| 1303 | sg_dma_len(&s->sg_tx), | ||
| 1304 | port->state->xmit.buf, sg_dma_address(&s->sg_tx)); | ||
| 1305 | |||
| 1306 | s->sg_len_tx = nent; | ||
| 1307 | |||
| 1308 | INIT_WORK(&s->work_tx, work_fn_tx); | ||
| 1309 | } | ||
| 1310 | |||
| 1311 | param = &s->param_rx; | ||
| 1312 | |||
| 1313 | /* Slave ID, e.g., SHDMA_SLAVE_SCIF0_RX */ | ||
| 1314 | param->slave_id = s->slave_rx; | ||
| 1315 | param->dma_dev = s->dma_dev; | ||
| 1316 | |||
| 1317 | chan = dma_request_channel(mask, filter, param); | ||
| 1318 | dev_dbg(port->dev, "%s: RX: got channel %p\n", __func__, chan); | ||
| 1319 | if (chan) { | ||
| 1320 | dma_addr_t dma[2]; | ||
| 1321 | void *buf[2]; | ||
| 1322 | int i; | ||
| 1323 | |||
| 1324 | s->chan_rx = chan; | ||
| 1325 | |||
| 1326 | s->buf_len_rx = 2 * max(16, (int)port->fifosize); | ||
| 1327 | buf[0] = dma_alloc_coherent(port->dev, s->buf_len_rx * 2, | ||
| 1328 | &dma[0], GFP_KERNEL); | ||
| 1329 | |||
| 1330 | if (!buf[0]) { | ||
| 1331 | dev_warn(port->dev, | ||
| 1332 | "failed to allocate dma buffer, using PIO\n"); | ||
| 1333 | sci_rx_dma_release(s, true); | ||
| 1334 | return; | ||
| 1335 | } | ||
| 1336 | |||
| 1337 | buf[1] = buf[0] + s->buf_len_rx; | ||
| 1338 | dma[1] = dma[0] + s->buf_len_rx; | ||
| 1339 | |||
| 1340 | for (i = 0; i < 2; i++) { | ||
| 1341 | struct scatterlist *sg = &s->sg_rx[i]; | ||
| 1342 | |||
| 1343 | sg_init_table(sg, 1); | ||
| 1344 | sg_set_page(sg, virt_to_page(buf[i]), s->buf_len_rx, | ||
| 1345 | (int)buf[i] & ~PAGE_MASK); | ||
| 1346 | sg->dma_address = dma[i]; | ||
| 1347 | sg->dma_length = sg->length; | ||
| 1348 | } | ||
| 1349 | |||
| 1350 | INIT_WORK(&s->work_rx, work_fn_rx); | ||
| 1351 | setup_timer(&s->rx_timer, rx_timer_fn, (unsigned long)s); | ||
| 1352 | |||
| 1353 | sci_submit_rx(s); | ||
| 1354 | } | ||
| 1355 | } | ||
| 1356 | |||
| 1357 | static void sci_free_dma(struct uart_port *port) | ||
| 1358 | { | ||
| 1359 | struct sci_port *s = to_sci_port(port); | ||
| 1360 | |||
| 1361 | if (!s->dma_dev) | ||
| 1362 | return; | ||
| 1363 | |||
| 1364 | if (s->chan_tx) | ||
| 1365 | sci_tx_dma_release(s, false); | ||
| 1366 | if (s->chan_rx) | ||
| 1367 | sci_rx_dma_release(s, false); | ||
| 1368 | } | ||
| 1369 | #endif | ||
| 1370 | |||
| 871 | static int sci_startup(struct uart_port *port) | 1371 | static int sci_startup(struct uart_port *port) |
| 872 | { | 1372 | { |
| 873 | struct sci_port *s = to_sci_port(port); | 1373 | struct sci_port *s = to_sci_port(port); |
| 874 | 1374 | ||
| 1375 | dev_dbg(port->dev, "%s(%d)\n", __func__, port->line); | ||
| 1376 | |||
| 875 | if (s->enable) | 1377 | if (s->enable) |
| 876 | s->enable(port); | 1378 | s->enable(port); |
| 877 | 1379 | ||
| 878 | sci_request_irq(s); | 1380 | sci_request_irq(s); |
| 1381 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 1382 | sci_request_dma(port); | ||
| 1383 | #endif | ||
| 879 | sci_start_tx(port); | 1384 | sci_start_tx(port); |
| 880 | sci_start_rx(port, 1); | 1385 | sci_start_rx(port); |
| 881 | 1386 | ||
| 882 | return 0; | 1387 | return 0; |
| 883 | } | 1388 | } |
| @@ -886,8 +1391,13 @@ static void sci_shutdown(struct uart_port *port) | |||
| 886 | { | 1391 | { |
| 887 | struct sci_port *s = to_sci_port(port); | 1392 | struct sci_port *s = to_sci_port(port); |
| 888 | 1393 | ||
| 1394 | dev_dbg(port->dev, "%s(%d)\n", __func__, port->line); | ||
| 1395 | |||
| 889 | sci_stop_rx(port); | 1396 | sci_stop_rx(port); |
| 890 | sci_stop_tx(port); | 1397 | sci_stop_tx(port); |
| 1398 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 1399 | sci_free_dma(port); | ||
| 1400 | #endif | ||
| 891 | sci_free_irq(s); | 1401 | sci_free_irq(s); |
| 892 | 1402 | ||
| 893 | if (s->disable) | 1403 | if (s->disable) |
| @@ -937,6 +1447,9 @@ static void sci_set_termios(struct uart_port *port, struct ktermios *termios, | |||
| 937 | 1447 | ||
| 938 | sci_out(port, SCSMR, smr_val); | 1448 | sci_out(port, SCSMR, smr_val); |
| 939 | 1449 | ||
| 1450 | dev_dbg(port->dev, "%s: SMR %x, t %x, SCSCR %x\n", __func__, smr_val, t, | ||
| 1451 | SCSCR_INIT(port)); | ||
| 1452 | |||
| 940 | if (t > 0) { | 1453 | if (t > 0) { |
| 941 | if (t >= 256) { | 1454 | if (t >= 256) { |
| 942 | sci_out(port, SCSMR, (sci_in(port, SCSMR) & ~3) | 1); | 1455 | sci_out(port, SCSMR, (sci_in(port, SCSMR) & ~3) | 1); |
| @@ -954,7 +1467,7 @@ static void sci_set_termios(struct uart_port *port, struct ktermios *termios, | |||
| 954 | sci_out(port, SCSCR, SCSCR_INIT(port)); | 1467 | sci_out(port, SCSCR, SCSCR_INIT(port)); |
| 955 | 1468 | ||
| 956 | if ((termios->c_cflag & CREAD) != 0) | 1469 | if ((termios->c_cflag & CREAD) != 0) |
| 957 | sci_start_rx(port, 0); | 1470 | sci_start_rx(port); |
| 958 | } | 1471 | } |
| 959 | 1472 | ||
| 960 | static const char *sci_type(struct uart_port *port) | 1473 | static const char *sci_type(struct uart_port *port) |
| @@ -1049,19 +1562,21 @@ static void __devinit sci_init_single(struct platform_device *dev, | |||
| 1049 | unsigned int index, | 1562 | unsigned int index, |
| 1050 | struct plat_sci_port *p) | 1563 | struct plat_sci_port *p) |
| 1051 | { | 1564 | { |
| 1052 | sci_port->port.ops = &sci_uart_ops; | 1565 | struct uart_port *port = &sci_port->port; |
| 1053 | sci_port->port.iotype = UPIO_MEM; | 1566 | |
| 1054 | sci_port->port.line = index; | 1567 | port->ops = &sci_uart_ops; |
| 1568 | port->iotype = UPIO_MEM; | ||
| 1569 | port->line = index; | ||
| 1055 | 1570 | ||
| 1056 | switch (p->type) { | 1571 | switch (p->type) { |
| 1057 | case PORT_SCIFA: | 1572 | case PORT_SCIFA: |
| 1058 | sci_port->port.fifosize = 64; | 1573 | port->fifosize = 64; |
| 1059 | break; | 1574 | break; |
| 1060 | case PORT_SCIF: | 1575 | case PORT_SCIF: |
| 1061 | sci_port->port.fifosize = 16; | 1576 | port->fifosize = 16; |
| 1062 | break; | 1577 | break; |
| 1063 | default: | 1578 | default: |
| 1064 | sci_port->port.fifosize = 1; | 1579 | port->fifosize = 1; |
| 1065 | break; | 1580 | break; |
| 1066 | } | 1581 | } |
| 1067 | 1582 | ||
| @@ -1070,19 +1585,28 @@ static void __devinit sci_init_single(struct platform_device *dev, | |||
| 1070 | sci_port->dclk = clk_get(&dev->dev, "peripheral_clk"); | 1585 | sci_port->dclk = clk_get(&dev->dev, "peripheral_clk"); |
| 1071 | sci_port->enable = sci_clk_enable; | 1586 | sci_port->enable = sci_clk_enable; |
| 1072 | sci_port->disable = sci_clk_disable; | 1587 | sci_port->disable = sci_clk_disable; |
| 1073 | sci_port->port.dev = &dev->dev; | 1588 | port->dev = &dev->dev; |
| 1074 | } | 1589 | } |
| 1075 | 1590 | ||
| 1076 | sci_port->break_timer.data = (unsigned long)sci_port; | 1591 | sci_port->break_timer.data = (unsigned long)sci_port; |
| 1077 | sci_port->break_timer.function = sci_break_timer; | 1592 | sci_port->break_timer.function = sci_break_timer; |
| 1078 | init_timer(&sci_port->break_timer); | 1593 | init_timer(&sci_port->break_timer); |
| 1079 | 1594 | ||
| 1080 | sci_port->port.mapbase = p->mapbase; | 1595 | port->mapbase = p->mapbase; |
| 1081 | sci_port->port.membase = p->membase; | 1596 | port->membase = p->membase; |
| 1082 | 1597 | ||
| 1083 | sci_port->port.irq = p->irqs[SCIx_TXI_IRQ]; | 1598 | port->irq = p->irqs[SCIx_TXI_IRQ]; |
| 1084 | sci_port->port.flags = p->flags; | 1599 | port->flags = p->flags; |
| 1085 | sci_port->type = sci_port->port.type = p->type; | 1600 | sci_port->type = port->type = p->type; |
| 1601 | |||
| 1602 | #ifdef CONFIG_SERIAL_SH_SCI_DMA | ||
| 1603 | sci_port->dma_dev = p->dma_dev; | ||
| 1604 | sci_port->slave_tx = p->dma_slave_tx; | ||
| 1605 | sci_port->slave_rx = p->dma_slave_rx; | ||
| 1606 | |||
| 1607 | dev_dbg(port->dev, "%s: DMA device %p, tx %d, rx %d\n", __func__, | ||
| 1608 | p->dma_dev, p->dma_slave_tx, p->dma_slave_rx); | ||
| 1609 | #endif | ||
| 1086 | 1610 | ||
| 1087 | memcpy(&sci_port->irqs, &p->irqs, sizeof(p->irqs)); | 1611 | memcpy(&sci_port->irqs, &p->irqs, sizeof(p->irqs)); |
| 1088 | } | 1612 | } |
diff --git a/include/linux/serial_sci.h b/include/linux/serial_sci.h index 1c297ddc9d5a..1b177d29a7f0 100644 --- a/include/linux/serial_sci.h +++ b/include/linux/serial_sci.h | |||
| @@ -2,6 +2,7 @@ | |||
| 2 | #define __LINUX_SERIAL_SCI_H | 2 | #define __LINUX_SERIAL_SCI_H |
| 3 | 3 | ||
| 4 | #include <linux/serial_core.h> | 4 | #include <linux/serial_core.h> |
| 5 | #include <asm/dmaengine.h> | ||
| 5 | 6 | ||
| 6 | /* | 7 | /* |
| 7 | * Generic header for SuperH SCI(F) (used by sh/sh64/h8300 and related parts) | 8 | * Generic header for SuperH SCI(F) (used by sh/sh64/h8300 and related parts) |
| @@ -16,6 +17,8 @@ enum { | |||
| 16 | SCIx_NR_IRQS, | 17 | SCIx_NR_IRQS, |
| 17 | }; | 18 | }; |
| 18 | 19 | ||
| 20 | struct device; | ||
| 21 | |||
| 19 | /* | 22 | /* |
| 20 | * Platform device specific platform_data struct | 23 | * Platform device specific platform_data struct |
| 21 | */ | 24 | */ |
| @@ -26,6 +29,9 @@ struct plat_sci_port { | |||
| 26 | unsigned int type; /* SCI / SCIF / IRDA */ | 29 | unsigned int type; /* SCI / SCIF / IRDA */ |
| 27 | upf_t flags; /* UPF_* flags */ | 30 | upf_t flags; /* UPF_* flags */ |
| 28 | char *clk; /* clock string */ | 31 | char *clk; /* clock string */ |
| 32 | struct device *dma_dev; | ||
| 33 | enum sh_dmae_slave_chan_id dma_slave_tx; | ||
| 34 | enum sh_dmae_slave_chan_id dma_slave_rx; | ||
| 29 | }; | 35 | }; |
| 30 | 36 | ||
| 31 | #endif /* __LINUX_SERIAL_SCI_H */ | 37 | #endif /* __LINUX_SERIAL_SCI_H */ |
diff --git a/sound/soc/sh/siu.h b/sound/soc/sh/siu.h index 9cc04ab2bce7..c0bfab8fed3d 100644 --- a/sound/soc/sh/siu.h +++ b/sound/soc/sh/siu.h | |||
| @@ -72,7 +72,7 @@ struct siu_firmware { | |||
| 72 | #include <linux/interrupt.h> | 72 | #include <linux/interrupt.h> |
| 73 | #include <linux/io.h> | 73 | #include <linux/io.h> |
| 74 | 74 | ||
| 75 | #include <asm/dma-sh.h> | 75 | #include <asm/dmaengine.h> |
| 76 | 76 | ||
| 77 | #include <sound/core.h> | 77 | #include <sound/core.h> |
| 78 | #include <sound/pcm.h> | 78 | #include <sound/pcm.h> |
diff --git a/sound/soc/sh/siu_pcm.c b/sound/soc/sh/siu_pcm.c index c5efc30f0136..ba7f8d05d977 100644 --- a/sound/soc/sh/siu_pcm.c +++ b/sound/soc/sh/siu_pcm.c | |||
| @@ -32,7 +32,7 @@ | |||
| 32 | #include <sound/pcm_params.h> | 32 | #include <sound/pcm_params.h> |
| 33 | #include <sound/soc-dai.h> | 33 | #include <sound/soc-dai.h> |
| 34 | 34 | ||
| 35 | #include <asm/dma-sh.h> | 35 | #include <asm/dmaengine.h> |
| 36 | #include <asm/siu.h> | 36 | #include <asm/siu.h> |
| 37 | 37 | ||
| 38 | #include "siu.h" | 38 | #include "siu.h" |
