diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 18:34:57 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 18:34:57 -0400 |
commit | ef08e78268423fc4d7fbc3e54bd9a67fc8da7cc5 (patch) | |
tree | d0561d3ef89c9cd277a38168e33850666cbd33c4 /drivers/dma | |
parent | 71db34fc4330f7c784397acb9f1e6ee7f7b32eb2 (diff) | |
parent | 5b2e02e401deb44e7f5befe19404d8b2688efea4 (diff) |
Merge branch 'next' of git://git.infradead.org/users/vkoul/slave-dma
Pull slave-dmaengine update from Vinod Koul:
"This includes the cookie cleanup by Russell, the addition of context
parameter for dmaengine APIs, more arm dmaengine driver cleanup by
moving code to dmaengine, this time for imx by Javier and pl330 by
Boojin along with the usual driver fixes."
Fix up some fairly trivial conflicts with various other cleanups.
* 'next' of git://git.infradead.org/users/vkoul/slave-dma: (67 commits)
dmaengine: imx: fix the build failure on x86_64
dmaengine: i.MX: Fix merge of cookie branch.
dmaengine: i.MX: Add support for interleaved transfers.
dmaengine: imx-dma: use 'dev_dbg' and 'dev_warn' for messages.
dmaengine: imx-dma: remove 'imx_dmav1_baseaddr' and 'dma_clk'.
dmaengine: imx-dma: remove unused arg of imxdma_sg_next.
dmaengine: imx-dma: remove internal structure.
dmaengine: imx-dma: remove 'resbytes' field of 'internal' structure.
dmaengine: imx-dma: remove 'in_use' field of 'internal' structure.
dmaengine: imx-dma: remove sg member from internal structure.
dmaengine: imx-dma: remove 'imxdma_setup_sg_hw' function.
dmaengine: imx-dma: remove 'imxdma_config_channel_hw' function.
dmaengine: imx-dma: remove 'imxdma_setup_mem2mem_hw' function.
dmaengine: imx-dma: remove dma_mode member of internal structure.
dmaengine: imx-dma: remove data member from internal structure.
dmaengine: imx-dma: merge old dma-v1.c with imx-dma.c
dmaengine: at_hdmac: add slave config operation
dmaengine: add context parameter to prep_slave_sg and prep_dma_cyclic
dmaengine/dma_slave: introduce inline wrappers
dma: imx-sdma: Treat firmware messages as warnings instead of erros
...
Diffstat (limited to 'drivers/dma')
37 files changed, 3547 insertions, 930 deletions
diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig index 4a6c46dea8a0..cf9da362d64f 100644 --- a/drivers/dma/Kconfig +++ b/drivers/dma/Kconfig | |||
@@ -201,7 +201,6 @@ config PL330_DMA | |||
201 | tristate "DMA API Driver for PL330" | 201 | tristate "DMA API Driver for PL330" |
202 | select DMA_ENGINE | 202 | select DMA_ENGINE |
203 | depends on ARM_AMBA | 203 | depends on ARM_AMBA |
204 | select PL330 | ||
205 | help | 204 | help |
206 | Select if your platform has one or more PL330 DMACs. | 205 | Select if your platform has one or more PL330 DMACs. |
207 | You need to provide platform specific settings via | 206 | You need to provide platform specific settings via |
@@ -231,7 +230,7 @@ config IMX_SDMA | |||
231 | 230 | ||
232 | config IMX_DMA | 231 | config IMX_DMA |
233 | tristate "i.MX DMA support" | 232 | tristate "i.MX DMA support" |
234 | depends on IMX_HAVE_DMA_V1 | 233 | depends on ARCH_MXC |
235 | select DMA_ENGINE | 234 | select DMA_ENGINE |
236 | help | 235 | help |
237 | Support the i.MX DMA engine. This engine is integrated into | 236 | Support the i.MX DMA engine. This engine is integrated into |
diff --git a/drivers/dma/amba-pl08x.c b/drivers/dma/amba-pl08x.c index 8a281584458b..c301a8ec31aa 100644 --- a/drivers/dma/amba-pl08x.c +++ b/drivers/dma/amba-pl08x.c | |||
@@ -85,6 +85,8 @@ | |||
85 | #include <linux/slab.h> | 85 | #include <linux/slab.h> |
86 | #include <asm/hardware/pl080.h> | 86 | #include <asm/hardware/pl080.h> |
87 | 87 | ||
88 | #include "dmaengine.h" | ||
89 | |||
88 | #define DRIVER_NAME "pl08xdmac" | 90 | #define DRIVER_NAME "pl08xdmac" |
89 | 91 | ||
90 | static struct amba_driver pl08x_amba_driver; | 92 | static struct amba_driver pl08x_amba_driver; |
@@ -649,7 +651,7 @@ static int pl08x_fill_llis_for_desc(struct pl08x_driver_data *pl08x, | |||
649 | } | 651 | } |
650 | 652 | ||
651 | if ((bd.srcbus.addr % bd.srcbus.buswidth) || | 653 | if ((bd.srcbus.addr % bd.srcbus.buswidth) || |
652 | (bd.srcbus.addr % bd.srcbus.buswidth)) { | 654 | (bd.dstbus.addr % bd.dstbus.buswidth)) { |
653 | dev_err(&pl08x->adev->dev, | 655 | dev_err(&pl08x->adev->dev, |
654 | "%s src & dst address must be aligned to src" | 656 | "%s src & dst address must be aligned to src" |
655 | " & dst width if peripheral is flow controller", | 657 | " & dst width if peripheral is flow controller", |
@@ -919,13 +921,10 @@ static dma_cookie_t pl08x_tx_submit(struct dma_async_tx_descriptor *tx) | |||
919 | struct pl08x_dma_chan *plchan = to_pl08x_chan(tx->chan); | 921 | struct pl08x_dma_chan *plchan = to_pl08x_chan(tx->chan); |
920 | struct pl08x_txd *txd = to_pl08x_txd(tx); | 922 | struct pl08x_txd *txd = to_pl08x_txd(tx); |
921 | unsigned long flags; | 923 | unsigned long flags; |
924 | dma_cookie_t cookie; | ||
922 | 925 | ||
923 | spin_lock_irqsave(&plchan->lock, flags); | 926 | spin_lock_irqsave(&plchan->lock, flags); |
924 | 927 | cookie = dma_cookie_assign(tx); | |
925 | plchan->chan.cookie += 1; | ||
926 | if (plchan->chan.cookie < 0) | ||
927 | plchan->chan.cookie = 1; | ||
928 | tx->cookie = plchan->chan.cookie; | ||
929 | 928 | ||
930 | /* Put this onto the pending list */ | 929 | /* Put this onto the pending list */ |
931 | list_add_tail(&txd->node, &plchan->pend_list); | 930 | list_add_tail(&txd->node, &plchan->pend_list); |
@@ -945,7 +944,7 @@ static dma_cookie_t pl08x_tx_submit(struct dma_async_tx_descriptor *tx) | |||
945 | 944 | ||
946 | spin_unlock_irqrestore(&plchan->lock, flags); | 945 | spin_unlock_irqrestore(&plchan->lock, flags); |
947 | 946 | ||
948 | return tx->cookie; | 947 | return cookie; |
949 | } | 948 | } |
950 | 949 | ||
951 | static struct dma_async_tx_descriptor *pl08x_prep_dma_interrupt( | 950 | static struct dma_async_tx_descriptor *pl08x_prep_dma_interrupt( |
@@ -965,31 +964,17 @@ static enum dma_status pl08x_dma_tx_status(struct dma_chan *chan, | |||
965 | dma_cookie_t cookie, struct dma_tx_state *txstate) | 964 | dma_cookie_t cookie, struct dma_tx_state *txstate) |
966 | { | 965 | { |
967 | struct pl08x_dma_chan *plchan = to_pl08x_chan(chan); | 966 | struct pl08x_dma_chan *plchan = to_pl08x_chan(chan); |
968 | dma_cookie_t last_used; | ||
969 | dma_cookie_t last_complete; | ||
970 | enum dma_status ret; | 967 | enum dma_status ret; |
971 | u32 bytesleft = 0; | ||
972 | 968 | ||
973 | last_used = plchan->chan.cookie; | 969 | ret = dma_cookie_status(chan, cookie, txstate); |
974 | last_complete = plchan->lc; | 970 | if (ret == DMA_SUCCESS) |
975 | |||
976 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
977 | if (ret == DMA_SUCCESS) { | ||
978 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
979 | return ret; | 971 | return ret; |
980 | } | ||
981 | 972 | ||
982 | /* | 973 | /* |
983 | * This cookie not complete yet | 974 | * This cookie not complete yet |
975 | * Get number of bytes left in the active transactions and queue | ||
984 | */ | 976 | */ |
985 | last_used = plchan->chan.cookie; | 977 | dma_set_residue(txstate, pl08x_getbytes_chan(plchan)); |
986 | last_complete = plchan->lc; | ||
987 | |||
988 | /* Get number of bytes left in the active transactions and queue */ | ||
989 | bytesleft = pl08x_getbytes_chan(plchan); | ||
990 | |||
991 | dma_set_tx_state(txstate, last_complete, last_used, | ||
992 | bytesleft); | ||
993 | 978 | ||
994 | if (plchan->state == PL08X_CHAN_PAUSED) | 979 | if (plchan->state == PL08X_CHAN_PAUSED) |
995 | return DMA_PAUSED; | 980 | return DMA_PAUSED; |
@@ -1139,6 +1124,8 @@ static int dma_set_runtime_config(struct dma_chan *chan, | |||
1139 | cctl |= burst << PL080_CONTROL_SB_SIZE_SHIFT; | 1124 | cctl |= burst << PL080_CONTROL_SB_SIZE_SHIFT; |
1140 | cctl |= burst << PL080_CONTROL_DB_SIZE_SHIFT; | 1125 | cctl |= burst << PL080_CONTROL_DB_SIZE_SHIFT; |
1141 | 1126 | ||
1127 | plchan->device_fc = config->device_fc; | ||
1128 | |||
1142 | if (plchan->runtime_direction == DMA_DEV_TO_MEM) { | 1129 | if (plchan->runtime_direction == DMA_DEV_TO_MEM) { |
1143 | plchan->src_addr = config->src_addr; | 1130 | plchan->src_addr = config->src_addr; |
1144 | plchan->src_cctl = pl08x_cctl(cctl) | PL080_CONTROL_DST_INCR | | 1131 | plchan->src_cctl = pl08x_cctl(cctl) | PL080_CONTROL_DST_INCR | |
@@ -1326,7 +1313,7 @@ static struct dma_async_tx_descriptor *pl08x_prep_dma_memcpy( | |||
1326 | static struct dma_async_tx_descriptor *pl08x_prep_slave_sg( | 1313 | static struct dma_async_tx_descriptor *pl08x_prep_slave_sg( |
1327 | struct dma_chan *chan, struct scatterlist *sgl, | 1314 | struct dma_chan *chan, struct scatterlist *sgl, |
1328 | unsigned int sg_len, enum dma_transfer_direction direction, | 1315 | unsigned int sg_len, enum dma_transfer_direction direction, |
1329 | unsigned long flags) | 1316 | unsigned long flags, void *context) |
1330 | { | 1317 | { |
1331 | struct pl08x_dma_chan *plchan = to_pl08x_chan(chan); | 1318 | struct pl08x_dma_chan *plchan = to_pl08x_chan(chan); |
1332 | struct pl08x_driver_data *pl08x = plchan->host; | 1319 | struct pl08x_driver_data *pl08x = plchan->host; |
@@ -1370,7 +1357,7 @@ static struct dma_async_tx_descriptor *pl08x_prep_slave_sg( | |||
1370 | return NULL; | 1357 | return NULL; |
1371 | } | 1358 | } |
1372 | 1359 | ||
1373 | if (plchan->cd->device_fc) | 1360 | if (plchan->device_fc) |
1374 | tmp = (direction == DMA_MEM_TO_DEV) ? PL080_FLOW_MEM2PER_PER : | 1361 | tmp = (direction == DMA_MEM_TO_DEV) ? PL080_FLOW_MEM2PER_PER : |
1375 | PL080_FLOW_PER2MEM_PER; | 1362 | PL080_FLOW_PER2MEM_PER; |
1376 | else | 1363 | else |
@@ -1541,7 +1528,7 @@ static void pl08x_tasklet(unsigned long data) | |||
1541 | 1528 | ||
1542 | if (txd) { | 1529 | if (txd) { |
1543 | /* Update last completed */ | 1530 | /* Update last completed */ |
1544 | plchan->lc = txd->tx.cookie; | 1531 | dma_cookie_complete(&txd->tx); |
1545 | } | 1532 | } |
1546 | 1533 | ||
1547 | /* If a new descriptor is queued, set it up plchan->at is NULL here */ | 1534 | /* If a new descriptor is queued, set it up plchan->at is NULL here */ |
@@ -1722,8 +1709,7 @@ static int pl08x_dma_init_virtual_channels(struct pl08x_driver_data *pl08x, | |||
1722 | chan->name); | 1709 | chan->name); |
1723 | 1710 | ||
1724 | chan->chan.device = dmadev; | 1711 | chan->chan.device = dmadev; |
1725 | chan->chan.cookie = 0; | 1712 | dma_cookie_init(&chan->chan); |
1726 | chan->lc = 0; | ||
1727 | 1713 | ||
1728 | spin_lock_init(&chan->lock); | 1714 | spin_lock_init(&chan->lock); |
1729 | INIT_LIST_HEAD(&chan->pend_list); | 1715 | INIT_LIST_HEAD(&chan->pend_list); |
diff --git a/drivers/dma/at_hdmac.c b/drivers/dma/at_hdmac.c index f4aed5fc2cb6..7aa58d204892 100644 --- a/drivers/dma/at_hdmac.c +++ b/drivers/dma/at_hdmac.c | |||
@@ -27,6 +27,7 @@ | |||
27 | #include <linux/of_device.h> | 27 | #include <linux/of_device.h> |
28 | 28 | ||
29 | #include "at_hdmac_regs.h" | 29 | #include "at_hdmac_regs.h" |
30 | #include "dmaengine.h" | ||
30 | 31 | ||
31 | /* | 32 | /* |
32 | * Glossary | 33 | * Glossary |
@@ -192,27 +193,6 @@ static void atc_desc_chain(struct at_desc **first, struct at_desc **prev, | |||
192 | } | 193 | } |
193 | 194 | ||
194 | /** | 195 | /** |
195 | * atc_assign_cookie - compute and assign new cookie | ||
196 | * @atchan: channel we work on | ||
197 | * @desc: descriptor to assign cookie for | ||
198 | * | ||
199 | * Called with atchan->lock held and bh disabled | ||
200 | */ | ||
201 | static dma_cookie_t | ||
202 | atc_assign_cookie(struct at_dma_chan *atchan, struct at_desc *desc) | ||
203 | { | ||
204 | dma_cookie_t cookie = atchan->chan_common.cookie; | ||
205 | |||
206 | if (++cookie < 0) | ||
207 | cookie = 1; | ||
208 | |||
209 | atchan->chan_common.cookie = cookie; | ||
210 | desc->txd.cookie = cookie; | ||
211 | |||
212 | return cookie; | ||
213 | } | ||
214 | |||
215 | /** | ||
216 | * atc_dostart - starts the DMA engine for real | 196 | * atc_dostart - starts the DMA engine for real |
217 | * @atchan: the channel we want to start | 197 | * @atchan: the channel we want to start |
218 | * @first: first descriptor in the list we want to begin with | 198 | * @first: first descriptor in the list we want to begin with |
@@ -269,7 +249,7 @@ atc_chain_complete(struct at_dma_chan *atchan, struct at_desc *desc) | |||
269 | dev_vdbg(chan2dev(&atchan->chan_common), | 249 | dev_vdbg(chan2dev(&atchan->chan_common), |
270 | "descriptor %u complete\n", txd->cookie); | 250 | "descriptor %u complete\n", txd->cookie); |
271 | 251 | ||
272 | atchan->completed_cookie = txd->cookie; | 252 | dma_cookie_complete(txd); |
273 | 253 | ||
274 | /* move children to free_list */ | 254 | /* move children to free_list */ |
275 | list_splice_init(&desc->tx_list, &atchan->free_list); | 255 | list_splice_init(&desc->tx_list, &atchan->free_list); |
@@ -547,7 +527,7 @@ static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx) | |||
547 | unsigned long flags; | 527 | unsigned long flags; |
548 | 528 | ||
549 | spin_lock_irqsave(&atchan->lock, flags); | 529 | spin_lock_irqsave(&atchan->lock, flags); |
550 | cookie = atc_assign_cookie(atchan, desc); | 530 | cookie = dma_cookie_assign(tx); |
551 | 531 | ||
552 | if (list_empty(&atchan->active_list)) { | 532 | if (list_empty(&atchan->active_list)) { |
553 | dev_vdbg(chan2dev(tx->chan), "tx_submit: started %u\n", | 533 | dev_vdbg(chan2dev(tx->chan), "tx_submit: started %u\n", |
@@ -659,14 +639,16 @@ err_desc_get: | |||
659 | * @sg_len: number of entries in @scatterlist | 639 | * @sg_len: number of entries in @scatterlist |
660 | * @direction: DMA direction | 640 | * @direction: DMA direction |
661 | * @flags: tx descriptor status flags | 641 | * @flags: tx descriptor status flags |
642 | * @context: transaction context (ignored) | ||
662 | */ | 643 | */ |
663 | static struct dma_async_tx_descriptor * | 644 | static struct dma_async_tx_descriptor * |
664 | atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 645 | atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
665 | unsigned int sg_len, enum dma_transfer_direction direction, | 646 | unsigned int sg_len, enum dma_transfer_direction direction, |
666 | unsigned long flags) | 647 | unsigned long flags, void *context) |
667 | { | 648 | { |
668 | struct at_dma_chan *atchan = to_at_dma_chan(chan); | 649 | struct at_dma_chan *atchan = to_at_dma_chan(chan); |
669 | struct at_dma_slave *atslave = chan->private; | 650 | struct at_dma_slave *atslave = chan->private; |
651 | struct dma_slave_config *sconfig = &atchan->dma_sconfig; | ||
670 | struct at_desc *first = NULL; | 652 | struct at_desc *first = NULL; |
671 | struct at_desc *prev = NULL; | 653 | struct at_desc *prev = NULL; |
672 | u32 ctrla; | 654 | u32 ctrla; |
@@ -688,19 +670,18 @@ atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | |||
688 | return NULL; | 670 | return NULL; |
689 | } | 671 | } |
690 | 672 | ||
691 | reg_width = atslave->reg_width; | ||
692 | |||
693 | ctrla = ATC_DEFAULT_CTRLA | atslave->ctrla; | 673 | ctrla = ATC_DEFAULT_CTRLA | atslave->ctrla; |
694 | ctrlb = ATC_IEN; | 674 | ctrlb = ATC_IEN; |
695 | 675 | ||
696 | switch (direction) { | 676 | switch (direction) { |
697 | case DMA_MEM_TO_DEV: | 677 | case DMA_MEM_TO_DEV: |
678 | reg_width = convert_buswidth(sconfig->dst_addr_width); | ||
698 | ctrla |= ATC_DST_WIDTH(reg_width); | 679 | ctrla |= ATC_DST_WIDTH(reg_width); |
699 | ctrlb |= ATC_DST_ADDR_MODE_FIXED | 680 | ctrlb |= ATC_DST_ADDR_MODE_FIXED |
700 | | ATC_SRC_ADDR_MODE_INCR | 681 | | ATC_SRC_ADDR_MODE_INCR |
701 | | ATC_FC_MEM2PER | 682 | | ATC_FC_MEM2PER |
702 | | ATC_SIF(AT_DMA_MEM_IF) | ATC_DIF(AT_DMA_PER_IF); | 683 | | ATC_SIF(AT_DMA_MEM_IF) | ATC_DIF(AT_DMA_PER_IF); |
703 | reg = atslave->tx_reg; | 684 | reg = sconfig->dst_addr; |
704 | for_each_sg(sgl, sg, sg_len, i) { | 685 | for_each_sg(sgl, sg, sg_len, i) { |
705 | struct at_desc *desc; | 686 | struct at_desc *desc; |
706 | u32 len; | 687 | u32 len; |
@@ -728,13 +709,14 @@ atc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | |||
728 | } | 709 | } |
729 | break; | 710 | break; |
730 | case DMA_DEV_TO_MEM: | 711 | case DMA_DEV_TO_MEM: |
712 | reg_width = convert_buswidth(sconfig->src_addr_width); | ||
731 | ctrla |= ATC_SRC_WIDTH(reg_width); | 713 | ctrla |= ATC_SRC_WIDTH(reg_width); |
732 | ctrlb |= ATC_DST_ADDR_MODE_INCR | 714 | ctrlb |= ATC_DST_ADDR_MODE_INCR |
733 | | ATC_SRC_ADDR_MODE_FIXED | 715 | | ATC_SRC_ADDR_MODE_FIXED |
734 | | ATC_FC_PER2MEM | 716 | | ATC_FC_PER2MEM |
735 | | ATC_SIF(AT_DMA_PER_IF) | ATC_DIF(AT_DMA_MEM_IF); | 717 | | ATC_SIF(AT_DMA_PER_IF) | ATC_DIF(AT_DMA_MEM_IF); |
736 | 718 | ||
737 | reg = atslave->rx_reg; | 719 | reg = sconfig->src_addr; |
738 | for_each_sg(sgl, sg, sg_len, i) { | 720 | for_each_sg(sgl, sg, sg_len, i) { |
739 | struct at_desc *desc; | 721 | struct at_desc *desc; |
740 | u32 len; | 722 | u32 len; |
@@ -810,12 +792,15 @@ err_out: | |||
810 | * atc_dma_cyclic_fill_desc - Fill one period decriptor | 792 | * atc_dma_cyclic_fill_desc - Fill one period decriptor |
811 | */ | 793 | */ |
812 | static int | 794 | static int |
813 | atc_dma_cyclic_fill_desc(struct at_dma_slave *atslave, struct at_desc *desc, | 795 | atc_dma_cyclic_fill_desc(struct dma_chan *chan, struct at_desc *desc, |
814 | unsigned int period_index, dma_addr_t buf_addr, | 796 | unsigned int period_index, dma_addr_t buf_addr, |
815 | size_t period_len, enum dma_transfer_direction direction) | 797 | unsigned int reg_width, size_t period_len, |
798 | enum dma_transfer_direction direction) | ||
816 | { | 799 | { |
817 | u32 ctrla; | 800 | struct at_dma_chan *atchan = to_at_dma_chan(chan); |
818 | unsigned int reg_width = atslave->reg_width; | 801 | struct at_dma_slave *atslave = chan->private; |
802 | struct dma_slave_config *sconfig = &atchan->dma_sconfig; | ||
803 | u32 ctrla; | ||
819 | 804 | ||
820 | /* prepare common CRTLA value */ | 805 | /* prepare common CRTLA value */ |
821 | ctrla = ATC_DEFAULT_CTRLA | atslave->ctrla | 806 | ctrla = ATC_DEFAULT_CTRLA | atslave->ctrla |
@@ -826,7 +811,7 @@ atc_dma_cyclic_fill_desc(struct at_dma_slave *atslave, struct at_desc *desc, | |||
826 | switch (direction) { | 811 | switch (direction) { |
827 | case DMA_MEM_TO_DEV: | 812 | case DMA_MEM_TO_DEV: |
828 | desc->lli.saddr = buf_addr + (period_len * period_index); | 813 | desc->lli.saddr = buf_addr + (period_len * period_index); |
829 | desc->lli.daddr = atslave->tx_reg; | 814 | desc->lli.daddr = sconfig->dst_addr; |
830 | desc->lli.ctrla = ctrla; | 815 | desc->lli.ctrla = ctrla; |
831 | desc->lli.ctrlb = ATC_DST_ADDR_MODE_FIXED | 816 | desc->lli.ctrlb = ATC_DST_ADDR_MODE_FIXED |
832 | | ATC_SRC_ADDR_MODE_INCR | 817 | | ATC_SRC_ADDR_MODE_INCR |
@@ -836,7 +821,7 @@ atc_dma_cyclic_fill_desc(struct at_dma_slave *atslave, struct at_desc *desc, | |||
836 | break; | 821 | break; |
837 | 822 | ||
838 | case DMA_DEV_TO_MEM: | 823 | case DMA_DEV_TO_MEM: |
839 | desc->lli.saddr = atslave->rx_reg; | 824 | desc->lli.saddr = sconfig->src_addr; |
840 | desc->lli.daddr = buf_addr + (period_len * period_index); | 825 | desc->lli.daddr = buf_addr + (period_len * period_index); |
841 | desc->lli.ctrla = ctrla; | 826 | desc->lli.ctrla = ctrla; |
842 | desc->lli.ctrlb = ATC_DST_ADDR_MODE_INCR | 827 | desc->lli.ctrlb = ATC_DST_ADDR_MODE_INCR |
@@ -860,16 +845,20 @@ atc_dma_cyclic_fill_desc(struct at_dma_slave *atslave, struct at_desc *desc, | |||
860 | * @buf_len: total number of bytes for the entire buffer | 845 | * @buf_len: total number of bytes for the entire buffer |
861 | * @period_len: number of bytes for each period | 846 | * @period_len: number of bytes for each period |
862 | * @direction: transfer direction, to or from device | 847 | * @direction: transfer direction, to or from device |
848 | * @context: transfer context (ignored) | ||
863 | */ | 849 | */ |
864 | static struct dma_async_tx_descriptor * | 850 | static struct dma_async_tx_descriptor * |
865 | atc_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t buf_addr, size_t buf_len, | 851 | atc_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t buf_addr, size_t buf_len, |
866 | size_t period_len, enum dma_transfer_direction direction) | 852 | size_t period_len, enum dma_transfer_direction direction, |
853 | void *context) | ||
867 | { | 854 | { |
868 | struct at_dma_chan *atchan = to_at_dma_chan(chan); | 855 | struct at_dma_chan *atchan = to_at_dma_chan(chan); |
869 | struct at_dma_slave *atslave = chan->private; | 856 | struct at_dma_slave *atslave = chan->private; |
857 | struct dma_slave_config *sconfig = &atchan->dma_sconfig; | ||
870 | struct at_desc *first = NULL; | 858 | struct at_desc *first = NULL; |
871 | struct at_desc *prev = NULL; | 859 | struct at_desc *prev = NULL; |
872 | unsigned long was_cyclic; | 860 | unsigned long was_cyclic; |
861 | unsigned int reg_width; | ||
873 | unsigned int periods = buf_len / period_len; | 862 | unsigned int periods = buf_len / period_len; |
874 | unsigned int i; | 863 | unsigned int i; |
875 | 864 | ||
@@ -889,8 +878,13 @@ atc_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t buf_addr, size_t buf_len, | |||
889 | return NULL; | 878 | return NULL; |
890 | } | 879 | } |
891 | 880 | ||
881 | if (sconfig->direction == DMA_MEM_TO_DEV) | ||
882 | reg_width = convert_buswidth(sconfig->dst_addr_width); | ||
883 | else | ||
884 | reg_width = convert_buswidth(sconfig->src_addr_width); | ||
885 | |||
892 | /* Check for too big/unaligned periods and unaligned DMA buffer */ | 886 | /* Check for too big/unaligned periods and unaligned DMA buffer */ |
893 | if (atc_dma_cyclic_check_values(atslave->reg_width, buf_addr, | 887 | if (atc_dma_cyclic_check_values(reg_width, buf_addr, |
894 | period_len, direction)) | 888 | period_len, direction)) |
895 | goto err_out; | 889 | goto err_out; |
896 | 890 | ||
@@ -902,8 +896,8 @@ atc_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t buf_addr, size_t buf_len, | |||
902 | if (!desc) | 896 | if (!desc) |
903 | goto err_desc_get; | 897 | goto err_desc_get; |
904 | 898 | ||
905 | if (atc_dma_cyclic_fill_desc(atslave, desc, i, buf_addr, | 899 | if (atc_dma_cyclic_fill_desc(chan, desc, i, buf_addr, |
906 | period_len, direction)) | 900 | reg_width, period_len, direction)) |
907 | goto err_desc_get; | 901 | goto err_desc_get; |
908 | 902 | ||
909 | atc_desc_chain(&first, &prev, desc); | 903 | atc_desc_chain(&first, &prev, desc); |
@@ -926,6 +920,23 @@ err_out: | |||
926 | return NULL; | 920 | return NULL; |
927 | } | 921 | } |
928 | 922 | ||
923 | static int set_runtime_config(struct dma_chan *chan, | ||
924 | struct dma_slave_config *sconfig) | ||
925 | { | ||
926 | struct at_dma_chan *atchan = to_at_dma_chan(chan); | ||
927 | |||
928 | /* Check if it is chan is configured for slave transfers */ | ||
929 | if (!chan->private) | ||
930 | return -EINVAL; | ||
931 | |||
932 | memcpy(&atchan->dma_sconfig, sconfig, sizeof(*sconfig)); | ||
933 | |||
934 | convert_burst(&atchan->dma_sconfig.src_maxburst); | ||
935 | convert_burst(&atchan->dma_sconfig.dst_maxburst); | ||
936 | |||
937 | return 0; | ||
938 | } | ||
939 | |||
929 | 940 | ||
930 | static int atc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | 941 | static int atc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, |
931 | unsigned long arg) | 942 | unsigned long arg) |
@@ -986,6 +997,8 @@ static int atc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | |||
986 | clear_bit(ATC_IS_CYCLIC, &atchan->status); | 997 | clear_bit(ATC_IS_CYCLIC, &atchan->status); |
987 | 998 | ||
988 | spin_unlock_irqrestore(&atchan->lock, flags); | 999 | spin_unlock_irqrestore(&atchan->lock, flags); |
1000 | } else if (cmd == DMA_SLAVE_CONFIG) { | ||
1001 | return set_runtime_config(chan, (struct dma_slave_config *)arg); | ||
989 | } else { | 1002 | } else { |
990 | return -ENXIO; | 1003 | return -ENXIO; |
991 | } | 1004 | } |
@@ -1016,26 +1029,20 @@ atc_tx_status(struct dma_chan *chan, | |||
1016 | 1029 | ||
1017 | spin_lock_irqsave(&atchan->lock, flags); | 1030 | spin_lock_irqsave(&atchan->lock, flags); |
1018 | 1031 | ||
1019 | last_complete = atchan->completed_cookie; | 1032 | ret = dma_cookie_status(chan, cookie, txstate); |
1020 | last_used = chan->cookie; | ||
1021 | |||
1022 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
1023 | if (ret != DMA_SUCCESS) { | 1033 | if (ret != DMA_SUCCESS) { |
1024 | atc_cleanup_descriptors(atchan); | 1034 | atc_cleanup_descriptors(atchan); |
1025 | 1035 | ||
1026 | last_complete = atchan->completed_cookie; | 1036 | ret = dma_cookie_status(chan, cookie, txstate); |
1027 | last_used = chan->cookie; | ||
1028 | |||
1029 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
1030 | } | 1037 | } |
1031 | 1038 | ||
1039 | last_complete = chan->completed_cookie; | ||
1040 | last_used = chan->cookie; | ||
1041 | |||
1032 | spin_unlock_irqrestore(&atchan->lock, flags); | 1042 | spin_unlock_irqrestore(&atchan->lock, flags); |
1033 | 1043 | ||
1034 | if (ret != DMA_SUCCESS) | 1044 | if (ret != DMA_SUCCESS) |
1035 | dma_set_tx_state(txstate, last_complete, last_used, | 1045 | dma_set_residue(txstate, atc_first_active(atchan)->len); |
1036 | atc_first_active(atchan)->len); | ||
1037 | else | ||
1038 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
1039 | 1046 | ||
1040 | if (atc_chan_is_paused(atchan)) | 1047 | if (atc_chan_is_paused(atchan)) |
1041 | ret = DMA_PAUSED; | 1048 | ret = DMA_PAUSED; |
@@ -1129,7 +1136,7 @@ static int atc_alloc_chan_resources(struct dma_chan *chan) | |||
1129 | spin_lock_irqsave(&atchan->lock, flags); | 1136 | spin_lock_irqsave(&atchan->lock, flags); |
1130 | atchan->descs_allocated = i; | 1137 | atchan->descs_allocated = i; |
1131 | list_splice(&tmp_list, &atchan->free_list); | 1138 | list_splice(&tmp_list, &atchan->free_list); |
1132 | atchan->completed_cookie = chan->cookie = 1; | 1139 | dma_cookie_init(chan); |
1133 | spin_unlock_irqrestore(&atchan->lock, flags); | 1140 | spin_unlock_irqrestore(&atchan->lock, flags); |
1134 | 1141 | ||
1135 | /* channel parameters */ | 1142 | /* channel parameters */ |
@@ -1329,7 +1336,7 @@ static int __init at_dma_probe(struct platform_device *pdev) | |||
1329 | struct at_dma_chan *atchan = &atdma->chan[i]; | 1336 | struct at_dma_chan *atchan = &atdma->chan[i]; |
1330 | 1337 | ||
1331 | atchan->chan_common.device = &atdma->dma_common; | 1338 | atchan->chan_common.device = &atdma->dma_common; |
1332 | atchan->chan_common.cookie = atchan->completed_cookie = 1; | 1339 | dma_cookie_init(&atchan->chan_common); |
1333 | list_add_tail(&atchan->chan_common.device_node, | 1340 | list_add_tail(&atchan->chan_common.device_node, |
1334 | &atdma->dma_common.channels); | 1341 | &atdma->dma_common.channels); |
1335 | 1342 | ||
diff --git a/drivers/dma/at_hdmac_regs.h b/drivers/dma/at_hdmac_regs.h index a8d3277d60b5..897a8bcaec90 100644 --- a/drivers/dma/at_hdmac_regs.h +++ b/drivers/dma/at_hdmac_regs.h | |||
@@ -207,8 +207,8 @@ enum atc_status { | |||
207 | * @save_cfg: configuration register that is saved on suspend/resume cycle | 207 | * @save_cfg: configuration register that is saved on suspend/resume cycle |
208 | * @save_dscr: for cyclic operations, preserve next descriptor address in | 208 | * @save_dscr: for cyclic operations, preserve next descriptor address in |
209 | * the cyclic list on suspend/resume cycle | 209 | * the cyclic list on suspend/resume cycle |
210 | * @dma_sconfig: configuration for slave transfers, passed via DMA_SLAVE_CONFIG | ||
210 | * @lock: serializes enqueue/dequeue operations to descriptors lists | 211 | * @lock: serializes enqueue/dequeue operations to descriptors lists |
211 | * @completed_cookie: identifier for the most recently completed operation | ||
212 | * @active_list: list of descriptors dmaengine is being running on | 212 | * @active_list: list of descriptors dmaengine is being running on |
213 | * @queue: list of descriptors ready to be submitted to engine | 213 | * @queue: list of descriptors ready to be submitted to engine |
214 | * @free_list: list of descriptors usable by the channel | 214 | * @free_list: list of descriptors usable by the channel |
@@ -223,11 +223,11 @@ struct at_dma_chan { | |||
223 | struct tasklet_struct tasklet; | 223 | struct tasklet_struct tasklet; |
224 | u32 save_cfg; | 224 | u32 save_cfg; |
225 | u32 save_dscr; | 225 | u32 save_dscr; |
226 | struct dma_slave_config dma_sconfig; | ||
226 | 227 | ||
227 | spinlock_t lock; | 228 | spinlock_t lock; |
228 | 229 | ||
229 | /* these other elements are all protected by lock */ | 230 | /* these other elements are all protected by lock */ |
230 | dma_cookie_t completed_cookie; | ||
231 | struct list_head active_list; | 231 | struct list_head active_list; |
232 | struct list_head queue; | 232 | struct list_head queue; |
233 | struct list_head free_list; | 233 | struct list_head free_list; |
@@ -245,6 +245,36 @@ static inline struct at_dma_chan *to_at_dma_chan(struct dma_chan *dchan) | |||
245 | return container_of(dchan, struct at_dma_chan, chan_common); | 245 | return container_of(dchan, struct at_dma_chan, chan_common); |
246 | } | 246 | } |
247 | 247 | ||
248 | /* | ||
249 | * Fix sconfig's burst size according to at_hdmac. We need to convert them as: | ||
250 | * 1 -> 0, 4 -> 1, 8 -> 2, 16 -> 3, 32 -> 4, 64 -> 5, 128 -> 6, 256 -> 7. | ||
251 | * | ||
252 | * This can be done by finding most significant bit set. | ||
253 | */ | ||
254 | static inline void convert_burst(u32 *maxburst) | ||
255 | { | ||
256 | if (*maxburst > 1) | ||
257 | *maxburst = fls(*maxburst) - 2; | ||
258 | else | ||
259 | *maxburst = 0; | ||
260 | } | ||
261 | |||
262 | /* | ||
263 | * Fix sconfig's bus width according to at_hdmac. | ||
264 | * 1 byte -> 0, 2 bytes -> 1, 4 bytes -> 2. | ||
265 | */ | ||
266 | static inline u8 convert_buswidth(enum dma_slave_buswidth addr_width) | ||
267 | { | ||
268 | switch (addr_width) { | ||
269 | case DMA_SLAVE_BUSWIDTH_2_BYTES: | ||
270 | return 1; | ||
271 | case DMA_SLAVE_BUSWIDTH_4_BYTES: | ||
272 | return 2; | ||
273 | default: | ||
274 | /* For 1 byte width or fallback */ | ||
275 | return 0; | ||
276 | } | ||
277 | } | ||
248 | 278 | ||
249 | /*-- Controller ------------------------------------------------------*/ | 279 | /*-- Controller ------------------------------------------------------*/ |
250 | 280 | ||
diff --git a/drivers/dma/coh901318.c b/drivers/dma/coh901318.c index d65a718c0f9b..dc89455f5550 100644 --- a/drivers/dma/coh901318.c +++ b/drivers/dma/coh901318.c | |||
@@ -24,6 +24,7 @@ | |||
24 | #include <mach/coh901318.h> | 24 | #include <mach/coh901318.h> |
25 | 25 | ||
26 | #include "coh901318_lli.h" | 26 | #include "coh901318_lli.h" |
27 | #include "dmaengine.h" | ||
27 | 28 | ||
28 | #define COHC_2_DEV(cohc) (&cohc->chan.dev->device) | 29 | #define COHC_2_DEV(cohc) (&cohc->chan.dev->device) |
29 | 30 | ||
@@ -59,7 +60,6 @@ struct coh901318_base { | |||
59 | struct coh901318_chan { | 60 | struct coh901318_chan { |
60 | spinlock_t lock; | 61 | spinlock_t lock; |
61 | int allocated; | 62 | int allocated; |
62 | int completed; | ||
63 | int id; | 63 | int id; |
64 | int stopped; | 64 | int stopped; |
65 | 65 | ||
@@ -318,20 +318,6 @@ static int coh901318_prep_linked_list(struct coh901318_chan *cohc, | |||
318 | 318 | ||
319 | return 0; | 319 | return 0; |
320 | } | 320 | } |
321 | static dma_cookie_t | ||
322 | coh901318_assign_cookie(struct coh901318_chan *cohc, | ||
323 | struct coh901318_desc *cohd) | ||
324 | { | ||
325 | dma_cookie_t cookie = cohc->chan.cookie; | ||
326 | |||
327 | if (++cookie < 0) | ||
328 | cookie = 1; | ||
329 | |||
330 | cohc->chan.cookie = cookie; | ||
331 | cohd->desc.cookie = cookie; | ||
332 | |||
333 | return cookie; | ||
334 | } | ||
335 | 321 | ||
336 | static struct coh901318_desc * | 322 | static struct coh901318_desc * |
337 | coh901318_desc_get(struct coh901318_chan *cohc) | 323 | coh901318_desc_get(struct coh901318_chan *cohc) |
@@ -705,7 +691,7 @@ static void dma_tasklet(unsigned long data) | |||
705 | callback_param = cohd_fin->desc.callback_param; | 691 | callback_param = cohd_fin->desc.callback_param; |
706 | 692 | ||
707 | /* sign this job as completed on the channel */ | 693 | /* sign this job as completed on the channel */ |
708 | cohc->completed = cohd_fin->desc.cookie; | 694 | dma_cookie_complete(&cohd_fin->desc); |
709 | 695 | ||
710 | /* release the lli allocation and remove the descriptor */ | 696 | /* release the lli allocation and remove the descriptor */ |
711 | coh901318_lli_free(&cohc->base->pool, &cohd_fin->lli); | 697 | coh901318_lli_free(&cohc->base->pool, &cohd_fin->lli); |
@@ -929,7 +915,7 @@ static int coh901318_alloc_chan_resources(struct dma_chan *chan) | |||
929 | coh901318_config(cohc, NULL); | 915 | coh901318_config(cohc, NULL); |
930 | 916 | ||
931 | cohc->allocated = 1; | 917 | cohc->allocated = 1; |
932 | cohc->completed = chan->cookie = 1; | 918 | dma_cookie_init(chan); |
933 | 919 | ||
934 | spin_unlock_irqrestore(&cohc->lock, flags); | 920 | spin_unlock_irqrestore(&cohc->lock, flags); |
935 | 921 | ||
@@ -966,16 +952,16 @@ coh901318_tx_submit(struct dma_async_tx_descriptor *tx) | |||
966 | desc); | 952 | desc); |
967 | struct coh901318_chan *cohc = to_coh901318_chan(tx->chan); | 953 | struct coh901318_chan *cohc = to_coh901318_chan(tx->chan); |
968 | unsigned long flags; | 954 | unsigned long flags; |
955 | dma_cookie_t cookie; | ||
969 | 956 | ||
970 | spin_lock_irqsave(&cohc->lock, flags); | 957 | spin_lock_irqsave(&cohc->lock, flags); |
971 | 958 | cookie = dma_cookie_assign(tx); | |
972 | tx->cookie = coh901318_assign_cookie(cohc, cohd); | ||
973 | 959 | ||
974 | coh901318_desc_queue(cohc, cohd); | 960 | coh901318_desc_queue(cohc, cohd); |
975 | 961 | ||
976 | spin_unlock_irqrestore(&cohc->lock, flags); | 962 | spin_unlock_irqrestore(&cohc->lock, flags); |
977 | 963 | ||
978 | return tx->cookie; | 964 | return cookie; |
979 | } | 965 | } |
980 | 966 | ||
981 | static struct dma_async_tx_descriptor * | 967 | static struct dma_async_tx_descriptor * |
@@ -1035,7 +1021,7 @@ coh901318_prep_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src, | |||
1035 | static struct dma_async_tx_descriptor * | 1021 | static struct dma_async_tx_descriptor * |
1036 | coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 1022 | coh901318_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
1037 | unsigned int sg_len, enum dma_transfer_direction direction, | 1023 | unsigned int sg_len, enum dma_transfer_direction direction, |
1038 | unsigned long flags) | 1024 | unsigned long flags, void *context) |
1039 | { | 1025 | { |
1040 | struct coh901318_chan *cohc = to_coh901318_chan(chan); | 1026 | struct coh901318_chan *cohc = to_coh901318_chan(chan); |
1041 | struct coh901318_lli *lli; | 1027 | struct coh901318_lli *lli; |
@@ -1165,17 +1151,12 @@ coh901318_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | |||
1165 | struct dma_tx_state *txstate) | 1151 | struct dma_tx_state *txstate) |
1166 | { | 1152 | { |
1167 | struct coh901318_chan *cohc = to_coh901318_chan(chan); | 1153 | struct coh901318_chan *cohc = to_coh901318_chan(chan); |
1168 | dma_cookie_t last_used; | 1154 | enum dma_status ret; |
1169 | dma_cookie_t last_complete; | ||
1170 | int ret; | ||
1171 | |||
1172 | last_complete = cohc->completed; | ||
1173 | last_used = chan->cookie; | ||
1174 | 1155 | ||
1175 | ret = dma_async_is_complete(cookie, last_complete, last_used); | 1156 | ret = dma_cookie_status(chan, cookie, txstate); |
1157 | /* FIXME: should be conditional on ret != DMA_SUCCESS? */ | ||
1158 | dma_set_residue(txstate, coh901318_get_bytes_left(chan)); | ||
1176 | 1159 | ||
1177 | dma_set_tx_state(txstate, last_complete, last_used, | ||
1178 | coh901318_get_bytes_left(chan)); | ||
1179 | if (ret == DMA_IN_PROGRESS && cohc->stopped) | 1160 | if (ret == DMA_IN_PROGRESS && cohc->stopped) |
1180 | ret = DMA_PAUSED; | 1161 | ret = DMA_PAUSED; |
1181 | 1162 | ||
diff --git a/drivers/dma/dmaengine.c b/drivers/dma/dmaengine.c index a6c6051ec858..767bcc31b365 100644 --- a/drivers/dma/dmaengine.c +++ b/drivers/dma/dmaengine.c | |||
@@ -510,8 +510,8 @@ struct dma_chan *__dma_request_channel(dma_cap_mask_t *mask, dma_filter_fn fn, v | |||
510 | dma_chan_name(chan)); | 510 | dma_chan_name(chan)); |
511 | list_del_rcu(&device->global_node); | 511 | list_del_rcu(&device->global_node); |
512 | } else if (err) | 512 | } else if (err) |
513 | pr_debug("dmaengine: failed to get %s: (%d)\n", | 513 | pr_debug("%s: failed to get %s: (%d)\n", |
514 | dma_chan_name(chan), err); | 514 | __func__, dma_chan_name(chan), err); |
515 | else | 515 | else |
516 | break; | 516 | break; |
517 | if (--device->privatecnt == 0) | 517 | if (--device->privatecnt == 0) |
@@ -564,8 +564,8 @@ void dmaengine_get(void) | |||
564 | list_del_rcu(&device->global_node); | 564 | list_del_rcu(&device->global_node); |
565 | break; | 565 | break; |
566 | } else if (err) | 566 | } else if (err) |
567 | pr_err("dmaengine: failed to get %s: (%d)\n", | 567 | pr_err("%s: failed to get %s: (%d)\n", |
568 | dma_chan_name(chan), err); | 568 | __func__, dma_chan_name(chan), err); |
569 | } | 569 | } |
570 | } | 570 | } |
571 | 571 | ||
diff --git a/drivers/dma/dmaengine.h b/drivers/dma/dmaengine.h new file mode 100644 index 000000000000..17f983a4e9ba --- /dev/null +++ b/drivers/dma/dmaengine.h | |||
@@ -0,0 +1,89 @@ | |||
1 | /* | ||
2 | * The contents of this file are private to DMA engine drivers, and is not | ||
3 | * part of the API to be used by DMA engine users. | ||
4 | */ | ||
5 | #ifndef DMAENGINE_H | ||
6 | #define DMAENGINE_H | ||
7 | |||
8 | #include <linux/bug.h> | ||
9 | #include <linux/dmaengine.h> | ||
10 | |||
11 | /** | ||
12 | * dma_cookie_init - initialize the cookies for a DMA channel | ||
13 | * @chan: dma channel to initialize | ||
14 | */ | ||
15 | static inline void dma_cookie_init(struct dma_chan *chan) | ||
16 | { | ||
17 | chan->cookie = DMA_MIN_COOKIE; | ||
18 | chan->completed_cookie = DMA_MIN_COOKIE; | ||
19 | } | ||
20 | |||
21 | /** | ||
22 | * dma_cookie_assign - assign a DMA engine cookie to the descriptor | ||
23 | * @tx: descriptor needing cookie | ||
24 | * | ||
25 | * Assign a unique non-zero per-channel cookie to the descriptor. | ||
26 | * Note: caller is expected to hold a lock to prevent concurrency. | ||
27 | */ | ||
28 | static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx) | ||
29 | { | ||
30 | struct dma_chan *chan = tx->chan; | ||
31 | dma_cookie_t cookie; | ||
32 | |||
33 | cookie = chan->cookie + 1; | ||
34 | if (cookie < DMA_MIN_COOKIE) | ||
35 | cookie = DMA_MIN_COOKIE; | ||
36 | tx->cookie = chan->cookie = cookie; | ||
37 | |||
38 | return cookie; | ||
39 | } | ||
40 | |||
41 | /** | ||
42 | * dma_cookie_complete - complete a descriptor | ||
43 | * @tx: descriptor to complete | ||
44 | * | ||
45 | * Mark this descriptor complete by updating the channels completed | ||
46 | * cookie marker. Zero the descriptors cookie to prevent accidental | ||
47 | * repeated completions. | ||
48 | * | ||
49 | * Note: caller is expected to hold a lock to prevent concurrency. | ||
50 | */ | ||
51 | static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx) | ||
52 | { | ||
53 | BUG_ON(tx->cookie < DMA_MIN_COOKIE); | ||
54 | tx->chan->completed_cookie = tx->cookie; | ||
55 | tx->cookie = 0; | ||
56 | } | ||
57 | |||
58 | /** | ||
59 | * dma_cookie_status - report cookie status | ||
60 | * @chan: dma channel | ||
61 | * @cookie: cookie we are interested in | ||
62 | * @state: dma_tx_state structure to return last/used cookies | ||
63 | * | ||
64 | * Report the status of the cookie, filling in the state structure if | ||
65 | * non-NULL. No locking is required. | ||
66 | */ | ||
67 | static inline enum dma_status dma_cookie_status(struct dma_chan *chan, | ||
68 | dma_cookie_t cookie, struct dma_tx_state *state) | ||
69 | { | ||
70 | dma_cookie_t used, complete; | ||
71 | |||
72 | used = chan->cookie; | ||
73 | complete = chan->completed_cookie; | ||
74 | barrier(); | ||
75 | if (state) { | ||
76 | state->last = complete; | ||
77 | state->used = used; | ||
78 | state->residue = 0; | ||
79 | } | ||
80 | return dma_async_is_complete(cookie, complete, used); | ||
81 | } | ||
82 | |||
83 | static inline void dma_set_residue(struct dma_tx_state *state, u32 residue) | ||
84 | { | ||
85 | if (state) | ||
86 | state->residue = residue; | ||
87 | } | ||
88 | |||
89 | #endif | ||
diff --git a/drivers/dma/dw_dmac.c b/drivers/dma/dw_dmac.c index 9b592b02b5f4..7439079f5eed 100644 --- a/drivers/dma/dw_dmac.c +++ b/drivers/dma/dw_dmac.c | |||
@@ -9,6 +9,7 @@ | |||
9 | * it under the terms of the GNU General Public License version 2 as | 9 | * it under the terms of the GNU General Public License version 2 as |
10 | * published by the Free Software Foundation. | 10 | * published by the Free Software Foundation. |
11 | */ | 11 | */ |
12 | #include <linux/bitops.h> | ||
12 | #include <linux/clk.h> | 13 | #include <linux/clk.h> |
13 | #include <linux/delay.h> | 14 | #include <linux/delay.h> |
14 | #include <linux/dmaengine.h> | 15 | #include <linux/dmaengine.h> |
@@ -22,6 +23,7 @@ | |||
22 | #include <linux/slab.h> | 23 | #include <linux/slab.h> |
23 | 24 | ||
24 | #include "dw_dmac_regs.h" | 25 | #include "dw_dmac_regs.h" |
26 | #include "dmaengine.h" | ||
25 | 27 | ||
26 | /* | 28 | /* |
27 | * This supports the Synopsys "DesignWare AHB Central DMA Controller", | 29 | * This supports the Synopsys "DesignWare AHB Central DMA Controller", |
@@ -33,19 +35,23 @@ | |||
33 | * which does not support descriptor writeback. | 35 | * which does not support descriptor writeback. |
34 | */ | 36 | */ |
35 | 37 | ||
36 | #define DWC_DEFAULT_CTLLO(private) ({ \ | 38 | #define DWC_DEFAULT_CTLLO(_chan) ({ \ |
37 | struct dw_dma_slave *__slave = (private); \ | 39 | struct dw_dma_slave *__slave = (_chan->private); \ |
38 | int dms = __slave ? __slave->dst_master : 0; \ | 40 | struct dw_dma_chan *_dwc = to_dw_dma_chan(_chan); \ |
39 | int sms = __slave ? __slave->src_master : 1; \ | 41 | struct dma_slave_config *_sconfig = &_dwc->dma_sconfig; \ |
40 | u8 smsize = __slave ? __slave->src_msize : DW_DMA_MSIZE_16; \ | 42 | int _dms = __slave ? __slave->dst_master : 0; \ |
41 | u8 dmsize = __slave ? __slave->dst_msize : DW_DMA_MSIZE_16; \ | 43 | int _sms = __slave ? __slave->src_master : 1; \ |
44 | u8 _smsize = __slave ? _sconfig->src_maxburst : \ | ||
45 | DW_DMA_MSIZE_16; \ | ||
46 | u8 _dmsize = __slave ? _sconfig->dst_maxburst : \ | ||
47 | DW_DMA_MSIZE_16; \ | ||
42 | \ | 48 | \ |
43 | (DWC_CTLL_DST_MSIZE(dmsize) \ | 49 | (DWC_CTLL_DST_MSIZE(_dmsize) \ |
44 | | DWC_CTLL_SRC_MSIZE(smsize) \ | 50 | | DWC_CTLL_SRC_MSIZE(_smsize) \ |
45 | | DWC_CTLL_LLP_D_EN \ | 51 | | DWC_CTLL_LLP_D_EN \ |
46 | | DWC_CTLL_LLP_S_EN \ | 52 | | DWC_CTLL_LLP_S_EN \ |
47 | | DWC_CTLL_DMS(dms) \ | 53 | | DWC_CTLL_DMS(_dms) \ |
48 | | DWC_CTLL_SMS(sms)); \ | 54 | | DWC_CTLL_SMS(_sms)); \ |
49 | }) | 55 | }) |
50 | 56 | ||
51 | /* | 57 | /* |
@@ -151,21 +157,6 @@ static void dwc_desc_put(struct dw_dma_chan *dwc, struct dw_desc *desc) | |||
151 | } | 157 | } |
152 | } | 158 | } |
153 | 159 | ||
154 | /* Called with dwc->lock held and bh disabled */ | ||
155 | static dma_cookie_t | ||
156 | dwc_assign_cookie(struct dw_dma_chan *dwc, struct dw_desc *desc) | ||
157 | { | ||
158 | dma_cookie_t cookie = dwc->chan.cookie; | ||
159 | |||
160 | if (++cookie < 0) | ||
161 | cookie = 1; | ||
162 | |||
163 | dwc->chan.cookie = cookie; | ||
164 | desc->txd.cookie = cookie; | ||
165 | |||
166 | return cookie; | ||
167 | } | ||
168 | |||
169 | static void dwc_initialize(struct dw_dma_chan *dwc) | 160 | static void dwc_initialize(struct dw_dma_chan *dwc) |
170 | { | 161 | { |
171 | struct dw_dma *dw = to_dw_dma(dwc->chan.device); | 162 | struct dw_dma *dw = to_dw_dma(dwc->chan.device); |
@@ -192,7 +183,6 @@ static void dwc_initialize(struct dw_dma_chan *dwc) | |||
192 | 183 | ||
193 | /* Enable interrupts */ | 184 | /* Enable interrupts */ |
194 | channel_set_bit(dw, MASK.XFER, dwc->mask); | 185 | channel_set_bit(dw, MASK.XFER, dwc->mask); |
195 | channel_set_bit(dw, MASK.BLOCK, dwc->mask); | ||
196 | channel_set_bit(dw, MASK.ERROR, dwc->mask); | 186 | channel_set_bit(dw, MASK.ERROR, dwc->mask); |
197 | 187 | ||
198 | dwc->initialized = true; | 188 | dwc->initialized = true; |
@@ -245,7 +235,7 @@ dwc_descriptor_complete(struct dw_dma_chan *dwc, struct dw_desc *desc, | |||
245 | dev_vdbg(chan2dev(&dwc->chan), "descriptor %u complete\n", txd->cookie); | 235 | dev_vdbg(chan2dev(&dwc->chan), "descriptor %u complete\n", txd->cookie); |
246 | 236 | ||
247 | spin_lock_irqsave(&dwc->lock, flags); | 237 | spin_lock_irqsave(&dwc->lock, flags); |
248 | dwc->completed = txd->cookie; | 238 | dma_cookie_complete(txd); |
249 | if (callback_required) { | 239 | if (callback_required) { |
250 | callback = txd->callback; | 240 | callback = txd->callback; |
251 | param = txd->callback_param; | 241 | param = txd->callback_param; |
@@ -329,12 +319,6 @@ static void dwc_scan_descriptors(struct dw_dma *dw, struct dw_dma_chan *dwc) | |||
329 | unsigned long flags; | 319 | unsigned long flags; |
330 | 320 | ||
331 | spin_lock_irqsave(&dwc->lock, flags); | 321 | spin_lock_irqsave(&dwc->lock, flags); |
332 | /* | ||
333 | * Clear block interrupt flag before scanning so that we don't | ||
334 | * miss any, and read LLP before RAW_XFER to ensure it is | ||
335 | * valid if we decide to scan the list. | ||
336 | */ | ||
337 | dma_writel(dw, CLEAR.BLOCK, dwc->mask); | ||
338 | llp = channel_readl(dwc, LLP); | 322 | llp = channel_readl(dwc, LLP); |
339 | status_xfer = dma_readl(dw, RAW.XFER); | 323 | status_xfer = dma_readl(dw, RAW.XFER); |
340 | 324 | ||
@@ -470,17 +454,16 @@ EXPORT_SYMBOL(dw_dma_get_dst_addr); | |||
470 | 454 | ||
471 | /* called with dwc->lock held and all DMAC interrupts disabled */ | 455 | /* called with dwc->lock held and all DMAC interrupts disabled */ |
472 | static void dwc_handle_cyclic(struct dw_dma *dw, struct dw_dma_chan *dwc, | 456 | static void dwc_handle_cyclic(struct dw_dma *dw, struct dw_dma_chan *dwc, |
473 | u32 status_block, u32 status_err, u32 status_xfer) | 457 | u32 status_err, u32 status_xfer) |
474 | { | 458 | { |
475 | unsigned long flags; | 459 | unsigned long flags; |
476 | 460 | ||
477 | if (status_block & dwc->mask) { | 461 | if (dwc->mask) { |
478 | void (*callback)(void *param); | 462 | void (*callback)(void *param); |
479 | void *callback_param; | 463 | void *callback_param; |
480 | 464 | ||
481 | dev_vdbg(chan2dev(&dwc->chan), "new cyclic period llp 0x%08x\n", | 465 | dev_vdbg(chan2dev(&dwc->chan), "new cyclic period llp 0x%08x\n", |
482 | channel_readl(dwc, LLP)); | 466 | channel_readl(dwc, LLP)); |
483 | dma_writel(dw, CLEAR.BLOCK, dwc->mask); | ||
484 | 467 | ||
485 | callback = dwc->cdesc->period_callback; | 468 | callback = dwc->cdesc->period_callback; |
486 | callback_param = dwc->cdesc->period_callback_param; | 469 | callback_param = dwc->cdesc->period_callback_param; |
@@ -520,7 +503,6 @@ static void dwc_handle_cyclic(struct dw_dma *dw, struct dw_dma_chan *dwc, | |||
520 | channel_writel(dwc, CTL_LO, 0); | 503 | channel_writel(dwc, CTL_LO, 0); |
521 | channel_writel(dwc, CTL_HI, 0); | 504 | channel_writel(dwc, CTL_HI, 0); |
522 | 505 | ||
523 | dma_writel(dw, CLEAR.BLOCK, dwc->mask); | ||
524 | dma_writel(dw, CLEAR.ERROR, dwc->mask); | 506 | dma_writel(dw, CLEAR.ERROR, dwc->mask); |
525 | dma_writel(dw, CLEAR.XFER, dwc->mask); | 507 | dma_writel(dw, CLEAR.XFER, dwc->mask); |
526 | 508 | ||
@@ -537,36 +519,29 @@ static void dw_dma_tasklet(unsigned long data) | |||
537 | { | 519 | { |
538 | struct dw_dma *dw = (struct dw_dma *)data; | 520 | struct dw_dma *dw = (struct dw_dma *)data; |
539 | struct dw_dma_chan *dwc; | 521 | struct dw_dma_chan *dwc; |
540 | u32 status_block; | ||
541 | u32 status_xfer; | 522 | u32 status_xfer; |
542 | u32 status_err; | 523 | u32 status_err; |
543 | int i; | 524 | int i; |
544 | 525 | ||
545 | status_block = dma_readl(dw, RAW.BLOCK); | ||
546 | status_xfer = dma_readl(dw, RAW.XFER); | 526 | status_xfer = dma_readl(dw, RAW.XFER); |
547 | status_err = dma_readl(dw, RAW.ERROR); | 527 | status_err = dma_readl(dw, RAW.ERROR); |
548 | 528 | ||
549 | dev_vdbg(dw->dma.dev, "tasklet: status_block=%x status_err=%x\n", | 529 | dev_vdbg(dw->dma.dev, "tasklet: status_err=%x\n", status_err); |
550 | status_block, status_err); | ||
551 | 530 | ||
552 | for (i = 0; i < dw->dma.chancnt; i++) { | 531 | for (i = 0; i < dw->dma.chancnt; i++) { |
553 | dwc = &dw->chan[i]; | 532 | dwc = &dw->chan[i]; |
554 | if (test_bit(DW_DMA_IS_CYCLIC, &dwc->flags)) | 533 | if (test_bit(DW_DMA_IS_CYCLIC, &dwc->flags)) |
555 | dwc_handle_cyclic(dw, dwc, status_block, status_err, | 534 | dwc_handle_cyclic(dw, dwc, status_err, status_xfer); |
556 | status_xfer); | ||
557 | else if (status_err & (1 << i)) | 535 | else if (status_err & (1 << i)) |
558 | dwc_handle_error(dw, dwc); | 536 | dwc_handle_error(dw, dwc); |
559 | else if ((status_block | status_xfer) & (1 << i)) | 537 | else if (status_xfer & (1 << i)) |
560 | dwc_scan_descriptors(dw, dwc); | 538 | dwc_scan_descriptors(dw, dwc); |
561 | } | 539 | } |
562 | 540 | ||
563 | /* | 541 | /* |
564 | * Re-enable interrupts. Block Complete interrupts are only | 542 | * Re-enable interrupts. |
565 | * enabled if the INT_EN bit in the descriptor is set. This | ||
566 | * will trigger a scan before the whole list is done. | ||
567 | */ | 543 | */ |
568 | channel_set_bit(dw, MASK.XFER, dw->all_chan_mask); | 544 | channel_set_bit(dw, MASK.XFER, dw->all_chan_mask); |
569 | channel_set_bit(dw, MASK.BLOCK, dw->all_chan_mask); | ||
570 | channel_set_bit(dw, MASK.ERROR, dw->all_chan_mask); | 545 | channel_set_bit(dw, MASK.ERROR, dw->all_chan_mask); |
571 | } | 546 | } |
572 | 547 | ||
@@ -583,7 +558,6 @@ static irqreturn_t dw_dma_interrupt(int irq, void *dev_id) | |||
583 | * softirq handler. | 558 | * softirq handler. |
584 | */ | 559 | */ |
585 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); | 560 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); |
586 | channel_clear_bit(dw, MASK.BLOCK, dw->all_chan_mask); | ||
587 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); | 561 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); |
588 | 562 | ||
589 | status = dma_readl(dw, STATUS_INT); | 563 | status = dma_readl(dw, STATUS_INT); |
@@ -594,7 +568,6 @@ static irqreturn_t dw_dma_interrupt(int irq, void *dev_id) | |||
594 | 568 | ||
595 | /* Try to recover */ | 569 | /* Try to recover */ |
596 | channel_clear_bit(dw, MASK.XFER, (1 << 8) - 1); | 570 | channel_clear_bit(dw, MASK.XFER, (1 << 8) - 1); |
597 | channel_clear_bit(dw, MASK.BLOCK, (1 << 8) - 1); | ||
598 | channel_clear_bit(dw, MASK.SRC_TRAN, (1 << 8) - 1); | 571 | channel_clear_bit(dw, MASK.SRC_TRAN, (1 << 8) - 1); |
599 | channel_clear_bit(dw, MASK.DST_TRAN, (1 << 8) - 1); | 572 | channel_clear_bit(dw, MASK.DST_TRAN, (1 << 8) - 1); |
600 | channel_clear_bit(dw, MASK.ERROR, (1 << 8) - 1); | 573 | channel_clear_bit(dw, MASK.ERROR, (1 << 8) - 1); |
@@ -615,7 +588,7 @@ static dma_cookie_t dwc_tx_submit(struct dma_async_tx_descriptor *tx) | |||
615 | unsigned long flags; | 588 | unsigned long flags; |
616 | 589 | ||
617 | spin_lock_irqsave(&dwc->lock, flags); | 590 | spin_lock_irqsave(&dwc->lock, flags); |
618 | cookie = dwc_assign_cookie(dwc, desc); | 591 | cookie = dma_cookie_assign(tx); |
619 | 592 | ||
620 | /* | 593 | /* |
621 | * REVISIT: We should attempt to chain as many descriptors as | 594 | * REVISIT: We should attempt to chain as many descriptors as |
@@ -674,7 +647,7 @@ dwc_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src, | |||
674 | else | 647 | else |
675 | src_width = dst_width = 0; | 648 | src_width = dst_width = 0; |
676 | 649 | ||
677 | ctllo = DWC_DEFAULT_CTLLO(chan->private) | 650 | ctllo = DWC_DEFAULT_CTLLO(chan) |
678 | | DWC_CTLL_DST_WIDTH(dst_width) | 651 | | DWC_CTLL_DST_WIDTH(dst_width) |
679 | | DWC_CTLL_SRC_WIDTH(src_width) | 652 | | DWC_CTLL_SRC_WIDTH(src_width) |
680 | | DWC_CTLL_DST_INC | 653 | | DWC_CTLL_DST_INC |
@@ -731,10 +704,11 @@ err_desc_get: | |||
731 | static struct dma_async_tx_descriptor * | 704 | static struct dma_async_tx_descriptor * |
732 | dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 705 | dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
733 | unsigned int sg_len, enum dma_transfer_direction direction, | 706 | unsigned int sg_len, enum dma_transfer_direction direction, |
734 | unsigned long flags) | 707 | unsigned long flags, void *context) |
735 | { | 708 | { |
736 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); | 709 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); |
737 | struct dw_dma_slave *dws = chan->private; | 710 | struct dw_dma_slave *dws = chan->private; |
711 | struct dma_slave_config *sconfig = &dwc->dma_sconfig; | ||
738 | struct dw_desc *prev; | 712 | struct dw_desc *prev; |
739 | struct dw_desc *first; | 713 | struct dw_desc *first; |
740 | u32 ctllo; | 714 | u32 ctllo; |
@@ -750,25 +724,34 @@ dwc_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | |||
750 | if (unlikely(!dws || !sg_len)) | 724 | if (unlikely(!dws || !sg_len)) |
751 | return NULL; | 725 | return NULL; |
752 | 726 | ||
753 | reg_width = dws->reg_width; | ||
754 | prev = first = NULL; | 727 | prev = first = NULL; |
755 | 728 | ||
756 | switch (direction) { | 729 | switch (direction) { |
757 | case DMA_MEM_TO_DEV: | 730 | case DMA_MEM_TO_DEV: |
758 | ctllo = (DWC_DEFAULT_CTLLO(chan->private) | 731 | reg_width = __fls(sconfig->dst_addr_width); |
732 | reg = sconfig->dst_addr; | ||
733 | ctllo = (DWC_DEFAULT_CTLLO(chan) | ||
759 | | DWC_CTLL_DST_WIDTH(reg_width) | 734 | | DWC_CTLL_DST_WIDTH(reg_width) |
760 | | DWC_CTLL_DST_FIX | 735 | | DWC_CTLL_DST_FIX |
761 | | DWC_CTLL_SRC_INC | 736 | | DWC_CTLL_SRC_INC); |
762 | | DWC_CTLL_FC(dws->fc)); | 737 | |
763 | reg = dws->tx_reg; | 738 | ctllo |= sconfig->device_fc ? DWC_CTLL_FC(DW_DMA_FC_P_M2P) : |
739 | DWC_CTLL_FC(DW_DMA_FC_D_M2P); | ||
740 | |||
764 | for_each_sg(sgl, sg, sg_len, i) { | 741 | for_each_sg(sgl, sg, sg_len, i) { |
765 | struct dw_desc *desc; | 742 | struct dw_desc *desc; |
766 | u32 len, dlen, mem; | 743 | u32 len, dlen, mem; |
767 | 744 | ||
768 | mem = sg_phys(sg); | 745 | mem = sg_phys(sg); |
769 | len = sg_dma_len(sg); | 746 | len = sg_dma_len(sg); |
770 | mem_width = 2; | 747 | |
771 | if (unlikely(mem & 3 || len & 3)) | 748 | if (!((mem | len) & 7)) |
749 | mem_width = 3; | ||
750 | else if (!((mem | len) & 3)) | ||
751 | mem_width = 2; | ||
752 | else if (!((mem | len) & 1)) | ||
753 | mem_width = 1; | ||
754 | else | ||
772 | mem_width = 0; | 755 | mem_width = 0; |
773 | 756 | ||
774 | slave_sg_todev_fill_desc: | 757 | slave_sg_todev_fill_desc: |
@@ -812,21 +795,30 @@ slave_sg_todev_fill_desc: | |||
812 | } | 795 | } |
813 | break; | 796 | break; |
814 | case DMA_DEV_TO_MEM: | 797 | case DMA_DEV_TO_MEM: |
815 | ctllo = (DWC_DEFAULT_CTLLO(chan->private) | 798 | reg_width = __fls(sconfig->src_addr_width); |
799 | reg = sconfig->src_addr; | ||
800 | ctllo = (DWC_DEFAULT_CTLLO(chan) | ||
816 | | DWC_CTLL_SRC_WIDTH(reg_width) | 801 | | DWC_CTLL_SRC_WIDTH(reg_width) |
817 | | DWC_CTLL_DST_INC | 802 | | DWC_CTLL_DST_INC |
818 | | DWC_CTLL_SRC_FIX | 803 | | DWC_CTLL_SRC_FIX); |
819 | | DWC_CTLL_FC(dws->fc)); | 804 | |
805 | ctllo |= sconfig->device_fc ? DWC_CTLL_FC(DW_DMA_FC_P_P2M) : | ||
806 | DWC_CTLL_FC(DW_DMA_FC_D_P2M); | ||
820 | 807 | ||
821 | reg = dws->rx_reg; | ||
822 | for_each_sg(sgl, sg, sg_len, i) { | 808 | for_each_sg(sgl, sg, sg_len, i) { |
823 | struct dw_desc *desc; | 809 | struct dw_desc *desc; |
824 | u32 len, dlen, mem; | 810 | u32 len, dlen, mem; |
825 | 811 | ||
826 | mem = sg_phys(sg); | 812 | mem = sg_phys(sg); |
827 | len = sg_dma_len(sg); | 813 | len = sg_dma_len(sg); |
828 | mem_width = 2; | 814 | |
829 | if (unlikely(mem & 3 || len & 3)) | 815 | if (!((mem | len) & 7)) |
816 | mem_width = 3; | ||
817 | else if (!((mem | len) & 3)) | ||
818 | mem_width = 2; | ||
819 | else if (!((mem | len) & 1)) | ||
820 | mem_width = 1; | ||
821 | else | ||
830 | mem_width = 0; | 822 | mem_width = 0; |
831 | 823 | ||
832 | slave_sg_fromdev_fill_desc: | 824 | slave_sg_fromdev_fill_desc: |
@@ -890,6 +882,39 @@ err_desc_get: | |||
890 | return NULL; | 882 | return NULL; |
891 | } | 883 | } |
892 | 884 | ||
885 | /* | ||
886 | * Fix sconfig's burst size according to dw_dmac. We need to convert them as: | ||
887 | * 1 -> 0, 4 -> 1, 8 -> 2, 16 -> 3. | ||
888 | * | ||
889 | * NOTE: burst size 2 is not supported by controller. | ||
890 | * | ||
891 | * This can be done by finding least significant bit set: n & (n - 1) | ||
892 | */ | ||
893 | static inline void convert_burst(u32 *maxburst) | ||
894 | { | ||
895 | if (*maxburst > 1) | ||
896 | *maxburst = fls(*maxburst) - 2; | ||
897 | else | ||
898 | *maxburst = 0; | ||
899 | } | ||
900 | |||
901 | static int | ||
902 | set_runtime_config(struct dma_chan *chan, struct dma_slave_config *sconfig) | ||
903 | { | ||
904 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); | ||
905 | |||
906 | /* Check if it is chan is configured for slave transfers */ | ||
907 | if (!chan->private) | ||
908 | return -EINVAL; | ||
909 | |||
910 | memcpy(&dwc->dma_sconfig, sconfig, sizeof(*sconfig)); | ||
911 | |||
912 | convert_burst(&dwc->dma_sconfig.src_maxburst); | ||
913 | convert_burst(&dwc->dma_sconfig.dst_maxburst); | ||
914 | |||
915 | return 0; | ||
916 | } | ||
917 | |||
893 | static int dwc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | 918 | static int dwc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, |
894 | unsigned long arg) | 919 | unsigned long arg) |
895 | { | 920 | { |
@@ -939,8 +964,11 @@ static int dwc_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | |||
939 | /* Flush all pending and queued descriptors */ | 964 | /* Flush all pending and queued descriptors */ |
940 | list_for_each_entry_safe(desc, _desc, &list, desc_node) | 965 | list_for_each_entry_safe(desc, _desc, &list, desc_node) |
941 | dwc_descriptor_complete(dwc, desc, false); | 966 | dwc_descriptor_complete(dwc, desc, false); |
942 | } else | 967 | } else if (cmd == DMA_SLAVE_CONFIG) { |
968 | return set_runtime_config(chan, (struct dma_slave_config *)arg); | ||
969 | } else { | ||
943 | return -ENXIO; | 970 | return -ENXIO; |
971 | } | ||
944 | 972 | ||
945 | return 0; | 973 | return 0; |
946 | } | 974 | } |
@@ -951,28 +979,17 @@ dwc_tx_status(struct dma_chan *chan, | |||
951 | struct dma_tx_state *txstate) | 979 | struct dma_tx_state *txstate) |
952 | { | 980 | { |
953 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); | 981 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); |
954 | dma_cookie_t last_used; | 982 | enum dma_status ret; |
955 | dma_cookie_t last_complete; | ||
956 | int ret; | ||
957 | |||
958 | last_complete = dwc->completed; | ||
959 | last_used = chan->cookie; | ||
960 | 983 | ||
961 | ret = dma_async_is_complete(cookie, last_complete, last_used); | 984 | ret = dma_cookie_status(chan, cookie, txstate); |
962 | if (ret != DMA_SUCCESS) { | 985 | if (ret != DMA_SUCCESS) { |
963 | dwc_scan_descriptors(to_dw_dma(chan->device), dwc); | 986 | dwc_scan_descriptors(to_dw_dma(chan->device), dwc); |
964 | 987 | ||
965 | last_complete = dwc->completed; | 988 | ret = dma_cookie_status(chan, cookie, txstate); |
966 | last_used = chan->cookie; | ||
967 | |||
968 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
969 | } | 989 | } |
970 | 990 | ||
971 | if (ret != DMA_SUCCESS) | 991 | if (ret != DMA_SUCCESS) |
972 | dma_set_tx_state(txstate, last_complete, last_used, | 992 | dma_set_residue(txstate, dwc_first_active(dwc)->len); |
973 | dwc_first_active(dwc)->len); | ||
974 | else | ||
975 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
976 | 993 | ||
977 | if (dwc->paused) | 994 | if (dwc->paused) |
978 | return DMA_PAUSED; | 995 | return DMA_PAUSED; |
@@ -1004,7 +1021,7 @@ static int dwc_alloc_chan_resources(struct dma_chan *chan) | |||
1004 | return -EIO; | 1021 | return -EIO; |
1005 | } | 1022 | } |
1006 | 1023 | ||
1007 | dwc->completed = chan->cookie = 1; | 1024 | dma_cookie_init(chan); |
1008 | 1025 | ||
1009 | /* | 1026 | /* |
1010 | * NOTE: some controllers may have additional features that we | 1027 | * NOTE: some controllers may have additional features that we |
@@ -1068,7 +1085,6 @@ static void dwc_free_chan_resources(struct dma_chan *chan) | |||
1068 | 1085 | ||
1069 | /* Disable interrupts */ | 1086 | /* Disable interrupts */ |
1070 | channel_clear_bit(dw, MASK.XFER, dwc->mask); | 1087 | channel_clear_bit(dw, MASK.XFER, dwc->mask); |
1071 | channel_clear_bit(dw, MASK.BLOCK, dwc->mask); | ||
1072 | channel_clear_bit(dw, MASK.ERROR, dwc->mask); | 1088 | channel_clear_bit(dw, MASK.ERROR, dwc->mask); |
1073 | 1089 | ||
1074 | spin_unlock_irqrestore(&dwc->lock, flags); | 1090 | spin_unlock_irqrestore(&dwc->lock, flags); |
@@ -1120,7 +1136,6 @@ int dw_dma_cyclic_start(struct dma_chan *chan) | |||
1120 | return -EBUSY; | 1136 | return -EBUSY; |
1121 | } | 1137 | } |
1122 | 1138 | ||
1123 | dma_writel(dw, CLEAR.BLOCK, dwc->mask); | ||
1124 | dma_writel(dw, CLEAR.ERROR, dwc->mask); | 1139 | dma_writel(dw, CLEAR.ERROR, dwc->mask); |
1125 | dma_writel(dw, CLEAR.XFER, dwc->mask); | 1140 | dma_writel(dw, CLEAR.XFER, dwc->mask); |
1126 | 1141 | ||
@@ -1175,11 +1190,11 @@ struct dw_cyclic_desc *dw_dma_cyclic_prep(struct dma_chan *chan, | |||
1175 | enum dma_transfer_direction direction) | 1190 | enum dma_transfer_direction direction) |
1176 | { | 1191 | { |
1177 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); | 1192 | struct dw_dma_chan *dwc = to_dw_dma_chan(chan); |
1193 | struct dma_slave_config *sconfig = &dwc->dma_sconfig; | ||
1178 | struct dw_cyclic_desc *cdesc; | 1194 | struct dw_cyclic_desc *cdesc; |
1179 | struct dw_cyclic_desc *retval = NULL; | 1195 | struct dw_cyclic_desc *retval = NULL; |
1180 | struct dw_desc *desc; | 1196 | struct dw_desc *desc; |
1181 | struct dw_desc *last = NULL; | 1197 | struct dw_desc *last = NULL; |
1182 | struct dw_dma_slave *dws = chan->private; | ||
1183 | unsigned long was_cyclic; | 1198 | unsigned long was_cyclic; |
1184 | unsigned int reg_width; | 1199 | unsigned int reg_width; |
1185 | unsigned int periods; | 1200 | unsigned int periods; |
@@ -1203,7 +1218,12 @@ struct dw_cyclic_desc *dw_dma_cyclic_prep(struct dma_chan *chan, | |||
1203 | } | 1218 | } |
1204 | 1219 | ||
1205 | retval = ERR_PTR(-EINVAL); | 1220 | retval = ERR_PTR(-EINVAL); |
1206 | reg_width = dws->reg_width; | 1221 | |
1222 | if (direction == DMA_MEM_TO_DEV) | ||
1223 | reg_width = __ffs(sconfig->dst_addr_width); | ||
1224 | else | ||
1225 | reg_width = __ffs(sconfig->src_addr_width); | ||
1226 | |||
1207 | periods = buf_len / period_len; | 1227 | periods = buf_len / period_len; |
1208 | 1228 | ||
1209 | /* Check for too big/unaligned periods and unaligned DMA buffer. */ | 1229 | /* Check for too big/unaligned periods and unaligned DMA buffer. */ |
@@ -1236,26 +1256,34 @@ struct dw_cyclic_desc *dw_dma_cyclic_prep(struct dma_chan *chan, | |||
1236 | 1256 | ||
1237 | switch (direction) { | 1257 | switch (direction) { |
1238 | case DMA_MEM_TO_DEV: | 1258 | case DMA_MEM_TO_DEV: |
1239 | desc->lli.dar = dws->tx_reg; | 1259 | desc->lli.dar = sconfig->dst_addr; |
1240 | desc->lli.sar = buf_addr + (period_len * i); | 1260 | desc->lli.sar = buf_addr + (period_len * i); |
1241 | desc->lli.ctllo = (DWC_DEFAULT_CTLLO(chan->private) | 1261 | desc->lli.ctllo = (DWC_DEFAULT_CTLLO(chan) |
1242 | | DWC_CTLL_DST_WIDTH(reg_width) | 1262 | | DWC_CTLL_DST_WIDTH(reg_width) |
1243 | | DWC_CTLL_SRC_WIDTH(reg_width) | 1263 | | DWC_CTLL_SRC_WIDTH(reg_width) |
1244 | | DWC_CTLL_DST_FIX | 1264 | | DWC_CTLL_DST_FIX |
1245 | | DWC_CTLL_SRC_INC | 1265 | | DWC_CTLL_SRC_INC |
1246 | | DWC_CTLL_FC(dws->fc) | ||
1247 | | DWC_CTLL_INT_EN); | 1266 | | DWC_CTLL_INT_EN); |
1267 | |||
1268 | desc->lli.ctllo |= sconfig->device_fc ? | ||
1269 | DWC_CTLL_FC(DW_DMA_FC_P_M2P) : | ||
1270 | DWC_CTLL_FC(DW_DMA_FC_D_M2P); | ||
1271 | |||
1248 | break; | 1272 | break; |
1249 | case DMA_DEV_TO_MEM: | 1273 | case DMA_DEV_TO_MEM: |
1250 | desc->lli.dar = buf_addr + (period_len * i); | 1274 | desc->lli.dar = buf_addr + (period_len * i); |
1251 | desc->lli.sar = dws->rx_reg; | 1275 | desc->lli.sar = sconfig->src_addr; |
1252 | desc->lli.ctllo = (DWC_DEFAULT_CTLLO(chan->private) | 1276 | desc->lli.ctllo = (DWC_DEFAULT_CTLLO(chan) |
1253 | | DWC_CTLL_SRC_WIDTH(reg_width) | 1277 | | DWC_CTLL_SRC_WIDTH(reg_width) |
1254 | | DWC_CTLL_DST_WIDTH(reg_width) | 1278 | | DWC_CTLL_DST_WIDTH(reg_width) |
1255 | | DWC_CTLL_DST_INC | 1279 | | DWC_CTLL_DST_INC |
1256 | | DWC_CTLL_SRC_FIX | 1280 | | DWC_CTLL_SRC_FIX |
1257 | | DWC_CTLL_FC(dws->fc) | ||
1258 | | DWC_CTLL_INT_EN); | 1281 | | DWC_CTLL_INT_EN); |
1282 | |||
1283 | desc->lli.ctllo |= sconfig->device_fc ? | ||
1284 | DWC_CTLL_FC(DW_DMA_FC_P_P2M) : | ||
1285 | DWC_CTLL_FC(DW_DMA_FC_D_P2M); | ||
1286 | |||
1259 | break; | 1287 | break; |
1260 | default: | 1288 | default: |
1261 | break; | 1289 | break; |
@@ -1322,7 +1350,6 @@ void dw_dma_cyclic_free(struct dma_chan *chan) | |||
1322 | while (dma_readl(dw, CH_EN) & dwc->mask) | 1350 | while (dma_readl(dw, CH_EN) & dwc->mask) |
1323 | cpu_relax(); | 1351 | cpu_relax(); |
1324 | 1352 | ||
1325 | dma_writel(dw, CLEAR.BLOCK, dwc->mask); | ||
1326 | dma_writel(dw, CLEAR.ERROR, dwc->mask); | 1353 | dma_writel(dw, CLEAR.ERROR, dwc->mask); |
1327 | dma_writel(dw, CLEAR.XFER, dwc->mask); | 1354 | dma_writel(dw, CLEAR.XFER, dwc->mask); |
1328 | 1355 | ||
@@ -1347,7 +1374,6 @@ static void dw_dma_off(struct dw_dma *dw) | |||
1347 | dma_writel(dw, CFG, 0); | 1374 | dma_writel(dw, CFG, 0); |
1348 | 1375 | ||
1349 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); | 1376 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); |
1350 | channel_clear_bit(dw, MASK.BLOCK, dw->all_chan_mask); | ||
1351 | channel_clear_bit(dw, MASK.SRC_TRAN, dw->all_chan_mask); | 1377 | channel_clear_bit(dw, MASK.SRC_TRAN, dw->all_chan_mask); |
1352 | channel_clear_bit(dw, MASK.DST_TRAN, dw->all_chan_mask); | 1378 | channel_clear_bit(dw, MASK.DST_TRAN, dw->all_chan_mask); |
1353 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); | 1379 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); |
@@ -1369,7 +1395,7 @@ static int __init dw_probe(struct platform_device *pdev) | |||
1369 | int err; | 1395 | int err; |
1370 | int i; | 1396 | int i; |
1371 | 1397 | ||
1372 | pdata = pdev->dev.platform_data; | 1398 | pdata = dev_get_platdata(&pdev->dev); |
1373 | if (!pdata || pdata->nr_channels > DW_DMA_MAX_NR_CHANNELS) | 1399 | if (!pdata || pdata->nr_channels > DW_DMA_MAX_NR_CHANNELS) |
1374 | return -EINVAL; | 1400 | return -EINVAL; |
1375 | 1401 | ||
@@ -1423,7 +1449,7 @@ static int __init dw_probe(struct platform_device *pdev) | |||
1423 | struct dw_dma_chan *dwc = &dw->chan[i]; | 1449 | struct dw_dma_chan *dwc = &dw->chan[i]; |
1424 | 1450 | ||
1425 | dwc->chan.device = &dw->dma; | 1451 | dwc->chan.device = &dw->dma; |
1426 | dwc->chan.cookie = dwc->completed = 1; | 1452 | dma_cookie_init(&dwc->chan); |
1427 | if (pdata->chan_allocation_order == CHAN_ALLOCATION_ASCENDING) | 1453 | if (pdata->chan_allocation_order == CHAN_ALLOCATION_ASCENDING) |
1428 | list_add_tail(&dwc->chan.device_node, | 1454 | list_add_tail(&dwc->chan.device_node, |
1429 | &dw->dma.channels); | 1455 | &dw->dma.channels); |
@@ -1432,7 +1458,7 @@ static int __init dw_probe(struct platform_device *pdev) | |||
1432 | 1458 | ||
1433 | /* 7 is highest priority & 0 is lowest. */ | 1459 | /* 7 is highest priority & 0 is lowest. */ |
1434 | if (pdata->chan_priority == CHAN_PRIORITY_ASCENDING) | 1460 | if (pdata->chan_priority == CHAN_PRIORITY_ASCENDING) |
1435 | dwc->priority = 7 - i; | 1461 | dwc->priority = pdata->nr_channels - i - 1; |
1436 | else | 1462 | else |
1437 | dwc->priority = i; | 1463 | dwc->priority = i; |
1438 | 1464 | ||
@@ -1449,13 +1475,11 @@ static int __init dw_probe(struct platform_device *pdev) | |||
1449 | 1475 | ||
1450 | /* Clear/disable all interrupts on all channels. */ | 1476 | /* Clear/disable all interrupts on all channels. */ |
1451 | dma_writel(dw, CLEAR.XFER, dw->all_chan_mask); | 1477 | dma_writel(dw, CLEAR.XFER, dw->all_chan_mask); |
1452 | dma_writel(dw, CLEAR.BLOCK, dw->all_chan_mask); | ||
1453 | dma_writel(dw, CLEAR.SRC_TRAN, dw->all_chan_mask); | 1478 | dma_writel(dw, CLEAR.SRC_TRAN, dw->all_chan_mask); |
1454 | dma_writel(dw, CLEAR.DST_TRAN, dw->all_chan_mask); | 1479 | dma_writel(dw, CLEAR.DST_TRAN, dw->all_chan_mask); |
1455 | dma_writel(dw, CLEAR.ERROR, dw->all_chan_mask); | 1480 | dma_writel(dw, CLEAR.ERROR, dw->all_chan_mask); |
1456 | 1481 | ||
1457 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); | 1482 | channel_clear_bit(dw, MASK.XFER, dw->all_chan_mask); |
1458 | channel_clear_bit(dw, MASK.BLOCK, dw->all_chan_mask); | ||
1459 | channel_clear_bit(dw, MASK.SRC_TRAN, dw->all_chan_mask); | 1483 | channel_clear_bit(dw, MASK.SRC_TRAN, dw->all_chan_mask); |
1460 | channel_clear_bit(dw, MASK.DST_TRAN, dw->all_chan_mask); | 1484 | channel_clear_bit(dw, MASK.DST_TRAN, dw->all_chan_mask); |
1461 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); | 1485 | channel_clear_bit(dw, MASK.ERROR, dw->all_chan_mask); |
@@ -1562,6 +1586,10 @@ static int dw_resume_noirq(struct device *dev) | |||
1562 | static const struct dev_pm_ops dw_dev_pm_ops = { | 1586 | static const struct dev_pm_ops dw_dev_pm_ops = { |
1563 | .suspend_noirq = dw_suspend_noirq, | 1587 | .suspend_noirq = dw_suspend_noirq, |
1564 | .resume_noirq = dw_resume_noirq, | 1588 | .resume_noirq = dw_resume_noirq, |
1589 | .freeze_noirq = dw_suspend_noirq, | ||
1590 | .thaw_noirq = dw_resume_noirq, | ||
1591 | .restore_noirq = dw_resume_noirq, | ||
1592 | .poweroff_noirq = dw_suspend_noirq, | ||
1565 | }; | 1593 | }; |
1566 | 1594 | ||
1567 | static struct platform_driver dw_driver = { | 1595 | static struct platform_driver dw_driver = { |
diff --git a/drivers/dma/dw_dmac_regs.h b/drivers/dma/dw_dmac_regs.h index 5eef6946a367..f298f69ecbf9 100644 --- a/drivers/dma/dw_dmac_regs.h +++ b/drivers/dma/dw_dmac_regs.h | |||
@@ -13,6 +13,18 @@ | |||
13 | 13 | ||
14 | #define DW_DMA_MAX_NR_CHANNELS 8 | 14 | #define DW_DMA_MAX_NR_CHANNELS 8 |
15 | 15 | ||
16 | /* flow controller */ | ||
17 | enum dw_dma_fc { | ||
18 | DW_DMA_FC_D_M2M, | ||
19 | DW_DMA_FC_D_M2P, | ||
20 | DW_DMA_FC_D_P2M, | ||
21 | DW_DMA_FC_D_P2P, | ||
22 | DW_DMA_FC_P_P2M, | ||
23 | DW_DMA_FC_SP_P2P, | ||
24 | DW_DMA_FC_P_M2P, | ||
25 | DW_DMA_FC_DP_P2P, | ||
26 | }; | ||
27 | |||
16 | /* | 28 | /* |
17 | * Redefine this macro to handle differences between 32- and 64-bit | 29 | * Redefine this macro to handle differences between 32- and 64-bit |
18 | * addressing, big vs. little endian, etc. | 30 | * addressing, big vs. little endian, etc. |
@@ -146,13 +158,15 @@ struct dw_dma_chan { | |||
146 | 158 | ||
147 | /* these other elements are all protected by lock */ | 159 | /* these other elements are all protected by lock */ |
148 | unsigned long flags; | 160 | unsigned long flags; |
149 | dma_cookie_t completed; | ||
150 | struct list_head active_list; | 161 | struct list_head active_list; |
151 | struct list_head queue; | 162 | struct list_head queue; |
152 | struct list_head free_list; | 163 | struct list_head free_list; |
153 | struct dw_cyclic_desc *cdesc; | 164 | struct dw_cyclic_desc *cdesc; |
154 | 165 | ||
155 | unsigned int descs_allocated; | 166 | unsigned int descs_allocated; |
167 | |||
168 | /* configuration passed via DMA_SLAVE_CONFIG */ | ||
169 | struct dma_slave_config dma_sconfig; | ||
156 | }; | 170 | }; |
157 | 171 | ||
158 | static inline struct dw_dma_chan_regs __iomem * | 172 | static inline struct dw_dma_chan_regs __iomem * |
diff --git a/drivers/dma/ep93xx_dma.c b/drivers/dma/ep93xx_dma.c index 59e7a965772b..e6f133b78dc2 100644 --- a/drivers/dma/ep93xx_dma.c +++ b/drivers/dma/ep93xx_dma.c | |||
@@ -28,6 +28,8 @@ | |||
28 | 28 | ||
29 | #include <mach/dma.h> | 29 | #include <mach/dma.h> |
30 | 30 | ||
31 | #include "dmaengine.h" | ||
32 | |||
31 | /* M2P registers */ | 33 | /* M2P registers */ |
32 | #define M2P_CONTROL 0x0000 | 34 | #define M2P_CONTROL 0x0000 |
33 | #define M2P_CONTROL_STALLINT BIT(0) | 35 | #define M2P_CONTROL_STALLINT BIT(0) |
@@ -122,7 +124,6 @@ struct ep93xx_dma_desc { | |||
122 | * @lock: lock protecting the fields following | 124 | * @lock: lock protecting the fields following |
123 | * @flags: flags for the channel | 125 | * @flags: flags for the channel |
124 | * @buffer: which buffer to use next (0/1) | 126 | * @buffer: which buffer to use next (0/1) |
125 | * @last_completed: last completed cookie value | ||
126 | * @active: flattened chain of descriptors currently being processed | 127 | * @active: flattened chain of descriptors currently being processed |
127 | * @queue: pending descriptors which are handled next | 128 | * @queue: pending descriptors which are handled next |
128 | * @free_list: list of free descriptors which can be used | 129 | * @free_list: list of free descriptors which can be used |
@@ -157,7 +158,6 @@ struct ep93xx_dma_chan { | |||
157 | #define EP93XX_DMA_IS_CYCLIC 0 | 158 | #define EP93XX_DMA_IS_CYCLIC 0 |
158 | 159 | ||
159 | int buffer; | 160 | int buffer; |
160 | dma_cookie_t last_completed; | ||
161 | struct list_head active; | 161 | struct list_head active; |
162 | struct list_head queue; | 162 | struct list_head queue; |
163 | struct list_head free_list; | 163 | struct list_head free_list; |
@@ -703,7 +703,7 @@ static void ep93xx_dma_tasklet(unsigned long data) | |||
703 | desc = ep93xx_dma_get_active(edmac); | 703 | desc = ep93xx_dma_get_active(edmac); |
704 | if (desc) { | 704 | if (desc) { |
705 | if (desc->complete) { | 705 | if (desc->complete) { |
706 | edmac->last_completed = desc->txd.cookie; | 706 | dma_cookie_complete(&desc->txd); |
707 | list_splice_init(&edmac->active, &list); | 707 | list_splice_init(&edmac->active, &list); |
708 | } | 708 | } |
709 | callback = desc->txd.callback; | 709 | callback = desc->txd.callback; |
@@ -783,17 +783,10 @@ static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
783 | unsigned long flags; | 783 | unsigned long flags; |
784 | 784 | ||
785 | spin_lock_irqsave(&edmac->lock, flags); | 785 | spin_lock_irqsave(&edmac->lock, flags); |
786 | 786 | cookie = dma_cookie_assign(tx); | |
787 | cookie = edmac->chan.cookie; | ||
788 | |||
789 | if (++cookie < 0) | ||
790 | cookie = 1; | ||
791 | 787 | ||
792 | desc = container_of(tx, struct ep93xx_dma_desc, txd); | 788 | desc = container_of(tx, struct ep93xx_dma_desc, txd); |
793 | 789 | ||
794 | edmac->chan.cookie = cookie; | ||
795 | desc->txd.cookie = cookie; | ||
796 | |||
797 | /* | 790 | /* |
798 | * If nothing is currently prosessed, we push this descriptor | 791 | * If nothing is currently prosessed, we push this descriptor |
799 | * directly to the hardware. Otherwise we put the descriptor | 792 | * directly to the hardware. Otherwise we put the descriptor |
@@ -861,8 +854,7 @@ static int ep93xx_dma_alloc_chan_resources(struct dma_chan *chan) | |||
861 | goto fail_clk_disable; | 854 | goto fail_clk_disable; |
862 | 855 | ||
863 | spin_lock_irq(&edmac->lock); | 856 | spin_lock_irq(&edmac->lock); |
864 | edmac->last_completed = 1; | 857 | dma_cookie_init(&edmac->chan); |
865 | edmac->chan.cookie = 1; | ||
866 | ret = edmac->edma->hw_setup(edmac); | 858 | ret = edmac->edma->hw_setup(edmac); |
867 | spin_unlock_irq(&edmac->lock); | 859 | spin_unlock_irq(&edmac->lock); |
868 | 860 | ||
@@ -983,13 +975,14 @@ fail: | |||
983 | * @sg_len: number of entries in @sgl | 975 | * @sg_len: number of entries in @sgl |
984 | * @dir: direction of tha DMA transfer | 976 | * @dir: direction of tha DMA transfer |
985 | * @flags: flags for the descriptor | 977 | * @flags: flags for the descriptor |
978 | * @context: operation context (ignored) | ||
986 | * | 979 | * |
987 | * Returns a valid DMA descriptor or %NULL in case of failure. | 980 | * Returns a valid DMA descriptor or %NULL in case of failure. |
988 | */ | 981 | */ |
989 | static struct dma_async_tx_descriptor * | 982 | static struct dma_async_tx_descriptor * |
990 | ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 983 | ep93xx_dma_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
991 | unsigned int sg_len, enum dma_transfer_direction dir, | 984 | unsigned int sg_len, enum dma_transfer_direction dir, |
992 | unsigned long flags) | 985 | unsigned long flags, void *context) |
993 | { | 986 | { |
994 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); | 987 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); |
995 | struct ep93xx_dma_desc *desc, *first; | 988 | struct ep93xx_dma_desc *desc, *first; |
@@ -1056,6 +1049,7 @@ fail: | |||
1056 | * @buf_len: length of the buffer (in bytes) | 1049 | * @buf_len: length of the buffer (in bytes) |
1057 | * @period_len: lenght of a single period | 1050 | * @period_len: lenght of a single period |
1058 | * @dir: direction of the operation | 1051 | * @dir: direction of the operation |
1052 | * @context: operation context (ignored) | ||
1059 | * | 1053 | * |
1060 | * Prepares a descriptor for cyclic DMA operation. This means that once the | 1054 | * Prepares a descriptor for cyclic DMA operation. This means that once the |
1061 | * descriptor is submitted, we will be submitting in a @period_len sized | 1055 | * descriptor is submitted, we will be submitting in a @period_len sized |
@@ -1068,7 +1062,7 @@ fail: | |||
1068 | static struct dma_async_tx_descriptor * | 1062 | static struct dma_async_tx_descriptor * |
1069 | ep93xx_dma_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr, | 1063 | ep93xx_dma_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr, |
1070 | size_t buf_len, size_t period_len, | 1064 | size_t buf_len, size_t period_len, |
1071 | enum dma_transfer_direction dir) | 1065 | enum dma_transfer_direction dir, void *context) |
1072 | { | 1066 | { |
1073 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); | 1067 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); |
1074 | struct ep93xx_dma_desc *desc, *first; | 1068 | struct ep93xx_dma_desc *desc, *first; |
@@ -1248,18 +1242,13 @@ static enum dma_status ep93xx_dma_tx_status(struct dma_chan *chan, | |||
1248 | struct dma_tx_state *state) | 1242 | struct dma_tx_state *state) |
1249 | { | 1243 | { |
1250 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); | 1244 | struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(chan); |
1251 | dma_cookie_t last_used, last_completed; | ||
1252 | enum dma_status ret; | 1245 | enum dma_status ret; |
1253 | unsigned long flags; | 1246 | unsigned long flags; |
1254 | 1247 | ||
1255 | spin_lock_irqsave(&edmac->lock, flags); | 1248 | spin_lock_irqsave(&edmac->lock, flags); |
1256 | last_used = chan->cookie; | 1249 | ret = dma_cookie_status(chan, cookie, state); |
1257 | last_completed = edmac->last_completed; | ||
1258 | spin_unlock_irqrestore(&edmac->lock, flags); | 1250 | spin_unlock_irqrestore(&edmac->lock, flags); |
1259 | 1251 | ||
1260 | ret = dma_async_is_complete(cookie, last_completed, last_used); | ||
1261 | dma_set_tx_state(state, last_completed, last_used, 0); | ||
1262 | |||
1263 | return ret; | 1252 | return ret; |
1264 | } | 1253 | } |
1265 | 1254 | ||
diff --git a/drivers/dma/fsldma.c b/drivers/dma/fsldma.c index b98070c33ca9..8f84761f98ba 100644 --- a/drivers/dma/fsldma.c +++ b/drivers/dma/fsldma.c | |||
@@ -35,6 +35,7 @@ | |||
35 | #include <linux/dmapool.h> | 35 | #include <linux/dmapool.h> |
36 | #include <linux/of_platform.h> | 36 | #include <linux/of_platform.h> |
37 | 37 | ||
38 | #include "dmaengine.h" | ||
38 | #include "fsldma.h" | 39 | #include "fsldma.h" |
39 | 40 | ||
40 | #define chan_dbg(chan, fmt, arg...) \ | 41 | #define chan_dbg(chan, fmt, arg...) \ |
@@ -413,17 +414,10 @@ static dma_cookie_t fsl_dma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
413 | * assign cookies to all of the software descriptors | 414 | * assign cookies to all of the software descriptors |
414 | * that make up this transaction | 415 | * that make up this transaction |
415 | */ | 416 | */ |
416 | cookie = chan->common.cookie; | ||
417 | list_for_each_entry(child, &desc->tx_list, node) { | 417 | list_for_each_entry(child, &desc->tx_list, node) { |
418 | cookie++; | 418 | cookie = dma_cookie_assign(&child->async_tx); |
419 | if (cookie < DMA_MIN_COOKIE) | ||
420 | cookie = DMA_MIN_COOKIE; | ||
421 | |||
422 | child->async_tx.cookie = cookie; | ||
423 | } | 419 | } |
424 | 420 | ||
425 | chan->common.cookie = cookie; | ||
426 | |||
427 | /* put this transaction onto the tail of the pending queue */ | 421 | /* put this transaction onto the tail of the pending queue */ |
428 | append_ld_queue(chan, desc); | 422 | append_ld_queue(chan, desc); |
429 | 423 | ||
@@ -765,6 +759,7 @@ fail: | |||
765 | * @sg_len: number of entries in @scatterlist | 759 | * @sg_len: number of entries in @scatterlist |
766 | * @direction: DMA direction | 760 | * @direction: DMA direction |
767 | * @flags: DMAEngine flags | 761 | * @flags: DMAEngine flags |
762 | * @context: transaction context (ignored) | ||
768 | * | 763 | * |
769 | * Prepare a set of descriptors for a DMA_SLAVE transaction. Following the | 764 | * Prepare a set of descriptors for a DMA_SLAVE transaction. Following the |
770 | * DMA_SLAVE API, this gets the device-specific information from the | 765 | * DMA_SLAVE API, this gets the device-specific information from the |
@@ -772,7 +767,8 @@ fail: | |||
772 | */ | 767 | */ |
773 | static struct dma_async_tx_descriptor *fsl_dma_prep_slave_sg( | 768 | static struct dma_async_tx_descriptor *fsl_dma_prep_slave_sg( |
774 | struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, | 769 | struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, |
775 | enum dma_transfer_direction direction, unsigned long flags) | 770 | enum dma_transfer_direction direction, unsigned long flags, |
771 | void *context) | ||
776 | { | 772 | { |
777 | /* | 773 | /* |
778 | * This operation is not supported on the Freescale DMA controller | 774 | * This operation is not supported on the Freescale DMA controller |
@@ -984,19 +980,14 @@ static enum dma_status fsl_tx_status(struct dma_chan *dchan, | |||
984 | struct dma_tx_state *txstate) | 980 | struct dma_tx_state *txstate) |
985 | { | 981 | { |
986 | struct fsldma_chan *chan = to_fsl_chan(dchan); | 982 | struct fsldma_chan *chan = to_fsl_chan(dchan); |
987 | dma_cookie_t last_complete; | 983 | enum dma_status ret; |
988 | dma_cookie_t last_used; | ||
989 | unsigned long flags; | 984 | unsigned long flags; |
990 | 985 | ||
991 | spin_lock_irqsave(&chan->desc_lock, flags); | 986 | spin_lock_irqsave(&chan->desc_lock, flags); |
992 | 987 | ret = dma_cookie_status(dchan, cookie, txstate); | |
993 | last_complete = chan->completed_cookie; | ||
994 | last_used = dchan->cookie; | ||
995 | |||
996 | spin_unlock_irqrestore(&chan->desc_lock, flags); | 988 | spin_unlock_irqrestore(&chan->desc_lock, flags); |
997 | 989 | ||
998 | dma_set_tx_state(txstate, last_complete, last_used, 0); | 990 | return ret; |
999 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
1000 | } | 991 | } |
1001 | 992 | ||
1002 | /*----------------------------------------------------------------------------*/ | 993 | /*----------------------------------------------------------------------------*/ |
@@ -1087,8 +1078,8 @@ static void dma_do_tasklet(unsigned long data) | |||
1087 | 1078 | ||
1088 | desc = to_fsl_desc(chan->ld_running.prev); | 1079 | desc = to_fsl_desc(chan->ld_running.prev); |
1089 | cookie = desc->async_tx.cookie; | 1080 | cookie = desc->async_tx.cookie; |
1081 | dma_cookie_complete(&desc->async_tx); | ||
1090 | 1082 | ||
1091 | chan->completed_cookie = cookie; | ||
1092 | chan_dbg(chan, "completed_cookie=%d\n", cookie); | 1083 | chan_dbg(chan, "completed_cookie=%d\n", cookie); |
1093 | } | 1084 | } |
1094 | 1085 | ||
@@ -1303,6 +1294,7 @@ static int __devinit fsl_dma_chan_probe(struct fsldma_device *fdev, | |||
1303 | chan->idle = true; | 1294 | chan->idle = true; |
1304 | 1295 | ||
1305 | chan->common.device = &fdev->common; | 1296 | chan->common.device = &fdev->common; |
1297 | dma_cookie_init(&chan->common); | ||
1306 | 1298 | ||
1307 | /* find the IRQ line, if it exists in the device tree */ | 1299 | /* find the IRQ line, if it exists in the device tree */ |
1308 | chan->irq = irq_of_parse_and_map(node, 0); | 1300 | chan->irq = irq_of_parse_and_map(node, 0); |
diff --git a/drivers/dma/fsldma.h b/drivers/dma/fsldma.h index 9cb5aa57c677..f5c38791fc74 100644 --- a/drivers/dma/fsldma.h +++ b/drivers/dma/fsldma.h | |||
@@ -137,7 +137,6 @@ struct fsldma_device { | |||
137 | struct fsldma_chan { | 137 | struct fsldma_chan { |
138 | char name[8]; /* Channel name */ | 138 | char name[8]; /* Channel name */ |
139 | struct fsldma_chan_regs __iomem *regs; | 139 | struct fsldma_chan_regs __iomem *regs; |
140 | dma_cookie_t completed_cookie; /* The maximum cookie completed */ | ||
141 | spinlock_t desc_lock; /* Descriptor operation lock */ | 140 | spinlock_t desc_lock; /* Descriptor operation lock */ |
142 | struct list_head ld_pending; /* Link descriptors queue */ | 141 | struct list_head ld_pending; /* Link descriptors queue */ |
143 | struct list_head ld_running; /* Link descriptors queue */ | 142 | struct list_head ld_running; /* Link descriptors queue */ |
diff --git a/drivers/dma/imx-dma.c b/drivers/dma/imx-dma.c index 38586ba8da91..a45b5d2a5987 100644 --- a/drivers/dma/imx-dma.c +++ b/drivers/dma/imx-dma.c | |||
@@ -5,6 +5,7 @@ | |||
5 | * found on i.MX1/21/27 | 5 | * found on i.MX1/21/27 |
6 | * | 6 | * |
7 | * Copyright 2010 Sascha Hauer, Pengutronix <s.hauer@pengutronix.de> | 7 | * Copyright 2010 Sascha Hauer, Pengutronix <s.hauer@pengutronix.de> |
8 | * Copyright 2012 Javier Martin, Vista Silicon <javier.martin@vista-silicon.com> | ||
8 | * | 9 | * |
9 | * The code contained herein is licensed under the GNU General Public | 10 | * The code contained herein is licensed under the GNU General Public |
10 | * License. You may obtain a copy of the GNU General Public License | 11 | * License. You may obtain a copy of the GNU General Public License |
@@ -22,37 +23,159 @@ | |||
22 | #include <linux/dma-mapping.h> | 23 | #include <linux/dma-mapping.h> |
23 | #include <linux/slab.h> | 24 | #include <linux/slab.h> |
24 | #include <linux/platform_device.h> | 25 | #include <linux/platform_device.h> |
26 | #include <linux/clk.h> | ||
25 | #include <linux/dmaengine.h> | 27 | #include <linux/dmaengine.h> |
26 | #include <linux/module.h> | 28 | #include <linux/module.h> |
27 | 29 | ||
28 | #include <asm/irq.h> | 30 | #include <asm/irq.h> |
29 | #include <mach/dma-v1.h> | 31 | #include <mach/dma.h> |
30 | #include <mach/hardware.h> | 32 | #include <mach/hardware.h> |
31 | 33 | ||
34 | #include "dmaengine.h" | ||
35 | #define IMXDMA_MAX_CHAN_DESCRIPTORS 16 | ||
36 | #define IMX_DMA_CHANNELS 16 | ||
37 | |||
38 | #define IMX_DMA_2D_SLOTS 2 | ||
39 | #define IMX_DMA_2D_SLOT_A 0 | ||
40 | #define IMX_DMA_2D_SLOT_B 1 | ||
41 | |||
42 | #define IMX_DMA_LENGTH_LOOP ((unsigned int)-1) | ||
43 | #define IMX_DMA_MEMSIZE_32 (0 << 4) | ||
44 | #define IMX_DMA_MEMSIZE_8 (1 << 4) | ||
45 | #define IMX_DMA_MEMSIZE_16 (2 << 4) | ||
46 | #define IMX_DMA_TYPE_LINEAR (0 << 10) | ||
47 | #define IMX_DMA_TYPE_2D (1 << 10) | ||
48 | #define IMX_DMA_TYPE_FIFO (2 << 10) | ||
49 | |||
50 | #define IMX_DMA_ERR_BURST (1 << 0) | ||
51 | #define IMX_DMA_ERR_REQUEST (1 << 1) | ||
52 | #define IMX_DMA_ERR_TRANSFER (1 << 2) | ||
53 | #define IMX_DMA_ERR_BUFFER (1 << 3) | ||
54 | #define IMX_DMA_ERR_TIMEOUT (1 << 4) | ||
55 | |||
56 | #define DMA_DCR 0x00 /* Control Register */ | ||
57 | #define DMA_DISR 0x04 /* Interrupt status Register */ | ||
58 | #define DMA_DIMR 0x08 /* Interrupt mask Register */ | ||
59 | #define DMA_DBTOSR 0x0c /* Burst timeout status Register */ | ||
60 | #define DMA_DRTOSR 0x10 /* Request timeout Register */ | ||
61 | #define DMA_DSESR 0x14 /* Transfer Error Status Register */ | ||
62 | #define DMA_DBOSR 0x18 /* Buffer overflow status Register */ | ||
63 | #define DMA_DBTOCR 0x1c /* Burst timeout control Register */ | ||
64 | #define DMA_WSRA 0x40 /* W-Size Register A */ | ||
65 | #define DMA_XSRA 0x44 /* X-Size Register A */ | ||
66 | #define DMA_YSRA 0x48 /* Y-Size Register A */ | ||
67 | #define DMA_WSRB 0x4c /* W-Size Register B */ | ||
68 | #define DMA_XSRB 0x50 /* X-Size Register B */ | ||
69 | #define DMA_YSRB 0x54 /* Y-Size Register B */ | ||
70 | #define DMA_SAR(x) (0x80 + ((x) << 6)) /* Source Address Registers */ | ||
71 | #define DMA_DAR(x) (0x84 + ((x) << 6)) /* Destination Address Registers */ | ||
72 | #define DMA_CNTR(x) (0x88 + ((x) << 6)) /* Count Registers */ | ||
73 | #define DMA_CCR(x) (0x8c + ((x) << 6)) /* Control Registers */ | ||
74 | #define DMA_RSSR(x) (0x90 + ((x) << 6)) /* Request source select Registers */ | ||
75 | #define DMA_BLR(x) (0x94 + ((x) << 6)) /* Burst length Registers */ | ||
76 | #define DMA_RTOR(x) (0x98 + ((x) << 6)) /* Request timeout Registers */ | ||
77 | #define DMA_BUCR(x) (0x98 + ((x) << 6)) /* Bus Utilization Registers */ | ||
78 | #define DMA_CCNR(x) (0x9C + ((x) << 6)) /* Channel counter Registers */ | ||
79 | |||
80 | #define DCR_DRST (1<<1) | ||
81 | #define DCR_DEN (1<<0) | ||
82 | #define DBTOCR_EN (1<<15) | ||
83 | #define DBTOCR_CNT(x) ((x) & 0x7fff) | ||
84 | #define CNTR_CNT(x) ((x) & 0xffffff) | ||
85 | #define CCR_ACRPT (1<<14) | ||
86 | #define CCR_DMOD_LINEAR (0x0 << 12) | ||
87 | #define CCR_DMOD_2D (0x1 << 12) | ||
88 | #define CCR_DMOD_FIFO (0x2 << 12) | ||
89 | #define CCR_DMOD_EOBFIFO (0x3 << 12) | ||
90 | #define CCR_SMOD_LINEAR (0x0 << 10) | ||
91 | #define CCR_SMOD_2D (0x1 << 10) | ||
92 | #define CCR_SMOD_FIFO (0x2 << 10) | ||
93 | #define CCR_SMOD_EOBFIFO (0x3 << 10) | ||
94 | #define CCR_MDIR_DEC (1<<9) | ||
95 | #define CCR_MSEL_B (1<<8) | ||
96 | #define CCR_DSIZ_32 (0x0 << 6) | ||
97 | #define CCR_DSIZ_8 (0x1 << 6) | ||
98 | #define CCR_DSIZ_16 (0x2 << 6) | ||
99 | #define CCR_SSIZ_32 (0x0 << 4) | ||
100 | #define CCR_SSIZ_8 (0x1 << 4) | ||
101 | #define CCR_SSIZ_16 (0x2 << 4) | ||
102 | #define CCR_REN (1<<3) | ||
103 | #define CCR_RPT (1<<2) | ||
104 | #define CCR_FRC (1<<1) | ||
105 | #define CCR_CEN (1<<0) | ||
106 | #define RTOR_EN (1<<15) | ||
107 | #define RTOR_CLK (1<<14) | ||
108 | #define RTOR_PSC (1<<13) | ||
109 | |||
110 | enum imxdma_prep_type { | ||
111 | IMXDMA_DESC_MEMCPY, | ||
112 | IMXDMA_DESC_INTERLEAVED, | ||
113 | IMXDMA_DESC_SLAVE_SG, | ||
114 | IMXDMA_DESC_CYCLIC, | ||
115 | }; | ||
116 | |||
117 | struct imx_dma_2d_config { | ||
118 | u16 xsr; | ||
119 | u16 ysr; | ||
120 | u16 wsr; | ||
121 | int count; | ||
122 | }; | ||
123 | |||
124 | struct imxdma_desc { | ||
125 | struct list_head node; | ||
126 | struct dma_async_tx_descriptor desc; | ||
127 | enum dma_status status; | ||
128 | dma_addr_t src; | ||
129 | dma_addr_t dest; | ||
130 | size_t len; | ||
131 | enum dma_transfer_direction direction; | ||
132 | enum imxdma_prep_type type; | ||
133 | /* For memcpy and interleaved */ | ||
134 | unsigned int config_port; | ||
135 | unsigned int config_mem; | ||
136 | /* For interleaved transfers */ | ||
137 | unsigned int x; | ||
138 | unsigned int y; | ||
139 | unsigned int w; | ||
140 | /* For slave sg and cyclic */ | ||
141 | struct scatterlist *sg; | ||
142 | unsigned int sgcount; | ||
143 | }; | ||
144 | |||
32 | struct imxdma_channel { | 145 | struct imxdma_channel { |
146 | int hw_chaining; | ||
147 | struct timer_list watchdog; | ||
33 | struct imxdma_engine *imxdma; | 148 | struct imxdma_engine *imxdma; |
34 | unsigned int channel; | 149 | unsigned int channel; |
35 | unsigned int imxdma_channel; | ||
36 | 150 | ||
151 | struct tasklet_struct dma_tasklet; | ||
152 | struct list_head ld_free; | ||
153 | struct list_head ld_queue; | ||
154 | struct list_head ld_active; | ||
155 | int descs_allocated; | ||
37 | enum dma_slave_buswidth word_size; | 156 | enum dma_slave_buswidth word_size; |
38 | dma_addr_t per_address; | 157 | dma_addr_t per_address; |
39 | u32 watermark_level; | 158 | u32 watermark_level; |
40 | struct dma_chan chan; | 159 | struct dma_chan chan; |
41 | spinlock_t lock; | ||
42 | struct dma_async_tx_descriptor desc; | 160 | struct dma_async_tx_descriptor desc; |
43 | dma_cookie_t last_completed; | ||
44 | enum dma_status status; | 161 | enum dma_status status; |
45 | int dma_request; | 162 | int dma_request; |
46 | struct scatterlist *sg_list; | 163 | struct scatterlist *sg_list; |
164 | u32 ccr_from_device; | ||
165 | u32 ccr_to_device; | ||
166 | bool enabled_2d; | ||
167 | int slot_2d; | ||
47 | }; | 168 | }; |
48 | 169 | ||
49 | #define MAX_DMA_CHANNELS 8 | ||
50 | |||
51 | struct imxdma_engine { | 170 | struct imxdma_engine { |
52 | struct device *dev; | 171 | struct device *dev; |
53 | struct device_dma_parameters dma_parms; | 172 | struct device_dma_parameters dma_parms; |
54 | struct dma_device dma_device; | 173 | struct dma_device dma_device; |
55 | struct imxdma_channel channel[MAX_DMA_CHANNELS]; | 174 | void __iomem *base; |
175 | struct clk *dma_clk; | ||
176 | spinlock_t lock; | ||
177 | struct imx_dma_2d_config slots_2d[IMX_DMA_2D_SLOTS]; | ||
178 | struct imxdma_channel channel[IMX_DMA_CHANNELS]; | ||
56 | }; | 179 | }; |
57 | 180 | ||
58 | static struct imxdma_channel *to_imxdma_chan(struct dma_chan *chan) | 181 | static struct imxdma_channel *to_imxdma_chan(struct dma_chan *chan) |
@@ -60,36 +183,418 @@ static struct imxdma_channel *to_imxdma_chan(struct dma_chan *chan) | |||
60 | return container_of(chan, struct imxdma_channel, chan); | 183 | return container_of(chan, struct imxdma_channel, chan); |
61 | } | 184 | } |
62 | 185 | ||
63 | static void imxdma_handle(struct imxdma_channel *imxdmac) | 186 | static inline bool imxdma_chan_is_doing_cyclic(struct imxdma_channel *imxdmac) |
187 | { | ||
188 | struct imxdma_desc *desc; | ||
189 | |||
190 | if (!list_empty(&imxdmac->ld_active)) { | ||
191 | desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, | ||
192 | node); | ||
193 | if (desc->type == IMXDMA_DESC_CYCLIC) | ||
194 | return true; | ||
195 | } | ||
196 | return false; | ||
197 | } | ||
198 | |||
199 | |||
200 | |||
201 | static void imx_dmav1_writel(struct imxdma_engine *imxdma, unsigned val, | ||
202 | unsigned offset) | ||
203 | { | ||
204 | __raw_writel(val, imxdma->base + offset); | ||
205 | } | ||
206 | |||
207 | static unsigned imx_dmav1_readl(struct imxdma_engine *imxdma, unsigned offset) | ||
208 | { | ||
209 | return __raw_readl(imxdma->base + offset); | ||
210 | } | ||
211 | |||
212 | static int imxdma_hw_chain(struct imxdma_channel *imxdmac) | ||
213 | { | ||
214 | if (cpu_is_mx27()) | ||
215 | return imxdmac->hw_chaining; | ||
216 | else | ||
217 | return 0; | ||
218 | } | ||
219 | |||
220 | /* | ||
221 | * imxdma_sg_next - prepare next chunk for scatter-gather DMA emulation | ||
222 | */ | ||
223 | static inline int imxdma_sg_next(struct imxdma_desc *d) | ||
64 | { | 224 | { |
65 | if (imxdmac->desc.callback) | 225 | struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); |
66 | imxdmac->desc.callback(imxdmac->desc.callback_param); | 226 | struct imxdma_engine *imxdma = imxdmac->imxdma; |
67 | imxdmac->last_completed = imxdmac->desc.cookie; | 227 | struct scatterlist *sg = d->sg; |
228 | unsigned long now; | ||
229 | |||
230 | now = min(d->len, sg->length); | ||
231 | if (d->len != IMX_DMA_LENGTH_LOOP) | ||
232 | d->len -= now; | ||
233 | |||
234 | if (d->direction == DMA_DEV_TO_MEM) | ||
235 | imx_dmav1_writel(imxdma, sg->dma_address, | ||
236 | DMA_DAR(imxdmac->channel)); | ||
237 | else | ||
238 | imx_dmav1_writel(imxdma, sg->dma_address, | ||
239 | DMA_SAR(imxdmac->channel)); | ||
240 | |||
241 | imx_dmav1_writel(imxdma, now, DMA_CNTR(imxdmac->channel)); | ||
242 | |||
243 | dev_dbg(imxdma->dev, " %s channel: %d dst 0x%08x, src 0x%08x, " | ||
244 | "size 0x%08x\n", __func__, imxdmac->channel, | ||
245 | imx_dmav1_readl(imxdma, DMA_DAR(imxdmac->channel)), | ||
246 | imx_dmav1_readl(imxdma, DMA_SAR(imxdmac->channel)), | ||
247 | imx_dmav1_readl(imxdma, DMA_CNTR(imxdmac->channel))); | ||
248 | |||
249 | return now; | ||
68 | } | 250 | } |
69 | 251 | ||
70 | static void imxdma_irq_handler(int channel, void *data) | 252 | static void imxdma_enable_hw(struct imxdma_desc *d) |
71 | { | 253 | { |
72 | struct imxdma_channel *imxdmac = data; | 254 | struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); |
255 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
256 | int channel = imxdmac->channel; | ||
257 | unsigned long flags; | ||
258 | |||
259 | dev_dbg(imxdma->dev, "%s channel %d\n", __func__, channel); | ||
260 | |||
261 | local_irq_save(flags); | ||
262 | |||
263 | imx_dmav1_writel(imxdma, 1 << channel, DMA_DISR); | ||
264 | imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_DIMR) & | ||
265 | ~(1 << channel), DMA_DIMR); | ||
266 | imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_CCR(channel)) | | ||
267 | CCR_CEN | CCR_ACRPT, DMA_CCR(channel)); | ||
268 | |||
269 | if ((cpu_is_mx21() || cpu_is_mx27()) && | ||
270 | d->sg && imxdma_hw_chain(imxdmac)) { | ||
271 | d->sg = sg_next(d->sg); | ||
272 | if (d->sg) { | ||
273 | u32 tmp; | ||
274 | imxdma_sg_next(d); | ||
275 | tmp = imx_dmav1_readl(imxdma, DMA_CCR(channel)); | ||
276 | imx_dmav1_writel(imxdma, tmp | CCR_RPT | CCR_ACRPT, | ||
277 | DMA_CCR(channel)); | ||
278 | } | ||
279 | } | ||
280 | |||
281 | local_irq_restore(flags); | ||
282 | } | ||
283 | |||
284 | static void imxdma_disable_hw(struct imxdma_channel *imxdmac) | ||
285 | { | ||
286 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
287 | int channel = imxdmac->channel; | ||
288 | unsigned long flags; | ||
289 | |||
290 | dev_dbg(imxdma->dev, "%s channel %d\n", __func__, channel); | ||
291 | |||
292 | if (imxdma_hw_chain(imxdmac)) | ||
293 | del_timer(&imxdmac->watchdog); | ||
294 | |||
295 | local_irq_save(flags); | ||
296 | imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_DIMR) | | ||
297 | (1 << channel), DMA_DIMR); | ||
298 | imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_CCR(channel)) & | ||
299 | ~CCR_CEN, DMA_CCR(channel)); | ||
300 | imx_dmav1_writel(imxdma, 1 << channel, DMA_DISR); | ||
301 | local_irq_restore(flags); | ||
302 | } | ||
303 | |||
304 | static void imxdma_watchdog(unsigned long data) | ||
305 | { | ||
306 | struct imxdma_channel *imxdmac = (struct imxdma_channel *)data; | ||
307 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
308 | int channel = imxdmac->channel; | ||
309 | |||
310 | imx_dmav1_writel(imxdma, 0, DMA_CCR(channel)); | ||
73 | 311 | ||
74 | imxdmac->status = DMA_SUCCESS; | 312 | /* Tasklet watchdog error handler */ |
75 | imxdma_handle(imxdmac); | 313 | tasklet_schedule(&imxdmac->dma_tasklet); |
314 | dev_dbg(imxdma->dev, "channel %d: watchdog timeout!\n", | ||
315 | imxdmac->channel); | ||
76 | } | 316 | } |
77 | 317 | ||
78 | static void imxdma_err_handler(int channel, void *data, int error) | 318 | static irqreturn_t imxdma_err_handler(int irq, void *dev_id) |
79 | { | 319 | { |
80 | struct imxdma_channel *imxdmac = data; | 320 | struct imxdma_engine *imxdma = dev_id; |
321 | unsigned int err_mask; | ||
322 | int i, disr; | ||
323 | int errcode; | ||
324 | |||
325 | disr = imx_dmav1_readl(imxdma, DMA_DISR); | ||
326 | |||
327 | err_mask = imx_dmav1_readl(imxdma, DMA_DBTOSR) | | ||
328 | imx_dmav1_readl(imxdma, DMA_DRTOSR) | | ||
329 | imx_dmav1_readl(imxdma, DMA_DSESR) | | ||
330 | imx_dmav1_readl(imxdma, DMA_DBOSR); | ||
331 | |||
332 | if (!err_mask) | ||
333 | return IRQ_HANDLED; | ||
334 | |||
335 | imx_dmav1_writel(imxdma, disr & err_mask, DMA_DISR); | ||
336 | |||
337 | for (i = 0; i < IMX_DMA_CHANNELS; i++) { | ||
338 | if (!(err_mask & (1 << i))) | ||
339 | continue; | ||
340 | errcode = 0; | ||
341 | |||
342 | if (imx_dmav1_readl(imxdma, DMA_DBTOSR) & (1 << i)) { | ||
343 | imx_dmav1_writel(imxdma, 1 << i, DMA_DBTOSR); | ||
344 | errcode |= IMX_DMA_ERR_BURST; | ||
345 | } | ||
346 | if (imx_dmav1_readl(imxdma, DMA_DRTOSR) & (1 << i)) { | ||
347 | imx_dmav1_writel(imxdma, 1 << i, DMA_DRTOSR); | ||
348 | errcode |= IMX_DMA_ERR_REQUEST; | ||
349 | } | ||
350 | if (imx_dmav1_readl(imxdma, DMA_DSESR) & (1 << i)) { | ||
351 | imx_dmav1_writel(imxdma, 1 << i, DMA_DSESR); | ||
352 | errcode |= IMX_DMA_ERR_TRANSFER; | ||
353 | } | ||
354 | if (imx_dmav1_readl(imxdma, DMA_DBOSR) & (1 << i)) { | ||
355 | imx_dmav1_writel(imxdma, 1 << i, DMA_DBOSR); | ||
356 | errcode |= IMX_DMA_ERR_BUFFER; | ||
357 | } | ||
358 | /* Tasklet error handler */ | ||
359 | tasklet_schedule(&imxdma->channel[i].dma_tasklet); | ||
360 | |||
361 | printk(KERN_WARNING | ||
362 | "DMA timeout on channel %d -%s%s%s%s\n", i, | ||
363 | errcode & IMX_DMA_ERR_BURST ? " burst" : "", | ||
364 | errcode & IMX_DMA_ERR_REQUEST ? " request" : "", | ||
365 | errcode & IMX_DMA_ERR_TRANSFER ? " transfer" : "", | ||
366 | errcode & IMX_DMA_ERR_BUFFER ? " buffer" : ""); | ||
367 | } | ||
368 | return IRQ_HANDLED; | ||
369 | } | ||
370 | |||
371 | static void dma_irq_handle_channel(struct imxdma_channel *imxdmac) | ||
372 | { | ||
373 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
374 | int chno = imxdmac->channel; | ||
375 | struct imxdma_desc *desc; | ||
376 | |||
377 | spin_lock(&imxdma->lock); | ||
378 | if (list_empty(&imxdmac->ld_active)) { | ||
379 | spin_unlock(&imxdma->lock); | ||
380 | goto out; | ||
381 | } | ||
382 | |||
383 | desc = list_first_entry(&imxdmac->ld_active, | ||
384 | struct imxdma_desc, | ||
385 | node); | ||
386 | spin_unlock(&imxdma->lock); | ||
387 | |||
388 | if (desc->sg) { | ||
389 | u32 tmp; | ||
390 | desc->sg = sg_next(desc->sg); | ||
391 | |||
392 | if (desc->sg) { | ||
393 | imxdma_sg_next(desc); | ||
394 | |||
395 | tmp = imx_dmav1_readl(imxdma, DMA_CCR(chno)); | ||
396 | |||
397 | if (imxdma_hw_chain(imxdmac)) { | ||
398 | /* FIXME: The timeout should probably be | ||
399 | * configurable | ||
400 | */ | ||
401 | mod_timer(&imxdmac->watchdog, | ||
402 | jiffies + msecs_to_jiffies(500)); | ||
403 | |||
404 | tmp |= CCR_CEN | CCR_RPT | CCR_ACRPT; | ||
405 | imx_dmav1_writel(imxdma, tmp, DMA_CCR(chno)); | ||
406 | } else { | ||
407 | imx_dmav1_writel(imxdma, tmp & ~CCR_CEN, | ||
408 | DMA_CCR(chno)); | ||
409 | tmp |= CCR_CEN; | ||
410 | } | ||
411 | |||
412 | imx_dmav1_writel(imxdma, tmp, DMA_CCR(chno)); | ||
413 | |||
414 | if (imxdma_chan_is_doing_cyclic(imxdmac)) | ||
415 | /* Tasklet progression */ | ||
416 | tasklet_schedule(&imxdmac->dma_tasklet); | ||
417 | |||
418 | return; | ||
419 | } | ||
420 | |||
421 | if (imxdma_hw_chain(imxdmac)) { | ||
422 | del_timer(&imxdmac->watchdog); | ||
423 | return; | ||
424 | } | ||
425 | } | ||
426 | |||
427 | out: | ||
428 | imx_dmav1_writel(imxdma, 0, DMA_CCR(chno)); | ||
429 | /* Tasklet irq */ | ||
430 | tasklet_schedule(&imxdmac->dma_tasklet); | ||
431 | } | ||
432 | |||
433 | static irqreturn_t dma_irq_handler(int irq, void *dev_id) | ||
434 | { | ||
435 | struct imxdma_engine *imxdma = dev_id; | ||
436 | int i, disr; | ||
437 | |||
438 | if (cpu_is_mx21() || cpu_is_mx27()) | ||
439 | imxdma_err_handler(irq, dev_id); | ||
440 | |||
441 | disr = imx_dmav1_readl(imxdma, DMA_DISR); | ||
442 | |||
443 | dev_dbg(imxdma->dev, "%s called, disr=0x%08x\n", __func__, disr); | ||
444 | |||
445 | imx_dmav1_writel(imxdma, disr, DMA_DISR); | ||
446 | for (i = 0; i < IMX_DMA_CHANNELS; i++) { | ||
447 | if (disr & (1 << i)) | ||
448 | dma_irq_handle_channel(&imxdma->channel[i]); | ||
449 | } | ||
450 | |||
451 | return IRQ_HANDLED; | ||
452 | } | ||
453 | |||
454 | static int imxdma_xfer_desc(struct imxdma_desc *d) | ||
455 | { | ||
456 | struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); | ||
457 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
458 | unsigned long flags; | ||
459 | int slot = -1; | ||
460 | int i; | ||
461 | |||
462 | /* Configure and enable */ | ||
463 | switch (d->type) { | ||
464 | case IMXDMA_DESC_INTERLEAVED: | ||
465 | /* Try to get a free 2D slot */ | ||
466 | spin_lock_irqsave(&imxdma->lock, flags); | ||
467 | for (i = 0; i < IMX_DMA_2D_SLOTS; i++) { | ||
468 | if ((imxdma->slots_2d[i].count > 0) && | ||
469 | ((imxdma->slots_2d[i].xsr != d->x) || | ||
470 | (imxdma->slots_2d[i].ysr != d->y) || | ||
471 | (imxdma->slots_2d[i].wsr != d->w))) | ||
472 | continue; | ||
473 | slot = i; | ||
474 | break; | ||
475 | } | ||
476 | if (slot < 0) | ||
477 | return -EBUSY; | ||
478 | |||
479 | imxdma->slots_2d[slot].xsr = d->x; | ||
480 | imxdma->slots_2d[slot].ysr = d->y; | ||
481 | imxdma->slots_2d[slot].wsr = d->w; | ||
482 | imxdma->slots_2d[slot].count++; | ||
483 | |||
484 | imxdmac->slot_2d = slot; | ||
485 | imxdmac->enabled_2d = true; | ||
486 | spin_unlock_irqrestore(&imxdma->lock, flags); | ||
487 | |||
488 | if (slot == IMX_DMA_2D_SLOT_A) { | ||
489 | d->config_mem &= ~CCR_MSEL_B; | ||
490 | d->config_port &= ~CCR_MSEL_B; | ||
491 | imx_dmav1_writel(imxdma, d->x, DMA_XSRA); | ||
492 | imx_dmav1_writel(imxdma, d->y, DMA_YSRA); | ||
493 | imx_dmav1_writel(imxdma, d->w, DMA_WSRA); | ||
494 | } else { | ||
495 | d->config_mem |= CCR_MSEL_B; | ||
496 | d->config_port |= CCR_MSEL_B; | ||
497 | imx_dmav1_writel(imxdma, d->x, DMA_XSRB); | ||
498 | imx_dmav1_writel(imxdma, d->y, DMA_YSRB); | ||
499 | imx_dmav1_writel(imxdma, d->w, DMA_WSRB); | ||
500 | } | ||
501 | /* | ||
502 | * We fall-through here intentionally, since a 2D transfer is | ||
503 | * similar to MEMCPY just adding the 2D slot configuration. | ||
504 | */ | ||
505 | case IMXDMA_DESC_MEMCPY: | ||
506 | imx_dmav1_writel(imxdma, d->src, DMA_SAR(imxdmac->channel)); | ||
507 | imx_dmav1_writel(imxdma, d->dest, DMA_DAR(imxdmac->channel)); | ||
508 | imx_dmav1_writel(imxdma, d->config_mem | (d->config_port << 2), | ||
509 | DMA_CCR(imxdmac->channel)); | ||
510 | |||
511 | imx_dmav1_writel(imxdma, d->len, DMA_CNTR(imxdmac->channel)); | ||
512 | |||
513 | dev_dbg(imxdma->dev, "%s channel: %d dest=0x%08x src=0x%08x " | ||
514 | "dma_length=%d\n", __func__, imxdmac->channel, | ||
515 | d->dest, d->src, d->len); | ||
516 | |||
517 | break; | ||
518 | /* Cyclic transfer is the same as slave_sg with special sg configuration. */ | ||
519 | case IMXDMA_DESC_CYCLIC: | ||
520 | case IMXDMA_DESC_SLAVE_SG: | ||
521 | if (d->direction == DMA_DEV_TO_MEM) { | ||
522 | imx_dmav1_writel(imxdma, imxdmac->per_address, | ||
523 | DMA_SAR(imxdmac->channel)); | ||
524 | imx_dmav1_writel(imxdma, imxdmac->ccr_from_device, | ||
525 | DMA_CCR(imxdmac->channel)); | ||
526 | |||
527 | dev_dbg(imxdma->dev, "%s channel: %d sg=%p sgcount=%d " | ||
528 | "total length=%d dev_addr=0x%08x (dev2mem)\n", | ||
529 | __func__, imxdmac->channel, d->sg, d->sgcount, | ||
530 | d->len, imxdmac->per_address); | ||
531 | } else if (d->direction == DMA_MEM_TO_DEV) { | ||
532 | imx_dmav1_writel(imxdma, imxdmac->per_address, | ||
533 | DMA_DAR(imxdmac->channel)); | ||
534 | imx_dmav1_writel(imxdma, imxdmac->ccr_to_device, | ||
535 | DMA_CCR(imxdmac->channel)); | ||
536 | |||
537 | dev_dbg(imxdma->dev, "%s channel: %d sg=%p sgcount=%d " | ||
538 | "total length=%d dev_addr=0x%08x (mem2dev)\n", | ||
539 | __func__, imxdmac->channel, d->sg, d->sgcount, | ||
540 | d->len, imxdmac->per_address); | ||
541 | } else { | ||
542 | dev_err(imxdma->dev, "%s channel: %d bad dma mode\n", | ||
543 | __func__, imxdmac->channel); | ||
544 | return -EINVAL; | ||
545 | } | ||
546 | |||
547 | imxdma_sg_next(d); | ||
81 | 548 | ||
82 | imxdmac->status = DMA_ERROR; | 549 | break; |
83 | imxdma_handle(imxdmac); | 550 | default: |
551 | return -EINVAL; | ||
552 | } | ||
553 | imxdma_enable_hw(d); | ||
554 | return 0; | ||
84 | } | 555 | } |
85 | 556 | ||
86 | static void imxdma_progression(int channel, void *data, | 557 | static void imxdma_tasklet(unsigned long data) |
87 | struct scatterlist *sg) | ||
88 | { | 558 | { |
89 | struct imxdma_channel *imxdmac = data; | 559 | struct imxdma_channel *imxdmac = (void *)data; |
560 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
561 | struct imxdma_desc *desc; | ||
90 | 562 | ||
91 | imxdmac->status = DMA_SUCCESS; | 563 | spin_lock(&imxdma->lock); |
92 | imxdma_handle(imxdmac); | 564 | |
565 | if (list_empty(&imxdmac->ld_active)) { | ||
566 | /* Someone might have called terminate all */ | ||
567 | goto out; | ||
568 | } | ||
569 | desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, node); | ||
570 | |||
571 | if (desc->desc.callback) | ||
572 | desc->desc.callback(desc->desc.callback_param); | ||
573 | |||
574 | dma_cookie_complete(&desc->desc); | ||
575 | |||
576 | /* If we are dealing with a cyclic descriptor keep it on ld_active */ | ||
577 | if (imxdma_chan_is_doing_cyclic(imxdmac)) | ||
578 | goto out; | ||
579 | |||
580 | /* Free 2D slot if it was an interleaved transfer */ | ||
581 | if (imxdmac->enabled_2d) { | ||
582 | imxdma->slots_2d[imxdmac->slot_2d].count--; | ||
583 | imxdmac->enabled_2d = false; | ||
584 | } | ||
585 | |||
586 | list_move_tail(imxdmac->ld_active.next, &imxdmac->ld_free); | ||
587 | |||
588 | if (!list_empty(&imxdmac->ld_queue)) { | ||
589 | desc = list_first_entry(&imxdmac->ld_queue, struct imxdma_desc, | ||
590 | node); | ||
591 | list_move_tail(imxdmac->ld_queue.next, &imxdmac->ld_active); | ||
592 | if (imxdma_xfer_desc(desc) < 0) | ||
593 | dev_warn(imxdma->dev, "%s: channel: %d couldn't xfer desc\n", | ||
594 | __func__, imxdmac->channel); | ||
595 | } | ||
596 | out: | ||
597 | spin_unlock(&imxdma->lock); | ||
93 | } | 598 | } |
94 | 599 | ||
95 | static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | 600 | static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, |
@@ -97,13 +602,18 @@ static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | |||
97 | { | 602 | { |
98 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 603 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
99 | struct dma_slave_config *dmaengine_cfg = (void *)arg; | 604 | struct dma_slave_config *dmaengine_cfg = (void *)arg; |
100 | int ret; | 605 | struct imxdma_engine *imxdma = imxdmac->imxdma; |
606 | unsigned long flags; | ||
101 | unsigned int mode = 0; | 607 | unsigned int mode = 0; |
102 | 608 | ||
103 | switch (cmd) { | 609 | switch (cmd) { |
104 | case DMA_TERMINATE_ALL: | 610 | case DMA_TERMINATE_ALL: |
105 | imxdmac->status = DMA_ERROR; | 611 | imxdma_disable_hw(imxdmac); |
106 | imx_dma_disable(imxdmac->imxdma_channel); | 612 | |
613 | spin_lock_irqsave(&imxdma->lock, flags); | ||
614 | list_splice_tail_init(&imxdmac->ld_active, &imxdmac->ld_free); | ||
615 | list_splice_tail_init(&imxdmac->ld_queue, &imxdmac->ld_free); | ||
616 | spin_unlock_irqrestore(&imxdma->lock, flags); | ||
107 | return 0; | 617 | return 0; |
108 | case DMA_SLAVE_CONFIG: | 618 | case DMA_SLAVE_CONFIG: |
109 | if (dmaengine_cfg->direction == DMA_DEV_TO_MEM) { | 619 | if (dmaengine_cfg->direction == DMA_DEV_TO_MEM) { |
@@ -128,16 +638,22 @@ static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, | |||
128 | mode = IMX_DMA_MEMSIZE_32; | 638 | mode = IMX_DMA_MEMSIZE_32; |
129 | break; | 639 | break; |
130 | } | 640 | } |
131 | ret = imx_dma_config_channel(imxdmac->imxdma_channel, | ||
132 | mode | IMX_DMA_TYPE_FIFO, | ||
133 | IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR, | ||
134 | imxdmac->dma_request, 1); | ||
135 | |||
136 | if (ret) | ||
137 | return ret; | ||
138 | 641 | ||
139 | imx_dma_config_burstlen(imxdmac->imxdma_channel, | 642 | imxdmac->hw_chaining = 1; |
140 | imxdmac->watermark_level * imxdmac->word_size); | 643 | if (!imxdma_hw_chain(imxdmac)) |
644 | return -EINVAL; | ||
645 | imxdmac->ccr_from_device = (mode | IMX_DMA_TYPE_FIFO) | | ||
646 | ((IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR) << 2) | | ||
647 | CCR_REN; | ||
648 | imxdmac->ccr_to_device = | ||
649 | (IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR) | | ||
650 | ((mode | IMX_DMA_TYPE_FIFO) << 2) | CCR_REN; | ||
651 | imx_dmav1_writel(imxdma, imxdmac->dma_request, | ||
652 | DMA_RSSR(imxdmac->channel)); | ||
653 | |||
654 | /* Set burst length */ | ||
655 | imx_dmav1_writel(imxdma, imxdmac->watermark_level * | ||
656 | imxdmac->word_size, DMA_BLR(imxdmac->channel)); | ||
141 | 657 | ||
142 | return 0; | 658 | return 0; |
143 | default: | 659 | default: |
@@ -151,43 +667,20 @@ static enum dma_status imxdma_tx_status(struct dma_chan *chan, | |||
151 | dma_cookie_t cookie, | 667 | dma_cookie_t cookie, |
152 | struct dma_tx_state *txstate) | 668 | struct dma_tx_state *txstate) |
153 | { | 669 | { |
154 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 670 | return dma_cookie_status(chan, cookie, txstate); |
155 | dma_cookie_t last_used; | ||
156 | enum dma_status ret; | ||
157 | |||
158 | last_used = chan->cookie; | ||
159 | |||
160 | ret = dma_async_is_complete(cookie, imxdmac->last_completed, last_used); | ||
161 | dma_set_tx_state(txstate, imxdmac->last_completed, last_used, 0); | ||
162 | |||
163 | return ret; | ||
164 | } | ||
165 | |||
166 | static dma_cookie_t imxdma_assign_cookie(struct imxdma_channel *imxdma) | ||
167 | { | ||
168 | dma_cookie_t cookie = imxdma->chan.cookie; | ||
169 | |||
170 | if (++cookie < 0) | ||
171 | cookie = 1; | ||
172 | |||
173 | imxdma->chan.cookie = cookie; | ||
174 | imxdma->desc.cookie = cookie; | ||
175 | |||
176 | return cookie; | ||
177 | } | 671 | } |
178 | 672 | ||
179 | static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) | 673 | static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) |
180 | { | 674 | { |
181 | struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan); | 675 | struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan); |
676 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
182 | dma_cookie_t cookie; | 677 | dma_cookie_t cookie; |
678 | unsigned long flags; | ||
183 | 679 | ||
184 | spin_lock_irq(&imxdmac->lock); | 680 | spin_lock_irqsave(&imxdma->lock, flags); |
185 | 681 | list_move_tail(imxdmac->ld_free.next, &imxdmac->ld_queue); | |
186 | cookie = imxdma_assign_cookie(imxdmac); | 682 | cookie = dma_cookie_assign(tx); |
187 | 683 | spin_unlock_irqrestore(&imxdma->lock, flags); | |
188 | imx_dma_enable(imxdmac->imxdma_channel); | ||
189 | |||
190 | spin_unlock_irq(&imxdmac->lock); | ||
191 | 684 | ||
192 | return cookie; | 685 | return cookie; |
193 | } | 686 | } |
@@ -197,23 +690,52 @@ static int imxdma_alloc_chan_resources(struct dma_chan *chan) | |||
197 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 690 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
198 | struct imx_dma_data *data = chan->private; | 691 | struct imx_dma_data *data = chan->private; |
199 | 692 | ||
200 | imxdmac->dma_request = data->dma_request; | 693 | if (data != NULL) |
694 | imxdmac->dma_request = data->dma_request; | ||
201 | 695 | ||
202 | dma_async_tx_descriptor_init(&imxdmac->desc, chan); | 696 | while (imxdmac->descs_allocated < IMXDMA_MAX_CHAN_DESCRIPTORS) { |
203 | imxdmac->desc.tx_submit = imxdma_tx_submit; | 697 | struct imxdma_desc *desc; |
204 | /* txd.flags will be overwritten in prep funcs */ | ||
205 | imxdmac->desc.flags = DMA_CTRL_ACK; | ||
206 | 698 | ||
207 | imxdmac->status = DMA_SUCCESS; | 699 | desc = kzalloc(sizeof(*desc), GFP_KERNEL); |
700 | if (!desc) | ||
701 | break; | ||
702 | __memzero(&desc->desc, sizeof(struct dma_async_tx_descriptor)); | ||
703 | dma_async_tx_descriptor_init(&desc->desc, chan); | ||
704 | desc->desc.tx_submit = imxdma_tx_submit; | ||
705 | /* txd.flags will be overwritten in prep funcs */ | ||
706 | desc->desc.flags = DMA_CTRL_ACK; | ||
707 | desc->status = DMA_SUCCESS; | ||
708 | |||
709 | list_add_tail(&desc->node, &imxdmac->ld_free); | ||
710 | imxdmac->descs_allocated++; | ||
711 | } | ||
208 | 712 | ||
209 | return 0; | 713 | if (!imxdmac->descs_allocated) |
714 | return -ENOMEM; | ||
715 | |||
716 | return imxdmac->descs_allocated; | ||
210 | } | 717 | } |
211 | 718 | ||
212 | static void imxdma_free_chan_resources(struct dma_chan *chan) | 719 | static void imxdma_free_chan_resources(struct dma_chan *chan) |
213 | { | 720 | { |
214 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 721 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
722 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
723 | struct imxdma_desc *desc, *_desc; | ||
724 | unsigned long flags; | ||
725 | |||
726 | spin_lock_irqsave(&imxdma->lock, flags); | ||
727 | |||
728 | imxdma_disable_hw(imxdmac); | ||
729 | list_splice_tail_init(&imxdmac->ld_active, &imxdmac->ld_free); | ||
730 | list_splice_tail_init(&imxdmac->ld_queue, &imxdmac->ld_free); | ||
215 | 731 | ||
216 | imx_dma_disable(imxdmac->imxdma_channel); | 732 | spin_unlock_irqrestore(&imxdma->lock, flags); |
733 | |||
734 | list_for_each_entry_safe(desc, _desc, &imxdmac->ld_free, node) { | ||
735 | kfree(desc); | ||
736 | imxdmac->descs_allocated--; | ||
737 | } | ||
738 | INIT_LIST_HEAD(&imxdmac->ld_free); | ||
217 | 739 | ||
218 | if (imxdmac->sg_list) { | 740 | if (imxdmac->sg_list) { |
219 | kfree(imxdmac->sg_list); | 741 | kfree(imxdmac->sg_list); |
@@ -224,27 +746,23 @@ static void imxdma_free_chan_resources(struct dma_chan *chan) | |||
224 | static struct dma_async_tx_descriptor *imxdma_prep_slave_sg( | 746 | static struct dma_async_tx_descriptor *imxdma_prep_slave_sg( |
225 | struct dma_chan *chan, struct scatterlist *sgl, | 747 | struct dma_chan *chan, struct scatterlist *sgl, |
226 | unsigned int sg_len, enum dma_transfer_direction direction, | 748 | unsigned int sg_len, enum dma_transfer_direction direction, |
227 | unsigned long flags) | 749 | unsigned long flags, void *context) |
228 | { | 750 | { |
229 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 751 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
230 | struct scatterlist *sg; | 752 | struct scatterlist *sg; |
231 | int i, ret, dma_length = 0; | 753 | int i, dma_length = 0; |
232 | unsigned int dmamode; | 754 | struct imxdma_desc *desc; |
233 | 755 | ||
234 | if (imxdmac->status == DMA_IN_PROGRESS) | 756 | if (list_empty(&imxdmac->ld_free) || |
757 | imxdma_chan_is_doing_cyclic(imxdmac)) | ||
235 | return NULL; | 758 | return NULL; |
236 | 759 | ||
237 | imxdmac->status = DMA_IN_PROGRESS; | 760 | desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); |
238 | 761 | ||
239 | for_each_sg(sgl, sg, sg_len, i) { | 762 | for_each_sg(sgl, sg, sg_len, i) { |
240 | dma_length += sg->length; | 763 | dma_length += sg->length; |
241 | } | 764 | } |
242 | 765 | ||
243 | if (direction == DMA_DEV_TO_MEM) | ||
244 | dmamode = DMA_MODE_READ; | ||
245 | else | ||
246 | dmamode = DMA_MODE_WRITE; | ||
247 | |||
248 | switch (imxdmac->word_size) { | 766 | switch (imxdmac->word_size) { |
249 | case DMA_SLAVE_BUSWIDTH_4_BYTES: | 767 | case DMA_SLAVE_BUSWIDTH_4_BYTES: |
250 | if (sgl->length & 3 || sgl->dma_address & 3) | 768 | if (sgl->length & 3 || sgl->dma_address & 3) |
@@ -260,37 +778,41 @@ static struct dma_async_tx_descriptor *imxdma_prep_slave_sg( | |||
260 | return NULL; | 778 | return NULL; |
261 | } | 779 | } |
262 | 780 | ||
263 | ret = imx_dma_setup_sg(imxdmac->imxdma_channel, sgl, sg_len, | 781 | desc->type = IMXDMA_DESC_SLAVE_SG; |
264 | dma_length, imxdmac->per_address, dmamode); | 782 | desc->sg = sgl; |
265 | if (ret) | 783 | desc->sgcount = sg_len; |
266 | return NULL; | 784 | desc->len = dma_length; |
785 | desc->direction = direction; | ||
786 | if (direction == DMA_DEV_TO_MEM) { | ||
787 | desc->src = imxdmac->per_address; | ||
788 | } else { | ||
789 | desc->dest = imxdmac->per_address; | ||
790 | } | ||
791 | desc->desc.callback = NULL; | ||
792 | desc->desc.callback_param = NULL; | ||
267 | 793 | ||
268 | return &imxdmac->desc; | 794 | return &desc->desc; |
269 | } | 795 | } |
270 | 796 | ||
271 | static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic( | 797 | static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic( |
272 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, | 798 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, |
273 | size_t period_len, enum dma_transfer_direction direction) | 799 | size_t period_len, enum dma_transfer_direction direction, |
800 | void *context) | ||
274 | { | 801 | { |
275 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | 802 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
276 | struct imxdma_engine *imxdma = imxdmac->imxdma; | 803 | struct imxdma_engine *imxdma = imxdmac->imxdma; |
277 | int i, ret; | 804 | struct imxdma_desc *desc; |
805 | int i; | ||
278 | unsigned int periods = buf_len / period_len; | 806 | unsigned int periods = buf_len / period_len; |
279 | unsigned int dmamode; | ||
280 | 807 | ||
281 | dev_dbg(imxdma->dev, "%s channel: %d buf_len=%d period_len=%d\n", | 808 | dev_dbg(imxdma->dev, "%s channel: %d buf_len=%d period_len=%d\n", |
282 | __func__, imxdmac->channel, buf_len, period_len); | 809 | __func__, imxdmac->channel, buf_len, period_len); |
283 | 810 | ||
284 | if (imxdmac->status == DMA_IN_PROGRESS) | 811 | if (list_empty(&imxdmac->ld_free) || |
812 | imxdma_chan_is_doing_cyclic(imxdmac)) | ||
285 | return NULL; | 813 | return NULL; |
286 | imxdmac->status = DMA_IN_PROGRESS; | ||
287 | 814 | ||
288 | ret = imx_dma_setup_progression_handler(imxdmac->imxdma_channel, | 815 | desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); |
289 | imxdma_progression); | ||
290 | if (ret) { | ||
291 | dev_err(imxdma->dev, "Failed to setup the DMA handler\n"); | ||
292 | return NULL; | ||
293 | } | ||
294 | 816 | ||
295 | if (imxdmac->sg_list) | 817 | if (imxdmac->sg_list) |
296 | kfree(imxdmac->sg_list); | 818 | kfree(imxdmac->sg_list); |
@@ -316,62 +838,221 @@ static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic( | |||
316 | imxdmac->sg_list[periods].page_link = | 838 | imxdmac->sg_list[periods].page_link = |
317 | ((unsigned long)imxdmac->sg_list | 0x01) & ~0x02; | 839 | ((unsigned long)imxdmac->sg_list | 0x01) & ~0x02; |
318 | 840 | ||
319 | if (direction == DMA_DEV_TO_MEM) | 841 | desc->type = IMXDMA_DESC_CYCLIC; |
320 | dmamode = DMA_MODE_READ; | 842 | desc->sg = imxdmac->sg_list; |
321 | else | 843 | desc->sgcount = periods; |
322 | dmamode = DMA_MODE_WRITE; | 844 | desc->len = IMX_DMA_LENGTH_LOOP; |
845 | desc->direction = direction; | ||
846 | if (direction == DMA_DEV_TO_MEM) { | ||
847 | desc->src = imxdmac->per_address; | ||
848 | } else { | ||
849 | desc->dest = imxdmac->per_address; | ||
850 | } | ||
851 | desc->desc.callback = NULL; | ||
852 | desc->desc.callback_param = NULL; | ||
853 | |||
854 | return &desc->desc; | ||
855 | } | ||
856 | |||
857 | static struct dma_async_tx_descriptor *imxdma_prep_dma_memcpy( | ||
858 | struct dma_chan *chan, dma_addr_t dest, | ||
859 | dma_addr_t src, size_t len, unsigned long flags) | ||
860 | { | ||
861 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | ||
862 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
863 | struct imxdma_desc *desc; | ||
323 | 864 | ||
324 | ret = imx_dma_setup_sg(imxdmac->imxdma_channel, imxdmac->sg_list, periods, | 865 | dev_dbg(imxdma->dev, "%s channel: %d src=0x%x dst=0x%x len=%d\n", |
325 | IMX_DMA_LENGTH_LOOP, imxdmac->per_address, dmamode); | 866 | __func__, imxdmac->channel, src, dest, len); |
326 | if (ret) | 867 | |
868 | if (list_empty(&imxdmac->ld_free) || | ||
869 | imxdma_chan_is_doing_cyclic(imxdmac)) | ||
327 | return NULL; | 870 | return NULL; |
328 | 871 | ||
329 | return &imxdmac->desc; | 872 | desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); |
873 | |||
874 | desc->type = IMXDMA_DESC_MEMCPY; | ||
875 | desc->src = src; | ||
876 | desc->dest = dest; | ||
877 | desc->len = len; | ||
878 | desc->direction = DMA_MEM_TO_MEM; | ||
879 | desc->config_port = IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR; | ||
880 | desc->config_mem = IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR; | ||
881 | desc->desc.callback = NULL; | ||
882 | desc->desc.callback_param = NULL; | ||
883 | |||
884 | return &desc->desc; | ||
885 | } | ||
886 | |||
887 | static struct dma_async_tx_descriptor *imxdma_prep_dma_interleaved( | ||
888 | struct dma_chan *chan, struct dma_interleaved_template *xt, | ||
889 | unsigned long flags) | ||
890 | { | ||
891 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); | ||
892 | struct imxdma_engine *imxdma = imxdmac->imxdma; | ||
893 | struct imxdma_desc *desc; | ||
894 | |||
895 | dev_dbg(imxdma->dev, "%s channel: %d src_start=0x%x dst_start=0x%x\n" | ||
896 | " src_sgl=%s dst_sgl=%s numf=%d frame_size=%d\n", __func__, | ||
897 | imxdmac->channel, xt->src_start, xt->dst_start, | ||
898 | xt->src_sgl ? "true" : "false", xt->dst_sgl ? "true" : "false", | ||
899 | xt->numf, xt->frame_size); | ||
900 | |||
901 | if (list_empty(&imxdmac->ld_free) || | ||
902 | imxdma_chan_is_doing_cyclic(imxdmac)) | ||
903 | return NULL; | ||
904 | |||
905 | if (xt->frame_size != 1 || xt->numf <= 0 || xt->dir != DMA_MEM_TO_MEM) | ||
906 | return NULL; | ||
907 | |||
908 | desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); | ||
909 | |||
910 | desc->type = IMXDMA_DESC_INTERLEAVED; | ||
911 | desc->src = xt->src_start; | ||
912 | desc->dest = xt->dst_start; | ||
913 | desc->x = xt->sgl[0].size; | ||
914 | desc->y = xt->numf; | ||
915 | desc->w = xt->sgl[0].icg + desc->x; | ||
916 | desc->len = desc->x * desc->y; | ||
917 | desc->direction = DMA_MEM_TO_MEM; | ||
918 | desc->config_port = IMX_DMA_MEMSIZE_32; | ||
919 | desc->config_mem = IMX_DMA_MEMSIZE_32; | ||
920 | if (xt->src_sgl) | ||
921 | desc->config_mem |= IMX_DMA_TYPE_2D; | ||
922 | if (xt->dst_sgl) | ||
923 | desc->config_port |= IMX_DMA_TYPE_2D; | ||
924 | desc->desc.callback = NULL; | ||
925 | desc->desc.callback_param = NULL; | ||
926 | |||
927 | return &desc->desc; | ||
330 | } | 928 | } |
331 | 929 | ||
332 | static void imxdma_issue_pending(struct dma_chan *chan) | 930 | static void imxdma_issue_pending(struct dma_chan *chan) |
333 | { | 931 | { |
334 | /* | 932 | struct imxdma_channel *imxdmac = to_imxdma_chan(chan); |
335 | * Nothing to do. We only have a single descriptor | 933 | struct imxdma_engine *imxdma = imxdmac->imxdma; |
336 | */ | 934 | struct imxdma_desc *desc; |
935 | unsigned long flags; | ||
936 | |||
937 | spin_lock_irqsave(&imxdma->lock, flags); | ||
938 | if (list_empty(&imxdmac->ld_active) && | ||
939 | !list_empty(&imxdmac->ld_queue)) { | ||
940 | desc = list_first_entry(&imxdmac->ld_queue, | ||
941 | struct imxdma_desc, node); | ||
942 | |||
943 | if (imxdma_xfer_desc(desc) < 0) { | ||
944 | dev_warn(imxdma->dev, | ||
945 | "%s: channel: %d couldn't issue DMA xfer\n", | ||
946 | __func__, imxdmac->channel); | ||
947 | } else { | ||
948 | list_move_tail(imxdmac->ld_queue.next, | ||
949 | &imxdmac->ld_active); | ||
950 | } | ||
951 | } | ||
952 | spin_unlock_irqrestore(&imxdma->lock, flags); | ||
337 | } | 953 | } |
338 | 954 | ||
339 | static int __init imxdma_probe(struct platform_device *pdev) | 955 | static int __init imxdma_probe(struct platform_device *pdev) |
340 | { | 956 | { |
341 | struct imxdma_engine *imxdma; | 957 | struct imxdma_engine *imxdma; |
342 | int ret, i; | 958 | int ret, i; |
343 | 959 | ||
960 | |||
344 | imxdma = kzalloc(sizeof(*imxdma), GFP_KERNEL); | 961 | imxdma = kzalloc(sizeof(*imxdma), GFP_KERNEL); |
345 | if (!imxdma) | 962 | if (!imxdma) |
346 | return -ENOMEM; | 963 | return -ENOMEM; |
347 | 964 | ||
965 | if (cpu_is_mx1()) { | ||
966 | imxdma->base = MX1_IO_ADDRESS(MX1_DMA_BASE_ADDR); | ||
967 | } else if (cpu_is_mx21()) { | ||
968 | imxdma->base = MX21_IO_ADDRESS(MX21_DMA_BASE_ADDR); | ||
969 | } else if (cpu_is_mx27()) { | ||
970 | imxdma->base = MX27_IO_ADDRESS(MX27_DMA_BASE_ADDR); | ||
971 | } else { | ||
972 | kfree(imxdma); | ||
973 | return 0; | ||
974 | } | ||
975 | |||
976 | imxdma->dma_clk = clk_get(NULL, "dma"); | ||
977 | if (IS_ERR(imxdma->dma_clk)) | ||
978 | return PTR_ERR(imxdma->dma_clk); | ||
979 | clk_enable(imxdma->dma_clk); | ||
980 | |||
981 | /* reset DMA module */ | ||
982 | imx_dmav1_writel(imxdma, DCR_DRST, DMA_DCR); | ||
983 | |||
984 | if (cpu_is_mx1()) { | ||
985 | ret = request_irq(MX1_DMA_INT, dma_irq_handler, 0, "DMA", imxdma); | ||
986 | if (ret) { | ||
987 | dev_warn(imxdma->dev, "Can't register IRQ for DMA\n"); | ||
988 | kfree(imxdma); | ||
989 | return ret; | ||
990 | } | ||
991 | |||
992 | ret = request_irq(MX1_DMA_ERR, imxdma_err_handler, 0, "DMA", imxdma); | ||
993 | if (ret) { | ||
994 | dev_warn(imxdma->dev, "Can't register ERRIRQ for DMA\n"); | ||
995 | free_irq(MX1_DMA_INT, NULL); | ||
996 | kfree(imxdma); | ||
997 | return ret; | ||
998 | } | ||
999 | } | ||
1000 | |||
1001 | /* enable DMA module */ | ||
1002 | imx_dmav1_writel(imxdma, DCR_DEN, DMA_DCR); | ||
1003 | |||
1004 | /* clear all interrupts */ | ||
1005 | imx_dmav1_writel(imxdma, (1 << IMX_DMA_CHANNELS) - 1, DMA_DISR); | ||
1006 | |||
1007 | /* disable interrupts */ | ||
1008 | imx_dmav1_writel(imxdma, (1 << IMX_DMA_CHANNELS) - 1, DMA_DIMR); | ||
1009 | |||
348 | INIT_LIST_HEAD(&imxdma->dma_device.channels); | 1010 | INIT_LIST_HEAD(&imxdma->dma_device.channels); |
349 | 1011 | ||
350 | dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); | 1012 | dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); |
351 | dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); | 1013 | dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); |
1014 | dma_cap_set(DMA_MEMCPY, imxdma->dma_device.cap_mask); | ||
1015 | dma_cap_set(DMA_INTERLEAVE, imxdma->dma_device.cap_mask); | ||
1016 | |||
1017 | /* Initialize 2D global parameters */ | ||
1018 | for (i = 0; i < IMX_DMA_2D_SLOTS; i++) | ||
1019 | imxdma->slots_2d[i].count = 0; | ||
1020 | |||
1021 | spin_lock_init(&imxdma->lock); | ||
352 | 1022 | ||
353 | /* Initialize channel parameters */ | 1023 | /* Initialize channel parameters */ |
354 | for (i = 0; i < MAX_DMA_CHANNELS; i++) { | 1024 | for (i = 0; i < IMX_DMA_CHANNELS; i++) { |
355 | struct imxdma_channel *imxdmac = &imxdma->channel[i]; | 1025 | struct imxdma_channel *imxdmac = &imxdma->channel[i]; |
356 | 1026 | ||
357 | imxdmac->imxdma_channel = imx_dma_request_by_prio("dmaengine", | 1027 | if (cpu_is_mx21() || cpu_is_mx27()) { |
358 | DMA_PRIO_MEDIUM); | 1028 | ret = request_irq(MX2x_INT_DMACH0 + i, |
359 | if ((int)imxdmac->channel < 0) { | 1029 | dma_irq_handler, 0, "DMA", imxdma); |
360 | ret = -ENODEV; | 1030 | if (ret) { |
361 | goto err_init; | 1031 | dev_warn(imxdma->dev, "Can't register IRQ %d " |
1032 | "for DMA channel %d\n", | ||
1033 | MX2x_INT_DMACH0 + i, i); | ||
1034 | goto err_init; | ||
1035 | } | ||
1036 | init_timer(&imxdmac->watchdog); | ||
1037 | imxdmac->watchdog.function = &imxdma_watchdog; | ||
1038 | imxdmac->watchdog.data = (unsigned long)imxdmac; | ||
362 | } | 1039 | } |
363 | 1040 | ||
364 | imx_dma_setup_handlers(imxdmac->imxdma_channel, | ||
365 | imxdma_irq_handler, imxdma_err_handler, imxdmac); | ||
366 | |||
367 | imxdmac->imxdma = imxdma; | 1041 | imxdmac->imxdma = imxdma; |
368 | spin_lock_init(&imxdmac->lock); | ||
369 | 1042 | ||
1043 | INIT_LIST_HEAD(&imxdmac->ld_queue); | ||
1044 | INIT_LIST_HEAD(&imxdmac->ld_free); | ||
1045 | INIT_LIST_HEAD(&imxdmac->ld_active); | ||
1046 | |||
1047 | tasklet_init(&imxdmac->dma_tasklet, imxdma_tasklet, | ||
1048 | (unsigned long)imxdmac); | ||
370 | imxdmac->chan.device = &imxdma->dma_device; | 1049 | imxdmac->chan.device = &imxdma->dma_device; |
1050 | dma_cookie_init(&imxdmac->chan); | ||
371 | imxdmac->channel = i; | 1051 | imxdmac->channel = i; |
372 | 1052 | ||
373 | /* Add the channel to the DMAC list */ | 1053 | /* Add the channel to the DMAC list */ |
374 | list_add_tail(&imxdmac->chan.device_node, &imxdma->dma_device.channels); | 1054 | list_add_tail(&imxdmac->chan.device_node, |
1055 | &imxdma->dma_device.channels); | ||
375 | } | 1056 | } |
376 | 1057 | ||
377 | imxdma->dev = &pdev->dev; | 1058 | imxdma->dev = &pdev->dev; |
@@ -382,11 +1063,14 @@ static int __init imxdma_probe(struct platform_device *pdev) | |||
382 | imxdma->dma_device.device_tx_status = imxdma_tx_status; | 1063 | imxdma->dma_device.device_tx_status = imxdma_tx_status; |
383 | imxdma->dma_device.device_prep_slave_sg = imxdma_prep_slave_sg; | 1064 | imxdma->dma_device.device_prep_slave_sg = imxdma_prep_slave_sg; |
384 | imxdma->dma_device.device_prep_dma_cyclic = imxdma_prep_dma_cyclic; | 1065 | imxdma->dma_device.device_prep_dma_cyclic = imxdma_prep_dma_cyclic; |
1066 | imxdma->dma_device.device_prep_dma_memcpy = imxdma_prep_dma_memcpy; | ||
1067 | imxdma->dma_device.device_prep_interleaved_dma = imxdma_prep_dma_interleaved; | ||
385 | imxdma->dma_device.device_control = imxdma_control; | 1068 | imxdma->dma_device.device_control = imxdma_control; |
386 | imxdma->dma_device.device_issue_pending = imxdma_issue_pending; | 1069 | imxdma->dma_device.device_issue_pending = imxdma_issue_pending; |
387 | 1070 | ||
388 | platform_set_drvdata(pdev, imxdma); | 1071 | platform_set_drvdata(pdev, imxdma); |
389 | 1072 | ||
1073 | imxdma->dma_device.copy_align = 2; /* 2^2 = 4 bytes alignment */ | ||
390 | imxdma->dma_device.dev->dma_parms = &imxdma->dma_parms; | 1074 | imxdma->dma_device.dev->dma_parms = &imxdma->dma_parms; |
391 | dma_set_max_seg_size(imxdma->dma_device.dev, 0xffffff); | 1075 | dma_set_max_seg_size(imxdma->dma_device.dev, 0xffffff); |
392 | 1076 | ||
@@ -399,9 +1083,13 @@ static int __init imxdma_probe(struct platform_device *pdev) | |||
399 | return 0; | 1083 | return 0; |
400 | 1084 | ||
401 | err_init: | 1085 | err_init: |
402 | while (--i >= 0) { | 1086 | |
403 | struct imxdma_channel *imxdmac = &imxdma->channel[i]; | 1087 | if (cpu_is_mx21() || cpu_is_mx27()) { |
404 | imx_dma_free(imxdmac->imxdma_channel); | 1088 | while (--i >= 0) |
1089 | free_irq(MX2x_INT_DMACH0 + i, NULL); | ||
1090 | } else if cpu_is_mx1() { | ||
1091 | free_irq(MX1_DMA_INT, NULL); | ||
1092 | free_irq(MX1_DMA_ERR, NULL); | ||
405 | } | 1093 | } |
406 | 1094 | ||
407 | kfree(imxdma); | 1095 | kfree(imxdma); |
@@ -415,10 +1103,12 @@ static int __exit imxdma_remove(struct platform_device *pdev) | |||
415 | 1103 | ||
416 | dma_async_device_unregister(&imxdma->dma_device); | 1104 | dma_async_device_unregister(&imxdma->dma_device); |
417 | 1105 | ||
418 | for (i = 0; i < MAX_DMA_CHANNELS; i++) { | 1106 | if (cpu_is_mx21() || cpu_is_mx27()) { |
419 | struct imxdma_channel *imxdmac = &imxdma->channel[i]; | 1107 | for (i = 0; i < IMX_DMA_CHANNELS; i++) |
420 | 1108 | free_irq(MX2x_INT_DMACH0 + i, NULL); | |
421 | imx_dma_free(imxdmac->imxdma_channel); | 1109 | } else if cpu_is_mx1() { |
1110 | free_irq(MX1_DMA_INT, NULL); | ||
1111 | free_irq(MX1_DMA_ERR, NULL); | ||
422 | } | 1112 | } |
423 | 1113 | ||
424 | kfree(imxdma); | 1114 | kfree(imxdma); |
diff --git a/drivers/dma/imx-sdma.c b/drivers/dma/imx-sdma.c index 63540d3e2153..d3e38e28bb6b 100644 --- a/drivers/dma/imx-sdma.c +++ b/drivers/dma/imx-sdma.c | |||
@@ -20,6 +20,7 @@ | |||
20 | #include <linux/init.h> | 20 | #include <linux/init.h> |
21 | #include <linux/module.h> | 21 | #include <linux/module.h> |
22 | #include <linux/types.h> | 22 | #include <linux/types.h> |
23 | #include <linux/bitops.h> | ||
23 | #include <linux/mm.h> | 24 | #include <linux/mm.h> |
24 | #include <linux/interrupt.h> | 25 | #include <linux/interrupt.h> |
25 | #include <linux/clk.h> | 26 | #include <linux/clk.h> |
@@ -41,6 +42,8 @@ | |||
41 | #include <mach/dma.h> | 42 | #include <mach/dma.h> |
42 | #include <mach/hardware.h> | 43 | #include <mach/hardware.h> |
43 | 44 | ||
45 | #include "dmaengine.h" | ||
46 | |||
44 | /* SDMA registers */ | 47 | /* SDMA registers */ |
45 | #define SDMA_H_C0PTR 0x000 | 48 | #define SDMA_H_C0PTR 0x000 |
46 | #define SDMA_H_INTR 0x004 | 49 | #define SDMA_H_INTR 0x004 |
@@ -259,19 +262,18 @@ struct sdma_channel { | |||
259 | unsigned int pc_from_device, pc_to_device; | 262 | unsigned int pc_from_device, pc_to_device; |
260 | unsigned long flags; | 263 | unsigned long flags; |
261 | dma_addr_t per_address; | 264 | dma_addr_t per_address; |
262 | u32 event_mask0, event_mask1; | 265 | unsigned long event_mask[2]; |
263 | u32 watermark_level; | 266 | unsigned long watermark_level; |
264 | u32 shp_addr, per_addr; | 267 | u32 shp_addr, per_addr; |
265 | struct dma_chan chan; | 268 | struct dma_chan chan; |
266 | spinlock_t lock; | 269 | spinlock_t lock; |
267 | struct dma_async_tx_descriptor desc; | 270 | struct dma_async_tx_descriptor desc; |
268 | dma_cookie_t last_completed; | ||
269 | enum dma_status status; | 271 | enum dma_status status; |
270 | unsigned int chn_count; | 272 | unsigned int chn_count; |
271 | unsigned int chn_real_count; | 273 | unsigned int chn_real_count; |
272 | }; | 274 | }; |
273 | 275 | ||
274 | #define IMX_DMA_SG_LOOP (1 << 0) | 276 | #define IMX_DMA_SG_LOOP BIT(0) |
275 | 277 | ||
276 | #define MAX_DMA_CHANNELS 32 | 278 | #define MAX_DMA_CHANNELS 32 |
277 | #define MXC_SDMA_DEFAULT_PRIORITY 1 | 279 | #define MXC_SDMA_DEFAULT_PRIORITY 1 |
@@ -345,9 +347,9 @@ static const struct of_device_id sdma_dt_ids[] = { | |||
345 | }; | 347 | }; |
346 | MODULE_DEVICE_TABLE(of, sdma_dt_ids); | 348 | MODULE_DEVICE_TABLE(of, sdma_dt_ids); |
347 | 349 | ||
348 | #define SDMA_H_CONFIG_DSPDMA (1 << 12) /* indicates if the DSPDMA is used */ | 350 | #define SDMA_H_CONFIG_DSPDMA BIT(12) /* indicates if the DSPDMA is used */ |
349 | #define SDMA_H_CONFIG_RTD_PINS (1 << 11) /* indicates if Real-Time Debug pins are enabled */ | 351 | #define SDMA_H_CONFIG_RTD_PINS BIT(11) /* indicates if Real-Time Debug pins are enabled */ |
350 | #define SDMA_H_CONFIG_ACR (1 << 4) /* indicates if AHB freq /core freq = 2 or 1 */ | 352 | #define SDMA_H_CONFIG_ACR BIT(4) /* indicates if AHB freq /core freq = 2 or 1 */ |
351 | #define SDMA_H_CONFIG_CSM (3) /* indicates which context switch mode is selected*/ | 353 | #define SDMA_H_CONFIG_CSM (3) /* indicates which context switch mode is selected*/ |
352 | 354 | ||
353 | static inline u32 chnenbl_ofs(struct sdma_engine *sdma, unsigned int event) | 355 | static inline u32 chnenbl_ofs(struct sdma_engine *sdma, unsigned int event) |
@@ -362,37 +364,42 @@ static int sdma_config_ownership(struct sdma_channel *sdmac, | |||
362 | { | 364 | { |
363 | struct sdma_engine *sdma = sdmac->sdma; | 365 | struct sdma_engine *sdma = sdmac->sdma; |
364 | int channel = sdmac->channel; | 366 | int channel = sdmac->channel; |
365 | u32 evt, mcu, dsp; | 367 | unsigned long evt, mcu, dsp; |
366 | 368 | ||
367 | if (event_override && mcu_override && dsp_override) | 369 | if (event_override && mcu_override && dsp_override) |
368 | return -EINVAL; | 370 | return -EINVAL; |
369 | 371 | ||
370 | evt = __raw_readl(sdma->regs + SDMA_H_EVTOVR); | 372 | evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR); |
371 | mcu = __raw_readl(sdma->regs + SDMA_H_HOSTOVR); | 373 | mcu = readl_relaxed(sdma->regs + SDMA_H_HOSTOVR); |
372 | dsp = __raw_readl(sdma->regs + SDMA_H_DSPOVR); | 374 | dsp = readl_relaxed(sdma->regs + SDMA_H_DSPOVR); |
373 | 375 | ||
374 | if (dsp_override) | 376 | if (dsp_override) |
375 | dsp &= ~(1 << channel); | 377 | __clear_bit(channel, &dsp); |
376 | else | 378 | else |
377 | dsp |= (1 << channel); | 379 | __set_bit(channel, &dsp); |
378 | 380 | ||
379 | if (event_override) | 381 | if (event_override) |
380 | evt &= ~(1 << channel); | 382 | __clear_bit(channel, &evt); |
381 | else | 383 | else |
382 | evt |= (1 << channel); | 384 | __set_bit(channel, &evt); |
383 | 385 | ||
384 | if (mcu_override) | 386 | if (mcu_override) |
385 | mcu &= ~(1 << channel); | 387 | __clear_bit(channel, &mcu); |
386 | else | 388 | else |
387 | mcu |= (1 << channel); | 389 | __set_bit(channel, &mcu); |
388 | 390 | ||
389 | __raw_writel(evt, sdma->regs + SDMA_H_EVTOVR); | 391 | writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR); |
390 | __raw_writel(mcu, sdma->regs + SDMA_H_HOSTOVR); | 392 | writel_relaxed(mcu, sdma->regs + SDMA_H_HOSTOVR); |
391 | __raw_writel(dsp, sdma->regs + SDMA_H_DSPOVR); | 393 | writel_relaxed(dsp, sdma->regs + SDMA_H_DSPOVR); |
392 | 394 | ||
393 | return 0; | 395 | return 0; |
394 | } | 396 | } |
395 | 397 | ||
398 | static void sdma_enable_channel(struct sdma_engine *sdma, int channel) | ||
399 | { | ||
400 | writel(BIT(channel), sdma->regs + SDMA_H_START); | ||
401 | } | ||
402 | |||
396 | /* | 403 | /* |
397 | * sdma_run_channel - run a channel and wait till it's done | 404 | * sdma_run_channel - run a channel and wait till it's done |
398 | */ | 405 | */ |
@@ -404,7 +411,7 @@ static int sdma_run_channel(struct sdma_channel *sdmac) | |||
404 | 411 | ||
405 | init_completion(&sdmac->done); | 412 | init_completion(&sdmac->done); |
406 | 413 | ||
407 | __raw_writel(1 << channel, sdma->regs + SDMA_H_START); | 414 | sdma_enable_channel(sdma, channel); |
408 | 415 | ||
409 | ret = wait_for_completion_timeout(&sdmac->done, HZ); | 416 | ret = wait_for_completion_timeout(&sdmac->done, HZ); |
410 | 417 | ||
@@ -451,12 +458,12 @@ static void sdma_event_enable(struct sdma_channel *sdmac, unsigned int event) | |||
451 | { | 458 | { |
452 | struct sdma_engine *sdma = sdmac->sdma; | 459 | struct sdma_engine *sdma = sdmac->sdma; |
453 | int channel = sdmac->channel; | 460 | int channel = sdmac->channel; |
454 | u32 val; | 461 | unsigned long val; |
455 | u32 chnenbl = chnenbl_ofs(sdma, event); | 462 | u32 chnenbl = chnenbl_ofs(sdma, event); |
456 | 463 | ||
457 | val = __raw_readl(sdma->regs + chnenbl); | 464 | val = readl_relaxed(sdma->regs + chnenbl); |
458 | val |= (1 << channel); | 465 | __set_bit(channel, &val); |
459 | __raw_writel(val, sdma->regs + chnenbl); | 466 | writel_relaxed(val, sdma->regs + chnenbl); |
460 | } | 467 | } |
461 | 468 | ||
462 | static void sdma_event_disable(struct sdma_channel *sdmac, unsigned int event) | 469 | static void sdma_event_disable(struct sdma_channel *sdmac, unsigned int event) |
@@ -464,11 +471,11 @@ static void sdma_event_disable(struct sdma_channel *sdmac, unsigned int event) | |||
464 | struct sdma_engine *sdma = sdmac->sdma; | 471 | struct sdma_engine *sdma = sdmac->sdma; |
465 | int channel = sdmac->channel; | 472 | int channel = sdmac->channel; |
466 | u32 chnenbl = chnenbl_ofs(sdma, event); | 473 | u32 chnenbl = chnenbl_ofs(sdma, event); |
467 | u32 val; | 474 | unsigned long val; |
468 | 475 | ||
469 | val = __raw_readl(sdma->regs + chnenbl); | 476 | val = readl_relaxed(sdma->regs + chnenbl); |
470 | val &= ~(1 << channel); | 477 | __clear_bit(channel, &val); |
471 | __raw_writel(val, sdma->regs + chnenbl); | 478 | writel_relaxed(val, sdma->regs + chnenbl); |
472 | } | 479 | } |
473 | 480 | ||
474 | static void sdma_handle_channel_loop(struct sdma_channel *sdmac) | 481 | static void sdma_handle_channel_loop(struct sdma_channel *sdmac) |
@@ -522,7 +529,7 @@ static void mxc_sdma_handle_channel_normal(struct sdma_channel *sdmac) | |||
522 | else | 529 | else |
523 | sdmac->status = DMA_SUCCESS; | 530 | sdmac->status = DMA_SUCCESS; |
524 | 531 | ||
525 | sdmac->last_completed = sdmac->desc.cookie; | 532 | dma_cookie_complete(&sdmac->desc); |
526 | if (sdmac->desc.callback) | 533 | if (sdmac->desc.callback) |
527 | sdmac->desc.callback(sdmac->desc.callback_param); | 534 | sdmac->desc.callback(sdmac->desc.callback_param); |
528 | } | 535 | } |
@@ -544,10 +551,10 @@ static void mxc_sdma_handle_channel(struct sdma_channel *sdmac) | |||
544 | static irqreturn_t sdma_int_handler(int irq, void *dev_id) | 551 | static irqreturn_t sdma_int_handler(int irq, void *dev_id) |
545 | { | 552 | { |
546 | struct sdma_engine *sdma = dev_id; | 553 | struct sdma_engine *sdma = dev_id; |
547 | u32 stat; | 554 | unsigned long stat; |
548 | 555 | ||
549 | stat = __raw_readl(sdma->regs + SDMA_H_INTR); | 556 | stat = readl_relaxed(sdma->regs + SDMA_H_INTR); |
550 | __raw_writel(stat, sdma->regs + SDMA_H_INTR); | 557 | writel_relaxed(stat, sdma->regs + SDMA_H_INTR); |
551 | 558 | ||
552 | while (stat) { | 559 | while (stat) { |
553 | int channel = fls(stat) - 1; | 560 | int channel = fls(stat) - 1; |
@@ -555,7 +562,7 @@ static irqreturn_t sdma_int_handler(int irq, void *dev_id) | |||
555 | 562 | ||
556 | mxc_sdma_handle_channel(sdmac); | 563 | mxc_sdma_handle_channel(sdmac); |
557 | 564 | ||
558 | stat &= ~(1 << channel); | 565 | __clear_bit(channel, &stat); |
559 | } | 566 | } |
560 | 567 | ||
561 | return IRQ_HANDLED; | 568 | return IRQ_HANDLED; |
@@ -663,11 +670,11 @@ static int sdma_load_context(struct sdma_channel *sdmac) | |||
663 | return load_address; | 670 | return load_address; |
664 | 671 | ||
665 | dev_dbg(sdma->dev, "load_address = %d\n", load_address); | 672 | dev_dbg(sdma->dev, "load_address = %d\n", load_address); |
666 | dev_dbg(sdma->dev, "wml = 0x%08x\n", sdmac->watermark_level); | 673 | dev_dbg(sdma->dev, "wml = 0x%08x\n", (u32)sdmac->watermark_level); |
667 | dev_dbg(sdma->dev, "shp_addr = 0x%08x\n", sdmac->shp_addr); | 674 | dev_dbg(sdma->dev, "shp_addr = 0x%08x\n", sdmac->shp_addr); |
668 | dev_dbg(sdma->dev, "per_addr = 0x%08x\n", sdmac->per_addr); | 675 | dev_dbg(sdma->dev, "per_addr = 0x%08x\n", sdmac->per_addr); |
669 | dev_dbg(sdma->dev, "event_mask0 = 0x%08x\n", sdmac->event_mask0); | 676 | dev_dbg(sdma->dev, "event_mask0 = 0x%08x\n", (u32)sdmac->event_mask[0]); |
670 | dev_dbg(sdma->dev, "event_mask1 = 0x%08x\n", sdmac->event_mask1); | 677 | dev_dbg(sdma->dev, "event_mask1 = 0x%08x\n", (u32)sdmac->event_mask[1]); |
671 | 678 | ||
672 | mutex_lock(&sdma->channel_0_lock); | 679 | mutex_lock(&sdma->channel_0_lock); |
673 | 680 | ||
@@ -677,8 +684,8 @@ static int sdma_load_context(struct sdma_channel *sdmac) | |||
677 | /* Send by context the event mask,base address for peripheral | 684 | /* Send by context the event mask,base address for peripheral |
678 | * and watermark level | 685 | * and watermark level |
679 | */ | 686 | */ |
680 | context->gReg[0] = sdmac->event_mask1; | 687 | context->gReg[0] = sdmac->event_mask[1]; |
681 | context->gReg[1] = sdmac->event_mask0; | 688 | context->gReg[1] = sdmac->event_mask[0]; |
682 | context->gReg[2] = sdmac->per_addr; | 689 | context->gReg[2] = sdmac->per_addr; |
683 | context->gReg[6] = sdmac->shp_addr; | 690 | context->gReg[6] = sdmac->shp_addr; |
684 | context->gReg[7] = sdmac->watermark_level; | 691 | context->gReg[7] = sdmac->watermark_level; |
@@ -701,7 +708,7 @@ static void sdma_disable_channel(struct sdma_channel *sdmac) | |||
701 | struct sdma_engine *sdma = sdmac->sdma; | 708 | struct sdma_engine *sdma = sdmac->sdma; |
702 | int channel = sdmac->channel; | 709 | int channel = sdmac->channel; |
703 | 710 | ||
704 | __raw_writel(1 << channel, sdma->regs + SDMA_H_STATSTOP); | 711 | writel_relaxed(BIT(channel), sdma->regs + SDMA_H_STATSTOP); |
705 | sdmac->status = DMA_ERROR; | 712 | sdmac->status = DMA_ERROR; |
706 | } | 713 | } |
707 | 714 | ||
@@ -711,13 +718,13 @@ static int sdma_config_channel(struct sdma_channel *sdmac) | |||
711 | 718 | ||
712 | sdma_disable_channel(sdmac); | 719 | sdma_disable_channel(sdmac); |
713 | 720 | ||
714 | sdmac->event_mask0 = 0; | 721 | sdmac->event_mask[0] = 0; |
715 | sdmac->event_mask1 = 0; | 722 | sdmac->event_mask[1] = 0; |
716 | sdmac->shp_addr = 0; | 723 | sdmac->shp_addr = 0; |
717 | sdmac->per_addr = 0; | 724 | sdmac->per_addr = 0; |
718 | 725 | ||
719 | if (sdmac->event_id0) { | 726 | if (sdmac->event_id0) { |
720 | if (sdmac->event_id0 > 32) | 727 | if (sdmac->event_id0 >= sdmac->sdma->num_events) |
721 | return -EINVAL; | 728 | return -EINVAL; |
722 | sdma_event_enable(sdmac, sdmac->event_id0); | 729 | sdma_event_enable(sdmac, sdmac->event_id0); |
723 | } | 730 | } |
@@ -740,15 +747,14 @@ static int sdma_config_channel(struct sdma_channel *sdmac) | |||
740 | (sdmac->peripheral_type != IMX_DMATYPE_DSP)) { | 747 | (sdmac->peripheral_type != IMX_DMATYPE_DSP)) { |
741 | /* Handle multiple event channels differently */ | 748 | /* Handle multiple event channels differently */ |
742 | if (sdmac->event_id1) { | 749 | if (sdmac->event_id1) { |
743 | sdmac->event_mask1 = 1 << (sdmac->event_id1 % 32); | 750 | sdmac->event_mask[1] = BIT(sdmac->event_id1 % 32); |
744 | if (sdmac->event_id1 > 31) | 751 | if (sdmac->event_id1 > 31) |
745 | sdmac->watermark_level |= 1 << 31; | 752 | __set_bit(31, &sdmac->watermark_level); |
746 | sdmac->event_mask0 = 1 << (sdmac->event_id0 % 32); | 753 | sdmac->event_mask[0] = BIT(sdmac->event_id0 % 32); |
747 | if (sdmac->event_id0 > 31) | 754 | if (sdmac->event_id0 > 31) |
748 | sdmac->watermark_level |= 1 << 30; | 755 | __set_bit(30, &sdmac->watermark_level); |
749 | } else { | 756 | } else { |
750 | sdmac->event_mask0 = 1 << sdmac->event_id0; | 757 | __set_bit(sdmac->event_id0, sdmac->event_mask); |
751 | sdmac->event_mask1 = 1 << (sdmac->event_id0 - 32); | ||
752 | } | 758 | } |
753 | /* Watermark Level */ | 759 | /* Watermark Level */ |
754 | sdmac->watermark_level |= sdmac->watermark_level; | 760 | sdmac->watermark_level |= sdmac->watermark_level; |
@@ -774,7 +780,7 @@ static int sdma_set_channel_priority(struct sdma_channel *sdmac, | |||
774 | return -EINVAL; | 780 | return -EINVAL; |
775 | } | 781 | } |
776 | 782 | ||
777 | __raw_writel(priority, sdma->regs + SDMA_CHNPRI_0 + 4 * channel); | 783 | writel_relaxed(priority, sdma->regs + SDMA_CHNPRI_0 + 4 * channel); |
778 | 784 | ||
779 | return 0; | 785 | return 0; |
780 | } | 786 | } |
@@ -796,8 +802,6 @@ static int sdma_request_channel(struct sdma_channel *sdmac) | |||
796 | sdma->channel_control[channel].base_bd_ptr = sdmac->bd_phys; | 802 | sdma->channel_control[channel].base_bd_ptr = sdmac->bd_phys; |
797 | sdma->channel_control[channel].current_bd_ptr = sdmac->bd_phys; | 803 | sdma->channel_control[channel].current_bd_ptr = sdmac->bd_phys; |
798 | 804 | ||
799 | clk_enable(sdma->clk); | ||
800 | |||
801 | sdma_set_channel_priority(sdmac, MXC_SDMA_DEFAULT_PRIORITY); | 805 | sdma_set_channel_priority(sdmac, MXC_SDMA_DEFAULT_PRIORITY); |
802 | 806 | ||
803 | init_completion(&sdmac->done); | 807 | init_completion(&sdmac->done); |
@@ -810,24 +814,6 @@ out: | |||
810 | return ret; | 814 | return ret; |
811 | } | 815 | } |
812 | 816 | ||
813 | static void sdma_enable_channel(struct sdma_engine *sdma, int channel) | ||
814 | { | ||
815 | __raw_writel(1 << channel, sdma->regs + SDMA_H_START); | ||
816 | } | ||
817 | |||
818 | static dma_cookie_t sdma_assign_cookie(struct sdma_channel *sdmac) | ||
819 | { | ||
820 | dma_cookie_t cookie = sdmac->chan.cookie; | ||
821 | |||
822 | if (++cookie < 0) | ||
823 | cookie = 1; | ||
824 | |||
825 | sdmac->chan.cookie = cookie; | ||
826 | sdmac->desc.cookie = cookie; | ||
827 | |||
828 | return cookie; | ||
829 | } | ||
830 | |||
831 | static struct sdma_channel *to_sdma_chan(struct dma_chan *chan) | 817 | static struct sdma_channel *to_sdma_chan(struct dma_chan *chan) |
832 | { | 818 | { |
833 | return container_of(chan, struct sdma_channel, chan); | 819 | return container_of(chan, struct sdma_channel, chan); |
@@ -837,14 +823,11 @@ static dma_cookie_t sdma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
837 | { | 823 | { |
838 | unsigned long flags; | 824 | unsigned long flags; |
839 | struct sdma_channel *sdmac = to_sdma_chan(tx->chan); | 825 | struct sdma_channel *sdmac = to_sdma_chan(tx->chan); |
840 | struct sdma_engine *sdma = sdmac->sdma; | ||
841 | dma_cookie_t cookie; | 826 | dma_cookie_t cookie; |
842 | 827 | ||
843 | spin_lock_irqsave(&sdmac->lock, flags); | 828 | spin_lock_irqsave(&sdmac->lock, flags); |
844 | 829 | ||
845 | cookie = sdma_assign_cookie(sdmac); | 830 | cookie = dma_cookie_assign(tx); |
846 | |||
847 | sdma_enable_channel(sdma, sdmac->channel); | ||
848 | 831 | ||
849 | spin_unlock_irqrestore(&sdmac->lock, flags); | 832 | spin_unlock_irqrestore(&sdmac->lock, flags); |
850 | 833 | ||
@@ -875,11 +858,14 @@ static int sdma_alloc_chan_resources(struct dma_chan *chan) | |||
875 | 858 | ||
876 | sdmac->peripheral_type = data->peripheral_type; | 859 | sdmac->peripheral_type = data->peripheral_type; |
877 | sdmac->event_id0 = data->dma_request; | 860 | sdmac->event_id0 = data->dma_request; |
878 | ret = sdma_set_channel_priority(sdmac, prio); | 861 | |
862 | clk_enable(sdmac->sdma->clk); | ||
863 | |||
864 | ret = sdma_request_channel(sdmac); | ||
879 | if (ret) | 865 | if (ret) |
880 | return ret; | 866 | return ret; |
881 | 867 | ||
882 | ret = sdma_request_channel(sdmac); | 868 | ret = sdma_set_channel_priority(sdmac, prio); |
883 | if (ret) | 869 | if (ret) |
884 | return ret; | 870 | return ret; |
885 | 871 | ||
@@ -916,7 +902,7 @@ static void sdma_free_chan_resources(struct dma_chan *chan) | |||
916 | static struct dma_async_tx_descriptor *sdma_prep_slave_sg( | 902 | static struct dma_async_tx_descriptor *sdma_prep_slave_sg( |
917 | struct dma_chan *chan, struct scatterlist *sgl, | 903 | struct dma_chan *chan, struct scatterlist *sgl, |
918 | unsigned int sg_len, enum dma_transfer_direction direction, | 904 | unsigned int sg_len, enum dma_transfer_direction direction, |
919 | unsigned long flags) | 905 | unsigned long flags, void *context) |
920 | { | 906 | { |
921 | struct sdma_channel *sdmac = to_sdma_chan(chan); | 907 | struct sdma_channel *sdmac = to_sdma_chan(chan); |
922 | struct sdma_engine *sdma = sdmac->sdma; | 908 | struct sdma_engine *sdma = sdmac->sdma; |
@@ -1014,7 +1000,8 @@ err_out: | |||
1014 | 1000 | ||
1015 | static struct dma_async_tx_descriptor *sdma_prep_dma_cyclic( | 1001 | static struct dma_async_tx_descriptor *sdma_prep_dma_cyclic( |
1016 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, | 1002 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, |
1017 | size_t period_len, enum dma_transfer_direction direction) | 1003 | size_t period_len, enum dma_transfer_direction direction, |
1004 | void *context) | ||
1018 | { | 1005 | { |
1019 | struct sdma_channel *sdmac = to_sdma_chan(chan); | 1006 | struct sdma_channel *sdmac = to_sdma_chan(chan); |
1020 | struct sdma_engine *sdma = sdmac->sdma; | 1007 | struct sdma_engine *sdma = sdmac->sdma; |
@@ -1128,7 +1115,7 @@ static enum dma_status sdma_tx_status(struct dma_chan *chan, | |||
1128 | 1115 | ||
1129 | last_used = chan->cookie; | 1116 | last_used = chan->cookie; |
1130 | 1117 | ||
1131 | dma_set_tx_state(txstate, sdmac->last_completed, last_used, | 1118 | dma_set_tx_state(txstate, chan->completed_cookie, last_used, |
1132 | sdmac->chn_count - sdmac->chn_real_count); | 1119 | sdmac->chn_count - sdmac->chn_real_count); |
1133 | 1120 | ||
1134 | return sdmac->status; | 1121 | return sdmac->status; |
@@ -1136,9 +1123,11 @@ static enum dma_status sdma_tx_status(struct dma_chan *chan, | |||
1136 | 1123 | ||
1137 | static void sdma_issue_pending(struct dma_chan *chan) | 1124 | static void sdma_issue_pending(struct dma_chan *chan) |
1138 | { | 1125 | { |
1139 | /* | 1126 | struct sdma_channel *sdmac = to_sdma_chan(chan); |
1140 | * Nothing to do. We only have a single descriptor | 1127 | struct sdma_engine *sdma = sdmac->sdma; |
1141 | */ | 1128 | |
1129 | if (sdmac->status == DMA_IN_PROGRESS) | ||
1130 | sdma_enable_channel(sdma, sdmac->channel); | ||
1142 | } | 1131 | } |
1143 | 1132 | ||
1144 | #define SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1 34 | 1133 | #define SDMA_SCRIPT_ADDRS_ARRAY_SIZE_V1 34 |
@@ -1230,7 +1219,7 @@ static int __init sdma_init(struct sdma_engine *sdma) | |||
1230 | clk_enable(sdma->clk); | 1219 | clk_enable(sdma->clk); |
1231 | 1220 | ||
1232 | /* Be sure SDMA has not started yet */ | 1221 | /* Be sure SDMA has not started yet */ |
1233 | __raw_writel(0, sdma->regs + SDMA_H_C0PTR); | 1222 | writel_relaxed(0, sdma->regs + SDMA_H_C0PTR); |
1234 | 1223 | ||
1235 | sdma->channel_control = dma_alloc_coherent(NULL, | 1224 | sdma->channel_control = dma_alloc_coherent(NULL, |
1236 | MAX_DMA_CHANNELS * sizeof (struct sdma_channel_control) + | 1225 | MAX_DMA_CHANNELS * sizeof (struct sdma_channel_control) + |
@@ -1253,11 +1242,11 @@ static int __init sdma_init(struct sdma_engine *sdma) | |||
1253 | 1242 | ||
1254 | /* disable all channels */ | 1243 | /* disable all channels */ |
1255 | for (i = 0; i < sdma->num_events; i++) | 1244 | for (i = 0; i < sdma->num_events; i++) |
1256 | __raw_writel(0, sdma->regs + chnenbl_ofs(sdma, i)); | 1245 | writel_relaxed(0, sdma->regs + chnenbl_ofs(sdma, i)); |
1257 | 1246 | ||
1258 | /* All channels have priority 0 */ | 1247 | /* All channels have priority 0 */ |
1259 | for (i = 0; i < MAX_DMA_CHANNELS; i++) | 1248 | for (i = 0; i < MAX_DMA_CHANNELS; i++) |
1260 | __raw_writel(0, sdma->regs + SDMA_CHNPRI_0 + i * 4); | 1249 | writel_relaxed(0, sdma->regs + SDMA_CHNPRI_0 + i * 4); |
1261 | 1250 | ||
1262 | ret = sdma_request_channel(&sdma->channel[0]); | 1251 | ret = sdma_request_channel(&sdma->channel[0]); |
1263 | if (ret) | 1252 | if (ret) |
@@ -1266,16 +1255,16 @@ static int __init sdma_init(struct sdma_engine *sdma) | |||
1266 | sdma_config_ownership(&sdma->channel[0], false, true, false); | 1255 | sdma_config_ownership(&sdma->channel[0], false, true, false); |
1267 | 1256 | ||
1268 | /* Set Command Channel (Channel Zero) */ | 1257 | /* Set Command Channel (Channel Zero) */ |
1269 | __raw_writel(0x4050, sdma->regs + SDMA_CHN0ADDR); | 1258 | writel_relaxed(0x4050, sdma->regs + SDMA_CHN0ADDR); |
1270 | 1259 | ||
1271 | /* Set bits of CONFIG register but with static context switching */ | 1260 | /* Set bits of CONFIG register but with static context switching */ |
1272 | /* FIXME: Check whether to set ACR bit depending on clock ratios */ | 1261 | /* FIXME: Check whether to set ACR bit depending on clock ratios */ |
1273 | __raw_writel(0, sdma->regs + SDMA_H_CONFIG); | 1262 | writel_relaxed(0, sdma->regs + SDMA_H_CONFIG); |
1274 | 1263 | ||
1275 | __raw_writel(ccb_phys, sdma->regs + SDMA_H_C0PTR); | 1264 | writel_relaxed(ccb_phys, sdma->regs + SDMA_H_C0PTR); |
1276 | 1265 | ||
1277 | /* Set bits of CONFIG register with given context switching mode */ | 1266 | /* Set bits of CONFIG register with given context switching mode */ |
1278 | __raw_writel(SDMA_H_CONFIG_CSM, sdma->regs + SDMA_H_CONFIG); | 1267 | writel_relaxed(SDMA_H_CONFIG_CSM, sdma->regs + SDMA_H_CONFIG); |
1279 | 1268 | ||
1280 | /* Initializes channel's priorities */ | 1269 | /* Initializes channel's priorities */ |
1281 | sdma_set_channel_priority(&sdma->channel[0], 7); | 1270 | sdma_set_channel_priority(&sdma->channel[0], 7); |
@@ -1367,6 +1356,7 @@ static int __init sdma_probe(struct platform_device *pdev) | |||
1367 | spin_lock_init(&sdmac->lock); | 1356 | spin_lock_init(&sdmac->lock); |
1368 | 1357 | ||
1369 | sdmac->chan.device = &sdma->dma_device; | 1358 | sdmac->chan.device = &sdma->dma_device; |
1359 | dma_cookie_init(&sdmac->chan); | ||
1370 | sdmac->channel = i; | 1360 | sdmac->channel = i; |
1371 | 1361 | ||
1372 | /* | 1362 | /* |
@@ -1387,7 +1377,9 @@ static int __init sdma_probe(struct platform_device *pdev) | |||
1387 | sdma_add_scripts(sdma, pdata->script_addrs); | 1377 | sdma_add_scripts(sdma, pdata->script_addrs); |
1388 | 1378 | ||
1389 | if (pdata) { | 1379 | if (pdata) { |
1390 | sdma_get_firmware(sdma, pdata->fw_name); | 1380 | ret = sdma_get_firmware(sdma, pdata->fw_name); |
1381 | if (ret) | ||
1382 | dev_warn(&pdev->dev, "failed to get firmware from platform data\n"); | ||
1391 | } else { | 1383 | } else { |
1392 | /* | 1384 | /* |
1393 | * Because that device tree does not encode ROM script address, | 1385 | * Because that device tree does not encode ROM script address, |
@@ -1396,15 +1388,12 @@ static int __init sdma_probe(struct platform_device *pdev) | |||
1396 | */ | 1388 | */ |
1397 | ret = of_property_read_string(np, "fsl,sdma-ram-script-name", | 1389 | ret = of_property_read_string(np, "fsl,sdma-ram-script-name", |
1398 | &fw_name); | 1390 | &fw_name); |
1399 | if (ret) { | 1391 | if (ret) |
1400 | dev_err(&pdev->dev, "failed to get firmware name\n"); | 1392 | dev_warn(&pdev->dev, "failed to get firmware name\n"); |
1401 | goto err_init; | 1393 | else { |
1402 | } | 1394 | ret = sdma_get_firmware(sdma, fw_name); |
1403 | 1395 | if (ret) | |
1404 | ret = sdma_get_firmware(sdma, fw_name); | 1396 | dev_warn(&pdev->dev, "failed to get firmware from device tree\n"); |
1405 | if (ret) { | ||
1406 | dev_err(&pdev->dev, "failed to get firmware\n"); | ||
1407 | goto err_init; | ||
1408 | } | 1397 | } |
1409 | } | 1398 | } |
1410 | 1399 | ||
diff --git a/drivers/dma/intel_mid_dma.c b/drivers/dma/intel_mid_dma.c index 74f70aadf9e4..c900ca7aaec4 100644 --- a/drivers/dma/intel_mid_dma.c +++ b/drivers/dma/intel_mid_dma.c | |||
@@ -29,6 +29,8 @@ | |||
29 | #include <linux/intel_mid_dma.h> | 29 | #include <linux/intel_mid_dma.h> |
30 | #include <linux/module.h> | 30 | #include <linux/module.h> |
31 | 31 | ||
32 | #include "dmaengine.h" | ||
33 | |||
32 | #define MAX_CHAN 4 /*max ch across controllers*/ | 34 | #define MAX_CHAN 4 /*max ch across controllers*/ |
33 | #include "intel_mid_dma_regs.h" | 35 | #include "intel_mid_dma_regs.h" |
34 | 36 | ||
@@ -288,7 +290,7 @@ static void midc_descriptor_complete(struct intel_mid_dma_chan *midc, | |||
288 | struct intel_mid_dma_lli *llitem; | 290 | struct intel_mid_dma_lli *llitem; |
289 | void *param_txd = NULL; | 291 | void *param_txd = NULL; |
290 | 292 | ||
291 | midc->completed = txd->cookie; | 293 | dma_cookie_complete(txd); |
292 | callback_txd = txd->callback; | 294 | callback_txd = txd->callback; |
293 | param_txd = txd->callback_param; | 295 | param_txd = txd->callback_param; |
294 | 296 | ||
@@ -434,14 +436,7 @@ static dma_cookie_t intel_mid_dma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
434 | dma_cookie_t cookie; | 436 | dma_cookie_t cookie; |
435 | 437 | ||
436 | spin_lock_bh(&midc->lock); | 438 | spin_lock_bh(&midc->lock); |
437 | cookie = midc->chan.cookie; | 439 | cookie = dma_cookie_assign(tx); |
438 | |||
439 | if (++cookie < 0) | ||
440 | cookie = 1; | ||
441 | |||
442 | midc->chan.cookie = cookie; | ||
443 | desc->txd.cookie = cookie; | ||
444 | |||
445 | 440 | ||
446 | if (list_empty(&midc->active_list)) | 441 | if (list_empty(&midc->active_list)) |
447 | list_add_tail(&desc->desc_node, &midc->active_list); | 442 | list_add_tail(&desc->desc_node, &midc->active_list); |
@@ -482,31 +477,18 @@ static enum dma_status intel_mid_dma_tx_status(struct dma_chan *chan, | |||
482 | dma_cookie_t cookie, | 477 | dma_cookie_t cookie, |
483 | struct dma_tx_state *txstate) | 478 | struct dma_tx_state *txstate) |
484 | { | 479 | { |
485 | struct intel_mid_dma_chan *midc = to_intel_mid_dma_chan(chan); | 480 | struct intel_mid_dma_chan *midc = to_intel_mid_dma_chan(chan); |
486 | dma_cookie_t last_used; | 481 | enum dma_status ret; |
487 | dma_cookie_t last_complete; | ||
488 | int ret; | ||
489 | 482 | ||
490 | last_complete = midc->completed; | 483 | ret = dma_cookie_status(chan, cookie, txstate); |
491 | last_used = chan->cookie; | ||
492 | |||
493 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
494 | if (ret != DMA_SUCCESS) { | 484 | if (ret != DMA_SUCCESS) { |
495 | spin_lock_bh(&midc->lock); | 485 | spin_lock_bh(&midc->lock); |
496 | midc_scan_descriptors(to_middma_device(chan->device), midc); | 486 | midc_scan_descriptors(to_middma_device(chan->device), midc); |
497 | spin_unlock_bh(&midc->lock); | 487 | spin_unlock_bh(&midc->lock); |
498 | 488 | ||
499 | last_complete = midc->completed; | 489 | ret = dma_cookie_status(chan, cookie, txstate); |
500 | last_used = chan->cookie; | ||
501 | |||
502 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
503 | } | 490 | } |
504 | 491 | ||
505 | if (txstate) { | ||
506 | txstate->last = last_complete; | ||
507 | txstate->used = last_used; | ||
508 | txstate->residue = 0; | ||
509 | } | ||
510 | return ret; | 492 | return ret; |
511 | } | 493 | } |
512 | 494 | ||
@@ -732,13 +714,14 @@ err_desc_get: | |||
732 | * @sg_len: length of sg txn | 714 | * @sg_len: length of sg txn |
733 | * @direction: DMA transfer dirtn | 715 | * @direction: DMA transfer dirtn |
734 | * @flags: DMA flags | 716 | * @flags: DMA flags |
717 | * @context: transfer context (ignored) | ||
735 | * | 718 | * |
736 | * Prepares LLI based periphral transfer | 719 | * Prepares LLI based periphral transfer |
737 | */ | 720 | */ |
738 | static struct dma_async_tx_descriptor *intel_mid_dma_prep_slave_sg( | 721 | static struct dma_async_tx_descriptor *intel_mid_dma_prep_slave_sg( |
739 | struct dma_chan *chan, struct scatterlist *sgl, | 722 | struct dma_chan *chan, struct scatterlist *sgl, |
740 | unsigned int sg_len, enum dma_transfer_direction direction, | 723 | unsigned int sg_len, enum dma_transfer_direction direction, |
741 | unsigned long flags) | 724 | unsigned long flags, void *context) |
742 | { | 725 | { |
743 | struct intel_mid_dma_chan *midc = NULL; | 726 | struct intel_mid_dma_chan *midc = NULL; |
744 | struct intel_mid_dma_slave *mids = NULL; | 727 | struct intel_mid_dma_slave *mids = NULL; |
@@ -832,7 +815,6 @@ static void intel_mid_dma_free_chan_resources(struct dma_chan *chan) | |||
832 | /*trying to free ch in use!!!!!*/ | 815 | /*trying to free ch in use!!!!!*/ |
833 | pr_err("ERR_MDMA: trying to free ch in use\n"); | 816 | pr_err("ERR_MDMA: trying to free ch in use\n"); |
834 | } | 817 | } |
835 | pm_runtime_put(&mid->pdev->dev); | ||
836 | spin_lock_bh(&midc->lock); | 818 | spin_lock_bh(&midc->lock); |
837 | midc->descs_allocated = 0; | 819 | midc->descs_allocated = 0; |
838 | list_for_each_entry_safe(desc, _desc, &midc->active_list, desc_node) { | 820 | list_for_each_entry_safe(desc, _desc, &midc->active_list, desc_node) { |
@@ -853,6 +835,7 @@ static void intel_mid_dma_free_chan_resources(struct dma_chan *chan) | |||
853 | /* Disable CH interrupts */ | 835 | /* Disable CH interrupts */ |
854 | iowrite32(MASK_INTR_REG(midc->ch_id), mid->dma_base + MASK_BLOCK); | 836 | iowrite32(MASK_INTR_REG(midc->ch_id), mid->dma_base + MASK_BLOCK); |
855 | iowrite32(MASK_INTR_REG(midc->ch_id), mid->dma_base + MASK_ERR); | 837 | iowrite32(MASK_INTR_REG(midc->ch_id), mid->dma_base + MASK_ERR); |
838 | pm_runtime_put(&mid->pdev->dev); | ||
856 | } | 839 | } |
857 | 840 | ||
858 | /** | 841 | /** |
@@ -886,7 +869,7 @@ static int intel_mid_dma_alloc_chan_resources(struct dma_chan *chan) | |||
886 | pm_runtime_put(&mid->pdev->dev); | 869 | pm_runtime_put(&mid->pdev->dev); |
887 | return -EIO; | 870 | return -EIO; |
888 | } | 871 | } |
889 | midc->completed = chan->cookie = 1; | 872 | dma_cookie_init(chan); |
890 | 873 | ||
891 | spin_lock_bh(&midc->lock); | 874 | spin_lock_bh(&midc->lock); |
892 | while (midc->descs_allocated < DESCS_PER_CHANNEL) { | 875 | while (midc->descs_allocated < DESCS_PER_CHANNEL) { |
@@ -1056,7 +1039,8 @@ static irqreturn_t intel_mid_dma_interrupt(int irq, void *data) | |||
1056 | } | 1039 | } |
1057 | err_status &= mid->intr_mask; | 1040 | err_status &= mid->intr_mask; |
1058 | if (err_status) { | 1041 | if (err_status) { |
1059 | iowrite32(MASK_INTR_REG(err_status), mid->dma_base + MASK_ERR); | 1042 | iowrite32((err_status << INT_MASK_WE), |
1043 | mid->dma_base + MASK_ERR); | ||
1060 | call_tasklet = 1; | 1044 | call_tasklet = 1; |
1061 | } | 1045 | } |
1062 | if (call_tasklet) | 1046 | if (call_tasklet) |
@@ -1118,7 +1102,7 @@ static int mid_setup_dma(struct pci_dev *pdev) | |||
1118 | struct intel_mid_dma_chan *midch = &dma->ch[i]; | 1102 | struct intel_mid_dma_chan *midch = &dma->ch[i]; |
1119 | 1103 | ||
1120 | midch->chan.device = &dma->common; | 1104 | midch->chan.device = &dma->common; |
1121 | midch->chan.cookie = 1; | 1105 | dma_cookie_init(&midch->chan); |
1122 | midch->ch_id = dma->chan_base + i; | 1106 | midch->ch_id = dma->chan_base + i; |
1123 | pr_debug("MDMA:Init CH %d, ID %d\n", i, midch->ch_id); | 1107 | pr_debug("MDMA:Init CH %d, ID %d\n", i, midch->ch_id); |
1124 | 1108 | ||
diff --git a/drivers/dma/intel_mid_dma_regs.h b/drivers/dma/intel_mid_dma_regs.h index c83d35b97bd8..1bfa9268feaf 100644 --- a/drivers/dma/intel_mid_dma_regs.h +++ b/drivers/dma/intel_mid_dma_regs.h | |||
@@ -165,7 +165,6 @@ union intel_mid_dma_cfg_hi { | |||
165 | * @dma_base: MMIO register space DMA engine base pointer | 165 | * @dma_base: MMIO register space DMA engine base pointer |
166 | * @ch_id: DMA channel id | 166 | * @ch_id: DMA channel id |
167 | * @lock: channel spinlock | 167 | * @lock: channel spinlock |
168 | * @completed: DMA cookie | ||
169 | * @active_list: current active descriptors | 168 | * @active_list: current active descriptors |
170 | * @queue: current queued up descriptors | 169 | * @queue: current queued up descriptors |
171 | * @free_list: current free descriptors | 170 | * @free_list: current free descriptors |
@@ -183,7 +182,6 @@ struct intel_mid_dma_chan { | |||
183 | void __iomem *dma_base; | 182 | void __iomem *dma_base; |
184 | int ch_id; | 183 | int ch_id; |
185 | spinlock_t lock; | 184 | spinlock_t lock; |
186 | dma_cookie_t completed; | ||
187 | struct list_head active_list; | 185 | struct list_head active_list; |
188 | struct list_head queue; | 186 | struct list_head queue; |
189 | struct list_head free_list; | 187 | struct list_head free_list; |
diff --git a/drivers/dma/ioat/dma.c b/drivers/dma/ioat/dma.c index a4d6cb0c0343..31493d80e0e9 100644 --- a/drivers/dma/ioat/dma.c +++ b/drivers/dma/ioat/dma.c | |||
@@ -40,6 +40,8 @@ | |||
40 | #include "registers.h" | 40 | #include "registers.h" |
41 | #include "hw.h" | 41 | #include "hw.h" |
42 | 42 | ||
43 | #include "../dmaengine.h" | ||
44 | |||
43 | int ioat_pending_level = 4; | 45 | int ioat_pending_level = 4; |
44 | module_param(ioat_pending_level, int, 0644); | 46 | module_param(ioat_pending_level, int, 0644); |
45 | MODULE_PARM_DESC(ioat_pending_level, | 47 | MODULE_PARM_DESC(ioat_pending_level, |
@@ -107,6 +109,7 @@ void ioat_init_channel(struct ioatdma_device *device, struct ioat_chan_common *c | |||
107 | chan->reg_base = device->reg_base + (0x80 * (idx + 1)); | 109 | chan->reg_base = device->reg_base + (0x80 * (idx + 1)); |
108 | spin_lock_init(&chan->cleanup_lock); | 110 | spin_lock_init(&chan->cleanup_lock); |
109 | chan->common.device = dma; | 111 | chan->common.device = dma; |
112 | dma_cookie_init(&chan->common); | ||
110 | list_add_tail(&chan->common.device_node, &dma->channels); | 113 | list_add_tail(&chan->common.device_node, &dma->channels); |
111 | device->idx[idx] = chan; | 114 | device->idx[idx] = chan; |
112 | init_timer(&chan->timer); | 115 | init_timer(&chan->timer); |
@@ -235,12 +238,7 @@ static dma_cookie_t ioat1_tx_submit(struct dma_async_tx_descriptor *tx) | |||
235 | 238 | ||
236 | spin_lock_bh(&ioat->desc_lock); | 239 | spin_lock_bh(&ioat->desc_lock); |
237 | /* cookie incr and addition to used_list must be atomic */ | 240 | /* cookie incr and addition to used_list must be atomic */ |
238 | cookie = c->cookie; | 241 | cookie = dma_cookie_assign(tx); |
239 | cookie++; | ||
240 | if (cookie < 0) | ||
241 | cookie = 1; | ||
242 | c->cookie = cookie; | ||
243 | tx->cookie = cookie; | ||
244 | dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie); | 242 | dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie); |
245 | 243 | ||
246 | /* write address into NextDescriptor field of last desc in chain */ | 244 | /* write address into NextDescriptor field of last desc in chain */ |
@@ -603,8 +601,7 @@ static void __cleanup(struct ioat_dma_chan *ioat, unsigned long phys_complete) | |||
603 | */ | 601 | */ |
604 | dump_desc_dbg(ioat, desc); | 602 | dump_desc_dbg(ioat, desc); |
605 | if (tx->cookie) { | 603 | if (tx->cookie) { |
606 | chan->completed_cookie = tx->cookie; | 604 | dma_cookie_complete(tx); |
607 | tx->cookie = 0; | ||
608 | ioat_dma_unmap(chan, tx->flags, desc->len, desc->hw); | 605 | ioat_dma_unmap(chan, tx->flags, desc->len, desc->hw); |
609 | ioat->active -= desc->hw->tx_cnt; | 606 | ioat->active -= desc->hw->tx_cnt; |
610 | if (tx->callback) { | 607 | if (tx->callback) { |
@@ -733,13 +730,15 @@ ioat_dma_tx_status(struct dma_chan *c, dma_cookie_t cookie, | |||
733 | { | 730 | { |
734 | struct ioat_chan_common *chan = to_chan_common(c); | 731 | struct ioat_chan_common *chan = to_chan_common(c); |
735 | struct ioatdma_device *device = chan->device; | 732 | struct ioatdma_device *device = chan->device; |
733 | enum dma_status ret; | ||
736 | 734 | ||
737 | if (ioat_tx_status(c, cookie, txstate) == DMA_SUCCESS) | 735 | ret = dma_cookie_status(c, cookie, txstate); |
738 | return DMA_SUCCESS; | 736 | if (ret == DMA_SUCCESS) |
737 | return ret; | ||
739 | 738 | ||
740 | device->cleanup_fn((unsigned long) c); | 739 | device->cleanup_fn((unsigned long) c); |
741 | 740 | ||
742 | return ioat_tx_status(c, cookie, txstate); | 741 | return dma_cookie_status(c, cookie, txstate); |
743 | } | 742 | } |
744 | 743 | ||
745 | static void ioat1_dma_start_null_desc(struct ioat_dma_chan *ioat) | 744 | static void ioat1_dma_start_null_desc(struct ioat_dma_chan *ioat) |
diff --git a/drivers/dma/ioat/dma.h b/drivers/dma/ioat/dma.h index 5216c8a92a21..c7888bccd974 100644 --- a/drivers/dma/ioat/dma.h +++ b/drivers/dma/ioat/dma.h | |||
@@ -90,7 +90,6 @@ struct ioat_chan_common { | |||
90 | void __iomem *reg_base; | 90 | void __iomem *reg_base; |
91 | unsigned long last_completion; | 91 | unsigned long last_completion; |
92 | spinlock_t cleanup_lock; | 92 | spinlock_t cleanup_lock; |
93 | dma_cookie_t completed_cookie; | ||
94 | unsigned long state; | 93 | unsigned long state; |
95 | #define IOAT_COMPLETION_PENDING 0 | 94 | #define IOAT_COMPLETION_PENDING 0 |
96 | #define IOAT_COMPLETION_ACK 1 | 95 | #define IOAT_COMPLETION_ACK 1 |
@@ -143,28 +142,6 @@ static inline struct ioat_dma_chan *to_ioat_chan(struct dma_chan *c) | |||
143 | return container_of(chan, struct ioat_dma_chan, base); | 142 | return container_of(chan, struct ioat_dma_chan, base); |
144 | } | 143 | } |
145 | 144 | ||
146 | /** | ||
147 | * ioat_tx_status - poll the status of an ioat transaction | ||
148 | * @c: channel handle | ||
149 | * @cookie: transaction identifier | ||
150 | * @txstate: if set, updated with the transaction state | ||
151 | */ | ||
152 | static inline enum dma_status | ||
153 | ioat_tx_status(struct dma_chan *c, dma_cookie_t cookie, | ||
154 | struct dma_tx_state *txstate) | ||
155 | { | ||
156 | struct ioat_chan_common *chan = to_chan_common(c); | ||
157 | dma_cookie_t last_used; | ||
158 | dma_cookie_t last_complete; | ||
159 | |||
160 | last_used = c->cookie; | ||
161 | last_complete = chan->completed_cookie; | ||
162 | |||
163 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
164 | |||
165 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
166 | } | ||
167 | |||
168 | /* wrapper around hardware descriptor format + additional software fields */ | 145 | /* wrapper around hardware descriptor format + additional software fields */ |
169 | 146 | ||
170 | /** | 147 | /** |
diff --git a/drivers/dma/ioat/dma_v2.c b/drivers/dma/ioat/dma_v2.c index 5d65f8377971..e8e110ff3d96 100644 --- a/drivers/dma/ioat/dma_v2.c +++ b/drivers/dma/ioat/dma_v2.c | |||
@@ -41,6 +41,8 @@ | |||
41 | #include "registers.h" | 41 | #include "registers.h" |
42 | #include "hw.h" | 42 | #include "hw.h" |
43 | 43 | ||
44 | #include "../dmaengine.h" | ||
45 | |||
44 | int ioat_ring_alloc_order = 8; | 46 | int ioat_ring_alloc_order = 8; |
45 | module_param(ioat_ring_alloc_order, int, 0644); | 47 | module_param(ioat_ring_alloc_order, int, 0644); |
46 | MODULE_PARM_DESC(ioat_ring_alloc_order, | 48 | MODULE_PARM_DESC(ioat_ring_alloc_order, |
@@ -147,8 +149,7 @@ static void __cleanup(struct ioat2_dma_chan *ioat, unsigned long phys_complete) | |||
147 | dump_desc_dbg(ioat, desc); | 149 | dump_desc_dbg(ioat, desc); |
148 | if (tx->cookie) { | 150 | if (tx->cookie) { |
149 | ioat_dma_unmap(chan, tx->flags, desc->len, desc->hw); | 151 | ioat_dma_unmap(chan, tx->flags, desc->len, desc->hw); |
150 | chan->completed_cookie = tx->cookie; | 152 | dma_cookie_complete(tx); |
151 | tx->cookie = 0; | ||
152 | if (tx->callback) { | 153 | if (tx->callback) { |
153 | tx->callback(tx->callback_param); | 154 | tx->callback(tx->callback_param); |
154 | tx->callback = NULL; | 155 | tx->callback = NULL; |
@@ -398,13 +399,9 @@ static dma_cookie_t ioat2_tx_submit_unlock(struct dma_async_tx_descriptor *tx) | |||
398 | struct dma_chan *c = tx->chan; | 399 | struct dma_chan *c = tx->chan; |
399 | struct ioat2_dma_chan *ioat = to_ioat2_chan(c); | 400 | struct ioat2_dma_chan *ioat = to_ioat2_chan(c); |
400 | struct ioat_chan_common *chan = &ioat->base; | 401 | struct ioat_chan_common *chan = &ioat->base; |
401 | dma_cookie_t cookie = c->cookie; | 402 | dma_cookie_t cookie; |
402 | 403 | ||
403 | cookie++; | 404 | cookie = dma_cookie_assign(tx); |
404 | if (cookie < 0) | ||
405 | cookie = 1; | ||
406 | tx->cookie = cookie; | ||
407 | c->cookie = cookie; | ||
408 | dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie); | 405 | dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie); |
409 | 406 | ||
410 | if (!test_and_set_bit(IOAT_COMPLETION_PENDING, &chan->state)) | 407 | if (!test_and_set_bit(IOAT_COMPLETION_PENDING, &chan->state)) |
diff --git a/drivers/dma/ioat/dma_v3.c b/drivers/dma/ioat/dma_v3.c index f519c93a61e7..2c4476c0e405 100644 --- a/drivers/dma/ioat/dma_v3.c +++ b/drivers/dma/ioat/dma_v3.c | |||
@@ -61,6 +61,7 @@ | |||
61 | #include <linux/dmaengine.h> | 61 | #include <linux/dmaengine.h> |
62 | #include <linux/dma-mapping.h> | 62 | #include <linux/dma-mapping.h> |
63 | #include <linux/prefetch.h> | 63 | #include <linux/prefetch.h> |
64 | #include "../dmaengine.h" | ||
64 | #include "registers.h" | 65 | #include "registers.h" |
65 | #include "hw.h" | 66 | #include "hw.h" |
66 | #include "dma.h" | 67 | #include "dma.h" |
@@ -277,9 +278,8 @@ static void __cleanup(struct ioat2_dma_chan *ioat, unsigned long phys_complete) | |||
277 | dump_desc_dbg(ioat, desc); | 278 | dump_desc_dbg(ioat, desc); |
278 | tx = &desc->txd; | 279 | tx = &desc->txd; |
279 | if (tx->cookie) { | 280 | if (tx->cookie) { |
280 | chan->completed_cookie = tx->cookie; | 281 | dma_cookie_complete(tx); |
281 | ioat3_dma_unmap(ioat, desc, idx + i); | 282 | ioat3_dma_unmap(ioat, desc, idx + i); |
282 | tx->cookie = 0; | ||
283 | if (tx->callback) { | 283 | if (tx->callback) { |
284 | tx->callback(tx->callback_param); | 284 | tx->callback(tx->callback_param); |
285 | tx->callback = NULL; | 285 | tx->callback = NULL; |
@@ -411,13 +411,15 @@ ioat3_tx_status(struct dma_chan *c, dma_cookie_t cookie, | |||
411 | struct dma_tx_state *txstate) | 411 | struct dma_tx_state *txstate) |
412 | { | 412 | { |
413 | struct ioat2_dma_chan *ioat = to_ioat2_chan(c); | 413 | struct ioat2_dma_chan *ioat = to_ioat2_chan(c); |
414 | enum dma_status ret; | ||
414 | 415 | ||
415 | if (ioat_tx_status(c, cookie, txstate) == DMA_SUCCESS) | 416 | ret = dma_cookie_status(c, cookie, txstate); |
416 | return DMA_SUCCESS; | 417 | if (ret == DMA_SUCCESS) |
418 | return ret; | ||
417 | 419 | ||
418 | ioat3_cleanup(ioat); | 420 | ioat3_cleanup(ioat); |
419 | 421 | ||
420 | return ioat_tx_status(c, cookie, txstate); | 422 | return dma_cookie_status(c, cookie, txstate); |
421 | } | 423 | } |
422 | 424 | ||
423 | static struct dma_async_tx_descriptor * | 425 | static struct dma_async_tx_descriptor * |
diff --git a/drivers/dma/iop-adma.c b/drivers/dma/iop-adma.c index faf88b7e1e71..da6c4c2c066a 100644 --- a/drivers/dma/iop-adma.c +++ b/drivers/dma/iop-adma.c | |||
@@ -36,6 +36,8 @@ | |||
36 | 36 | ||
37 | #include <mach/adma.h> | 37 | #include <mach/adma.h> |
38 | 38 | ||
39 | #include "dmaengine.h" | ||
40 | |||
39 | #define to_iop_adma_chan(chan) container_of(chan, struct iop_adma_chan, common) | 41 | #define to_iop_adma_chan(chan) container_of(chan, struct iop_adma_chan, common) |
40 | #define to_iop_adma_device(dev) \ | 42 | #define to_iop_adma_device(dev) \ |
41 | container_of(dev, struct iop_adma_device, common) | 43 | container_of(dev, struct iop_adma_device, common) |
@@ -317,7 +319,7 @@ static void __iop_adma_slot_cleanup(struct iop_adma_chan *iop_chan) | |||
317 | } | 319 | } |
318 | 320 | ||
319 | if (cookie > 0) { | 321 | if (cookie > 0) { |
320 | iop_chan->completed_cookie = cookie; | 322 | iop_chan->common.completed_cookie = cookie; |
321 | pr_debug("\tcompleted cookie %d\n", cookie); | 323 | pr_debug("\tcompleted cookie %d\n", cookie); |
322 | } | 324 | } |
323 | } | 325 | } |
@@ -438,18 +440,6 @@ retry: | |||
438 | return NULL; | 440 | return NULL; |
439 | } | 441 | } |
440 | 442 | ||
441 | static dma_cookie_t | ||
442 | iop_desc_assign_cookie(struct iop_adma_chan *iop_chan, | ||
443 | struct iop_adma_desc_slot *desc) | ||
444 | { | ||
445 | dma_cookie_t cookie = iop_chan->common.cookie; | ||
446 | cookie++; | ||
447 | if (cookie < 0) | ||
448 | cookie = 1; | ||
449 | iop_chan->common.cookie = desc->async_tx.cookie = cookie; | ||
450 | return cookie; | ||
451 | } | ||
452 | |||
453 | static void iop_adma_check_threshold(struct iop_adma_chan *iop_chan) | 443 | static void iop_adma_check_threshold(struct iop_adma_chan *iop_chan) |
454 | { | 444 | { |
455 | dev_dbg(iop_chan->device->common.dev, "pending: %d\n", | 445 | dev_dbg(iop_chan->device->common.dev, "pending: %d\n", |
@@ -477,7 +467,7 @@ iop_adma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
477 | slots_per_op = grp_start->slots_per_op; | 467 | slots_per_op = grp_start->slots_per_op; |
478 | 468 | ||
479 | spin_lock_bh(&iop_chan->lock); | 469 | spin_lock_bh(&iop_chan->lock); |
480 | cookie = iop_desc_assign_cookie(iop_chan, sw_desc); | 470 | cookie = dma_cookie_assign(tx); |
481 | 471 | ||
482 | old_chain_tail = list_entry(iop_chan->chain.prev, | 472 | old_chain_tail = list_entry(iop_chan->chain.prev, |
483 | struct iop_adma_desc_slot, chain_node); | 473 | struct iop_adma_desc_slot, chain_node); |
@@ -904,24 +894,15 @@ static enum dma_status iop_adma_status(struct dma_chan *chan, | |||
904 | struct dma_tx_state *txstate) | 894 | struct dma_tx_state *txstate) |
905 | { | 895 | { |
906 | struct iop_adma_chan *iop_chan = to_iop_adma_chan(chan); | 896 | struct iop_adma_chan *iop_chan = to_iop_adma_chan(chan); |
907 | dma_cookie_t last_used; | 897 | int ret; |
908 | dma_cookie_t last_complete; | 898 | |
909 | enum dma_status ret; | 899 | ret = dma_cookie_status(chan, cookie, txstate); |
910 | |||
911 | last_used = chan->cookie; | ||
912 | last_complete = iop_chan->completed_cookie; | ||
913 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
914 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
915 | if (ret == DMA_SUCCESS) | 900 | if (ret == DMA_SUCCESS) |
916 | return ret; | 901 | return ret; |
917 | 902 | ||
918 | iop_adma_slot_cleanup(iop_chan); | 903 | iop_adma_slot_cleanup(iop_chan); |
919 | 904 | ||
920 | last_used = chan->cookie; | 905 | return dma_cookie_status(chan, cookie, txstate); |
921 | last_complete = iop_chan->completed_cookie; | ||
922 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
923 | |||
924 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
925 | } | 906 | } |
926 | 907 | ||
927 | static irqreturn_t iop_adma_eot_handler(int irq, void *data) | 908 | static irqreturn_t iop_adma_eot_handler(int irq, void *data) |
@@ -1565,6 +1546,7 @@ static int __devinit iop_adma_probe(struct platform_device *pdev) | |||
1565 | INIT_LIST_HEAD(&iop_chan->chain); | 1546 | INIT_LIST_HEAD(&iop_chan->chain); |
1566 | INIT_LIST_HEAD(&iop_chan->all_slots); | 1547 | INIT_LIST_HEAD(&iop_chan->all_slots); |
1567 | iop_chan->common.device = dma_dev; | 1548 | iop_chan->common.device = dma_dev; |
1549 | dma_cookie_init(&iop_chan->common); | ||
1568 | list_add_tail(&iop_chan->common.device_node, &dma_dev->channels); | 1550 | list_add_tail(&iop_chan->common.device_node, &dma_dev->channels); |
1569 | 1551 | ||
1570 | if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { | 1552 | if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { |
@@ -1642,16 +1624,12 @@ static void iop_chan_start_null_memcpy(struct iop_adma_chan *iop_chan) | |||
1642 | iop_desc_set_dest_addr(grp_start, iop_chan, 0); | 1624 | iop_desc_set_dest_addr(grp_start, iop_chan, 0); |
1643 | iop_desc_set_memcpy_src_addr(grp_start, 0); | 1625 | iop_desc_set_memcpy_src_addr(grp_start, 0); |
1644 | 1626 | ||
1645 | cookie = iop_chan->common.cookie; | 1627 | cookie = dma_cookie_assign(&sw_desc->async_tx); |
1646 | cookie++; | ||
1647 | if (cookie <= 1) | ||
1648 | cookie = 2; | ||
1649 | 1628 | ||
1650 | /* initialize the completed cookie to be less than | 1629 | /* initialize the completed cookie to be less than |
1651 | * the most recently used cookie | 1630 | * the most recently used cookie |
1652 | */ | 1631 | */ |
1653 | iop_chan->completed_cookie = cookie - 1; | 1632 | iop_chan->common.completed_cookie = cookie - 1; |
1654 | iop_chan->common.cookie = sw_desc->async_tx.cookie = cookie; | ||
1655 | 1633 | ||
1656 | /* channel should not be busy */ | 1634 | /* channel should not be busy */ |
1657 | BUG_ON(iop_chan_is_busy(iop_chan)); | 1635 | BUG_ON(iop_chan_is_busy(iop_chan)); |
@@ -1699,16 +1677,12 @@ static void iop_chan_start_null_xor(struct iop_adma_chan *iop_chan) | |||
1699 | iop_desc_set_xor_src_addr(grp_start, 0, 0); | 1677 | iop_desc_set_xor_src_addr(grp_start, 0, 0); |
1700 | iop_desc_set_xor_src_addr(grp_start, 1, 0); | 1678 | iop_desc_set_xor_src_addr(grp_start, 1, 0); |
1701 | 1679 | ||
1702 | cookie = iop_chan->common.cookie; | 1680 | cookie = dma_cookie_assign(&sw_desc->async_tx); |
1703 | cookie++; | ||
1704 | if (cookie <= 1) | ||
1705 | cookie = 2; | ||
1706 | 1681 | ||
1707 | /* initialize the completed cookie to be less than | 1682 | /* initialize the completed cookie to be less than |
1708 | * the most recently used cookie | 1683 | * the most recently used cookie |
1709 | */ | 1684 | */ |
1710 | iop_chan->completed_cookie = cookie - 1; | 1685 | iop_chan->common.completed_cookie = cookie - 1; |
1711 | iop_chan->common.cookie = sw_desc->async_tx.cookie = cookie; | ||
1712 | 1686 | ||
1713 | /* channel should not be busy */ | 1687 | /* channel should not be busy */ |
1714 | BUG_ON(iop_chan_is_busy(iop_chan)); | 1688 | BUG_ON(iop_chan_is_busy(iop_chan)); |
diff --git a/drivers/dma/ipu/ipu_idmac.c b/drivers/dma/ipu/ipu_idmac.c index 6212b16e8cf2..62e3f8ec2461 100644 --- a/drivers/dma/ipu/ipu_idmac.c +++ b/drivers/dma/ipu/ipu_idmac.c | |||
@@ -25,6 +25,7 @@ | |||
25 | 25 | ||
26 | #include <mach/ipu.h> | 26 | #include <mach/ipu.h> |
27 | 27 | ||
28 | #include "../dmaengine.h" | ||
28 | #include "ipu_intern.h" | 29 | #include "ipu_intern.h" |
29 | 30 | ||
30 | #define FS_VF_IN_VALID 0x00000002 | 31 | #define FS_VF_IN_VALID 0x00000002 |
@@ -866,14 +867,7 @@ static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx) | |||
866 | 867 | ||
867 | dev_dbg(dev, "Submitting sg %p\n", &desc->sg[0]); | 868 | dev_dbg(dev, "Submitting sg %p\n", &desc->sg[0]); |
868 | 869 | ||
869 | cookie = ichan->dma_chan.cookie; | 870 | cookie = dma_cookie_assign(tx); |
870 | |||
871 | if (++cookie < 0) | ||
872 | cookie = 1; | ||
873 | |||
874 | /* from dmaengine.h: "last cookie value returned to client" */ | ||
875 | ichan->dma_chan.cookie = cookie; | ||
876 | tx->cookie = cookie; | ||
877 | 871 | ||
878 | /* ipu->lock can be taken under ichan->lock, but not v.v. */ | 872 | /* ipu->lock can be taken under ichan->lock, but not v.v. */ |
879 | spin_lock_irqsave(&ichan->lock, flags); | 873 | spin_lock_irqsave(&ichan->lock, flags); |
@@ -1295,7 +1289,7 @@ static irqreturn_t idmac_interrupt(int irq, void *dev_id) | |||
1295 | /* Flip the active buffer - even if update above failed */ | 1289 | /* Flip the active buffer - even if update above failed */ |
1296 | ichan->active_buffer = !ichan->active_buffer; | 1290 | ichan->active_buffer = !ichan->active_buffer; |
1297 | if (done) | 1291 | if (done) |
1298 | ichan->completed = desc->txd.cookie; | 1292 | dma_cookie_complete(&desc->txd); |
1299 | 1293 | ||
1300 | callback = desc->txd.callback; | 1294 | callback = desc->txd.callback; |
1301 | callback_param = desc->txd.callback_param; | 1295 | callback_param = desc->txd.callback_param; |
@@ -1341,7 +1335,8 @@ static void ipu_gc_tasklet(unsigned long arg) | |||
1341 | /* Allocate and initialise a transfer descriptor. */ | 1335 | /* Allocate and initialise a transfer descriptor. */ |
1342 | static struct dma_async_tx_descriptor *idmac_prep_slave_sg(struct dma_chan *chan, | 1336 | static struct dma_async_tx_descriptor *idmac_prep_slave_sg(struct dma_chan *chan, |
1343 | struct scatterlist *sgl, unsigned int sg_len, | 1337 | struct scatterlist *sgl, unsigned int sg_len, |
1344 | enum dma_transfer_direction direction, unsigned long tx_flags) | 1338 | enum dma_transfer_direction direction, unsigned long tx_flags, |
1339 | void *context) | ||
1345 | { | 1340 | { |
1346 | struct idmac_channel *ichan = to_idmac_chan(chan); | 1341 | struct idmac_channel *ichan = to_idmac_chan(chan); |
1347 | struct idmac_tx_desc *desc = NULL; | 1342 | struct idmac_tx_desc *desc = NULL; |
@@ -1510,8 +1505,7 @@ static int idmac_alloc_chan_resources(struct dma_chan *chan) | |||
1510 | BUG_ON(chan->client_count > 1); | 1505 | BUG_ON(chan->client_count > 1); |
1511 | WARN_ON(ichan->status != IPU_CHANNEL_FREE); | 1506 | WARN_ON(ichan->status != IPU_CHANNEL_FREE); |
1512 | 1507 | ||
1513 | chan->cookie = 1; | 1508 | dma_cookie_init(chan); |
1514 | ichan->completed = -ENXIO; | ||
1515 | 1509 | ||
1516 | ret = ipu_irq_map(chan->chan_id); | 1510 | ret = ipu_irq_map(chan->chan_id); |
1517 | if (ret < 0) | 1511 | if (ret < 0) |
@@ -1600,9 +1594,7 @@ static void idmac_free_chan_resources(struct dma_chan *chan) | |||
1600 | static enum dma_status idmac_tx_status(struct dma_chan *chan, | 1594 | static enum dma_status idmac_tx_status(struct dma_chan *chan, |
1601 | dma_cookie_t cookie, struct dma_tx_state *txstate) | 1595 | dma_cookie_t cookie, struct dma_tx_state *txstate) |
1602 | { | 1596 | { |
1603 | struct idmac_channel *ichan = to_idmac_chan(chan); | 1597 | dma_set_tx_state(txstate, chan->completed_cookie, chan->cookie, 0); |
1604 | |||
1605 | dma_set_tx_state(txstate, ichan->completed, chan->cookie, 0); | ||
1606 | if (cookie != chan->cookie) | 1598 | if (cookie != chan->cookie) |
1607 | return DMA_ERROR; | 1599 | return DMA_ERROR; |
1608 | return DMA_SUCCESS; | 1600 | return DMA_SUCCESS; |
@@ -1638,11 +1630,10 @@ static int __init ipu_idmac_init(struct ipu *ipu) | |||
1638 | 1630 | ||
1639 | ichan->status = IPU_CHANNEL_FREE; | 1631 | ichan->status = IPU_CHANNEL_FREE; |
1640 | ichan->sec_chan_en = false; | 1632 | ichan->sec_chan_en = false; |
1641 | ichan->completed = -ENXIO; | ||
1642 | snprintf(ichan->eof_name, sizeof(ichan->eof_name), "IDMAC EOF %d", i); | 1633 | snprintf(ichan->eof_name, sizeof(ichan->eof_name), "IDMAC EOF %d", i); |
1643 | 1634 | ||
1644 | dma_chan->device = &idmac->dma; | 1635 | dma_chan->device = &idmac->dma; |
1645 | dma_chan->cookie = 1; | 1636 | dma_cookie_init(dma_chan); |
1646 | dma_chan->chan_id = i; | 1637 | dma_chan->chan_id = i; |
1647 | list_add_tail(&dma_chan->device_node, &dma->channels); | 1638 | list_add_tail(&dma_chan->device_node, &dma->channels); |
1648 | } | 1639 | } |
diff --git a/drivers/dma/mpc512x_dma.c b/drivers/dma/mpc512x_dma.c index 4d6d4cf66949..2ab0a3d0eed5 100644 --- a/drivers/dma/mpc512x_dma.c +++ b/drivers/dma/mpc512x_dma.c | |||
@@ -44,6 +44,8 @@ | |||
44 | 44 | ||
45 | #include <linux/random.h> | 45 | #include <linux/random.h> |
46 | 46 | ||
47 | #include "dmaengine.h" | ||
48 | |||
47 | /* Number of DMA Transfer descriptors allocated per channel */ | 49 | /* Number of DMA Transfer descriptors allocated per channel */ |
48 | #define MPC_DMA_DESCRIPTORS 64 | 50 | #define MPC_DMA_DESCRIPTORS 64 |
49 | 51 | ||
@@ -188,7 +190,6 @@ struct mpc_dma_chan { | |||
188 | struct list_head completed; | 190 | struct list_head completed; |
189 | struct mpc_dma_tcd *tcd; | 191 | struct mpc_dma_tcd *tcd; |
190 | dma_addr_t tcd_paddr; | 192 | dma_addr_t tcd_paddr; |
191 | dma_cookie_t completed_cookie; | ||
192 | 193 | ||
193 | /* Lock for this structure */ | 194 | /* Lock for this structure */ |
194 | spinlock_t lock; | 195 | spinlock_t lock; |
@@ -365,7 +366,7 @@ static void mpc_dma_process_completed(struct mpc_dma *mdma) | |||
365 | /* Free descriptors */ | 366 | /* Free descriptors */ |
366 | spin_lock_irqsave(&mchan->lock, flags); | 367 | spin_lock_irqsave(&mchan->lock, flags); |
367 | list_splice_tail_init(&list, &mchan->free); | 368 | list_splice_tail_init(&list, &mchan->free); |
368 | mchan->completed_cookie = last_cookie; | 369 | mchan->chan.completed_cookie = last_cookie; |
369 | spin_unlock_irqrestore(&mchan->lock, flags); | 370 | spin_unlock_irqrestore(&mchan->lock, flags); |
370 | } | 371 | } |
371 | } | 372 | } |
@@ -438,13 +439,7 @@ static dma_cookie_t mpc_dma_tx_submit(struct dma_async_tx_descriptor *txd) | |||
438 | mpc_dma_execute(mchan); | 439 | mpc_dma_execute(mchan); |
439 | 440 | ||
440 | /* Update cookie */ | 441 | /* Update cookie */ |
441 | cookie = mchan->chan.cookie + 1; | 442 | cookie = dma_cookie_assign(txd); |
442 | if (cookie <= 0) | ||
443 | cookie = 1; | ||
444 | |||
445 | mchan->chan.cookie = cookie; | ||
446 | mdesc->desc.cookie = cookie; | ||
447 | |||
448 | spin_unlock_irqrestore(&mchan->lock, flags); | 443 | spin_unlock_irqrestore(&mchan->lock, flags); |
449 | 444 | ||
450 | return cookie; | 445 | return cookie; |
@@ -562,17 +557,14 @@ mpc_dma_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | |||
562 | struct dma_tx_state *txstate) | 557 | struct dma_tx_state *txstate) |
563 | { | 558 | { |
564 | struct mpc_dma_chan *mchan = dma_chan_to_mpc_dma_chan(chan); | 559 | struct mpc_dma_chan *mchan = dma_chan_to_mpc_dma_chan(chan); |
560 | enum dma_status ret; | ||
565 | unsigned long flags; | 561 | unsigned long flags; |
566 | dma_cookie_t last_used; | ||
567 | dma_cookie_t last_complete; | ||
568 | 562 | ||
569 | spin_lock_irqsave(&mchan->lock, flags); | 563 | spin_lock_irqsave(&mchan->lock, flags); |
570 | last_used = mchan->chan.cookie; | 564 | ret = dma_cookie_status(chan, cookie, txstate); |
571 | last_complete = mchan->completed_cookie; | ||
572 | spin_unlock_irqrestore(&mchan->lock, flags); | 565 | spin_unlock_irqrestore(&mchan->lock, flags); |
573 | 566 | ||
574 | dma_set_tx_state(txstate, last_complete, last_used, 0); | 567 | return ret; |
575 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
576 | } | 568 | } |
577 | 569 | ||
578 | /* Prepare descriptor for memory to memory copy */ | 570 | /* Prepare descriptor for memory to memory copy */ |
@@ -741,8 +733,7 @@ static int __devinit mpc_dma_probe(struct platform_device *op) | |||
741 | mchan = &mdma->channels[i]; | 733 | mchan = &mdma->channels[i]; |
742 | 734 | ||
743 | mchan->chan.device = dma; | 735 | mchan->chan.device = dma; |
744 | mchan->chan.cookie = 1; | 736 | dma_cookie_init(&mchan->chan); |
745 | mchan->completed_cookie = mchan->chan.cookie; | ||
746 | 737 | ||
747 | INIT_LIST_HEAD(&mchan->free); | 738 | INIT_LIST_HEAD(&mchan->free); |
748 | INIT_LIST_HEAD(&mchan->prepared); | 739 | INIT_LIST_HEAD(&mchan->prepared); |
diff --git a/drivers/dma/mv_xor.c b/drivers/dma/mv_xor.c index e779b434af45..fa5d55fea46c 100644 --- a/drivers/dma/mv_xor.c +++ b/drivers/dma/mv_xor.c | |||
@@ -26,6 +26,8 @@ | |||
26 | #include <linux/platform_device.h> | 26 | #include <linux/platform_device.h> |
27 | #include <linux/memory.h> | 27 | #include <linux/memory.h> |
28 | #include <plat/mv_xor.h> | 28 | #include <plat/mv_xor.h> |
29 | |||
30 | #include "dmaengine.h" | ||
29 | #include "mv_xor.h" | 31 | #include "mv_xor.h" |
30 | 32 | ||
31 | static void mv_xor_issue_pending(struct dma_chan *chan); | 33 | static void mv_xor_issue_pending(struct dma_chan *chan); |
@@ -435,7 +437,7 @@ static void __mv_xor_slot_cleanup(struct mv_xor_chan *mv_chan) | |||
435 | } | 437 | } |
436 | 438 | ||
437 | if (cookie > 0) | 439 | if (cookie > 0) |
438 | mv_chan->completed_cookie = cookie; | 440 | mv_chan->common.completed_cookie = cookie; |
439 | } | 441 | } |
440 | 442 | ||
441 | static void | 443 | static void |
@@ -534,18 +536,6 @@ retry: | |||
534 | return NULL; | 536 | return NULL; |
535 | } | 537 | } |
536 | 538 | ||
537 | static dma_cookie_t | ||
538 | mv_desc_assign_cookie(struct mv_xor_chan *mv_chan, | ||
539 | struct mv_xor_desc_slot *desc) | ||
540 | { | ||
541 | dma_cookie_t cookie = mv_chan->common.cookie; | ||
542 | |||
543 | if (++cookie < 0) | ||
544 | cookie = 1; | ||
545 | mv_chan->common.cookie = desc->async_tx.cookie = cookie; | ||
546 | return cookie; | ||
547 | } | ||
548 | |||
549 | /************************ DMA engine API functions ****************************/ | 539 | /************************ DMA engine API functions ****************************/ |
550 | static dma_cookie_t | 540 | static dma_cookie_t |
551 | mv_xor_tx_submit(struct dma_async_tx_descriptor *tx) | 541 | mv_xor_tx_submit(struct dma_async_tx_descriptor *tx) |
@@ -563,7 +553,7 @@ mv_xor_tx_submit(struct dma_async_tx_descriptor *tx) | |||
563 | grp_start = sw_desc->group_head; | 553 | grp_start = sw_desc->group_head; |
564 | 554 | ||
565 | spin_lock_bh(&mv_chan->lock); | 555 | spin_lock_bh(&mv_chan->lock); |
566 | cookie = mv_desc_assign_cookie(mv_chan, sw_desc); | 556 | cookie = dma_cookie_assign(tx); |
567 | 557 | ||
568 | if (list_empty(&mv_chan->chain)) | 558 | if (list_empty(&mv_chan->chain)) |
569 | list_splice_init(&sw_desc->tx_list, &mv_chan->chain); | 559 | list_splice_init(&sw_desc->tx_list, &mv_chan->chain); |
@@ -820,27 +810,16 @@ static enum dma_status mv_xor_status(struct dma_chan *chan, | |||
820 | struct dma_tx_state *txstate) | 810 | struct dma_tx_state *txstate) |
821 | { | 811 | { |
822 | struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan); | 812 | struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan); |
823 | dma_cookie_t last_used; | ||
824 | dma_cookie_t last_complete; | ||
825 | enum dma_status ret; | 813 | enum dma_status ret; |
826 | 814 | ||
827 | last_used = chan->cookie; | 815 | ret = dma_cookie_status(chan, cookie, txstate); |
828 | last_complete = mv_chan->completed_cookie; | ||
829 | mv_chan->is_complete_cookie = cookie; | ||
830 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
831 | |||
832 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
833 | if (ret == DMA_SUCCESS) { | 816 | if (ret == DMA_SUCCESS) { |
834 | mv_xor_clean_completed_slots(mv_chan); | 817 | mv_xor_clean_completed_slots(mv_chan); |
835 | return ret; | 818 | return ret; |
836 | } | 819 | } |
837 | mv_xor_slot_cleanup(mv_chan); | 820 | mv_xor_slot_cleanup(mv_chan); |
838 | 821 | ||
839 | last_used = chan->cookie; | 822 | return dma_cookie_status(chan, cookie, txstate); |
840 | last_complete = mv_chan->completed_cookie; | ||
841 | |||
842 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
843 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
844 | } | 823 | } |
845 | 824 | ||
846 | static void mv_dump_xor_regs(struct mv_xor_chan *chan) | 825 | static void mv_dump_xor_regs(struct mv_xor_chan *chan) |
@@ -1214,6 +1193,7 @@ static int __devinit mv_xor_probe(struct platform_device *pdev) | |||
1214 | INIT_LIST_HEAD(&mv_chan->completed_slots); | 1193 | INIT_LIST_HEAD(&mv_chan->completed_slots); |
1215 | INIT_LIST_HEAD(&mv_chan->all_slots); | 1194 | INIT_LIST_HEAD(&mv_chan->all_slots); |
1216 | mv_chan->common.device = dma_dev; | 1195 | mv_chan->common.device = dma_dev; |
1196 | dma_cookie_init(&mv_chan->common); | ||
1217 | 1197 | ||
1218 | list_add_tail(&mv_chan->common.device_node, &dma_dev->channels); | 1198 | list_add_tail(&mv_chan->common.device_node, &dma_dev->channels); |
1219 | 1199 | ||
diff --git a/drivers/dma/mv_xor.h b/drivers/dma/mv_xor.h index 977b592e976b..654876b7ba1d 100644 --- a/drivers/dma/mv_xor.h +++ b/drivers/dma/mv_xor.h | |||
@@ -78,7 +78,6 @@ struct mv_xor_device { | |||
78 | /** | 78 | /** |
79 | * struct mv_xor_chan - internal representation of a XOR channel | 79 | * struct mv_xor_chan - internal representation of a XOR channel |
80 | * @pending: allows batching of hardware operations | 80 | * @pending: allows batching of hardware operations |
81 | * @completed_cookie: identifier for the most recently completed operation | ||
82 | * @lock: serializes enqueue/dequeue operations to the descriptors pool | 81 | * @lock: serializes enqueue/dequeue operations to the descriptors pool |
83 | * @mmr_base: memory mapped register base | 82 | * @mmr_base: memory mapped register base |
84 | * @idx: the index of the xor channel | 83 | * @idx: the index of the xor channel |
@@ -93,7 +92,6 @@ struct mv_xor_device { | |||
93 | */ | 92 | */ |
94 | struct mv_xor_chan { | 93 | struct mv_xor_chan { |
95 | int pending; | 94 | int pending; |
96 | dma_cookie_t completed_cookie; | ||
97 | spinlock_t lock; /* protects the descriptor slot pool */ | 95 | spinlock_t lock; /* protects the descriptor slot pool */ |
98 | void __iomem *mmr_base; | 96 | void __iomem *mmr_base; |
99 | unsigned int idx; | 97 | unsigned int idx; |
@@ -109,7 +107,6 @@ struct mv_xor_chan { | |||
109 | #ifdef USE_TIMER | 107 | #ifdef USE_TIMER |
110 | unsigned long cleanup_time; | 108 | unsigned long cleanup_time; |
111 | u32 current_on_last_cleanup; | 109 | u32 current_on_last_cleanup; |
112 | dma_cookie_t is_complete_cookie; | ||
113 | #endif | 110 | #endif |
114 | }; | 111 | }; |
115 | 112 | ||
diff --git a/drivers/dma/mxs-dma.c b/drivers/dma/mxs-dma.c index b06cd4ca626f..65334c49b71e 100644 --- a/drivers/dma/mxs-dma.c +++ b/drivers/dma/mxs-dma.c | |||
@@ -28,6 +28,8 @@ | |||
28 | #include <mach/dma.h> | 28 | #include <mach/dma.h> |
29 | #include <mach/common.h> | 29 | #include <mach/common.h> |
30 | 30 | ||
31 | #include "dmaengine.h" | ||
32 | |||
31 | /* | 33 | /* |
32 | * NOTE: The term "PIO" throughout the mxs-dma implementation means | 34 | * NOTE: The term "PIO" throughout the mxs-dma implementation means |
33 | * PIO mode of mxs apbh-dma and apbx-dma. With this working mode, | 35 | * PIO mode of mxs apbh-dma and apbx-dma. With this working mode, |
@@ -111,7 +113,6 @@ struct mxs_dma_chan { | |||
111 | struct mxs_dma_ccw *ccw; | 113 | struct mxs_dma_ccw *ccw; |
112 | dma_addr_t ccw_phys; | 114 | dma_addr_t ccw_phys; |
113 | int desc_count; | 115 | int desc_count; |
114 | dma_cookie_t last_completed; | ||
115 | enum dma_status status; | 116 | enum dma_status status; |
116 | unsigned int flags; | 117 | unsigned int flags; |
117 | #define MXS_DMA_SG_LOOP (1 << 0) | 118 | #define MXS_DMA_SG_LOOP (1 << 0) |
@@ -193,19 +194,6 @@ static void mxs_dma_resume_chan(struct mxs_dma_chan *mxs_chan) | |||
193 | mxs_chan->status = DMA_IN_PROGRESS; | 194 | mxs_chan->status = DMA_IN_PROGRESS; |
194 | } | 195 | } |
195 | 196 | ||
196 | static dma_cookie_t mxs_dma_assign_cookie(struct mxs_dma_chan *mxs_chan) | ||
197 | { | ||
198 | dma_cookie_t cookie = mxs_chan->chan.cookie; | ||
199 | |||
200 | if (++cookie < 0) | ||
201 | cookie = 1; | ||
202 | |||
203 | mxs_chan->chan.cookie = cookie; | ||
204 | mxs_chan->desc.cookie = cookie; | ||
205 | |||
206 | return cookie; | ||
207 | } | ||
208 | |||
209 | static struct mxs_dma_chan *to_mxs_dma_chan(struct dma_chan *chan) | 197 | static struct mxs_dma_chan *to_mxs_dma_chan(struct dma_chan *chan) |
210 | { | 198 | { |
211 | return container_of(chan, struct mxs_dma_chan, chan); | 199 | return container_of(chan, struct mxs_dma_chan, chan); |
@@ -217,7 +205,7 @@ static dma_cookie_t mxs_dma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
217 | 205 | ||
218 | mxs_dma_enable_chan(mxs_chan); | 206 | mxs_dma_enable_chan(mxs_chan); |
219 | 207 | ||
220 | return mxs_dma_assign_cookie(mxs_chan); | 208 | return dma_cookie_assign(tx); |
221 | } | 209 | } |
222 | 210 | ||
223 | static void mxs_dma_tasklet(unsigned long data) | 211 | static void mxs_dma_tasklet(unsigned long data) |
@@ -274,7 +262,7 @@ static irqreturn_t mxs_dma_int_handler(int irq, void *dev_id) | |||
274 | stat1 &= ~(1 << channel); | 262 | stat1 &= ~(1 << channel); |
275 | 263 | ||
276 | if (mxs_chan->status == DMA_SUCCESS) | 264 | if (mxs_chan->status == DMA_SUCCESS) |
277 | mxs_chan->last_completed = mxs_chan->desc.cookie; | 265 | dma_cookie_complete(&mxs_chan->desc); |
278 | 266 | ||
279 | /* schedule tasklet on this channel */ | 267 | /* schedule tasklet on this channel */ |
280 | tasklet_schedule(&mxs_chan->tasklet); | 268 | tasklet_schedule(&mxs_chan->tasklet); |
@@ -352,7 +340,7 @@ static void mxs_dma_free_chan_resources(struct dma_chan *chan) | |||
352 | static struct dma_async_tx_descriptor *mxs_dma_prep_slave_sg( | 340 | static struct dma_async_tx_descriptor *mxs_dma_prep_slave_sg( |
353 | struct dma_chan *chan, struct scatterlist *sgl, | 341 | struct dma_chan *chan, struct scatterlist *sgl, |
354 | unsigned int sg_len, enum dma_transfer_direction direction, | 342 | unsigned int sg_len, enum dma_transfer_direction direction, |
355 | unsigned long append) | 343 | unsigned long append, void *context) |
356 | { | 344 | { |
357 | struct mxs_dma_chan *mxs_chan = to_mxs_dma_chan(chan); | 345 | struct mxs_dma_chan *mxs_chan = to_mxs_dma_chan(chan); |
358 | struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; | 346 | struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; |
@@ -447,7 +435,8 @@ err_out: | |||
447 | 435 | ||
448 | static struct dma_async_tx_descriptor *mxs_dma_prep_dma_cyclic( | 436 | static struct dma_async_tx_descriptor *mxs_dma_prep_dma_cyclic( |
449 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, | 437 | struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, |
450 | size_t period_len, enum dma_transfer_direction direction) | 438 | size_t period_len, enum dma_transfer_direction direction, |
439 | void *context) | ||
451 | { | 440 | { |
452 | struct mxs_dma_chan *mxs_chan = to_mxs_dma_chan(chan); | 441 | struct mxs_dma_chan *mxs_chan = to_mxs_dma_chan(chan); |
453 | struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; | 442 | struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; |
@@ -538,7 +527,7 @@ static enum dma_status mxs_dma_tx_status(struct dma_chan *chan, | |||
538 | dma_cookie_t last_used; | 527 | dma_cookie_t last_used; |
539 | 528 | ||
540 | last_used = chan->cookie; | 529 | last_used = chan->cookie; |
541 | dma_set_tx_state(txstate, mxs_chan->last_completed, last_used, 0); | 530 | dma_set_tx_state(txstate, chan->completed_cookie, last_used, 0); |
542 | 531 | ||
543 | return mxs_chan->status; | 532 | return mxs_chan->status; |
544 | } | 533 | } |
@@ -630,6 +619,7 @@ static int __init mxs_dma_probe(struct platform_device *pdev) | |||
630 | 619 | ||
631 | mxs_chan->mxs_dma = mxs_dma; | 620 | mxs_chan->mxs_dma = mxs_dma; |
632 | mxs_chan->chan.device = &mxs_dma->dma_device; | 621 | mxs_chan->chan.device = &mxs_dma->dma_device; |
622 | dma_cookie_init(&mxs_chan->chan); | ||
633 | 623 | ||
634 | tasklet_init(&mxs_chan->tasklet, mxs_dma_tasklet, | 624 | tasklet_init(&mxs_chan->tasklet, mxs_dma_tasklet, |
635 | (unsigned long) mxs_chan); | 625 | (unsigned long) mxs_chan); |
diff --git a/drivers/dma/pch_dma.c b/drivers/dma/pch_dma.c index 823f58179f9d..65c0495a6d40 100644 --- a/drivers/dma/pch_dma.c +++ b/drivers/dma/pch_dma.c | |||
@@ -25,6 +25,8 @@ | |||
25 | #include <linux/module.h> | 25 | #include <linux/module.h> |
26 | #include <linux/pch_dma.h> | 26 | #include <linux/pch_dma.h> |
27 | 27 | ||
28 | #include "dmaengine.h" | ||
29 | |||
28 | #define DRV_NAME "pch-dma" | 30 | #define DRV_NAME "pch-dma" |
29 | 31 | ||
30 | #define DMA_CTL0_DISABLE 0x0 | 32 | #define DMA_CTL0_DISABLE 0x0 |
@@ -105,7 +107,6 @@ struct pch_dma_chan { | |||
105 | 107 | ||
106 | spinlock_t lock; | 108 | spinlock_t lock; |
107 | 109 | ||
108 | dma_cookie_t completed_cookie; | ||
109 | struct list_head active_list; | 110 | struct list_head active_list; |
110 | struct list_head queue; | 111 | struct list_head queue; |
111 | struct list_head free_list; | 112 | struct list_head free_list; |
@@ -416,20 +417,6 @@ static void pdc_advance_work(struct pch_dma_chan *pd_chan) | |||
416 | } | 417 | } |
417 | } | 418 | } |
418 | 419 | ||
419 | static dma_cookie_t pdc_assign_cookie(struct pch_dma_chan *pd_chan, | ||
420 | struct pch_dma_desc *desc) | ||
421 | { | ||
422 | dma_cookie_t cookie = pd_chan->chan.cookie; | ||
423 | |||
424 | if (++cookie < 0) | ||
425 | cookie = 1; | ||
426 | |||
427 | pd_chan->chan.cookie = cookie; | ||
428 | desc->txd.cookie = cookie; | ||
429 | |||
430 | return cookie; | ||
431 | } | ||
432 | |||
433 | static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd) | 420 | static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd) |
434 | { | 421 | { |
435 | struct pch_dma_desc *desc = to_pd_desc(txd); | 422 | struct pch_dma_desc *desc = to_pd_desc(txd); |
@@ -437,7 +424,7 @@ static dma_cookie_t pd_tx_submit(struct dma_async_tx_descriptor *txd) | |||
437 | dma_cookie_t cookie; | 424 | dma_cookie_t cookie; |
438 | 425 | ||
439 | spin_lock(&pd_chan->lock); | 426 | spin_lock(&pd_chan->lock); |
440 | cookie = pdc_assign_cookie(pd_chan, desc); | 427 | cookie = dma_cookie_assign(txd); |
441 | 428 | ||
442 | if (list_empty(&pd_chan->active_list)) { | 429 | if (list_empty(&pd_chan->active_list)) { |
443 | list_add_tail(&desc->desc_node, &pd_chan->active_list); | 430 | list_add_tail(&desc->desc_node, &pd_chan->active_list); |
@@ -544,7 +531,7 @@ static int pd_alloc_chan_resources(struct dma_chan *chan) | |||
544 | spin_lock_irq(&pd_chan->lock); | 531 | spin_lock_irq(&pd_chan->lock); |
545 | list_splice(&tmp_list, &pd_chan->free_list); | 532 | list_splice(&tmp_list, &pd_chan->free_list); |
546 | pd_chan->descs_allocated = i; | 533 | pd_chan->descs_allocated = i; |
547 | pd_chan->completed_cookie = chan->cookie = 1; | 534 | dma_cookie_init(chan); |
548 | spin_unlock_irq(&pd_chan->lock); | 535 | spin_unlock_irq(&pd_chan->lock); |
549 | 536 | ||
550 | pdc_enable_irq(chan, 1); | 537 | pdc_enable_irq(chan, 1); |
@@ -578,19 +565,12 @@ static enum dma_status pd_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | |||
578 | struct dma_tx_state *txstate) | 565 | struct dma_tx_state *txstate) |
579 | { | 566 | { |
580 | struct pch_dma_chan *pd_chan = to_pd_chan(chan); | 567 | struct pch_dma_chan *pd_chan = to_pd_chan(chan); |
581 | dma_cookie_t last_used; | 568 | enum dma_status ret; |
582 | dma_cookie_t last_completed; | ||
583 | int ret; | ||
584 | 569 | ||
585 | spin_lock_irq(&pd_chan->lock); | 570 | spin_lock_irq(&pd_chan->lock); |
586 | last_completed = pd_chan->completed_cookie; | 571 | ret = dma_cookie_status(chan, cookie, txstate); |
587 | last_used = chan->cookie; | ||
588 | spin_unlock_irq(&pd_chan->lock); | 572 | spin_unlock_irq(&pd_chan->lock); |
589 | 573 | ||
590 | ret = dma_async_is_complete(cookie, last_completed, last_used); | ||
591 | |||
592 | dma_set_tx_state(txstate, last_completed, last_used, 0); | ||
593 | |||
594 | return ret; | 574 | return ret; |
595 | } | 575 | } |
596 | 576 | ||
@@ -607,7 +587,8 @@ static void pd_issue_pending(struct dma_chan *chan) | |||
607 | 587 | ||
608 | static struct dma_async_tx_descriptor *pd_prep_slave_sg(struct dma_chan *chan, | 588 | static struct dma_async_tx_descriptor *pd_prep_slave_sg(struct dma_chan *chan, |
609 | struct scatterlist *sgl, unsigned int sg_len, | 589 | struct scatterlist *sgl, unsigned int sg_len, |
610 | enum dma_transfer_direction direction, unsigned long flags) | 590 | enum dma_transfer_direction direction, unsigned long flags, |
591 | void *context) | ||
611 | { | 592 | { |
612 | struct pch_dma_chan *pd_chan = to_pd_chan(chan); | 593 | struct pch_dma_chan *pd_chan = to_pd_chan(chan); |
613 | struct pch_dma_slave *pd_slave = chan->private; | 594 | struct pch_dma_slave *pd_slave = chan->private; |
@@ -932,7 +913,7 @@ static int __devinit pch_dma_probe(struct pci_dev *pdev, | |||
932 | struct pch_dma_chan *pd_chan = &pd->channels[i]; | 913 | struct pch_dma_chan *pd_chan = &pd->channels[i]; |
933 | 914 | ||
934 | pd_chan->chan.device = &pd->dma; | 915 | pd_chan->chan.device = &pd->dma; |
935 | pd_chan->chan.cookie = 1; | 916 | dma_cookie_init(&pd_chan->chan); |
936 | 917 | ||
937 | pd_chan->membase = ®s->desc[i]; | 918 | pd_chan->membase = ®s->desc[i]; |
938 | 919 | ||
diff --git a/drivers/dma/pl330.c b/drivers/dma/pl330.c index 16b66c827f19..282caf118be8 100644 --- a/drivers/dma/pl330.c +++ b/drivers/dma/pl330.c | |||
@@ -1,4 +1,6 @@ | |||
1 | /* linux/drivers/dma/pl330.c | 1 | /* |
2 | * Copyright (c) 2012 Samsung Electronics Co., Ltd. | ||
3 | * http://www.samsung.com | ||
2 | * | 4 | * |
3 | * Copyright (C) 2010 Samsung Electronics Co. Ltd. | 5 | * Copyright (C) 2010 Samsung Electronics Co. Ltd. |
4 | * Jaswinder Singh <jassi.brar@samsung.com> | 6 | * Jaswinder Singh <jassi.brar@samsung.com> |
@@ -9,10 +11,15 @@ | |||
9 | * (at your option) any later version. | 11 | * (at your option) any later version. |
10 | */ | 12 | */ |
11 | 13 | ||
14 | #include <linux/kernel.h> | ||
12 | #include <linux/io.h> | 15 | #include <linux/io.h> |
13 | #include <linux/init.h> | 16 | #include <linux/init.h> |
14 | #include <linux/slab.h> | 17 | #include <linux/slab.h> |
15 | #include <linux/module.h> | 18 | #include <linux/module.h> |
19 | #include <linux/string.h> | ||
20 | #include <linux/delay.h> | ||
21 | #include <linux/interrupt.h> | ||
22 | #include <linux/dma-mapping.h> | ||
16 | #include <linux/dmaengine.h> | 23 | #include <linux/dmaengine.h> |
17 | #include <linux/interrupt.h> | 24 | #include <linux/interrupt.h> |
18 | #include <linux/amba/bus.h> | 25 | #include <linux/amba/bus.h> |
@@ -21,8 +28,497 @@ | |||
21 | #include <linux/scatterlist.h> | 28 | #include <linux/scatterlist.h> |
22 | #include <linux/of.h> | 29 | #include <linux/of.h> |
23 | 30 | ||
31 | #include "dmaengine.h" | ||
32 | #define PL330_MAX_CHAN 8 | ||
33 | #define PL330_MAX_IRQS 32 | ||
34 | #define PL330_MAX_PERI 32 | ||
35 | |||
36 | enum pl330_srccachectrl { | ||
37 | SCCTRL0, /* Noncacheable and nonbufferable */ | ||
38 | SCCTRL1, /* Bufferable only */ | ||
39 | SCCTRL2, /* Cacheable, but do not allocate */ | ||
40 | SCCTRL3, /* Cacheable and bufferable, but do not allocate */ | ||
41 | SINVALID1, | ||
42 | SINVALID2, | ||
43 | SCCTRL6, /* Cacheable write-through, allocate on reads only */ | ||
44 | SCCTRL7, /* Cacheable write-back, allocate on reads only */ | ||
45 | }; | ||
46 | |||
47 | enum pl330_dstcachectrl { | ||
48 | DCCTRL0, /* Noncacheable and nonbufferable */ | ||
49 | DCCTRL1, /* Bufferable only */ | ||
50 | DCCTRL2, /* Cacheable, but do not allocate */ | ||
51 | DCCTRL3, /* Cacheable and bufferable, but do not allocate */ | ||
52 | DINVALID1, /* AWCACHE = 0x1000 */ | ||
53 | DINVALID2, | ||
54 | DCCTRL6, /* Cacheable write-through, allocate on writes only */ | ||
55 | DCCTRL7, /* Cacheable write-back, allocate on writes only */ | ||
56 | }; | ||
57 | |||
58 | enum pl330_byteswap { | ||
59 | SWAP_NO, | ||
60 | SWAP_2, | ||
61 | SWAP_4, | ||
62 | SWAP_8, | ||
63 | SWAP_16, | ||
64 | }; | ||
65 | |||
66 | enum pl330_reqtype { | ||
67 | MEMTOMEM, | ||
68 | MEMTODEV, | ||
69 | DEVTOMEM, | ||
70 | DEVTODEV, | ||
71 | }; | ||
72 | |||
73 | /* Register and Bit field Definitions */ | ||
74 | #define DS 0x0 | ||
75 | #define DS_ST_STOP 0x0 | ||
76 | #define DS_ST_EXEC 0x1 | ||
77 | #define DS_ST_CMISS 0x2 | ||
78 | #define DS_ST_UPDTPC 0x3 | ||
79 | #define DS_ST_WFE 0x4 | ||
80 | #define DS_ST_ATBRR 0x5 | ||
81 | #define DS_ST_QBUSY 0x6 | ||
82 | #define DS_ST_WFP 0x7 | ||
83 | #define DS_ST_KILL 0x8 | ||
84 | #define DS_ST_CMPLT 0x9 | ||
85 | #define DS_ST_FLTCMP 0xe | ||
86 | #define DS_ST_FAULT 0xf | ||
87 | |||
88 | #define DPC 0x4 | ||
89 | #define INTEN 0x20 | ||
90 | #define ES 0x24 | ||
91 | #define INTSTATUS 0x28 | ||
92 | #define INTCLR 0x2c | ||
93 | #define FSM 0x30 | ||
94 | #define FSC 0x34 | ||
95 | #define FTM 0x38 | ||
96 | |||
97 | #define _FTC 0x40 | ||
98 | #define FTC(n) (_FTC + (n)*0x4) | ||
99 | |||
100 | #define _CS 0x100 | ||
101 | #define CS(n) (_CS + (n)*0x8) | ||
102 | #define CS_CNS (1 << 21) | ||
103 | |||
104 | #define _CPC 0x104 | ||
105 | #define CPC(n) (_CPC + (n)*0x8) | ||
106 | |||
107 | #define _SA 0x400 | ||
108 | #define SA(n) (_SA + (n)*0x20) | ||
109 | |||
110 | #define _DA 0x404 | ||
111 | #define DA(n) (_DA + (n)*0x20) | ||
112 | |||
113 | #define _CC 0x408 | ||
114 | #define CC(n) (_CC + (n)*0x20) | ||
115 | |||
116 | #define CC_SRCINC (1 << 0) | ||
117 | #define CC_DSTINC (1 << 14) | ||
118 | #define CC_SRCPRI (1 << 8) | ||
119 | #define CC_DSTPRI (1 << 22) | ||
120 | #define CC_SRCNS (1 << 9) | ||
121 | #define CC_DSTNS (1 << 23) | ||
122 | #define CC_SRCIA (1 << 10) | ||
123 | #define CC_DSTIA (1 << 24) | ||
124 | #define CC_SRCBRSTLEN_SHFT 4 | ||
125 | #define CC_DSTBRSTLEN_SHFT 18 | ||
126 | #define CC_SRCBRSTSIZE_SHFT 1 | ||
127 | #define CC_DSTBRSTSIZE_SHFT 15 | ||
128 | #define CC_SRCCCTRL_SHFT 11 | ||
129 | #define CC_SRCCCTRL_MASK 0x7 | ||
130 | #define CC_DSTCCTRL_SHFT 25 | ||
131 | #define CC_DRCCCTRL_MASK 0x7 | ||
132 | #define CC_SWAP_SHFT 28 | ||
133 | |||
134 | #define _LC0 0x40c | ||
135 | #define LC0(n) (_LC0 + (n)*0x20) | ||
136 | |||
137 | #define _LC1 0x410 | ||
138 | #define LC1(n) (_LC1 + (n)*0x20) | ||
139 | |||
140 | #define DBGSTATUS 0xd00 | ||
141 | #define DBG_BUSY (1 << 0) | ||
142 | |||
143 | #define DBGCMD 0xd04 | ||
144 | #define DBGINST0 0xd08 | ||
145 | #define DBGINST1 0xd0c | ||
146 | |||
147 | #define CR0 0xe00 | ||
148 | #define CR1 0xe04 | ||
149 | #define CR2 0xe08 | ||
150 | #define CR3 0xe0c | ||
151 | #define CR4 0xe10 | ||
152 | #define CRD 0xe14 | ||
153 | |||
154 | #define PERIPH_ID 0xfe0 | ||
155 | #define PERIPH_REV_SHIFT 20 | ||
156 | #define PERIPH_REV_MASK 0xf | ||
157 | #define PERIPH_REV_R0P0 0 | ||
158 | #define PERIPH_REV_R1P0 1 | ||
159 | #define PERIPH_REV_R1P1 2 | ||
160 | #define PCELL_ID 0xff0 | ||
161 | |||
162 | #define CR0_PERIPH_REQ_SET (1 << 0) | ||
163 | #define CR0_BOOT_EN_SET (1 << 1) | ||
164 | #define CR0_BOOT_MAN_NS (1 << 2) | ||
165 | #define CR0_NUM_CHANS_SHIFT 4 | ||
166 | #define CR0_NUM_CHANS_MASK 0x7 | ||
167 | #define CR0_NUM_PERIPH_SHIFT 12 | ||
168 | #define CR0_NUM_PERIPH_MASK 0x1f | ||
169 | #define CR0_NUM_EVENTS_SHIFT 17 | ||
170 | #define CR0_NUM_EVENTS_MASK 0x1f | ||
171 | |||
172 | #define CR1_ICACHE_LEN_SHIFT 0 | ||
173 | #define CR1_ICACHE_LEN_MASK 0x7 | ||
174 | #define CR1_NUM_ICACHELINES_SHIFT 4 | ||
175 | #define CR1_NUM_ICACHELINES_MASK 0xf | ||
176 | |||
177 | #define CRD_DATA_WIDTH_SHIFT 0 | ||
178 | #define CRD_DATA_WIDTH_MASK 0x7 | ||
179 | #define CRD_WR_CAP_SHIFT 4 | ||
180 | #define CRD_WR_CAP_MASK 0x7 | ||
181 | #define CRD_WR_Q_DEP_SHIFT 8 | ||
182 | #define CRD_WR_Q_DEP_MASK 0xf | ||
183 | #define CRD_RD_CAP_SHIFT 12 | ||
184 | #define CRD_RD_CAP_MASK 0x7 | ||
185 | #define CRD_RD_Q_DEP_SHIFT 16 | ||
186 | #define CRD_RD_Q_DEP_MASK 0xf | ||
187 | #define CRD_DATA_BUFF_SHIFT 20 | ||
188 | #define CRD_DATA_BUFF_MASK 0x3ff | ||
189 | |||
190 | #define PART 0x330 | ||
191 | #define DESIGNER 0x41 | ||
192 | #define REVISION 0x0 | ||
193 | #define INTEG_CFG 0x0 | ||
194 | #define PERIPH_ID_VAL ((PART << 0) | (DESIGNER << 12)) | ||
195 | |||
196 | #define PCELL_ID_VAL 0xb105f00d | ||
197 | |||
198 | #define PL330_STATE_STOPPED (1 << 0) | ||
199 | #define PL330_STATE_EXECUTING (1 << 1) | ||
200 | #define PL330_STATE_WFE (1 << 2) | ||
201 | #define PL330_STATE_FAULTING (1 << 3) | ||
202 | #define PL330_STATE_COMPLETING (1 << 4) | ||
203 | #define PL330_STATE_WFP (1 << 5) | ||
204 | #define PL330_STATE_KILLING (1 << 6) | ||
205 | #define PL330_STATE_FAULT_COMPLETING (1 << 7) | ||
206 | #define PL330_STATE_CACHEMISS (1 << 8) | ||
207 | #define PL330_STATE_UPDTPC (1 << 9) | ||
208 | #define PL330_STATE_ATBARRIER (1 << 10) | ||
209 | #define PL330_STATE_QUEUEBUSY (1 << 11) | ||
210 | #define PL330_STATE_INVALID (1 << 15) | ||
211 | |||
212 | #define PL330_STABLE_STATES (PL330_STATE_STOPPED | PL330_STATE_EXECUTING \ | ||
213 | | PL330_STATE_WFE | PL330_STATE_FAULTING) | ||
214 | |||
215 | #define CMD_DMAADDH 0x54 | ||
216 | #define CMD_DMAEND 0x00 | ||
217 | #define CMD_DMAFLUSHP 0x35 | ||
218 | #define CMD_DMAGO 0xa0 | ||
219 | #define CMD_DMALD 0x04 | ||
220 | #define CMD_DMALDP 0x25 | ||
221 | #define CMD_DMALP 0x20 | ||
222 | #define CMD_DMALPEND 0x28 | ||
223 | #define CMD_DMAKILL 0x01 | ||
224 | #define CMD_DMAMOV 0xbc | ||
225 | #define CMD_DMANOP 0x18 | ||
226 | #define CMD_DMARMB 0x12 | ||
227 | #define CMD_DMASEV 0x34 | ||
228 | #define CMD_DMAST 0x08 | ||
229 | #define CMD_DMASTP 0x29 | ||
230 | #define CMD_DMASTZ 0x0c | ||
231 | #define CMD_DMAWFE 0x36 | ||
232 | #define CMD_DMAWFP 0x30 | ||
233 | #define CMD_DMAWMB 0x13 | ||
234 | |||
235 | #define SZ_DMAADDH 3 | ||
236 | #define SZ_DMAEND 1 | ||
237 | #define SZ_DMAFLUSHP 2 | ||
238 | #define SZ_DMALD 1 | ||
239 | #define SZ_DMALDP 2 | ||
240 | #define SZ_DMALP 2 | ||
241 | #define SZ_DMALPEND 2 | ||
242 | #define SZ_DMAKILL 1 | ||
243 | #define SZ_DMAMOV 6 | ||
244 | #define SZ_DMANOP 1 | ||
245 | #define SZ_DMARMB 1 | ||
246 | #define SZ_DMASEV 2 | ||
247 | #define SZ_DMAST 1 | ||
248 | #define SZ_DMASTP 2 | ||
249 | #define SZ_DMASTZ 1 | ||
250 | #define SZ_DMAWFE 2 | ||
251 | #define SZ_DMAWFP 2 | ||
252 | #define SZ_DMAWMB 1 | ||
253 | #define SZ_DMAGO 6 | ||
254 | |||
255 | #define BRST_LEN(ccr) ((((ccr) >> CC_SRCBRSTLEN_SHFT) & 0xf) + 1) | ||
256 | #define BRST_SIZE(ccr) (1 << (((ccr) >> CC_SRCBRSTSIZE_SHFT) & 0x7)) | ||
257 | |||
258 | #define BYTE_TO_BURST(b, ccr) ((b) / BRST_SIZE(ccr) / BRST_LEN(ccr)) | ||
259 | #define BURST_TO_BYTE(c, ccr) ((c) * BRST_SIZE(ccr) * BRST_LEN(ccr)) | ||
260 | |||
261 | /* | ||
262 | * With 256 bytes, we can do more than 2.5MB and 5MB xfers per req | ||
263 | * at 1byte/burst for P<->M and M<->M respectively. | ||
264 | * For typical scenario, at 1word/burst, 10MB and 20MB xfers per req | ||
265 | * should be enough for P<->M and M<->M respectively. | ||
266 | */ | ||
267 | #define MCODE_BUFF_PER_REQ 256 | ||
268 | |||
269 | /* If the _pl330_req is available to the client */ | ||
270 | #define IS_FREE(req) (*((u8 *)((req)->mc_cpu)) == CMD_DMAEND) | ||
271 | |||
272 | /* Use this _only_ to wait on transient states */ | ||
273 | #define UNTIL(t, s) while (!(_state(t) & (s))) cpu_relax(); | ||
274 | |||
275 | #ifdef PL330_DEBUG_MCGEN | ||
276 | static unsigned cmd_line; | ||
277 | #define PL330_DBGCMD_DUMP(off, x...) do { \ | ||
278 | printk("%x:", cmd_line); \ | ||
279 | printk(x); \ | ||
280 | cmd_line += off; \ | ||
281 | } while (0) | ||
282 | #define PL330_DBGMC_START(addr) (cmd_line = addr) | ||
283 | #else | ||
284 | #define PL330_DBGCMD_DUMP(off, x...) do {} while (0) | ||
285 | #define PL330_DBGMC_START(addr) do {} while (0) | ||
286 | #endif | ||
287 | |||
288 | /* The number of default descriptors */ | ||
289 | |||
24 | #define NR_DEFAULT_DESC 16 | 290 | #define NR_DEFAULT_DESC 16 |
25 | 291 | ||
292 | /* Populated by the PL330 core driver for DMA API driver's info */ | ||
293 | struct pl330_config { | ||
294 | u32 periph_id; | ||
295 | u32 pcell_id; | ||
296 | #define DMAC_MODE_NS (1 << 0) | ||
297 | unsigned int mode; | ||
298 | unsigned int data_bus_width:10; /* In number of bits */ | ||
299 | unsigned int data_buf_dep:10; | ||
300 | unsigned int num_chan:4; | ||
301 | unsigned int num_peri:6; | ||
302 | u32 peri_ns; | ||
303 | unsigned int num_events:6; | ||
304 | u32 irq_ns; | ||
305 | }; | ||
306 | |||
307 | /* Handle to the DMAC provided to the PL330 core */ | ||
308 | struct pl330_info { | ||
309 | /* Owning device */ | ||
310 | struct device *dev; | ||
311 | /* Size of MicroCode buffers for each channel. */ | ||
312 | unsigned mcbufsz; | ||
313 | /* ioremap'ed address of PL330 registers. */ | ||
314 | void __iomem *base; | ||
315 | /* Client can freely use it. */ | ||
316 | void *client_data; | ||
317 | /* PL330 core data, Client must not touch it. */ | ||
318 | void *pl330_data; | ||
319 | /* Populated by the PL330 core driver during pl330_add */ | ||
320 | struct pl330_config pcfg; | ||
321 | /* | ||
322 | * If the DMAC has some reset mechanism, then the | ||
323 | * client may want to provide pointer to the method. | ||
324 | */ | ||
325 | void (*dmac_reset)(struct pl330_info *pi); | ||
326 | }; | ||
327 | |||
328 | /** | ||
329 | * Request Configuration. | ||
330 | * The PL330 core does not modify this and uses the last | ||
331 | * working configuration if the request doesn't provide any. | ||
332 | * | ||
333 | * The Client may want to provide this info only for the | ||
334 | * first request and a request with new settings. | ||
335 | */ | ||
336 | struct pl330_reqcfg { | ||
337 | /* Address Incrementing */ | ||
338 | unsigned dst_inc:1; | ||
339 | unsigned src_inc:1; | ||
340 | |||
341 | /* | ||
342 | * For now, the SRC & DST protection levels | ||
343 | * and burst size/length are assumed same. | ||
344 | */ | ||
345 | bool nonsecure; | ||
346 | bool privileged; | ||
347 | bool insnaccess; | ||
348 | unsigned brst_len:5; | ||
349 | unsigned brst_size:3; /* in power of 2 */ | ||
350 | |||
351 | enum pl330_dstcachectrl dcctl; | ||
352 | enum pl330_srccachectrl scctl; | ||
353 | enum pl330_byteswap swap; | ||
354 | struct pl330_config *pcfg; | ||
355 | }; | ||
356 | |||
357 | /* | ||
358 | * One cycle of DMAC operation. | ||
359 | * There may be more than one xfer in a request. | ||
360 | */ | ||
361 | struct pl330_xfer { | ||
362 | u32 src_addr; | ||
363 | u32 dst_addr; | ||
364 | /* Size to xfer */ | ||
365 | u32 bytes; | ||
366 | /* | ||
367 | * Pointer to next xfer in the list. | ||
368 | * The last xfer in the req must point to NULL. | ||
369 | */ | ||
370 | struct pl330_xfer *next; | ||
371 | }; | ||
372 | |||
373 | /* The xfer callbacks are made with one of these arguments. */ | ||
374 | enum pl330_op_err { | ||
375 | /* The all xfers in the request were success. */ | ||
376 | PL330_ERR_NONE, | ||
377 | /* If req aborted due to global error. */ | ||
378 | PL330_ERR_ABORT, | ||
379 | /* If req failed due to problem with Channel. */ | ||
380 | PL330_ERR_FAIL, | ||
381 | }; | ||
382 | |||
383 | /* A request defining Scatter-Gather List ending with NULL xfer. */ | ||
384 | struct pl330_req { | ||
385 | enum pl330_reqtype rqtype; | ||
386 | /* Index of peripheral for the xfer. */ | ||
387 | unsigned peri:5; | ||
388 | /* Unique token for this xfer, set by the client. */ | ||
389 | void *token; | ||
390 | /* Callback to be called after xfer. */ | ||
391 | void (*xfer_cb)(void *token, enum pl330_op_err err); | ||
392 | /* If NULL, req will be done at last set parameters. */ | ||
393 | struct pl330_reqcfg *cfg; | ||
394 | /* Pointer to first xfer in the request. */ | ||
395 | struct pl330_xfer *x; | ||
396 | }; | ||
397 | |||
398 | /* | ||
399 | * To know the status of the channel and DMAC, the client | ||
400 | * provides a pointer to this structure. The PL330 core | ||
401 | * fills it with current information. | ||
402 | */ | ||
403 | struct pl330_chanstatus { | ||
404 | /* | ||
405 | * If the DMAC engine halted due to some error, | ||
406 | * the client should remove-add DMAC. | ||
407 | */ | ||
408 | bool dmac_halted; | ||
409 | /* | ||
410 | * If channel is halted due to some error, | ||
411 | * the client should ABORT/FLUSH and START the channel. | ||
412 | */ | ||
413 | bool faulting; | ||
414 | /* Location of last load */ | ||
415 | u32 src_addr; | ||
416 | /* Location of last store */ | ||
417 | u32 dst_addr; | ||
418 | /* | ||
419 | * Pointer to the currently active req, NULL if channel is | ||
420 | * inactive, even though the requests may be present. | ||
421 | */ | ||
422 | struct pl330_req *top_req; | ||
423 | /* Pointer to req waiting second in the queue if any. */ | ||
424 | struct pl330_req *wait_req; | ||
425 | }; | ||
426 | |||
427 | enum pl330_chan_op { | ||
428 | /* Start the channel */ | ||
429 | PL330_OP_START, | ||
430 | /* Abort the active xfer */ | ||
431 | PL330_OP_ABORT, | ||
432 | /* Stop xfer and flush queue */ | ||
433 | PL330_OP_FLUSH, | ||
434 | }; | ||
435 | |||
436 | struct _xfer_spec { | ||
437 | u32 ccr; | ||
438 | struct pl330_req *r; | ||
439 | struct pl330_xfer *x; | ||
440 | }; | ||
441 | |||
442 | enum dmamov_dst { | ||
443 | SAR = 0, | ||
444 | CCR, | ||
445 | DAR, | ||
446 | }; | ||
447 | |||
448 | enum pl330_dst { | ||
449 | SRC = 0, | ||
450 | DST, | ||
451 | }; | ||
452 | |||
453 | enum pl330_cond { | ||
454 | SINGLE, | ||
455 | BURST, | ||
456 | ALWAYS, | ||
457 | }; | ||
458 | |||
459 | struct _pl330_req { | ||
460 | u32 mc_bus; | ||
461 | void *mc_cpu; | ||
462 | /* Number of bytes taken to setup MC for the req */ | ||
463 | u32 mc_len; | ||
464 | struct pl330_req *r; | ||
465 | /* Hook to attach to DMAC's list of reqs with due callback */ | ||
466 | struct list_head rqd; | ||
467 | }; | ||
468 | |||
469 | /* ToBeDone for tasklet */ | ||
470 | struct _pl330_tbd { | ||
471 | bool reset_dmac; | ||
472 | bool reset_mngr; | ||
473 | u8 reset_chan; | ||
474 | }; | ||
475 | |||
476 | /* A DMAC Thread */ | ||
477 | struct pl330_thread { | ||
478 | u8 id; | ||
479 | int ev; | ||
480 | /* If the channel is not yet acquired by any client */ | ||
481 | bool free; | ||
482 | /* Parent DMAC */ | ||
483 | struct pl330_dmac *dmac; | ||
484 | /* Only two at a time */ | ||
485 | struct _pl330_req req[2]; | ||
486 | /* Index of the last enqueued request */ | ||
487 | unsigned lstenq; | ||
488 | /* Index of the last submitted request or -1 if the DMA is stopped */ | ||
489 | int req_running; | ||
490 | }; | ||
491 | |||
492 | enum pl330_dmac_state { | ||
493 | UNINIT, | ||
494 | INIT, | ||
495 | DYING, | ||
496 | }; | ||
497 | |||
498 | /* A DMAC */ | ||
499 | struct pl330_dmac { | ||
500 | spinlock_t lock; | ||
501 | /* Holds list of reqs with due callbacks */ | ||
502 | struct list_head req_done; | ||
503 | /* Pointer to platform specific stuff */ | ||
504 | struct pl330_info *pinfo; | ||
505 | /* Maximum possible events/irqs */ | ||
506 | int events[32]; | ||
507 | /* BUS address of MicroCode buffer */ | ||
508 | u32 mcode_bus; | ||
509 | /* CPU address of MicroCode buffer */ | ||
510 | void *mcode_cpu; | ||
511 | /* List of all Channel threads */ | ||
512 | struct pl330_thread *channels; | ||
513 | /* Pointer to the MANAGER thread */ | ||
514 | struct pl330_thread *manager; | ||
515 | /* To handle bad news in interrupt */ | ||
516 | struct tasklet_struct tasks; | ||
517 | struct _pl330_tbd dmac_tbd; | ||
518 | /* State of DMAC operation */ | ||
519 | enum pl330_dmac_state state; | ||
520 | }; | ||
521 | |||
26 | enum desc_status { | 522 | enum desc_status { |
27 | /* In the DMAC pool */ | 523 | /* In the DMAC pool */ |
28 | FREE, | 524 | FREE, |
@@ -51,9 +547,6 @@ struct dma_pl330_chan { | |||
51 | /* DMA-Engine Channel */ | 547 | /* DMA-Engine Channel */ |
52 | struct dma_chan chan; | 548 | struct dma_chan chan; |
53 | 549 | ||
54 | /* Last completed cookie */ | ||
55 | dma_cookie_t completed; | ||
56 | |||
57 | /* List of to be xfered descriptors */ | 550 | /* List of to be xfered descriptors */ |
58 | struct list_head work_list; | 551 | struct list_head work_list; |
59 | 552 | ||
@@ -117,6 +610,1599 @@ struct dma_pl330_desc { | |||
117 | struct dma_pl330_chan *pchan; | 610 | struct dma_pl330_chan *pchan; |
118 | }; | 611 | }; |
119 | 612 | ||
613 | static inline void _callback(struct pl330_req *r, enum pl330_op_err err) | ||
614 | { | ||
615 | if (r && r->xfer_cb) | ||
616 | r->xfer_cb(r->token, err); | ||
617 | } | ||
618 | |||
619 | static inline bool _queue_empty(struct pl330_thread *thrd) | ||
620 | { | ||
621 | return (IS_FREE(&thrd->req[0]) && IS_FREE(&thrd->req[1])) | ||
622 | ? true : false; | ||
623 | } | ||
624 | |||
625 | static inline bool _queue_full(struct pl330_thread *thrd) | ||
626 | { | ||
627 | return (IS_FREE(&thrd->req[0]) || IS_FREE(&thrd->req[1])) | ||
628 | ? false : true; | ||
629 | } | ||
630 | |||
631 | static inline bool is_manager(struct pl330_thread *thrd) | ||
632 | { | ||
633 | struct pl330_dmac *pl330 = thrd->dmac; | ||
634 | |||
635 | /* MANAGER is indexed at the end */ | ||
636 | if (thrd->id == pl330->pinfo->pcfg.num_chan) | ||
637 | return true; | ||
638 | else | ||
639 | return false; | ||
640 | } | ||
641 | |||
642 | /* If manager of the thread is in Non-Secure mode */ | ||
643 | static inline bool _manager_ns(struct pl330_thread *thrd) | ||
644 | { | ||
645 | struct pl330_dmac *pl330 = thrd->dmac; | ||
646 | |||
647 | return (pl330->pinfo->pcfg.mode & DMAC_MODE_NS) ? true : false; | ||
648 | } | ||
649 | |||
650 | static inline u32 get_id(struct pl330_info *pi, u32 off) | ||
651 | { | ||
652 | void __iomem *regs = pi->base; | ||
653 | u32 id = 0; | ||
654 | |||
655 | id |= (readb(regs + off + 0x0) << 0); | ||
656 | id |= (readb(regs + off + 0x4) << 8); | ||
657 | id |= (readb(regs + off + 0x8) << 16); | ||
658 | id |= (readb(regs + off + 0xc) << 24); | ||
659 | |||
660 | return id; | ||
661 | } | ||
662 | |||
663 | static inline u32 get_revision(u32 periph_id) | ||
664 | { | ||
665 | return (periph_id >> PERIPH_REV_SHIFT) & PERIPH_REV_MASK; | ||
666 | } | ||
667 | |||
668 | static inline u32 _emit_ADDH(unsigned dry_run, u8 buf[], | ||
669 | enum pl330_dst da, u16 val) | ||
670 | { | ||
671 | if (dry_run) | ||
672 | return SZ_DMAADDH; | ||
673 | |||
674 | buf[0] = CMD_DMAADDH; | ||
675 | buf[0] |= (da << 1); | ||
676 | *((u16 *)&buf[1]) = val; | ||
677 | |||
678 | PL330_DBGCMD_DUMP(SZ_DMAADDH, "\tDMAADDH %s %u\n", | ||
679 | da == 1 ? "DA" : "SA", val); | ||
680 | |||
681 | return SZ_DMAADDH; | ||
682 | } | ||
683 | |||
684 | static inline u32 _emit_END(unsigned dry_run, u8 buf[]) | ||
685 | { | ||
686 | if (dry_run) | ||
687 | return SZ_DMAEND; | ||
688 | |||
689 | buf[0] = CMD_DMAEND; | ||
690 | |||
691 | PL330_DBGCMD_DUMP(SZ_DMAEND, "\tDMAEND\n"); | ||
692 | |||
693 | return SZ_DMAEND; | ||
694 | } | ||
695 | |||
696 | static inline u32 _emit_FLUSHP(unsigned dry_run, u8 buf[], u8 peri) | ||
697 | { | ||
698 | if (dry_run) | ||
699 | return SZ_DMAFLUSHP; | ||
700 | |||
701 | buf[0] = CMD_DMAFLUSHP; | ||
702 | |||
703 | peri &= 0x1f; | ||
704 | peri <<= 3; | ||
705 | buf[1] = peri; | ||
706 | |||
707 | PL330_DBGCMD_DUMP(SZ_DMAFLUSHP, "\tDMAFLUSHP %u\n", peri >> 3); | ||
708 | |||
709 | return SZ_DMAFLUSHP; | ||
710 | } | ||
711 | |||
712 | static inline u32 _emit_LD(unsigned dry_run, u8 buf[], enum pl330_cond cond) | ||
713 | { | ||
714 | if (dry_run) | ||
715 | return SZ_DMALD; | ||
716 | |||
717 | buf[0] = CMD_DMALD; | ||
718 | |||
719 | if (cond == SINGLE) | ||
720 | buf[0] |= (0 << 1) | (1 << 0); | ||
721 | else if (cond == BURST) | ||
722 | buf[0] |= (1 << 1) | (1 << 0); | ||
723 | |||
724 | PL330_DBGCMD_DUMP(SZ_DMALD, "\tDMALD%c\n", | ||
725 | cond == SINGLE ? 'S' : (cond == BURST ? 'B' : 'A')); | ||
726 | |||
727 | return SZ_DMALD; | ||
728 | } | ||
729 | |||
730 | static inline u32 _emit_LDP(unsigned dry_run, u8 buf[], | ||
731 | enum pl330_cond cond, u8 peri) | ||
732 | { | ||
733 | if (dry_run) | ||
734 | return SZ_DMALDP; | ||
735 | |||
736 | buf[0] = CMD_DMALDP; | ||
737 | |||
738 | if (cond == BURST) | ||
739 | buf[0] |= (1 << 1); | ||
740 | |||
741 | peri &= 0x1f; | ||
742 | peri <<= 3; | ||
743 | buf[1] = peri; | ||
744 | |||
745 | PL330_DBGCMD_DUMP(SZ_DMALDP, "\tDMALDP%c %u\n", | ||
746 | cond == SINGLE ? 'S' : 'B', peri >> 3); | ||
747 | |||
748 | return SZ_DMALDP; | ||
749 | } | ||
750 | |||
751 | static inline u32 _emit_LP(unsigned dry_run, u8 buf[], | ||
752 | unsigned loop, u8 cnt) | ||
753 | { | ||
754 | if (dry_run) | ||
755 | return SZ_DMALP; | ||
756 | |||
757 | buf[0] = CMD_DMALP; | ||
758 | |||
759 | if (loop) | ||
760 | buf[0] |= (1 << 1); | ||
761 | |||
762 | cnt--; /* DMAC increments by 1 internally */ | ||
763 | buf[1] = cnt; | ||
764 | |||
765 | PL330_DBGCMD_DUMP(SZ_DMALP, "\tDMALP_%c %u\n", loop ? '1' : '0', cnt); | ||
766 | |||
767 | return SZ_DMALP; | ||
768 | } | ||
769 | |||
770 | struct _arg_LPEND { | ||
771 | enum pl330_cond cond; | ||
772 | bool forever; | ||
773 | unsigned loop; | ||
774 | u8 bjump; | ||
775 | }; | ||
776 | |||
777 | static inline u32 _emit_LPEND(unsigned dry_run, u8 buf[], | ||
778 | const struct _arg_LPEND *arg) | ||
779 | { | ||
780 | enum pl330_cond cond = arg->cond; | ||
781 | bool forever = arg->forever; | ||
782 | unsigned loop = arg->loop; | ||
783 | u8 bjump = arg->bjump; | ||
784 | |||
785 | if (dry_run) | ||
786 | return SZ_DMALPEND; | ||
787 | |||
788 | buf[0] = CMD_DMALPEND; | ||
789 | |||
790 | if (loop) | ||
791 | buf[0] |= (1 << 2); | ||
792 | |||
793 | if (!forever) | ||
794 | buf[0] |= (1 << 4); | ||
795 | |||
796 | if (cond == SINGLE) | ||
797 | buf[0] |= (0 << 1) | (1 << 0); | ||
798 | else if (cond == BURST) | ||
799 | buf[0] |= (1 << 1) | (1 << 0); | ||
800 | |||
801 | buf[1] = bjump; | ||
802 | |||
803 | PL330_DBGCMD_DUMP(SZ_DMALPEND, "\tDMALP%s%c_%c bjmpto_%x\n", | ||
804 | forever ? "FE" : "END", | ||
805 | cond == SINGLE ? 'S' : (cond == BURST ? 'B' : 'A'), | ||
806 | loop ? '1' : '0', | ||
807 | bjump); | ||
808 | |||
809 | return SZ_DMALPEND; | ||
810 | } | ||
811 | |||
812 | static inline u32 _emit_KILL(unsigned dry_run, u8 buf[]) | ||
813 | { | ||
814 | if (dry_run) | ||
815 | return SZ_DMAKILL; | ||
816 | |||
817 | buf[0] = CMD_DMAKILL; | ||
818 | |||
819 | return SZ_DMAKILL; | ||
820 | } | ||
821 | |||
822 | static inline u32 _emit_MOV(unsigned dry_run, u8 buf[], | ||
823 | enum dmamov_dst dst, u32 val) | ||
824 | { | ||
825 | if (dry_run) | ||
826 | return SZ_DMAMOV; | ||
827 | |||
828 | buf[0] = CMD_DMAMOV; | ||
829 | buf[1] = dst; | ||
830 | *((u32 *)&buf[2]) = val; | ||
831 | |||
832 | PL330_DBGCMD_DUMP(SZ_DMAMOV, "\tDMAMOV %s 0x%x\n", | ||
833 | dst == SAR ? "SAR" : (dst == DAR ? "DAR" : "CCR"), val); | ||
834 | |||
835 | return SZ_DMAMOV; | ||
836 | } | ||
837 | |||
838 | static inline u32 _emit_NOP(unsigned dry_run, u8 buf[]) | ||
839 | { | ||
840 | if (dry_run) | ||
841 | return SZ_DMANOP; | ||
842 | |||
843 | buf[0] = CMD_DMANOP; | ||
844 | |||
845 | PL330_DBGCMD_DUMP(SZ_DMANOP, "\tDMANOP\n"); | ||
846 | |||
847 | return SZ_DMANOP; | ||
848 | } | ||
849 | |||
850 | static inline u32 _emit_RMB(unsigned dry_run, u8 buf[]) | ||
851 | { | ||
852 | if (dry_run) | ||
853 | return SZ_DMARMB; | ||
854 | |||
855 | buf[0] = CMD_DMARMB; | ||
856 | |||
857 | PL330_DBGCMD_DUMP(SZ_DMARMB, "\tDMARMB\n"); | ||
858 | |||
859 | return SZ_DMARMB; | ||
860 | } | ||
861 | |||
862 | static inline u32 _emit_SEV(unsigned dry_run, u8 buf[], u8 ev) | ||
863 | { | ||
864 | if (dry_run) | ||
865 | return SZ_DMASEV; | ||
866 | |||
867 | buf[0] = CMD_DMASEV; | ||
868 | |||
869 | ev &= 0x1f; | ||
870 | ev <<= 3; | ||
871 | buf[1] = ev; | ||
872 | |||
873 | PL330_DBGCMD_DUMP(SZ_DMASEV, "\tDMASEV %u\n", ev >> 3); | ||
874 | |||
875 | return SZ_DMASEV; | ||
876 | } | ||
877 | |||
878 | static inline u32 _emit_ST(unsigned dry_run, u8 buf[], enum pl330_cond cond) | ||
879 | { | ||
880 | if (dry_run) | ||
881 | return SZ_DMAST; | ||
882 | |||
883 | buf[0] = CMD_DMAST; | ||
884 | |||
885 | if (cond == SINGLE) | ||
886 | buf[0] |= (0 << 1) | (1 << 0); | ||
887 | else if (cond == BURST) | ||
888 | buf[0] |= (1 << 1) | (1 << 0); | ||
889 | |||
890 | PL330_DBGCMD_DUMP(SZ_DMAST, "\tDMAST%c\n", | ||
891 | cond == SINGLE ? 'S' : (cond == BURST ? 'B' : 'A')); | ||
892 | |||
893 | return SZ_DMAST; | ||
894 | } | ||
895 | |||
896 | static inline u32 _emit_STP(unsigned dry_run, u8 buf[], | ||
897 | enum pl330_cond cond, u8 peri) | ||
898 | { | ||
899 | if (dry_run) | ||
900 | return SZ_DMASTP; | ||
901 | |||
902 | buf[0] = CMD_DMASTP; | ||
903 | |||
904 | if (cond == BURST) | ||
905 | buf[0] |= (1 << 1); | ||
906 | |||
907 | peri &= 0x1f; | ||
908 | peri <<= 3; | ||
909 | buf[1] = peri; | ||
910 | |||
911 | PL330_DBGCMD_DUMP(SZ_DMASTP, "\tDMASTP%c %u\n", | ||
912 | cond == SINGLE ? 'S' : 'B', peri >> 3); | ||
913 | |||
914 | return SZ_DMASTP; | ||
915 | } | ||
916 | |||
917 | static inline u32 _emit_STZ(unsigned dry_run, u8 buf[]) | ||
918 | { | ||
919 | if (dry_run) | ||
920 | return SZ_DMASTZ; | ||
921 | |||
922 | buf[0] = CMD_DMASTZ; | ||
923 | |||
924 | PL330_DBGCMD_DUMP(SZ_DMASTZ, "\tDMASTZ\n"); | ||
925 | |||
926 | return SZ_DMASTZ; | ||
927 | } | ||
928 | |||
929 | static inline u32 _emit_WFE(unsigned dry_run, u8 buf[], u8 ev, | ||
930 | unsigned invalidate) | ||
931 | { | ||
932 | if (dry_run) | ||
933 | return SZ_DMAWFE; | ||
934 | |||
935 | buf[0] = CMD_DMAWFE; | ||
936 | |||
937 | ev &= 0x1f; | ||
938 | ev <<= 3; | ||
939 | buf[1] = ev; | ||
940 | |||
941 | if (invalidate) | ||
942 | buf[1] |= (1 << 1); | ||
943 | |||
944 | PL330_DBGCMD_DUMP(SZ_DMAWFE, "\tDMAWFE %u%s\n", | ||
945 | ev >> 3, invalidate ? ", I" : ""); | ||
946 | |||
947 | return SZ_DMAWFE; | ||
948 | } | ||
949 | |||
950 | static inline u32 _emit_WFP(unsigned dry_run, u8 buf[], | ||
951 | enum pl330_cond cond, u8 peri) | ||
952 | { | ||
953 | if (dry_run) | ||
954 | return SZ_DMAWFP; | ||
955 | |||
956 | buf[0] = CMD_DMAWFP; | ||
957 | |||
958 | if (cond == SINGLE) | ||
959 | buf[0] |= (0 << 1) | (0 << 0); | ||
960 | else if (cond == BURST) | ||
961 | buf[0] |= (1 << 1) | (0 << 0); | ||
962 | else | ||
963 | buf[0] |= (0 << 1) | (1 << 0); | ||
964 | |||
965 | peri &= 0x1f; | ||
966 | peri <<= 3; | ||
967 | buf[1] = peri; | ||
968 | |||
969 | PL330_DBGCMD_DUMP(SZ_DMAWFP, "\tDMAWFP%c %u\n", | ||
970 | cond == SINGLE ? 'S' : (cond == BURST ? 'B' : 'P'), peri >> 3); | ||
971 | |||
972 | return SZ_DMAWFP; | ||
973 | } | ||
974 | |||
975 | static inline u32 _emit_WMB(unsigned dry_run, u8 buf[]) | ||
976 | { | ||
977 | if (dry_run) | ||
978 | return SZ_DMAWMB; | ||
979 | |||
980 | buf[0] = CMD_DMAWMB; | ||
981 | |||
982 | PL330_DBGCMD_DUMP(SZ_DMAWMB, "\tDMAWMB\n"); | ||
983 | |||
984 | return SZ_DMAWMB; | ||
985 | } | ||
986 | |||
987 | struct _arg_GO { | ||
988 | u8 chan; | ||
989 | u32 addr; | ||
990 | unsigned ns; | ||
991 | }; | ||
992 | |||
993 | static inline u32 _emit_GO(unsigned dry_run, u8 buf[], | ||
994 | const struct _arg_GO *arg) | ||
995 | { | ||
996 | u8 chan = arg->chan; | ||
997 | u32 addr = arg->addr; | ||
998 | unsigned ns = arg->ns; | ||
999 | |||
1000 | if (dry_run) | ||
1001 | return SZ_DMAGO; | ||
1002 | |||
1003 | buf[0] = CMD_DMAGO; | ||
1004 | buf[0] |= (ns << 1); | ||
1005 | |||
1006 | buf[1] = chan & 0x7; | ||
1007 | |||
1008 | *((u32 *)&buf[2]) = addr; | ||
1009 | |||
1010 | return SZ_DMAGO; | ||
1011 | } | ||
1012 | |||
1013 | #define msecs_to_loops(t) (loops_per_jiffy / 1000 * HZ * t) | ||
1014 | |||
1015 | /* Returns Time-Out */ | ||
1016 | static bool _until_dmac_idle(struct pl330_thread *thrd) | ||
1017 | { | ||
1018 | void __iomem *regs = thrd->dmac->pinfo->base; | ||
1019 | unsigned long loops = msecs_to_loops(5); | ||
1020 | |||
1021 | do { | ||
1022 | /* Until Manager is Idle */ | ||
1023 | if (!(readl(regs + DBGSTATUS) & DBG_BUSY)) | ||
1024 | break; | ||
1025 | |||
1026 | cpu_relax(); | ||
1027 | } while (--loops); | ||
1028 | |||
1029 | if (!loops) | ||
1030 | return true; | ||
1031 | |||
1032 | return false; | ||
1033 | } | ||
1034 | |||
1035 | static inline void _execute_DBGINSN(struct pl330_thread *thrd, | ||
1036 | u8 insn[], bool as_manager) | ||
1037 | { | ||
1038 | void __iomem *regs = thrd->dmac->pinfo->base; | ||
1039 | u32 val; | ||
1040 | |||
1041 | val = (insn[0] << 16) | (insn[1] << 24); | ||
1042 | if (!as_manager) { | ||
1043 | val |= (1 << 0); | ||
1044 | val |= (thrd->id << 8); /* Channel Number */ | ||
1045 | } | ||
1046 | writel(val, regs + DBGINST0); | ||
1047 | |||
1048 | val = *((u32 *)&insn[2]); | ||
1049 | writel(val, regs + DBGINST1); | ||
1050 | |||
1051 | /* If timed out due to halted state-machine */ | ||
1052 | if (_until_dmac_idle(thrd)) { | ||
1053 | dev_err(thrd->dmac->pinfo->dev, "DMAC halted!\n"); | ||
1054 | return; | ||
1055 | } | ||
1056 | |||
1057 | /* Get going */ | ||
1058 | writel(0, regs + DBGCMD); | ||
1059 | } | ||
1060 | |||
1061 | /* | ||
1062 | * Mark a _pl330_req as free. | ||
1063 | * We do it by writing DMAEND as the first instruction | ||
1064 | * because no valid request is going to have DMAEND as | ||
1065 | * its first instruction to execute. | ||
1066 | */ | ||
1067 | static void mark_free(struct pl330_thread *thrd, int idx) | ||
1068 | { | ||
1069 | struct _pl330_req *req = &thrd->req[idx]; | ||
1070 | |||
1071 | _emit_END(0, req->mc_cpu); | ||
1072 | req->mc_len = 0; | ||
1073 | |||
1074 | thrd->req_running = -1; | ||
1075 | } | ||
1076 | |||
1077 | static inline u32 _state(struct pl330_thread *thrd) | ||
1078 | { | ||
1079 | void __iomem *regs = thrd->dmac->pinfo->base; | ||
1080 | u32 val; | ||
1081 | |||
1082 | if (is_manager(thrd)) | ||
1083 | val = readl(regs + DS) & 0xf; | ||
1084 | else | ||
1085 | val = readl(regs + CS(thrd->id)) & 0xf; | ||
1086 | |||
1087 | switch (val) { | ||
1088 | case DS_ST_STOP: | ||
1089 | return PL330_STATE_STOPPED; | ||
1090 | case DS_ST_EXEC: | ||
1091 | return PL330_STATE_EXECUTING; | ||
1092 | case DS_ST_CMISS: | ||
1093 | return PL330_STATE_CACHEMISS; | ||
1094 | case DS_ST_UPDTPC: | ||
1095 | return PL330_STATE_UPDTPC; | ||
1096 | case DS_ST_WFE: | ||
1097 | return PL330_STATE_WFE; | ||
1098 | case DS_ST_FAULT: | ||
1099 | return PL330_STATE_FAULTING; | ||
1100 | case DS_ST_ATBRR: | ||
1101 | if (is_manager(thrd)) | ||
1102 | return PL330_STATE_INVALID; | ||
1103 | else | ||
1104 | return PL330_STATE_ATBARRIER; | ||
1105 | case DS_ST_QBUSY: | ||
1106 | if (is_manager(thrd)) | ||
1107 | return PL330_STATE_INVALID; | ||
1108 | else | ||
1109 | return PL330_STATE_QUEUEBUSY; | ||
1110 | case DS_ST_WFP: | ||
1111 | if (is_manager(thrd)) | ||
1112 | return PL330_STATE_INVALID; | ||
1113 | else | ||
1114 | return PL330_STATE_WFP; | ||
1115 | case DS_ST_KILL: | ||
1116 | if (is_manager(thrd)) | ||
1117 | return PL330_STATE_INVALID; | ||
1118 | else | ||
1119 | return PL330_STATE_KILLING; | ||
1120 | case DS_ST_CMPLT: | ||
1121 | if (is_manager(thrd)) | ||
1122 | return PL330_STATE_INVALID; | ||
1123 | else | ||
1124 | return PL330_STATE_COMPLETING; | ||
1125 | case DS_ST_FLTCMP: | ||
1126 | if (is_manager(thrd)) | ||
1127 | return PL330_STATE_INVALID; | ||
1128 | else | ||
1129 | return PL330_STATE_FAULT_COMPLETING; | ||
1130 | default: | ||
1131 | return PL330_STATE_INVALID; | ||
1132 | } | ||
1133 | } | ||
1134 | |||
1135 | static void _stop(struct pl330_thread *thrd) | ||
1136 | { | ||
1137 | void __iomem *regs = thrd->dmac->pinfo->base; | ||
1138 | u8 insn[6] = {0, 0, 0, 0, 0, 0}; | ||
1139 | |||
1140 | if (_state(thrd) == PL330_STATE_FAULT_COMPLETING) | ||
1141 | UNTIL(thrd, PL330_STATE_FAULTING | PL330_STATE_KILLING); | ||
1142 | |||
1143 | /* Return if nothing needs to be done */ | ||
1144 | if (_state(thrd) == PL330_STATE_COMPLETING | ||
1145 | || _state(thrd) == PL330_STATE_KILLING | ||
1146 | || _state(thrd) == PL330_STATE_STOPPED) | ||
1147 | return; | ||
1148 | |||
1149 | _emit_KILL(0, insn); | ||
1150 | |||
1151 | /* Stop generating interrupts for SEV */ | ||
1152 | writel(readl(regs + INTEN) & ~(1 << thrd->ev), regs + INTEN); | ||
1153 | |||
1154 | _execute_DBGINSN(thrd, insn, is_manager(thrd)); | ||
1155 | } | ||
1156 | |||
1157 | /* Start doing req 'idx' of thread 'thrd' */ | ||
1158 | static bool _trigger(struct pl330_thread *thrd) | ||
1159 | { | ||
1160 | void __iomem *regs = thrd->dmac->pinfo->base; | ||
1161 | struct _pl330_req *req; | ||
1162 | struct pl330_req *r; | ||
1163 | struct _arg_GO go; | ||
1164 | unsigned ns; | ||
1165 | u8 insn[6] = {0, 0, 0, 0, 0, 0}; | ||
1166 | int idx; | ||
1167 | |||
1168 | /* Return if already ACTIVE */ | ||
1169 | if (_state(thrd) != PL330_STATE_STOPPED) | ||
1170 | return true; | ||
1171 | |||
1172 | idx = 1 - thrd->lstenq; | ||
1173 | if (!IS_FREE(&thrd->req[idx])) | ||
1174 | req = &thrd->req[idx]; | ||
1175 | else { | ||
1176 | idx = thrd->lstenq; | ||
1177 | if (!IS_FREE(&thrd->req[idx])) | ||
1178 | req = &thrd->req[idx]; | ||
1179 | else | ||
1180 | req = NULL; | ||
1181 | } | ||
1182 | |||
1183 | /* Return if no request */ | ||
1184 | if (!req || !req->r) | ||
1185 | return true; | ||
1186 | |||
1187 | r = req->r; | ||
1188 | |||
1189 | if (r->cfg) | ||
1190 | ns = r->cfg->nonsecure ? 1 : 0; | ||
1191 | else if (readl(regs + CS(thrd->id)) & CS_CNS) | ||
1192 | ns = 1; | ||
1193 | else | ||
1194 | ns = 0; | ||
1195 | |||
1196 | /* See 'Abort Sources' point-4 at Page 2-25 */ | ||
1197 | if (_manager_ns(thrd) && !ns) | ||
1198 | dev_info(thrd->dmac->pinfo->dev, "%s:%d Recipe for ABORT!\n", | ||
1199 | __func__, __LINE__); | ||
1200 | |||
1201 | go.chan = thrd->id; | ||
1202 | go.addr = req->mc_bus; | ||
1203 | go.ns = ns; | ||
1204 | _emit_GO(0, insn, &go); | ||
1205 | |||
1206 | /* Set to generate interrupts for SEV */ | ||
1207 | writel(readl(regs + INTEN) | (1 << thrd->ev), regs + INTEN); | ||
1208 | |||
1209 | /* Only manager can execute GO */ | ||
1210 | _execute_DBGINSN(thrd, insn, true); | ||
1211 | |||
1212 | thrd->req_running = idx; | ||
1213 | |||
1214 | return true; | ||
1215 | } | ||
1216 | |||
1217 | static bool _start(struct pl330_thread *thrd) | ||
1218 | { | ||
1219 | switch (_state(thrd)) { | ||
1220 | case PL330_STATE_FAULT_COMPLETING: | ||
1221 | UNTIL(thrd, PL330_STATE_FAULTING | PL330_STATE_KILLING); | ||
1222 | |||
1223 | if (_state(thrd) == PL330_STATE_KILLING) | ||
1224 | UNTIL(thrd, PL330_STATE_STOPPED) | ||
1225 | |||
1226 | case PL330_STATE_FAULTING: | ||
1227 | _stop(thrd); | ||
1228 | |||
1229 | case PL330_STATE_KILLING: | ||
1230 | case PL330_STATE_COMPLETING: | ||
1231 | UNTIL(thrd, PL330_STATE_STOPPED) | ||
1232 | |||
1233 | case PL330_STATE_STOPPED: | ||
1234 | return _trigger(thrd); | ||
1235 | |||
1236 | case PL330_STATE_WFP: | ||
1237 | case PL330_STATE_QUEUEBUSY: | ||
1238 | case PL330_STATE_ATBARRIER: | ||
1239 | case PL330_STATE_UPDTPC: | ||
1240 | case PL330_STATE_CACHEMISS: | ||
1241 | case PL330_STATE_EXECUTING: | ||
1242 | return true; | ||
1243 | |||
1244 | case PL330_STATE_WFE: /* For RESUME, nothing yet */ | ||
1245 | default: | ||
1246 | return false; | ||
1247 | } | ||
1248 | } | ||
1249 | |||
1250 | static inline int _ldst_memtomem(unsigned dry_run, u8 buf[], | ||
1251 | const struct _xfer_spec *pxs, int cyc) | ||
1252 | { | ||
1253 | int off = 0; | ||
1254 | struct pl330_config *pcfg = pxs->r->cfg->pcfg; | ||
1255 | |||
1256 | /* check lock-up free version */ | ||
1257 | if (get_revision(pcfg->periph_id) >= PERIPH_REV_R1P0) { | ||
1258 | while (cyc--) { | ||
1259 | off += _emit_LD(dry_run, &buf[off], ALWAYS); | ||
1260 | off += _emit_ST(dry_run, &buf[off], ALWAYS); | ||
1261 | } | ||
1262 | } else { | ||
1263 | while (cyc--) { | ||
1264 | off += _emit_LD(dry_run, &buf[off], ALWAYS); | ||
1265 | off += _emit_RMB(dry_run, &buf[off]); | ||
1266 | off += _emit_ST(dry_run, &buf[off], ALWAYS); | ||
1267 | off += _emit_WMB(dry_run, &buf[off]); | ||
1268 | } | ||
1269 | } | ||
1270 | |||
1271 | return off; | ||
1272 | } | ||
1273 | |||
1274 | static inline int _ldst_devtomem(unsigned dry_run, u8 buf[], | ||
1275 | const struct _xfer_spec *pxs, int cyc) | ||
1276 | { | ||
1277 | int off = 0; | ||
1278 | |||
1279 | while (cyc--) { | ||
1280 | off += _emit_WFP(dry_run, &buf[off], SINGLE, pxs->r->peri); | ||
1281 | off += _emit_LDP(dry_run, &buf[off], SINGLE, pxs->r->peri); | ||
1282 | off += _emit_ST(dry_run, &buf[off], ALWAYS); | ||
1283 | off += _emit_FLUSHP(dry_run, &buf[off], pxs->r->peri); | ||
1284 | } | ||
1285 | |||
1286 | return off; | ||
1287 | } | ||
1288 | |||
1289 | static inline int _ldst_memtodev(unsigned dry_run, u8 buf[], | ||
1290 | const struct _xfer_spec *pxs, int cyc) | ||
1291 | { | ||
1292 | int off = 0; | ||
1293 | |||
1294 | while (cyc--) { | ||
1295 | off += _emit_WFP(dry_run, &buf[off], SINGLE, pxs->r->peri); | ||
1296 | off += _emit_LD(dry_run, &buf[off], ALWAYS); | ||
1297 | off += _emit_STP(dry_run, &buf[off], SINGLE, pxs->r->peri); | ||
1298 | off += _emit_FLUSHP(dry_run, &buf[off], pxs->r->peri); | ||
1299 | } | ||
1300 | |||
1301 | return off; | ||
1302 | } | ||
1303 | |||
1304 | static int _bursts(unsigned dry_run, u8 buf[], | ||
1305 | const struct _xfer_spec *pxs, int cyc) | ||
1306 | { | ||
1307 | int off = 0; | ||
1308 | |||
1309 | switch (pxs->r->rqtype) { | ||
1310 | case MEMTODEV: | ||
1311 | off += _ldst_memtodev(dry_run, &buf[off], pxs, cyc); | ||
1312 | break; | ||
1313 | case DEVTOMEM: | ||
1314 | off += _ldst_devtomem(dry_run, &buf[off], pxs, cyc); | ||
1315 | break; | ||
1316 | case MEMTOMEM: | ||
1317 | off += _ldst_memtomem(dry_run, &buf[off], pxs, cyc); | ||
1318 | break; | ||
1319 | default: | ||
1320 | off += 0x40000000; /* Scare off the Client */ | ||
1321 | break; | ||
1322 | } | ||
1323 | |||
1324 | return off; | ||
1325 | } | ||
1326 | |||
1327 | /* Returns bytes consumed and updates bursts */ | ||
1328 | static inline int _loop(unsigned dry_run, u8 buf[], | ||
1329 | unsigned long *bursts, const struct _xfer_spec *pxs) | ||
1330 | { | ||
1331 | int cyc, cycmax, szlp, szlpend, szbrst, off; | ||
1332 | unsigned lcnt0, lcnt1, ljmp0, ljmp1; | ||
1333 | struct _arg_LPEND lpend; | ||
1334 | |||
1335 | /* Max iterations possible in DMALP is 256 */ | ||
1336 | if (*bursts >= 256*256) { | ||
1337 | lcnt1 = 256; | ||
1338 | lcnt0 = 256; | ||
1339 | cyc = *bursts / lcnt1 / lcnt0; | ||
1340 | } else if (*bursts > 256) { | ||
1341 | lcnt1 = 256; | ||
1342 | lcnt0 = *bursts / lcnt1; | ||
1343 | cyc = 1; | ||
1344 | } else { | ||
1345 | lcnt1 = *bursts; | ||
1346 | lcnt0 = 0; | ||
1347 | cyc = 1; | ||
1348 | } | ||
1349 | |||
1350 | szlp = _emit_LP(1, buf, 0, 0); | ||
1351 | szbrst = _bursts(1, buf, pxs, 1); | ||
1352 | |||
1353 | lpend.cond = ALWAYS; | ||
1354 | lpend.forever = false; | ||
1355 | lpend.loop = 0; | ||
1356 | lpend.bjump = 0; | ||
1357 | szlpend = _emit_LPEND(1, buf, &lpend); | ||
1358 | |||
1359 | if (lcnt0) { | ||
1360 | szlp *= 2; | ||
1361 | szlpend *= 2; | ||
1362 | } | ||
1363 | |||
1364 | /* | ||
1365 | * Max bursts that we can unroll due to limit on the | ||
1366 | * size of backward jump that can be encoded in DMALPEND | ||
1367 | * which is 8-bits and hence 255 | ||
1368 | */ | ||
1369 | cycmax = (255 - (szlp + szlpend)) / szbrst; | ||
1370 | |||
1371 | cyc = (cycmax < cyc) ? cycmax : cyc; | ||
1372 | |||
1373 | off = 0; | ||
1374 | |||
1375 | if (lcnt0) { | ||
1376 | off += _emit_LP(dry_run, &buf[off], 0, lcnt0); | ||
1377 | ljmp0 = off; | ||
1378 | } | ||
1379 | |||
1380 | off += _emit_LP(dry_run, &buf[off], 1, lcnt1); | ||
1381 | ljmp1 = off; | ||
1382 | |||
1383 | off += _bursts(dry_run, &buf[off], pxs, cyc); | ||
1384 | |||
1385 | lpend.cond = ALWAYS; | ||
1386 | lpend.forever = false; | ||
1387 | lpend.loop = 1; | ||
1388 | lpend.bjump = off - ljmp1; | ||
1389 | off += _emit_LPEND(dry_run, &buf[off], &lpend); | ||
1390 | |||
1391 | if (lcnt0) { | ||
1392 | lpend.cond = ALWAYS; | ||
1393 | lpend.forever = false; | ||
1394 | lpend.loop = 0; | ||
1395 | lpend.bjump = off - ljmp0; | ||
1396 | off += _emit_LPEND(dry_run, &buf[off], &lpend); | ||
1397 | } | ||
1398 | |||
1399 | *bursts = lcnt1 * cyc; | ||
1400 | if (lcnt0) | ||
1401 | *bursts *= lcnt0; | ||
1402 | |||
1403 | return off; | ||
1404 | } | ||
1405 | |||
1406 | static inline int _setup_loops(unsigned dry_run, u8 buf[], | ||
1407 | const struct _xfer_spec *pxs) | ||
1408 | { | ||
1409 | struct pl330_xfer *x = pxs->x; | ||
1410 | u32 ccr = pxs->ccr; | ||
1411 | unsigned long c, bursts = BYTE_TO_BURST(x->bytes, ccr); | ||
1412 | int off = 0; | ||
1413 | |||
1414 | while (bursts) { | ||
1415 | c = bursts; | ||
1416 | off += _loop(dry_run, &buf[off], &c, pxs); | ||
1417 | bursts -= c; | ||
1418 | } | ||
1419 | |||
1420 | return off; | ||
1421 | } | ||
1422 | |||
1423 | static inline int _setup_xfer(unsigned dry_run, u8 buf[], | ||
1424 | const struct _xfer_spec *pxs) | ||
1425 | { | ||
1426 | struct pl330_xfer *x = pxs->x; | ||
1427 | int off = 0; | ||
1428 | |||
1429 | /* DMAMOV SAR, x->src_addr */ | ||
1430 | off += _emit_MOV(dry_run, &buf[off], SAR, x->src_addr); | ||
1431 | /* DMAMOV DAR, x->dst_addr */ | ||
1432 | off += _emit_MOV(dry_run, &buf[off], DAR, x->dst_addr); | ||
1433 | |||
1434 | /* Setup Loop(s) */ | ||
1435 | off += _setup_loops(dry_run, &buf[off], pxs); | ||
1436 | |||
1437 | return off; | ||
1438 | } | ||
1439 | |||
1440 | /* | ||
1441 | * A req is a sequence of one or more xfer units. | ||
1442 | * Returns the number of bytes taken to setup the MC for the req. | ||
1443 | */ | ||
1444 | static int _setup_req(unsigned dry_run, struct pl330_thread *thrd, | ||
1445 | unsigned index, struct _xfer_spec *pxs) | ||
1446 | { | ||
1447 | struct _pl330_req *req = &thrd->req[index]; | ||
1448 | struct pl330_xfer *x; | ||
1449 | u8 *buf = req->mc_cpu; | ||
1450 | int off = 0; | ||
1451 | |||
1452 | PL330_DBGMC_START(req->mc_bus); | ||
1453 | |||
1454 | /* DMAMOV CCR, ccr */ | ||
1455 | off += _emit_MOV(dry_run, &buf[off], CCR, pxs->ccr); | ||
1456 | |||
1457 | x = pxs->r->x; | ||
1458 | do { | ||
1459 | /* Error if xfer length is not aligned at burst size */ | ||
1460 | if (x->bytes % (BRST_SIZE(pxs->ccr) * BRST_LEN(pxs->ccr))) | ||
1461 | return -EINVAL; | ||
1462 | |||
1463 | pxs->x = x; | ||
1464 | off += _setup_xfer(dry_run, &buf[off], pxs); | ||
1465 | |||
1466 | x = x->next; | ||
1467 | } while (x); | ||
1468 | |||
1469 | /* DMASEV peripheral/event */ | ||
1470 | off += _emit_SEV(dry_run, &buf[off], thrd->ev); | ||
1471 | /* DMAEND */ | ||
1472 | off += _emit_END(dry_run, &buf[off]); | ||
1473 | |||
1474 | return off; | ||
1475 | } | ||
1476 | |||
1477 | static inline u32 _prepare_ccr(const struct pl330_reqcfg *rqc) | ||
1478 | { | ||
1479 | u32 ccr = 0; | ||
1480 | |||
1481 | if (rqc->src_inc) | ||
1482 | ccr |= CC_SRCINC; | ||
1483 | |||
1484 | if (rqc->dst_inc) | ||
1485 | ccr |= CC_DSTINC; | ||
1486 | |||
1487 | /* We set same protection levels for Src and DST for now */ | ||
1488 | if (rqc->privileged) | ||
1489 | ccr |= CC_SRCPRI | CC_DSTPRI; | ||
1490 | if (rqc->nonsecure) | ||
1491 | ccr |= CC_SRCNS | CC_DSTNS; | ||
1492 | if (rqc->insnaccess) | ||
1493 | ccr |= CC_SRCIA | CC_DSTIA; | ||
1494 | |||
1495 | ccr |= (((rqc->brst_len - 1) & 0xf) << CC_SRCBRSTLEN_SHFT); | ||
1496 | ccr |= (((rqc->brst_len - 1) & 0xf) << CC_DSTBRSTLEN_SHFT); | ||
1497 | |||
1498 | ccr |= (rqc->brst_size << CC_SRCBRSTSIZE_SHFT); | ||
1499 | ccr |= (rqc->brst_size << CC_DSTBRSTSIZE_SHFT); | ||
1500 | |||
1501 | ccr |= (rqc->scctl << CC_SRCCCTRL_SHFT); | ||
1502 | ccr |= (rqc->dcctl << CC_DSTCCTRL_SHFT); | ||
1503 | |||
1504 | ccr |= (rqc->swap << CC_SWAP_SHFT); | ||
1505 | |||
1506 | return ccr; | ||
1507 | } | ||
1508 | |||
1509 | static inline bool _is_valid(u32 ccr) | ||
1510 | { | ||
1511 | enum pl330_dstcachectrl dcctl; | ||
1512 | enum pl330_srccachectrl scctl; | ||
1513 | |||
1514 | dcctl = (ccr >> CC_DSTCCTRL_SHFT) & CC_DRCCCTRL_MASK; | ||
1515 | scctl = (ccr >> CC_SRCCCTRL_SHFT) & CC_SRCCCTRL_MASK; | ||
1516 | |||
1517 | if (dcctl == DINVALID1 || dcctl == DINVALID2 | ||
1518 | || scctl == SINVALID1 || scctl == SINVALID2) | ||
1519 | return false; | ||
1520 | else | ||
1521 | return true; | ||
1522 | } | ||
1523 | |||
1524 | /* | ||
1525 | * Submit a list of xfers after which the client wants notification. | ||
1526 | * Client is not notified after each xfer unit, just once after all | ||
1527 | * xfer units are done or some error occurs. | ||
1528 | */ | ||
1529 | static int pl330_submit_req(void *ch_id, struct pl330_req *r) | ||
1530 | { | ||
1531 | struct pl330_thread *thrd = ch_id; | ||
1532 | struct pl330_dmac *pl330; | ||
1533 | struct pl330_info *pi; | ||
1534 | struct _xfer_spec xs; | ||
1535 | unsigned long flags; | ||
1536 | void __iomem *regs; | ||
1537 | unsigned idx; | ||
1538 | u32 ccr; | ||
1539 | int ret = 0; | ||
1540 | |||
1541 | /* No Req or Unacquired Channel or DMAC */ | ||
1542 | if (!r || !thrd || thrd->free) | ||
1543 | return -EINVAL; | ||
1544 | |||
1545 | pl330 = thrd->dmac; | ||
1546 | pi = pl330->pinfo; | ||
1547 | regs = pi->base; | ||
1548 | |||
1549 | if (pl330->state == DYING | ||
1550 | || pl330->dmac_tbd.reset_chan & (1 << thrd->id)) { | ||
1551 | dev_info(thrd->dmac->pinfo->dev, "%s:%d\n", | ||
1552 | __func__, __LINE__); | ||
1553 | return -EAGAIN; | ||
1554 | } | ||
1555 | |||
1556 | /* If request for non-existing peripheral */ | ||
1557 | if (r->rqtype != MEMTOMEM && r->peri >= pi->pcfg.num_peri) { | ||
1558 | dev_info(thrd->dmac->pinfo->dev, | ||
1559 | "%s:%d Invalid peripheral(%u)!\n", | ||
1560 | __func__, __LINE__, r->peri); | ||
1561 | return -EINVAL; | ||
1562 | } | ||
1563 | |||
1564 | spin_lock_irqsave(&pl330->lock, flags); | ||
1565 | |||
1566 | if (_queue_full(thrd)) { | ||
1567 | ret = -EAGAIN; | ||
1568 | goto xfer_exit; | ||
1569 | } | ||
1570 | |||
1571 | /* Prefer Secure Channel */ | ||
1572 | if (!_manager_ns(thrd)) | ||
1573 | r->cfg->nonsecure = 0; | ||
1574 | else | ||
1575 | r->cfg->nonsecure = 1; | ||
1576 | |||
1577 | /* Use last settings, if not provided */ | ||
1578 | if (r->cfg) | ||
1579 | ccr = _prepare_ccr(r->cfg); | ||
1580 | else | ||
1581 | ccr = readl(regs + CC(thrd->id)); | ||
1582 | |||
1583 | /* If this req doesn't have valid xfer settings */ | ||
1584 | if (!_is_valid(ccr)) { | ||
1585 | ret = -EINVAL; | ||
1586 | dev_info(thrd->dmac->pinfo->dev, "%s:%d Invalid CCR(%x)!\n", | ||
1587 | __func__, __LINE__, ccr); | ||
1588 | goto xfer_exit; | ||
1589 | } | ||
1590 | |||
1591 | idx = IS_FREE(&thrd->req[0]) ? 0 : 1; | ||
1592 | |||
1593 | xs.ccr = ccr; | ||
1594 | xs.r = r; | ||
1595 | |||
1596 | /* First dry run to check if req is acceptable */ | ||
1597 | ret = _setup_req(1, thrd, idx, &xs); | ||
1598 | if (ret < 0) | ||
1599 | goto xfer_exit; | ||
1600 | |||
1601 | if (ret > pi->mcbufsz / 2) { | ||
1602 | dev_info(thrd->dmac->pinfo->dev, | ||
1603 | "%s:%d Trying increasing mcbufsz\n", | ||
1604 | __func__, __LINE__); | ||
1605 | ret = -ENOMEM; | ||
1606 | goto xfer_exit; | ||
1607 | } | ||
1608 | |||
1609 | /* Hook the request */ | ||
1610 | thrd->lstenq = idx; | ||
1611 | thrd->req[idx].mc_len = _setup_req(0, thrd, idx, &xs); | ||
1612 | thrd->req[idx].r = r; | ||
1613 | |||
1614 | ret = 0; | ||
1615 | |||
1616 | xfer_exit: | ||
1617 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1618 | |||
1619 | return ret; | ||
1620 | } | ||
1621 | |||
1622 | static void pl330_dotask(unsigned long data) | ||
1623 | { | ||
1624 | struct pl330_dmac *pl330 = (struct pl330_dmac *) data; | ||
1625 | struct pl330_info *pi = pl330->pinfo; | ||
1626 | unsigned long flags; | ||
1627 | int i; | ||
1628 | |||
1629 | spin_lock_irqsave(&pl330->lock, flags); | ||
1630 | |||
1631 | /* The DMAC itself gone nuts */ | ||
1632 | if (pl330->dmac_tbd.reset_dmac) { | ||
1633 | pl330->state = DYING; | ||
1634 | /* Reset the manager too */ | ||
1635 | pl330->dmac_tbd.reset_mngr = true; | ||
1636 | /* Clear the reset flag */ | ||
1637 | pl330->dmac_tbd.reset_dmac = false; | ||
1638 | } | ||
1639 | |||
1640 | if (pl330->dmac_tbd.reset_mngr) { | ||
1641 | _stop(pl330->manager); | ||
1642 | /* Reset all channels */ | ||
1643 | pl330->dmac_tbd.reset_chan = (1 << pi->pcfg.num_chan) - 1; | ||
1644 | /* Clear the reset flag */ | ||
1645 | pl330->dmac_tbd.reset_mngr = false; | ||
1646 | } | ||
1647 | |||
1648 | for (i = 0; i < pi->pcfg.num_chan; i++) { | ||
1649 | |||
1650 | if (pl330->dmac_tbd.reset_chan & (1 << i)) { | ||
1651 | struct pl330_thread *thrd = &pl330->channels[i]; | ||
1652 | void __iomem *regs = pi->base; | ||
1653 | enum pl330_op_err err; | ||
1654 | |||
1655 | _stop(thrd); | ||
1656 | |||
1657 | if (readl(regs + FSC) & (1 << thrd->id)) | ||
1658 | err = PL330_ERR_FAIL; | ||
1659 | else | ||
1660 | err = PL330_ERR_ABORT; | ||
1661 | |||
1662 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1663 | |||
1664 | _callback(thrd->req[1 - thrd->lstenq].r, err); | ||
1665 | _callback(thrd->req[thrd->lstenq].r, err); | ||
1666 | |||
1667 | spin_lock_irqsave(&pl330->lock, flags); | ||
1668 | |||
1669 | thrd->req[0].r = NULL; | ||
1670 | thrd->req[1].r = NULL; | ||
1671 | mark_free(thrd, 0); | ||
1672 | mark_free(thrd, 1); | ||
1673 | |||
1674 | /* Clear the reset flag */ | ||
1675 | pl330->dmac_tbd.reset_chan &= ~(1 << i); | ||
1676 | } | ||
1677 | } | ||
1678 | |||
1679 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1680 | |||
1681 | return; | ||
1682 | } | ||
1683 | |||
1684 | /* Returns 1 if state was updated, 0 otherwise */ | ||
1685 | static int pl330_update(const struct pl330_info *pi) | ||
1686 | { | ||
1687 | struct _pl330_req *rqdone; | ||
1688 | struct pl330_dmac *pl330; | ||
1689 | unsigned long flags; | ||
1690 | void __iomem *regs; | ||
1691 | u32 val; | ||
1692 | int id, ev, ret = 0; | ||
1693 | |||
1694 | if (!pi || !pi->pl330_data) | ||
1695 | return 0; | ||
1696 | |||
1697 | regs = pi->base; | ||
1698 | pl330 = pi->pl330_data; | ||
1699 | |||
1700 | spin_lock_irqsave(&pl330->lock, flags); | ||
1701 | |||
1702 | val = readl(regs + FSM) & 0x1; | ||
1703 | if (val) | ||
1704 | pl330->dmac_tbd.reset_mngr = true; | ||
1705 | else | ||
1706 | pl330->dmac_tbd.reset_mngr = false; | ||
1707 | |||
1708 | val = readl(regs + FSC) & ((1 << pi->pcfg.num_chan) - 1); | ||
1709 | pl330->dmac_tbd.reset_chan |= val; | ||
1710 | if (val) { | ||
1711 | int i = 0; | ||
1712 | while (i < pi->pcfg.num_chan) { | ||
1713 | if (val & (1 << i)) { | ||
1714 | dev_info(pi->dev, | ||
1715 | "Reset Channel-%d\t CS-%x FTC-%x\n", | ||
1716 | i, readl(regs + CS(i)), | ||
1717 | readl(regs + FTC(i))); | ||
1718 | _stop(&pl330->channels[i]); | ||
1719 | } | ||
1720 | i++; | ||
1721 | } | ||
1722 | } | ||
1723 | |||
1724 | /* Check which event happened i.e, thread notified */ | ||
1725 | val = readl(regs + ES); | ||
1726 | if (pi->pcfg.num_events < 32 | ||
1727 | && val & ~((1 << pi->pcfg.num_events) - 1)) { | ||
1728 | pl330->dmac_tbd.reset_dmac = true; | ||
1729 | dev_err(pi->dev, "%s:%d Unexpected!\n", __func__, __LINE__); | ||
1730 | ret = 1; | ||
1731 | goto updt_exit; | ||
1732 | } | ||
1733 | |||
1734 | for (ev = 0; ev < pi->pcfg.num_events; ev++) { | ||
1735 | if (val & (1 << ev)) { /* Event occurred */ | ||
1736 | struct pl330_thread *thrd; | ||
1737 | u32 inten = readl(regs + INTEN); | ||
1738 | int active; | ||
1739 | |||
1740 | /* Clear the event */ | ||
1741 | if (inten & (1 << ev)) | ||
1742 | writel(1 << ev, regs + INTCLR); | ||
1743 | |||
1744 | ret = 1; | ||
1745 | |||
1746 | id = pl330->events[ev]; | ||
1747 | |||
1748 | thrd = &pl330->channels[id]; | ||
1749 | |||
1750 | active = thrd->req_running; | ||
1751 | if (active == -1) /* Aborted */ | ||
1752 | continue; | ||
1753 | |||
1754 | rqdone = &thrd->req[active]; | ||
1755 | mark_free(thrd, active); | ||
1756 | |||
1757 | /* Get going again ASAP */ | ||
1758 | _start(thrd); | ||
1759 | |||
1760 | /* For now, just make a list of callbacks to be done */ | ||
1761 | list_add_tail(&rqdone->rqd, &pl330->req_done); | ||
1762 | } | ||
1763 | } | ||
1764 | |||
1765 | /* Now that we are in no hurry, do the callbacks */ | ||
1766 | while (!list_empty(&pl330->req_done)) { | ||
1767 | struct pl330_req *r; | ||
1768 | |||
1769 | rqdone = container_of(pl330->req_done.next, | ||
1770 | struct _pl330_req, rqd); | ||
1771 | |||
1772 | list_del_init(&rqdone->rqd); | ||
1773 | |||
1774 | /* Detach the req */ | ||
1775 | r = rqdone->r; | ||
1776 | rqdone->r = NULL; | ||
1777 | |||
1778 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1779 | _callback(r, PL330_ERR_NONE); | ||
1780 | spin_lock_irqsave(&pl330->lock, flags); | ||
1781 | } | ||
1782 | |||
1783 | updt_exit: | ||
1784 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1785 | |||
1786 | if (pl330->dmac_tbd.reset_dmac | ||
1787 | || pl330->dmac_tbd.reset_mngr | ||
1788 | || pl330->dmac_tbd.reset_chan) { | ||
1789 | ret = 1; | ||
1790 | tasklet_schedule(&pl330->tasks); | ||
1791 | } | ||
1792 | |||
1793 | return ret; | ||
1794 | } | ||
1795 | |||
1796 | static int pl330_chan_ctrl(void *ch_id, enum pl330_chan_op op) | ||
1797 | { | ||
1798 | struct pl330_thread *thrd = ch_id; | ||
1799 | struct pl330_dmac *pl330; | ||
1800 | unsigned long flags; | ||
1801 | int ret = 0, active; | ||
1802 | |||
1803 | if (!thrd || thrd->free || thrd->dmac->state == DYING) | ||
1804 | return -EINVAL; | ||
1805 | |||
1806 | pl330 = thrd->dmac; | ||
1807 | active = thrd->req_running; | ||
1808 | |||
1809 | spin_lock_irqsave(&pl330->lock, flags); | ||
1810 | |||
1811 | switch (op) { | ||
1812 | case PL330_OP_FLUSH: | ||
1813 | /* Make sure the channel is stopped */ | ||
1814 | _stop(thrd); | ||
1815 | |||
1816 | thrd->req[0].r = NULL; | ||
1817 | thrd->req[1].r = NULL; | ||
1818 | mark_free(thrd, 0); | ||
1819 | mark_free(thrd, 1); | ||
1820 | break; | ||
1821 | |||
1822 | case PL330_OP_ABORT: | ||
1823 | /* Make sure the channel is stopped */ | ||
1824 | _stop(thrd); | ||
1825 | |||
1826 | /* ABORT is only for the active req */ | ||
1827 | if (active == -1) | ||
1828 | break; | ||
1829 | |||
1830 | thrd->req[active].r = NULL; | ||
1831 | mark_free(thrd, active); | ||
1832 | |||
1833 | /* Start the next */ | ||
1834 | case PL330_OP_START: | ||
1835 | if ((active == -1) && !_start(thrd)) | ||
1836 | ret = -EIO; | ||
1837 | break; | ||
1838 | |||
1839 | default: | ||
1840 | ret = -EINVAL; | ||
1841 | } | ||
1842 | |||
1843 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1844 | return ret; | ||
1845 | } | ||
1846 | |||
1847 | /* Reserve an event */ | ||
1848 | static inline int _alloc_event(struct pl330_thread *thrd) | ||
1849 | { | ||
1850 | struct pl330_dmac *pl330 = thrd->dmac; | ||
1851 | struct pl330_info *pi = pl330->pinfo; | ||
1852 | int ev; | ||
1853 | |||
1854 | for (ev = 0; ev < pi->pcfg.num_events; ev++) | ||
1855 | if (pl330->events[ev] == -1) { | ||
1856 | pl330->events[ev] = thrd->id; | ||
1857 | return ev; | ||
1858 | } | ||
1859 | |||
1860 | return -1; | ||
1861 | } | ||
1862 | |||
1863 | static bool _chan_ns(const struct pl330_info *pi, int i) | ||
1864 | { | ||
1865 | return pi->pcfg.irq_ns & (1 << i); | ||
1866 | } | ||
1867 | |||
1868 | /* Upon success, returns IdentityToken for the | ||
1869 | * allocated channel, NULL otherwise. | ||
1870 | */ | ||
1871 | static void *pl330_request_channel(const struct pl330_info *pi) | ||
1872 | { | ||
1873 | struct pl330_thread *thrd = NULL; | ||
1874 | struct pl330_dmac *pl330; | ||
1875 | unsigned long flags; | ||
1876 | int chans, i; | ||
1877 | |||
1878 | if (!pi || !pi->pl330_data) | ||
1879 | return NULL; | ||
1880 | |||
1881 | pl330 = pi->pl330_data; | ||
1882 | |||
1883 | if (pl330->state == DYING) | ||
1884 | return NULL; | ||
1885 | |||
1886 | chans = pi->pcfg.num_chan; | ||
1887 | |||
1888 | spin_lock_irqsave(&pl330->lock, flags); | ||
1889 | |||
1890 | for (i = 0; i < chans; i++) { | ||
1891 | thrd = &pl330->channels[i]; | ||
1892 | if ((thrd->free) && (!_manager_ns(thrd) || | ||
1893 | _chan_ns(pi, i))) { | ||
1894 | thrd->ev = _alloc_event(thrd); | ||
1895 | if (thrd->ev >= 0) { | ||
1896 | thrd->free = false; | ||
1897 | thrd->lstenq = 1; | ||
1898 | thrd->req[0].r = NULL; | ||
1899 | mark_free(thrd, 0); | ||
1900 | thrd->req[1].r = NULL; | ||
1901 | mark_free(thrd, 1); | ||
1902 | break; | ||
1903 | } | ||
1904 | } | ||
1905 | thrd = NULL; | ||
1906 | } | ||
1907 | |||
1908 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1909 | |||
1910 | return thrd; | ||
1911 | } | ||
1912 | |||
1913 | /* Release an event */ | ||
1914 | static inline void _free_event(struct pl330_thread *thrd, int ev) | ||
1915 | { | ||
1916 | struct pl330_dmac *pl330 = thrd->dmac; | ||
1917 | struct pl330_info *pi = pl330->pinfo; | ||
1918 | |||
1919 | /* If the event is valid and was held by the thread */ | ||
1920 | if (ev >= 0 && ev < pi->pcfg.num_events | ||
1921 | && pl330->events[ev] == thrd->id) | ||
1922 | pl330->events[ev] = -1; | ||
1923 | } | ||
1924 | |||
1925 | static void pl330_release_channel(void *ch_id) | ||
1926 | { | ||
1927 | struct pl330_thread *thrd = ch_id; | ||
1928 | struct pl330_dmac *pl330; | ||
1929 | unsigned long flags; | ||
1930 | |||
1931 | if (!thrd || thrd->free) | ||
1932 | return; | ||
1933 | |||
1934 | _stop(thrd); | ||
1935 | |||
1936 | _callback(thrd->req[1 - thrd->lstenq].r, PL330_ERR_ABORT); | ||
1937 | _callback(thrd->req[thrd->lstenq].r, PL330_ERR_ABORT); | ||
1938 | |||
1939 | pl330 = thrd->dmac; | ||
1940 | |||
1941 | spin_lock_irqsave(&pl330->lock, flags); | ||
1942 | _free_event(thrd, thrd->ev); | ||
1943 | thrd->free = true; | ||
1944 | spin_unlock_irqrestore(&pl330->lock, flags); | ||
1945 | } | ||
1946 | |||
1947 | /* Initialize the structure for PL330 configuration, that can be used | ||
1948 | * by the client driver the make best use of the DMAC | ||
1949 | */ | ||
1950 | static void read_dmac_config(struct pl330_info *pi) | ||
1951 | { | ||
1952 | void __iomem *regs = pi->base; | ||
1953 | u32 val; | ||
1954 | |||
1955 | val = readl(regs + CRD) >> CRD_DATA_WIDTH_SHIFT; | ||
1956 | val &= CRD_DATA_WIDTH_MASK; | ||
1957 | pi->pcfg.data_bus_width = 8 * (1 << val); | ||
1958 | |||
1959 | val = readl(regs + CRD) >> CRD_DATA_BUFF_SHIFT; | ||
1960 | val &= CRD_DATA_BUFF_MASK; | ||
1961 | pi->pcfg.data_buf_dep = val + 1; | ||
1962 | |||
1963 | val = readl(regs + CR0) >> CR0_NUM_CHANS_SHIFT; | ||
1964 | val &= CR0_NUM_CHANS_MASK; | ||
1965 | val += 1; | ||
1966 | pi->pcfg.num_chan = val; | ||
1967 | |||
1968 | val = readl(regs + CR0); | ||
1969 | if (val & CR0_PERIPH_REQ_SET) { | ||
1970 | val = (val >> CR0_NUM_PERIPH_SHIFT) & CR0_NUM_PERIPH_MASK; | ||
1971 | val += 1; | ||
1972 | pi->pcfg.num_peri = val; | ||
1973 | pi->pcfg.peri_ns = readl(regs + CR4); | ||
1974 | } else { | ||
1975 | pi->pcfg.num_peri = 0; | ||
1976 | } | ||
1977 | |||
1978 | val = readl(regs + CR0); | ||
1979 | if (val & CR0_BOOT_MAN_NS) | ||
1980 | pi->pcfg.mode |= DMAC_MODE_NS; | ||
1981 | else | ||
1982 | pi->pcfg.mode &= ~DMAC_MODE_NS; | ||
1983 | |||
1984 | val = readl(regs + CR0) >> CR0_NUM_EVENTS_SHIFT; | ||
1985 | val &= CR0_NUM_EVENTS_MASK; | ||
1986 | val += 1; | ||
1987 | pi->pcfg.num_events = val; | ||
1988 | |||
1989 | pi->pcfg.irq_ns = readl(regs + CR3); | ||
1990 | |||
1991 | pi->pcfg.periph_id = get_id(pi, PERIPH_ID); | ||
1992 | pi->pcfg.pcell_id = get_id(pi, PCELL_ID); | ||
1993 | } | ||
1994 | |||
1995 | static inline void _reset_thread(struct pl330_thread *thrd) | ||
1996 | { | ||
1997 | struct pl330_dmac *pl330 = thrd->dmac; | ||
1998 | struct pl330_info *pi = pl330->pinfo; | ||
1999 | |||
2000 | thrd->req[0].mc_cpu = pl330->mcode_cpu | ||
2001 | + (thrd->id * pi->mcbufsz); | ||
2002 | thrd->req[0].mc_bus = pl330->mcode_bus | ||
2003 | + (thrd->id * pi->mcbufsz); | ||
2004 | thrd->req[0].r = NULL; | ||
2005 | mark_free(thrd, 0); | ||
2006 | |||
2007 | thrd->req[1].mc_cpu = thrd->req[0].mc_cpu | ||
2008 | + pi->mcbufsz / 2; | ||
2009 | thrd->req[1].mc_bus = thrd->req[0].mc_bus | ||
2010 | + pi->mcbufsz / 2; | ||
2011 | thrd->req[1].r = NULL; | ||
2012 | mark_free(thrd, 1); | ||
2013 | } | ||
2014 | |||
2015 | static int dmac_alloc_threads(struct pl330_dmac *pl330) | ||
2016 | { | ||
2017 | struct pl330_info *pi = pl330->pinfo; | ||
2018 | int chans = pi->pcfg.num_chan; | ||
2019 | struct pl330_thread *thrd; | ||
2020 | int i; | ||
2021 | |||
2022 | /* Allocate 1 Manager and 'chans' Channel threads */ | ||
2023 | pl330->channels = kzalloc((1 + chans) * sizeof(*thrd), | ||
2024 | GFP_KERNEL); | ||
2025 | if (!pl330->channels) | ||
2026 | return -ENOMEM; | ||
2027 | |||
2028 | /* Init Channel threads */ | ||
2029 | for (i = 0; i < chans; i++) { | ||
2030 | thrd = &pl330->channels[i]; | ||
2031 | thrd->id = i; | ||
2032 | thrd->dmac = pl330; | ||
2033 | _reset_thread(thrd); | ||
2034 | thrd->free = true; | ||
2035 | } | ||
2036 | |||
2037 | /* MANAGER is indexed at the end */ | ||
2038 | thrd = &pl330->channels[chans]; | ||
2039 | thrd->id = chans; | ||
2040 | thrd->dmac = pl330; | ||
2041 | thrd->free = false; | ||
2042 | pl330->manager = thrd; | ||
2043 | |||
2044 | return 0; | ||
2045 | } | ||
2046 | |||
2047 | static int dmac_alloc_resources(struct pl330_dmac *pl330) | ||
2048 | { | ||
2049 | struct pl330_info *pi = pl330->pinfo; | ||
2050 | int chans = pi->pcfg.num_chan; | ||
2051 | int ret; | ||
2052 | |||
2053 | /* | ||
2054 | * Alloc MicroCode buffer for 'chans' Channel threads. | ||
2055 | * A channel's buffer offset is (Channel_Id * MCODE_BUFF_PERCHAN) | ||
2056 | */ | ||
2057 | pl330->mcode_cpu = dma_alloc_coherent(pi->dev, | ||
2058 | chans * pi->mcbufsz, | ||
2059 | &pl330->mcode_bus, GFP_KERNEL); | ||
2060 | if (!pl330->mcode_cpu) { | ||
2061 | dev_err(pi->dev, "%s:%d Can't allocate memory!\n", | ||
2062 | __func__, __LINE__); | ||
2063 | return -ENOMEM; | ||
2064 | } | ||
2065 | |||
2066 | ret = dmac_alloc_threads(pl330); | ||
2067 | if (ret) { | ||
2068 | dev_err(pi->dev, "%s:%d Can't to create channels for DMAC!\n", | ||
2069 | __func__, __LINE__); | ||
2070 | dma_free_coherent(pi->dev, | ||
2071 | chans * pi->mcbufsz, | ||
2072 | pl330->mcode_cpu, pl330->mcode_bus); | ||
2073 | return ret; | ||
2074 | } | ||
2075 | |||
2076 | return 0; | ||
2077 | } | ||
2078 | |||
2079 | static int pl330_add(struct pl330_info *pi) | ||
2080 | { | ||
2081 | struct pl330_dmac *pl330; | ||
2082 | void __iomem *regs; | ||
2083 | int i, ret; | ||
2084 | |||
2085 | if (!pi || !pi->dev) | ||
2086 | return -EINVAL; | ||
2087 | |||
2088 | /* If already added */ | ||
2089 | if (pi->pl330_data) | ||
2090 | return -EINVAL; | ||
2091 | |||
2092 | /* | ||
2093 | * If the SoC can perform reset on the DMAC, then do it | ||
2094 | * before reading its configuration. | ||
2095 | */ | ||
2096 | if (pi->dmac_reset) | ||
2097 | pi->dmac_reset(pi); | ||
2098 | |||
2099 | regs = pi->base; | ||
2100 | |||
2101 | /* Check if we can handle this DMAC */ | ||
2102 | if ((get_id(pi, PERIPH_ID) & 0xfffff) != PERIPH_ID_VAL | ||
2103 | || get_id(pi, PCELL_ID) != PCELL_ID_VAL) { | ||
2104 | dev_err(pi->dev, "PERIPH_ID 0x%x, PCELL_ID 0x%x !\n", | ||
2105 | get_id(pi, PERIPH_ID), get_id(pi, PCELL_ID)); | ||
2106 | return -EINVAL; | ||
2107 | } | ||
2108 | |||
2109 | /* Read the configuration of the DMAC */ | ||
2110 | read_dmac_config(pi); | ||
2111 | |||
2112 | if (pi->pcfg.num_events == 0) { | ||
2113 | dev_err(pi->dev, "%s:%d Can't work without events!\n", | ||
2114 | __func__, __LINE__); | ||
2115 | return -EINVAL; | ||
2116 | } | ||
2117 | |||
2118 | pl330 = kzalloc(sizeof(*pl330), GFP_KERNEL); | ||
2119 | if (!pl330) { | ||
2120 | dev_err(pi->dev, "%s:%d Can't allocate memory!\n", | ||
2121 | __func__, __LINE__); | ||
2122 | return -ENOMEM; | ||
2123 | } | ||
2124 | |||
2125 | /* Assign the info structure and private data */ | ||
2126 | pl330->pinfo = pi; | ||
2127 | pi->pl330_data = pl330; | ||
2128 | |||
2129 | spin_lock_init(&pl330->lock); | ||
2130 | |||
2131 | INIT_LIST_HEAD(&pl330->req_done); | ||
2132 | |||
2133 | /* Use default MC buffer size if not provided */ | ||
2134 | if (!pi->mcbufsz) | ||
2135 | pi->mcbufsz = MCODE_BUFF_PER_REQ * 2; | ||
2136 | |||
2137 | /* Mark all events as free */ | ||
2138 | for (i = 0; i < pi->pcfg.num_events; i++) | ||
2139 | pl330->events[i] = -1; | ||
2140 | |||
2141 | /* Allocate resources needed by the DMAC */ | ||
2142 | ret = dmac_alloc_resources(pl330); | ||
2143 | if (ret) { | ||
2144 | dev_err(pi->dev, "Unable to create channels for DMAC\n"); | ||
2145 | kfree(pl330); | ||
2146 | return ret; | ||
2147 | } | ||
2148 | |||
2149 | tasklet_init(&pl330->tasks, pl330_dotask, (unsigned long) pl330); | ||
2150 | |||
2151 | pl330->state = INIT; | ||
2152 | |||
2153 | return 0; | ||
2154 | } | ||
2155 | |||
2156 | static int dmac_free_threads(struct pl330_dmac *pl330) | ||
2157 | { | ||
2158 | struct pl330_info *pi = pl330->pinfo; | ||
2159 | int chans = pi->pcfg.num_chan; | ||
2160 | struct pl330_thread *thrd; | ||
2161 | int i; | ||
2162 | |||
2163 | /* Release Channel threads */ | ||
2164 | for (i = 0; i < chans; i++) { | ||
2165 | thrd = &pl330->channels[i]; | ||
2166 | pl330_release_channel((void *)thrd); | ||
2167 | } | ||
2168 | |||
2169 | /* Free memory */ | ||
2170 | kfree(pl330->channels); | ||
2171 | |||
2172 | return 0; | ||
2173 | } | ||
2174 | |||
2175 | static void dmac_free_resources(struct pl330_dmac *pl330) | ||
2176 | { | ||
2177 | struct pl330_info *pi = pl330->pinfo; | ||
2178 | int chans = pi->pcfg.num_chan; | ||
2179 | |||
2180 | dmac_free_threads(pl330); | ||
2181 | |||
2182 | dma_free_coherent(pi->dev, chans * pi->mcbufsz, | ||
2183 | pl330->mcode_cpu, pl330->mcode_bus); | ||
2184 | } | ||
2185 | |||
2186 | static void pl330_del(struct pl330_info *pi) | ||
2187 | { | ||
2188 | struct pl330_dmac *pl330; | ||
2189 | |||
2190 | if (!pi || !pi->pl330_data) | ||
2191 | return; | ||
2192 | |||
2193 | pl330 = pi->pl330_data; | ||
2194 | |||
2195 | pl330->state = UNINIT; | ||
2196 | |||
2197 | tasklet_kill(&pl330->tasks); | ||
2198 | |||
2199 | /* Free DMAC resources */ | ||
2200 | dmac_free_resources(pl330); | ||
2201 | |||
2202 | kfree(pl330); | ||
2203 | pi->pl330_data = NULL; | ||
2204 | } | ||
2205 | |||
120 | /* forward declaration */ | 2206 | /* forward declaration */ |
121 | static struct amba_driver pl330_driver; | 2207 | static struct amba_driver pl330_driver; |
122 | 2208 | ||
@@ -234,7 +2320,7 @@ static void pl330_tasklet(unsigned long data) | |||
234 | /* Pick up ripe tomatoes */ | 2320 | /* Pick up ripe tomatoes */ |
235 | list_for_each_entry_safe(desc, _dt, &pch->work_list, node) | 2321 | list_for_each_entry_safe(desc, _dt, &pch->work_list, node) |
236 | if (desc->status == DONE) { | 2322 | if (desc->status == DONE) { |
237 | pch->completed = desc->txd.cookie; | 2323 | dma_cookie_complete(&desc->txd); |
238 | list_move_tail(&desc->node, &list); | 2324 | list_move_tail(&desc->node, &list); |
239 | } | 2325 | } |
240 | 2326 | ||
@@ -305,7 +2391,7 @@ static int pl330_alloc_chan_resources(struct dma_chan *chan) | |||
305 | 2391 | ||
306 | spin_lock_irqsave(&pch->lock, flags); | 2392 | spin_lock_irqsave(&pch->lock, flags); |
307 | 2393 | ||
308 | pch->completed = chan->cookie = 1; | 2394 | dma_cookie_init(chan); |
309 | pch->cyclic = false; | 2395 | pch->cyclic = false; |
310 | 2396 | ||
311 | pch->pl330_chid = pl330_request_channel(&pdmac->pif); | 2397 | pch->pl330_chid = pl330_request_channel(&pdmac->pif); |
@@ -340,7 +2426,6 @@ static int pl330_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, unsigned | |||
340 | /* Mark all desc done */ | 2426 | /* Mark all desc done */ |
341 | list_for_each_entry_safe(desc, _dt, &pch->work_list , node) { | 2427 | list_for_each_entry_safe(desc, _dt, &pch->work_list , node) { |
342 | desc->status = DONE; | 2428 | desc->status = DONE; |
343 | pch->completed = desc->txd.cookie; | ||
344 | list_move_tail(&desc->node, &list); | 2429 | list_move_tail(&desc->node, &list); |
345 | } | 2430 | } |
346 | 2431 | ||
@@ -396,18 +2481,7 @@ static enum dma_status | |||
396 | pl330_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | 2481 | pl330_tx_status(struct dma_chan *chan, dma_cookie_t cookie, |
397 | struct dma_tx_state *txstate) | 2482 | struct dma_tx_state *txstate) |
398 | { | 2483 | { |
399 | struct dma_pl330_chan *pch = to_pchan(chan); | 2484 | return dma_cookie_status(chan, cookie, txstate); |
400 | dma_cookie_t last_done, last_used; | ||
401 | int ret; | ||
402 | |||
403 | last_done = pch->completed; | ||
404 | last_used = chan->cookie; | ||
405 | |||
406 | ret = dma_async_is_complete(cookie, last_done, last_used); | ||
407 | |||
408 | dma_set_tx_state(txstate, last_done, last_used, 0); | ||
409 | |||
410 | return ret; | ||
411 | } | 2485 | } |
412 | 2486 | ||
413 | static void pl330_issue_pending(struct dma_chan *chan) | 2487 | static void pl330_issue_pending(struct dma_chan *chan) |
@@ -430,26 +2504,16 @@ static dma_cookie_t pl330_tx_submit(struct dma_async_tx_descriptor *tx) | |||
430 | spin_lock_irqsave(&pch->lock, flags); | 2504 | spin_lock_irqsave(&pch->lock, flags); |
431 | 2505 | ||
432 | /* Assign cookies to all nodes */ | 2506 | /* Assign cookies to all nodes */ |
433 | cookie = tx->chan->cookie; | ||
434 | |||
435 | while (!list_empty(&last->node)) { | 2507 | while (!list_empty(&last->node)) { |
436 | desc = list_entry(last->node.next, struct dma_pl330_desc, node); | 2508 | desc = list_entry(last->node.next, struct dma_pl330_desc, node); |
437 | 2509 | ||
438 | if (++cookie < 0) | 2510 | dma_cookie_assign(&desc->txd); |
439 | cookie = 1; | ||
440 | desc->txd.cookie = cookie; | ||
441 | 2511 | ||
442 | list_move_tail(&desc->node, &pch->work_list); | 2512 | list_move_tail(&desc->node, &pch->work_list); |
443 | } | 2513 | } |
444 | 2514 | ||
445 | if (++cookie < 0) | 2515 | cookie = dma_cookie_assign(&last->txd); |
446 | cookie = 1; | ||
447 | last->txd.cookie = cookie; | ||
448 | |||
449 | list_add_tail(&last->node, &pch->work_list); | 2516 | list_add_tail(&last->node, &pch->work_list); |
450 | |||
451 | tx->chan->cookie = cookie; | ||
452 | |||
453 | spin_unlock_irqrestore(&pch->lock, flags); | 2517 | spin_unlock_irqrestore(&pch->lock, flags); |
454 | 2518 | ||
455 | return cookie; | 2519 | return cookie; |
@@ -553,6 +2617,7 @@ static struct dma_pl330_desc *pl330_get_desc(struct dma_pl330_chan *pch) | |||
553 | async_tx_ack(&desc->txd); | 2617 | async_tx_ack(&desc->txd); |
554 | 2618 | ||
555 | desc->req.peri = peri_id ? pch->chan.chan_id : 0; | 2619 | desc->req.peri = peri_id ? pch->chan.chan_id : 0; |
2620 | desc->rqcfg.pcfg = &pch->dmac->pif.pcfg; | ||
556 | 2621 | ||
557 | dma_async_tx_descriptor_init(&desc->txd, &pch->chan); | 2622 | dma_async_tx_descriptor_init(&desc->txd, &pch->chan); |
558 | 2623 | ||
@@ -621,7 +2686,8 @@ static inline int get_burst_len(struct dma_pl330_desc *desc, size_t len) | |||
621 | 2686 | ||
622 | static struct dma_async_tx_descriptor *pl330_prep_dma_cyclic( | 2687 | static struct dma_async_tx_descriptor *pl330_prep_dma_cyclic( |
623 | struct dma_chan *chan, dma_addr_t dma_addr, size_t len, | 2688 | struct dma_chan *chan, dma_addr_t dma_addr, size_t len, |
624 | size_t period_len, enum dma_transfer_direction direction) | 2689 | size_t period_len, enum dma_transfer_direction direction, |
2690 | void *context) | ||
625 | { | 2691 | { |
626 | struct dma_pl330_desc *desc; | 2692 | struct dma_pl330_desc *desc; |
627 | struct dma_pl330_chan *pch = to_pchan(chan); | 2693 | struct dma_pl330_chan *pch = to_pchan(chan); |
@@ -711,7 +2777,7 @@ pl330_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dst, | |||
711 | static struct dma_async_tx_descriptor * | 2777 | static struct dma_async_tx_descriptor * |
712 | pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 2778 | pl330_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
713 | unsigned int sg_len, enum dma_transfer_direction direction, | 2779 | unsigned int sg_len, enum dma_transfer_direction direction, |
714 | unsigned long flg) | 2780 | unsigned long flg, void *context) |
715 | { | 2781 | { |
716 | struct dma_pl330_desc *first, *desc = NULL; | 2782 | struct dma_pl330_desc *first, *desc = NULL; |
717 | struct dma_pl330_chan *pch = to_pchan(chan); | 2783 | struct dma_pl330_chan *pch = to_pchan(chan); |
@@ -829,7 +2895,7 @@ pl330_probe(struct amba_device *adev, const struct amba_id *id) | |||
829 | if (IS_ERR(pdmac->clk)) { | 2895 | if (IS_ERR(pdmac->clk)) { |
830 | dev_err(&adev->dev, "Cannot get operation clock.\n"); | 2896 | dev_err(&adev->dev, "Cannot get operation clock.\n"); |
831 | ret = -EINVAL; | 2897 | ret = -EINVAL; |
832 | goto probe_err1; | 2898 | goto probe_err2; |
833 | } | 2899 | } |
834 | 2900 | ||
835 | amba_set_drvdata(adev, pdmac); | 2901 | amba_set_drvdata(adev, pdmac); |
@@ -843,11 +2909,11 @@ pl330_probe(struct amba_device *adev, const struct amba_id *id) | |||
843 | ret = request_irq(irq, pl330_irq_handler, 0, | 2909 | ret = request_irq(irq, pl330_irq_handler, 0, |
844 | dev_name(&adev->dev), pi); | 2910 | dev_name(&adev->dev), pi); |
845 | if (ret) | 2911 | if (ret) |
846 | goto probe_err2; | 2912 | goto probe_err3; |
847 | 2913 | ||
848 | ret = pl330_add(pi); | 2914 | ret = pl330_add(pi); |
849 | if (ret) | 2915 | if (ret) |
850 | goto probe_err3; | 2916 | goto probe_err4; |
851 | 2917 | ||
852 | INIT_LIST_HEAD(&pdmac->desc_pool); | 2918 | INIT_LIST_HEAD(&pdmac->desc_pool); |
853 | spin_lock_init(&pdmac->pool_lock); | 2919 | spin_lock_init(&pdmac->pool_lock); |
@@ -904,7 +2970,7 @@ pl330_probe(struct amba_device *adev, const struct amba_id *id) | |||
904 | ret = dma_async_device_register(pd); | 2970 | ret = dma_async_device_register(pd); |
905 | if (ret) { | 2971 | if (ret) { |
906 | dev_err(&adev->dev, "unable to register DMAC\n"); | 2972 | dev_err(&adev->dev, "unable to register DMAC\n"); |
907 | goto probe_err4; | 2973 | goto probe_err5; |
908 | } | 2974 | } |
909 | 2975 | ||
910 | dev_info(&adev->dev, | 2976 | dev_info(&adev->dev, |
@@ -917,10 +2983,15 @@ pl330_probe(struct amba_device *adev, const struct amba_id *id) | |||
917 | 2983 | ||
918 | return 0; | 2984 | return 0; |
919 | 2985 | ||
920 | probe_err4: | 2986 | probe_err5: |
921 | pl330_del(pi); | 2987 | pl330_del(pi); |
922 | probe_err3: | 2988 | probe_err4: |
923 | free_irq(irq, pi); | 2989 | free_irq(irq, pi); |
2990 | probe_err3: | ||
2991 | #ifndef CONFIG_PM_RUNTIME | ||
2992 | clk_disable(pdmac->clk); | ||
2993 | #endif | ||
2994 | clk_put(pdmac->clk); | ||
924 | probe_err2: | 2995 | probe_err2: |
925 | iounmap(pi->base); | 2996 | iounmap(pi->base); |
926 | probe_err1: | 2997 | probe_err1: |
diff --git a/drivers/dma/ppc4xx/adma.c b/drivers/dma/ppc4xx/adma.c index fc457a7e8832..ced98826684a 100644 --- a/drivers/dma/ppc4xx/adma.c +++ b/drivers/dma/ppc4xx/adma.c | |||
@@ -46,6 +46,7 @@ | |||
46 | #include <asm/dcr.h> | 46 | #include <asm/dcr.h> |
47 | #include <asm/dcr-regs.h> | 47 | #include <asm/dcr-regs.h> |
48 | #include "adma.h" | 48 | #include "adma.h" |
49 | #include "../dmaengine.h" | ||
49 | 50 | ||
50 | enum ppc_adma_init_code { | 51 | enum ppc_adma_init_code { |
51 | PPC_ADMA_INIT_OK = 0, | 52 | PPC_ADMA_INIT_OK = 0, |
@@ -1930,7 +1931,7 @@ static void __ppc440spe_adma_slot_cleanup(struct ppc440spe_adma_chan *chan) | |||
1930 | if (end_of_chain && slot_cnt) { | 1931 | if (end_of_chain && slot_cnt) { |
1931 | /* Should wait for ZeroSum completion */ | 1932 | /* Should wait for ZeroSum completion */ |
1932 | if (cookie > 0) | 1933 | if (cookie > 0) |
1933 | chan->completed_cookie = cookie; | 1934 | chan->common.completed_cookie = cookie; |
1934 | return; | 1935 | return; |
1935 | } | 1936 | } |
1936 | 1937 | ||
@@ -1960,7 +1961,7 @@ static void __ppc440spe_adma_slot_cleanup(struct ppc440spe_adma_chan *chan) | |||
1960 | BUG_ON(!seen_current); | 1961 | BUG_ON(!seen_current); |
1961 | 1962 | ||
1962 | if (cookie > 0) { | 1963 | if (cookie > 0) { |
1963 | chan->completed_cookie = cookie; | 1964 | chan->common.completed_cookie = cookie; |
1964 | pr_debug("\tcompleted cookie %d\n", cookie); | 1965 | pr_debug("\tcompleted cookie %d\n", cookie); |
1965 | } | 1966 | } |
1966 | 1967 | ||
@@ -2150,22 +2151,6 @@ static int ppc440spe_adma_alloc_chan_resources(struct dma_chan *chan) | |||
2150 | } | 2151 | } |
2151 | 2152 | ||
2152 | /** | 2153 | /** |
2153 | * ppc440spe_desc_assign_cookie - assign a cookie | ||
2154 | */ | ||
2155 | static dma_cookie_t ppc440spe_desc_assign_cookie( | ||
2156 | struct ppc440spe_adma_chan *chan, | ||
2157 | struct ppc440spe_adma_desc_slot *desc) | ||
2158 | { | ||
2159 | dma_cookie_t cookie = chan->common.cookie; | ||
2160 | |||
2161 | cookie++; | ||
2162 | if (cookie < 0) | ||
2163 | cookie = 1; | ||
2164 | chan->common.cookie = desc->async_tx.cookie = cookie; | ||
2165 | return cookie; | ||
2166 | } | ||
2167 | |||
2168 | /** | ||
2169 | * ppc440spe_rxor_set_region_data - | 2154 | * ppc440spe_rxor_set_region_data - |
2170 | */ | 2155 | */ |
2171 | static void ppc440spe_rxor_set_region(struct ppc440spe_adma_desc_slot *desc, | 2156 | static void ppc440spe_rxor_set_region(struct ppc440spe_adma_desc_slot *desc, |
@@ -2235,8 +2220,7 @@ static dma_cookie_t ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx) | |||
2235 | slots_per_op = group_start->slots_per_op; | 2220 | slots_per_op = group_start->slots_per_op; |
2236 | 2221 | ||
2237 | spin_lock_bh(&chan->lock); | 2222 | spin_lock_bh(&chan->lock); |
2238 | 2223 | cookie = dma_cookie_assign(tx); | |
2239 | cookie = ppc440spe_desc_assign_cookie(chan, sw_desc); | ||
2240 | 2224 | ||
2241 | if (unlikely(list_empty(&chan->chain))) { | 2225 | if (unlikely(list_empty(&chan->chain))) { |
2242 | /* first peer */ | 2226 | /* first peer */ |
@@ -3944,28 +3928,16 @@ static enum dma_status ppc440spe_adma_tx_status(struct dma_chan *chan, | |||
3944 | dma_cookie_t cookie, struct dma_tx_state *txstate) | 3928 | dma_cookie_t cookie, struct dma_tx_state *txstate) |
3945 | { | 3929 | { |
3946 | struct ppc440spe_adma_chan *ppc440spe_chan; | 3930 | struct ppc440spe_adma_chan *ppc440spe_chan; |
3947 | dma_cookie_t last_used; | ||
3948 | dma_cookie_t last_complete; | ||
3949 | enum dma_status ret; | 3931 | enum dma_status ret; |
3950 | 3932 | ||
3951 | ppc440spe_chan = to_ppc440spe_adma_chan(chan); | 3933 | ppc440spe_chan = to_ppc440spe_adma_chan(chan); |
3952 | last_used = chan->cookie; | 3934 | ret = dma_cookie_status(chan, cookie, txstate); |
3953 | last_complete = ppc440spe_chan->completed_cookie; | ||
3954 | |||
3955 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
3956 | |||
3957 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
3958 | if (ret == DMA_SUCCESS) | 3935 | if (ret == DMA_SUCCESS) |
3959 | return ret; | 3936 | return ret; |
3960 | 3937 | ||
3961 | ppc440spe_adma_slot_cleanup(ppc440spe_chan); | 3938 | ppc440spe_adma_slot_cleanup(ppc440spe_chan); |
3962 | 3939 | ||
3963 | last_used = chan->cookie; | 3940 | return dma_cookie_status(chan, cookie, txstate); |
3964 | last_complete = ppc440spe_chan->completed_cookie; | ||
3965 | |||
3966 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
3967 | |||
3968 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
3969 | } | 3941 | } |
3970 | 3942 | ||
3971 | /** | 3943 | /** |
@@ -4050,16 +4022,12 @@ static void ppc440spe_chan_start_null_xor(struct ppc440spe_adma_chan *chan) | |||
4050 | async_tx_ack(&sw_desc->async_tx); | 4022 | async_tx_ack(&sw_desc->async_tx); |
4051 | ppc440spe_desc_init_null_xor(group_start); | 4023 | ppc440spe_desc_init_null_xor(group_start); |
4052 | 4024 | ||
4053 | cookie = chan->common.cookie; | 4025 | cookie = dma_cookie_assign(&sw_desc->async_tx); |
4054 | cookie++; | ||
4055 | if (cookie <= 1) | ||
4056 | cookie = 2; | ||
4057 | 4026 | ||
4058 | /* initialize the completed cookie to be less than | 4027 | /* initialize the completed cookie to be less than |
4059 | * the most recently used cookie | 4028 | * the most recently used cookie |
4060 | */ | 4029 | */ |
4061 | chan->completed_cookie = cookie - 1; | 4030 | chan->common.completed_cookie = cookie - 1; |
4062 | chan->common.cookie = sw_desc->async_tx.cookie = cookie; | ||
4063 | 4031 | ||
4064 | /* channel should not be busy */ | 4032 | /* channel should not be busy */ |
4065 | BUG_ON(ppc440spe_chan_is_busy(chan)); | 4033 | BUG_ON(ppc440spe_chan_is_busy(chan)); |
@@ -4529,6 +4497,7 @@ static int __devinit ppc440spe_adma_probe(struct platform_device *ofdev) | |||
4529 | INIT_LIST_HEAD(&chan->all_slots); | 4497 | INIT_LIST_HEAD(&chan->all_slots); |
4530 | chan->device = adev; | 4498 | chan->device = adev; |
4531 | chan->common.device = &adev->common; | 4499 | chan->common.device = &adev->common; |
4500 | dma_cookie_init(&chan->common); | ||
4532 | list_add_tail(&chan->common.device_node, &adev->common.channels); | 4501 | list_add_tail(&chan->common.device_node, &adev->common.channels); |
4533 | tasklet_init(&chan->irq_tasklet, ppc440spe_adma_tasklet, | 4502 | tasklet_init(&chan->irq_tasklet, ppc440spe_adma_tasklet, |
4534 | (unsigned long)chan); | 4503 | (unsigned long)chan); |
diff --git a/drivers/dma/ppc4xx/adma.h b/drivers/dma/ppc4xx/adma.h index 8ada5a812e3b..26b7a5ed9ac7 100644 --- a/drivers/dma/ppc4xx/adma.h +++ b/drivers/dma/ppc4xx/adma.h | |||
@@ -81,7 +81,6 @@ struct ppc440spe_adma_device { | |||
81 | * @common: common dmaengine channel object members | 81 | * @common: common dmaengine channel object members |
82 | * @all_slots: complete domain of slots usable by the channel | 82 | * @all_slots: complete domain of slots usable by the channel |
83 | * @pending: allows batching of hardware operations | 83 | * @pending: allows batching of hardware operations |
84 | * @completed_cookie: identifier for the most recently completed operation | ||
85 | * @slots_allocated: records the actual size of the descriptor slot pool | 84 | * @slots_allocated: records the actual size of the descriptor slot pool |
86 | * @hw_chain_inited: h/w descriptor chain initialization flag | 85 | * @hw_chain_inited: h/w descriptor chain initialization flag |
87 | * @irq_tasklet: bottom half where ppc440spe_adma_slot_cleanup runs | 86 | * @irq_tasklet: bottom half where ppc440spe_adma_slot_cleanup runs |
@@ -99,7 +98,6 @@ struct ppc440spe_adma_chan { | |||
99 | struct list_head all_slots; | 98 | struct list_head all_slots; |
100 | struct ppc440spe_adma_desc_slot *last_used; | 99 | struct ppc440spe_adma_desc_slot *last_used; |
101 | int pending; | 100 | int pending; |
102 | dma_cookie_t completed_cookie; | ||
103 | int slots_allocated; | 101 | int slots_allocated; |
104 | int hw_chain_inited; | 102 | int hw_chain_inited; |
105 | struct tasklet_struct irq_tasklet; | 103 | struct tasklet_struct irq_tasklet; |
diff --git a/drivers/dma/shdma.c b/drivers/dma/shdma.c index 812fd76e9c18..19d7a8d3975d 100644 --- a/drivers/dma/shdma.c +++ b/drivers/dma/shdma.c | |||
@@ -30,6 +30,8 @@ | |||
30 | #include <linux/kdebug.h> | 30 | #include <linux/kdebug.h> |
31 | #include <linux/spinlock.h> | 31 | #include <linux/spinlock.h> |
32 | #include <linux/rculist.h> | 32 | #include <linux/rculist.h> |
33 | |||
34 | #include "dmaengine.h" | ||
33 | #include "shdma.h" | 35 | #include "shdma.h" |
34 | 36 | ||
35 | /* DMA descriptor control */ | 37 | /* DMA descriptor control */ |
@@ -296,13 +298,7 @@ static dma_cookie_t sh_dmae_tx_submit(struct dma_async_tx_descriptor *tx) | |||
296 | else | 298 | else |
297 | power_up = false; | 299 | power_up = false; |
298 | 300 | ||
299 | cookie = sh_chan->common.cookie; | 301 | cookie = dma_cookie_assign(tx); |
300 | cookie++; | ||
301 | if (cookie < 0) | ||
302 | cookie = 1; | ||
303 | |||
304 | sh_chan->common.cookie = cookie; | ||
305 | tx->cookie = cookie; | ||
306 | 302 | ||
307 | /* Mark all chunks of this descriptor as submitted, move to the queue */ | 303 | /* Mark all chunks of this descriptor as submitted, move to the queue */ |
308 | list_for_each_entry_safe(chunk, c, desc->node.prev, node) { | 304 | list_for_each_entry_safe(chunk, c, desc->node.prev, node) { |
@@ -673,7 +669,8 @@ static struct dma_async_tx_descriptor *sh_dmae_prep_memcpy( | |||
673 | 669 | ||
674 | static struct dma_async_tx_descriptor *sh_dmae_prep_slave_sg( | 670 | static struct dma_async_tx_descriptor *sh_dmae_prep_slave_sg( |
675 | struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, | 671 | struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len, |
676 | enum dma_transfer_direction direction, unsigned long flags) | 672 | enum dma_transfer_direction direction, unsigned long flags, |
673 | void *context) | ||
677 | { | 674 | { |
678 | struct sh_dmae_slave *param; | 675 | struct sh_dmae_slave *param; |
679 | struct sh_dmae_chan *sh_chan; | 676 | struct sh_dmae_chan *sh_chan; |
@@ -764,12 +761,12 @@ static dma_async_tx_callback __ld_cleanup(struct sh_dmae_chan *sh_chan, bool all | |||
764 | cookie = tx->cookie; | 761 | cookie = tx->cookie; |
765 | 762 | ||
766 | if (desc->mark == DESC_COMPLETED && desc->chunks == 1) { | 763 | if (desc->mark == DESC_COMPLETED && desc->chunks == 1) { |
767 | if (sh_chan->completed_cookie != desc->cookie - 1) | 764 | if (sh_chan->common.completed_cookie != desc->cookie - 1) |
768 | dev_dbg(sh_chan->dev, | 765 | dev_dbg(sh_chan->dev, |
769 | "Completing cookie %d, expected %d\n", | 766 | "Completing cookie %d, expected %d\n", |
770 | desc->cookie, | 767 | desc->cookie, |
771 | sh_chan->completed_cookie + 1); | 768 | sh_chan->common.completed_cookie + 1); |
772 | sh_chan->completed_cookie = desc->cookie; | 769 | sh_chan->common.completed_cookie = desc->cookie; |
773 | } | 770 | } |
774 | 771 | ||
775 | /* Call callback on the last chunk */ | 772 | /* Call callback on the last chunk */ |
@@ -823,7 +820,7 @@ static dma_async_tx_callback __ld_cleanup(struct sh_dmae_chan *sh_chan, bool all | |||
823 | * Terminating and the loop completed normally: forgive | 820 | * Terminating and the loop completed normally: forgive |
824 | * uncompleted cookies | 821 | * uncompleted cookies |
825 | */ | 822 | */ |
826 | sh_chan->completed_cookie = sh_chan->common.cookie; | 823 | sh_chan->common.completed_cookie = sh_chan->common.cookie; |
827 | 824 | ||
828 | spin_unlock_irqrestore(&sh_chan->desc_lock, flags); | 825 | spin_unlock_irqrestore(&sh_chan->desc_lock, flags); |
829 | 826 | ||
@@ -883,23 +880,14 @@ static enum dma_status sh_dmae_tx_status(struct dma_chan *chan, | |||
883 | struct dma_tx_state *txstate) | 880 | struct dma_tx_state *txstate) |
884 | { | 881 | { |
885 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); | 882 | struct sh_dmae_chan *sh_chan = to_sh_chan(chan); |
886 | dma_cookie_t last_used; | ||
887 | dma_cookie_t last_complete; | ||
888 | enum dma_status status; | 883 | enum dma_status status; |
889 | unsigned long flags; | 884 | unsigned long flags; |
890 | 885 | ||
891 | sh_dmae_chan_ld_cleanup(sh_chan, false); | 886 | sh_dmae_chan_ld_cleanup(sh_chan, false); |
892 | 887 | ||
893 | /* First read completed cookie to avoid a skew */ | ||
894 | last_complete = sh_chan->completed_cookie; | ||
895 | rmb(); | ||
896 | last_used = chan->cookie; | ||
897 | BUG_ON(last_complete < 0); | ||
898 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
899 | |||
900 | spin_lock_irqsave(&sh_chan->desc_lock, flags); | 888 | spin_lock_irqsave(&sh_chan->desc_lock, flags); |
901 | 889 | ||
902 | status = dma_async_is_complete(cookie, last_complete, last_used); | 890 | status = dma_cookie_status(chan, cookie, txstate); |
903 | 891 | ||
904 | /* | 892 | /* |
905 | * If we don't find cookie on the queue, it has been aborted and we have | 893 | * If we don't find cookie on the queue, it has been aborted and we have |
@@ -1102,6 +1090,7 @@ static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id, | |||
1102 | 1090 | ||
1103 | /* reference struct dma_device */ | 1091 | /* reference struct dma_device */ |
1104 | new_sh_chan->common.device = &shdev->common; | 1092 | new_sh_chan->common.device = &shdev->common; |
1093 | dma_cookie_init(&new_sh_chan->common); | ||
1105 | 1094 | ||
1106 | new_sh_chan->dev = shdev->common.dev; | 1095 | new_sh_chan->dev = shdev->common.dev; |
1107 | new_sh_chan->id = id; | 1096 | new_sh_chan->id = id; |
diff --git a/drivers/dma/shdma.h b/drivers/dma/shdma.h index 2b55a276dc5b..0b1d2c105f02 100644 --- a/drivers/dma/shdma.h +++ b/drivers/dma/shdma.h | |||
@@ -30,7 +30,6 @@ enum dmae_pm_state { | |||
30 | }; | 30 | }; |
31 | 31 | ||
32 | struct sh_dmae_chan { | 32 | struct sh_dmae_chan { |
33 | dma_cookie_t completed_cookie; /* The maximum cookie completed */ | ||
34 | spinlock_t desc_lock; /* Descriptor operation lock */ | 33 | spinlock_t desc_lock; /* Descriptor operation lock */ |
35 | struct list_head ld_queue; /* Link descriptors queue */ | 34 | struct list_head ld_queue; /* Link descriptors queue */ |
36 | struct list_head ld_free; /* Link descriptors free */ | 35 | struct list_head ld_free; /* Link descriptors free */ |
diff --git a/drivers/dma/sirf-dma.c b/drivers/dma/sirf-dma.c index 2333810d1688..434ad31174f2 100644 --- a/drivers/dma/sirf-dma.c +++ b/drivers/dma/sirf-dma.c | |||
@@ -18,6 +18,8 @@ | |||
18 | #include <linux/of_platform.h> | 18 | #include <linux/of_platform.h> |
19 | #include <linux/sirfsoc_dma.h> | 19 | #include <linux/sirfsoc_dma.h> |
20 | 20 | ||
21 | #include "dmaengine.h" | ||
22 | |||
21 | #define SIRFSOC_DMA_DESCRIPTORS 16 | 23 | #define SIRFSOC_DMA_DESCRIPTORS 16 |
22 | #define SIRFSOC_DMA_CHANNELS 16 | 24 | #define SIRFSOC_DMA_CHANNELS 16 |
23 | 25 | ||
@@ -59,7 +61,6 @@ struct sirfsoc_dma_chan { | |||
59 | struct list_head queued; | 61 | struct list_head queued; |
60 | struct list_head active; | 62 | struct list_head active; |
61 | struct list_head completed; | 63 | struct list_head completed; |
62 | dma_cookie_t completed_cookie; | ||
63 | unsigned long happened_cyclic; | 64 | unsigned long happened_cyclic; |
64 | unsigned long completed_cyclic; | 65 | unsigned long completed_cyclic; |
65 | 66 | ||
@@ -208,7 +209,7 @@ static void sirfsoc_dma_process_completed(struct sirfsoc_dma *sdma) | |||
208 | /* Free descriptors */ | 209 | /* Free descriptors */ |
209 | spin_lock_irqsave(&schan->lock, flags); | 210 | spin_lock_irqsave(&schan->lock, flags); |
210 | list_splice_tail_init(&list, &schan->free); | 211 | list_splice_tail_init(&list, &schan->free); |
211 | schan->completed_cookie = last_cookie; | 212 | schan->chan.completed_cookie = last_cookie; |
212 | spin_unlock_irqrestore(&schan->lock, flags); | 213 | spin_unlock_irqrestore(&schan->lock, flags); |
213 | } else { | 214 | } else { |
214 | /* for cyclic channel, desc is always in active list */ | 215 | /* for cyclic channel, desc is always in active list */ |
@@ -258,13 +259,7 @@ static dma_cookie_t sirfsoc_dma_tx_submit(struct dma_async_tx_descriptor *txd) | |||
258 | /* Move descriptor to queue */ | 259 | /* Move descriptor to queue */ |
259 | list_move_tail(&sdesc->node, &schan->queued); | 260 | list_move_tail(&sdesc->node, &schan->queued); |
260 | 261 | ||
261 | /* Update cookie */ | 262 | cookie = dma_cookie_assign(txd); |
262 | cookie = schan->chan.cookie + 1; | ||
263 | if (cookie <= 0) | ||
264 | cookie = 1; | ||
265 | |||
266 | schan->chan.cookie = cookie; | ||
267 | sdesc->desc.cookie = cookie; | ||
268 | 263 | ||
269 | spin_unlock_irqrestore(&schan->lock, flags); | 264 | spin_unlock_irqrestore(&schan->lock, flags); |
270 | 265 | ||
@@ -414,16 +409,13 @@ sirfsoc_dma_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | |||
414 | { | 409 | { |
415 | struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(chan); | 410 | struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(chan); |
416 | unsigned long flags; | 411 | unsigned long flags; |
417 | dma_cookie_t last_used; | 412 | enum dma_status ret; |
418 | dma_cookie_t last_complete; | ||
419 | 413 | ||
420 | spin_lock_irqsave(&schan->lock, flags); | 414 | spin_lock_irqsave(&schan->lock, flags); |
421 | last_used = schan->chan.cookie; | 415 | ret = dma_cookie_status(chan, cookie, txstate); |
422 | last_complete = schan->completed_cookie; | ||
423 | spin_unlock_irqrestore(&schan->lock, flags); | 416 | spin_unlock_irqrestore(&schan->lock, flags); |
424 | 417 | ||
425 | dma_set_tx_state(txstate, last_complete, last_used, 0); | 418 | return ret; |
426 | return dma_async_is_complete(cookie, last_complete, last_used); | ||
427 | } | 419 | } |
428 | 420 | ||
429 | static struct dma_async_tx_descriptor *sirfsoc_dma_prep_interleaved( | 421 | static struct dma_async_tx_descriptor *sirfsoc_dma_prep_interleaved( |
@@ -497,7 +489,7 @@ err_dir: | |||
497 | static struct dma_async_tx_descriptor * | 489 | static struct dma_async_tx_descriptor * |
498 | sirfsoc_dma_prep_cyclic(struct dma_chan *chan, dma_addr_t addr, | 490 | sirfsoc_dma_prep_cyclic(struct dma_chan *chan, dma_addr_t addr, |
499 | size_t buf_len, size_t period_len, | 491 | size_t buf_len, size_t period_len, |
500 | enum dma_transfer_direction direction) | 492 | enum dma_transfer_direction direction, void *context) |
501 | { | 493 | { |
502 | struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(chan); | 494 | struct sirfsoc_dma_chan *schan = dma_chan_to_sirfsoc_dma_chan(chan); |
503 | struct sirfsoc_dma_desc *sdesc = NULL; | 495 | struct sirfsoc_dma_desc *sdesc = NULL; |
@@ -635,8 +627,7 @@ static int __devinit sirfsoc_dma_probe(struct platform_device *op) | |||
635 | schan = &sdma->channels[i]; | 627 | schan = &sdma->channels[i]; |
636 | 628 | ||
637 | schan->chan.device = dma; | 629 | schan->chan.device = dma; |
638 | schan->chan.cookie = 1; | 630 | dma_cookie_init(&schan->chan); |
639 | schan->completed_cookie = schan->chan.cookie; | ||
640 | 631 | ||
641 | INIT_LIST_HEAD(&schan->free); | 632 | INIT_LIST_HEAD(&schan->free); |
642 | INIT_LIST_HEAD(&schan->prepared); | 633 | INIT_LIST_HEAD(&schan->prepared); |
diff --git a/drivers/dma/ste_dma40.c b/drivers/dma/ste_dma40.c index cc5ecbc067a3..bdd41d4bfa8d 100644 --- a/drivers/dma/ste_dma40.c +++ b/drivers/dma/ste_dma40.c | |||
@@ -21,6 +21,7 @@ | |||
21 | 21 | ||
22 | #include <plat/ste_dma40.h> | 22 | #include <plat/ste_dma40.h> |
23 | 23 | ||
24 | #include "dmaengine.h" | ||
24 | #include "ste_dma40_ll.h" | 25 | #include "ste_dma40_ll.h" |
25 | 26 | ||
26 | #define D40_NAME "dma40" | 27 | #define D40_NAME "dma40" |
@@ -220,8 +221,6 @@ struct d40_base; | |||
220 | * | 221 | * |
221 | * @lock: A spinlock to protect this struct. | 222 | * @lock: A spinlock to protect this struct. |
222 | * @log_num: The logical number, if any of this channel. | 223 | * @log_num: The logical number, if any of this channel. |
223 | * @completed: Starts with 1, after first interrupt it is set to dma engine's | ||
224 | * current cookie. | ||
225 | * @pending_tx: The number of pending transfers. Used between interrupt handler | 224 | * @pending_tx: The number of pending transfers. Used between interrupt handler |
226 | * and tasklet. | 225 | * and tasklet. |
227 | * @busy: Set to true when transfer is ongoing on this channel. | 226 | * @busy: Set to true when transfer is ongoing on this channel. |
@@ -250,8 +249,6 @@ struct d40_base; | |||
250 | struct d40_chan { | 249 | struct d40_chan { |
251 | spinlock_t lock; | 250 | spinlock_t lock; |
252 | int log_num; | 251 | int log_num; |
253 | /* ID of the most recent completed transfer */ | ||
254 | int completed; | ||
255 | int pending_tx; | 252 | int pending_tx; |
256 | bool busy; | 253 | bool busy; |
257 | struct d40_phy_res *phy_chan; | 254 | struct d40_phy_res *phy_chan; |
@@ -1223,21 +1220,14 @@ static dma_cookie_t d40_tx_submit(struct dma_async_tx_descriptor *tx) | |||
1223 | chan); | 1220 | chan); |
1224 | struct d40_desc *d40d = container_of(tx, struct d40_desc, txd); | 1221 | struct d40_desc *d40d = container_of(tx, struct d40_desc, txd); |
1225 | unsigned long flags; | 1222 | unsigned long flags; |
1223 | dma_cookie_t cookie; | ||
1226 | 1224 | ||
1227 | spin_lock_irqsave(&d40c->lock, flags); | 1225 | spin_lock_irqsave(&d40c->lock, flags); |
1228 | 1226 | cookie = dma_cookie_assign(tx); | |
1229 | d40c->chan.cookie++; | ||
1230 | |||
1231 | if (d40c->chan.cookie < 0) | ||
1232 | d40c->chan.cookie = 1; | ||
1233 | |||
1234 | d40d->txd.cookie = d40c->chan.cookie; | ||
1235 | |||
1236 | d40_desc_queue(d40c, d40d); | 1227 | d40_desc_queue(d40c, d40d); |
1237 | |||
1238 | spin_unlock_irqrestore(&d40c->lock, flags); | 1228 | spin_unlock_irqrestore(&d40c->lock, flags); |
1239 | 1229 | ||
1240 | return tx->cookie; | 1230 | return cookie; |
1241 | } | 1231 | } |
1242 | 1232 | ||
1243 | static int d40_start(struct d40_chan *d40c) | 1233 | static int d40_start(struct d40_chan *d40c) |
@@ -1357,7 +1347,7 @@ static void dma_tasklet(unsigned long data) | |||
1357 | goto err; | 1347 | goto err; |
1358 | 1348 | ||
1359 | if (!d40d->cyclic) | 1349 | if (!d40d->cyclic) |
1360 | d40c->completed = d40d->txd.cookie; | 1350 | dma_cookie_complete(&d40d->txd); |
1361 | 1351 | ||
1362 | /* | 1352 | /* |
1363 | * If terminating a channel pending_tx is set to zero. | 1353 | * If terminating a channel pending_tx is set to zero. |
@@ -2182,7 +2172,7 @@ static int d40_alloc_chan_resources(struct dma_chan *chan) | |||
2182 | bool is_free_phy; | 2172 | bool is_free_phy; |
2183 | spin_lock_irqsave(&d40c->lock, flags); | 2173 | spin_lock_irqsave(&d40c->lock, flags); |
2184 | 2174 | ||
2185 | d40c->completed = chan->cookie = 1; | 2175 | dma_cookie_init(chan); |
2186 | 2176 | ||
2187 | /* If no dma configuration is set use default configuration (memcpy) */ | 2177 | /* If no dma configuration is set use default configuration (memcpy) */ |
2188 | if (!d40c->configured) { | 2178 | if (!d40c->configured) { |
@@ -2299,7 +2289,8 @@ static struct dma_async_tx_descriptor *d40_prep_slave_sg(struct dma_chan *chan, | |||
2299 | struct scatterlist *sgl, | 2289 | struct scatterlist *sgl, |
2300 | unsigned int sg_len, | 2290 | unsigned int sg_len, |
2301 | enum dma_transfer_direction direction, | 2291 | enum dma_transfer_direction direction, |
2302 | unsigned long dma_flags) | 2292 | unsigned long dma_flags, |
2293 | void *context) | ||
2303 | { | 2294 | { |
2304 | if (direction != DMA_DEV_TO_MEM && direction != DMA_MEM_TO_DEV) | 2295 | if (direction != DMA_DEV_TO_MEM && direction != DMA_MEM_TO_DEV) |
2305 | return NULL; | 2296 | return NULL; |
@@ -2310,7 +2301,7 @@ static struct dma_async_tx_descriptor *d40_prep_slave_sg(struct dma_chan *chan, | |||
2310 | static struct dma_async_tx_descriptor * | 2301 | static struct dma_async_tx_descriptor * |
2311 | dma40_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr, | 2302 | dma40_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr, |
2312 | size_t buf_len, size_t period_len, | 2303 | size_t buf_len, size_t period_len, |
2313 | enum dma_transfer_direction direction) | 2304 | enum dma_transfer_direction direction, void *context) |
2314 | { | 2305 | { |
2315 | unsigned int periods = buf_len / period_len; | 2306 | unsigned int periods = buf_len / period_len; |
2316 | struct dma_async_tx_descriptor *txd; | 2307 | struct dma_async_tx_descriptor *txd; |
@@ -2342,25 +2333,19 @@ static enum dma_status d40_tx_status(struct dma_chan *chan, | |||
2342 | struct dma_tx_state *txstate) | 2333 | struct dma_tx_state *txstate) |
2343 | { | 2334 | { |
2344 | struct d40_chan *d40c = container_of(chan, struct d40_chan, chan); | 2335 | struct d40_chan *d40c = container_of(chan, struct d40_chan, chan); |
2345 | dma_cookie_t last_used; | 2336 | enum dma_status ret; |
2346 | dma_cookie_t last_complete; | ||
2347 | int ret; | ||
2348 | 2337 | ||
2349 | if (d40c->phy_chan == NULL) { | 2338 | if (d40c->phy_chan == NULL) { |
2350 | chan_err(d40c, "Cannot read status of unallocated channel\n"); | 2339 | chan_err(d40c, "Cannot read status of unallocated channel\n"); |
2351 | return -EINVAL; | 2340 | return -EINVAL; |
2352 | } | 2341 | } |
2353 | 2342 | ||
2354 | last_complete = d40c->completed; | 2343 | ret = dma_cookie_status(chan, cookie, txstate); |
2355 | last_used = chan->cookie; | 2344 | if (ret != DMA_SUCCESS) |
2345 | dma_set_residue(txstate, stedma40_residue(chan)); | ||
2356 | 2346 | ||
2357 | if (d40_is_paused(d40c)) | 2347 | if (d40_is_paused(d40c)) |
2358 | ret = DMA_PAUSED; | 2348 | ret = DMA_PAUSED; |
2359 | else | ||
2360 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
2361 | |||
2362 | dma_set_tx_state(txstate, last_complete, last_used, | ||
2363 | stedma40_residue(chan)); | ||
2364 | 2349 | ||
2365 | return ret; | 2350 | return ret; |
2366 | } | 2351 | } |
diff --git a/drivers/dma/timb_dma.c b/drivers/dma/timb_dma.c index a6f9c1684a0f..4e0dff59901d 100644 --- a/drivers/dma/timb_dma.c +++ b/drivers/dma/timb_dma.c | |||
@@ -31,6 +31,8 @@ | |||
31 | 31 | ||
32 | #include <linux/timb_dma.h> | 32 | #include <linux/timb_dma.h> |
33 | 33 | ||
34 | #include "dmaengine.h" | ||
35 | |||
34 | #define DRIVER_NAME "timb-dma" | 36 | #define DRIVER_NAME "timb-dma" |
35 | 37 | ||
36 | /* Global DMA registers */ | 38 | /* Global DMA registers */ |
@@ -84,7 +86,6 @@ struct timb_dma_chan { | |||
84 | especially the lists and descriptors, | 86 | especially the lists and descriptors, |
85 | from races between the tasklet and calls | 87 | from races between the tasklet and calls |
86 | from above */ | 88 | from above */ |
87 | dma_cookie_t last_completed_cookie; | ||
88 | bool ongoing; | 89 | bool ongoing; |
89 | struct list_head active_list; | 90 | struct list_head active_list; |
90 | struct list_head queue; | 91 | struct list_head queue; |
@@ -284,7 +285,7 @@ static void __td_finish(struct timb_dma_chan *td_chan) | |||
284 | else | 285 | else |
285 | iowrite32(0, td_chan->membase + TIMBDMA_OFFS_TX_DLAR); | 286 | iowrite32(0, td_chan->membase + TIMBDMA_OFFS_TX_DLAR); |
286 | */ | 287 | */ |
287 | td_chan->last_completed_cookie = txd->cookie; | 288 | dma_cookie_complete(txd); |
288 | td_chan->ongoing = false; | 289 | td_chan->ongoing = false; |
289 | 290 | ||
290 | callback = txd->callback; | 291 | callback = txd->callback; |
@@ -349,12 +350,7 @@ static dma_cookie_t td_tx_submit(struct dma_async_tx_descriptor *txd) | |||
349 | dma_cookie_t cookie; | 350 | dma_cookie_t cookie; |
350 | 351 | ||
351 | spin_lock_bh(&td_chan->lock); | 352 | spin_lock_bh(&td_chan->lock); |
352 | 353 | cookie = dma_cookie_assign(txd); | |
353 | cookie = txd->chan->cookie; | ||
354 | if (++cookie < 0) | ||
355 | cookie = 1; | ||
356 | txd->chan->cookie = cookie; | ||
357 | txd->cookie = cookie; | ||
358 | 354 | ||
359 | if (list_empty(&td_chan->active_list)) { | 355 | if (list_empty(&td_chan->active_list)) { |
360 | dev_dbg(chan2dev(txd->chan), "%s: started %u\n", __func__, | 356 | dev_dbg(chan2dev(txd->chan), "%s: started %u\n", __func__, |
@@ -481,8 +477,7 @@ static int td_alloc_chan_resources(struct dma_chan *chan) | |||
481 | } | 477 | } |
482 | 478 | ||
483 | spin_lock_bh(&td_chan->lock); | 479 | spin_lock_bh(&td_chan->lock); |
484 | td_chan->last_completed_cookie = 1; | 480 | dma_cookie_init(chan); |
485 | chan->cookie = 1; | ||
486 | spin_unlock_bh(&td_chan->lock); | 481 | spin_unlock_bh(&td_chan->lock); |
487 | 482 | ||
488 | return 0; | 483 | return 0; |
@@ -515,24 +510,13 @@ static void td_free_chan_resources(struct dma_chan *chan) | |||
515 | static enum dma_status td_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | 510 | static enum dma_status td_tx_status(struct dma_chan *chan, dma_cookie_t cookie, |
516 | struct dma_tx_state *txstate) | 511 | struct dma_tx_state *txstate) |
517 | { | 512 | { |
518 | struct timb_dma_chan *td_chan = | 513 | enum dma_status ret; |
519 | container_of(chan, struct timb_dma_chan, chan); | ||
520 | dma_cookie_t last_used; | ||
521 | dma_cookie_t last_complete; | ||
522 | int ret; | ||
523 | 514 | ||
524 | dev_dbg(chan2dev(chan), "%s: Entry\n", __func__); | 515 | dev_dbg(chan2dev(chan), "%s: Entry\n", __func__); |
525 | 516 | ||
526 | last_complete = td_chan->last_completed_cookie; | 517 | ret = dma_cookie_status(chan, cookie, txstate); |
527 | last_used = chan->cookie; | ||
528 | |||
529 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
530 | |||
531 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
532 | 518 | ||
533 | dev_dbg(chan2dev(chan), | 519 | dev_dbg(chan2dev(chan), "%s: exit, ret: %d\n", __func__, ret); |
534 | "%s: exit, ret: %d, last_complete: %d, last_used: %d\n", | ||
535 | __func__, ret, last_complete, last_used); | ||
536 | 520 | ||
537 | return ret; | 521 | return ret; |
538 | } | 522 | } |
@@ -558,7 +542,8 @@ static void td_issue_pending(struct dma_chan *chan) | |||
558 | 542 | ||
559 | static struct dma_async_tx_descriptor *td_prep_slave_sg(struct dma_chan *chan, | 543 | static struct dma_async_tx_descriptor *td_prep_slave_sg(struct dma_chan *chan, |
560 | struct scatterlist *sgl, unsigned int sg_len, | 544 | struct scatterlist *sgl, unsigned int sg_len, |
561 | enum dma_transfer_direction direction, unsigned long flags) | 545 | enum dma_transfer_direction direction, unsigned long flags, |
546 | void *context) | ||
562 | { | 547 | { |
563 | struct timb_dma_chan *td_chan = | 548 | struct timb_dma_chan *td_chan = |
564 | container_of(chan, struct timb_dma_chan, chan); | 549 | container_of(chan, struct timb_dma_chan, chan); |
@@ -766,7 +751,7 @@ static int __devinit td_probe(struct platform_device *pdev) | |||
766 | } | 751 | } |
767 | 752 | ||
768 | td_chan->chan.device = &td->dma; | 753 | td_chan->chan.device = &td->dma; |
769 | td_chan->chan.cookie = 1; | 754 | dma_cookie_init(&td_chan->chan); |
770 | spin_lock_init(&td_chan->lock); | 755 | spin_lock_init(&td_chan->lock); |
771 | INIT_LIST_HEAD(&td_chan->active_list); | 756 | INIT_LIST_HEAD(&td_chan->active_list); |
772 | INIT_LIST_HEAD(&td_chan->queue); | 757 | INIT_LIST_HEAD(&td_chan->queue); |
diff --git a/drivers/dma/txx9dmac.c b/drivers/dma/txx9dmac.c index 6122c364cf11..913f55c76c99 100644 --- a/drivers/dma/txx9dmac.c +++ b/drivers/dma/txx9dmac.c | |||
@@ -15,6 +15,8 @@ | |||
15 | #include <linux/platform_device.h> | 15 | #include <linux/platform_device.h> |
16 | #include <linux/slab.h> | 16 | #include <linux/slab.h> |
17 | #include <linux/scatterlist.h> | 17 | #include <linux/scatterlist.h> |
18 | |||
19 | #include "dmaengine.h" | ||
18 | #include "txx9dmac.h" | 20 | #include "txx9dmac.h" |
19 | 21 | ||
20 | static struct txx9dmac_chan *to_txx9dmac_chan(struct dma_chan *chan) | 22 | static struct txx9dmac_chan *to_txx9dmac_chan(struct dma_chan *chan) |
@@ -279,21 +281,6 @@ static void txx9dmac_desc_put(struct txx9dmac_chan *dc, | |||
279 | } | 281 | } |
280 | } | 282 | } |
281 | 283 | ||
282 | /* Called with dc->lock held and bh disabled */ | ||
283 | static dma_cookie_t | ||
284 | txx9dmac_assign_cookie(struct txx9dmac_chan *dc, struct txx9dmac_desc *desc) | ||
285 | { | ||
286 | dma_cookie_t cookie = dc->chan.cookie; | ||
287 | |||
288 | if (++cookie < 0) | ||
289 | cookie = 1; | ||
290 | |||
291 | dc->chan.cookie = cookie; | ||
292 | desc->txd.cookie = cookie; | ||
293 | |||
294 | return cookie; | ||
295 | } | ||
296 | |||
297 | /*----------------------------------------------------------------------*/ | 284 | /*----------------------------------------------------------------------*/ |
298 | 285 | ||
299 | static void txx9dmac_dump_regs(struct txx9dmac_chan *dc) | 286 | static void txx9dmac_dump_regs(struct txx9dmac_chan *dc) |
@@ -424,7 +411,7 @@ txx9dmac_descriptor_complete(struct txx9dmac_chan *dc, | |||
424 | dev_vdbg(chan2dev(&dc->chan), "descriptor %u %p complete\n", | 411 | dev_vdbg(chan2dev(&dc->chan), "descriptor %u %p complete\n", |
425 | txd->cookie, desc); | 412 | txd->cookie, desc); |
426 | 413 | ||
427 | dc->completed = txd->cookie; | 414 | dma_cookie_complete(txd); |
428 | callback = txd->callback; | 415 | callback = txd->callback; |
429 | param = txd->callback_param; | 416 | param = txd->callback_param; |
430 | 417 | ||
@@ -738,7 +725,7 @@ static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx) | |||
738 | dma_cookie_t cookie; | 725 | dma_cookie_t cookie; |
739 | 726 | ||
740 | spin_lock_bh(&dc->lock); | 727 | spin_lock_bh(&dc->lock); |
741 | cookie = txx9dmac_assign_cookie(dc, desc); | 728 | cookie = dma_cookie_assign(tx); |
742 | 729 | ||
743 | dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u %p\n", | 730 | dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u %p\n", |
744 | desc->txd.cookie, desc); | 731 | desc->txd.cookie, desc); |
@@ -846,7 +833,7 @@ txx9dmac_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src, | |||
846 | static struct dma_async_tx_descriptor * | 833 | static struct dma_async_tx_descriptor * |
847 | txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, | 834 | txx9dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, |
848 | unsigned int sg_len, enum dma_transfer_direction direction, | 835 | unsigned int sg_len, enum dma_transfer_direction direction, |
849 | unsigned long flags) | 836 | unsigned long flags, void *context) |
850 | { | 837 | { |
851 | struct txx9dmac_chan *dc = to_txx9dmac_chan(chan); | 838 | struct txx9dmac_chan *dc = to_txx9dmac_chan(chan); |
852 | struct txx9dmac_dev *ddev = dc->ddev; | 839 | struct txx9dmac_dev *ddev = dc->ddev; |
@@ -972,27 +959,17 @@ txx9dmac_tx_status(struct dma_chan *chan, dma_cookie_t cookie, | |||
972 | struct dma_tx_state *txstate) | 959 | struct dma_tx_state *txstate) |
973 | { | 960 | { |
974 | struct txx9dmac_chan *dc = to_txx9dmac_chan(chan); | 961 | struct txx9dmac_chan *dc = to_txx9dmac_chan(chan); |
975 | dma_cookie_t last_used; | 962 | enum dma_status ret; |
976 | dma_cookie_t last_complete; | ||
977 | int ret; | ||
978 | 963 | ||
979 | last_complete = dc->completed; | 964 | ret = dma_cookie_status(chan, cookie, txstate); |
980 | last_used = chan->cookie; | ||
981 | |||
982 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
983 | if (ret != DMA_SUCCESS) { | 965 | if (ret != DMA_SUCCESS) { |
984 | spin_lock_bh(&dc->lock); | 966 | spin_lock_bh(&dc->lock); |
985 | txx9dmac_scan_descriptors(dc); | 967 | txx9dmac_scan_descriptors(dc); |
986 | spin_unlock_bh(&dc->lock); | 968 | spin_unlock_bh(&dc->lock); |
987 | 969 | ||
988 | last_complete = dc->completed; | 970 | ret = dma_cookie_status(chan, cookie, txstate); |
989 | last_used = chan->cookie; | ||
990 | |||
991 | ret = dma_async_is_complete(cookie, last_complete, last_used); | ||
992 | } | 971 | } |
993 | 972 | ||
994 | dma_set_tx_state(txstate, last_complete, last_used, 0); | ||
995 | |||
996 | return ret; | 973 | return ret; |
997 | } | 974 | } |
998 | 975 | ||
@@ -1057,7 +1034,7 @@ static int txx9dmac_alloc_chan_resources(struct dma_chan *chan) | |||
1057 | return -EIO; | 1034 | return -EIO; |
1058 | } | 1035 | } |
1059 | 1036 | ||
1060 | dc->completed = chan->cookie = 1; | 1037 | dma_cookie_init(chan); |
1061 | 1038 | ||
1062 | dc->ccr = TXX9_DMA_CCR_IMMCHN | TXX9_DMA_CCR_INTENE | CCR_LE; | 1039 | dc->ccr = TXX9_DMA_CCR_IMMCHN | TXX9_DMA_CCR_INTENE | CCR_LE; |
1063 | txx9dmac_chan_set_SMPCHN(dc); | 1040 | txx9dmac_chan_set_SMPCHN(dc); |
@@ -1186,7 +1163,7 @@ static int __init txx9dmac_chan_probe(struct platform_device *pdev) | |||
1186 | dc->ddev->chan[ch] = dc; | 1163 | dc->ddev->chan[ch] = dc; |
1187 | dc->chan.device = &dc->dma; | 1164 | dc->chan.device = &dc->dma; |
1188 | list_add_tail(&dc->chan.device_node, &dc->chan.device->channels); | 1165 | list_add_tail(&dc->chan.device_node, &dc->chan.device->channels); |
1189 | dc->chan.cookie = dc->completed = 1; | 1166 | dma_cookie_init(&dc->chan); |
1190 | 1167 | ||
1191 | if (is_dmac64(dc)) | 1168 | if (is_dmac64(dc)) |
1192 | dc->ch_regs = &__txx9dmac_regs(dc->ddev)->CHAN[ch]; | 1169 | dc->ch_regs = &__txx9dmac_regs(dc->ddev)->CHAN[ch]; |
diff --git a/drivers/dma/txx9dmac.h b/drivers/dma/txx9dmac.h index 365d42366b9f..f5a760598882 100644 --- a/drivers/dma/txx9dmac.h +++ b/drivers/dma/txx9dmac.h | |||
@@ -172,7 +172,6 @@ struct txx9dmac_chan { | |||
172 | spinlock_t lock; | 172 | spinlock_t lock; |
173 | 173 | ||
174 | /* these other elements are all protected by lock */ | 174 | /* these other elements are all protected by lock */ |
175 | dma_cookie_t completed; | ||
176 | struct list_head active_list; | 175 | struct list_head active_list; |
177 | struct list_head queue; | 176 | struct list_head queue; |
178 | struct list_head free_list; | 177 | struct list_head free_list; |