diff options
author | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2017-12-01 20:04:39 -0500 |
---|---|---|
committer | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2017-12-05 14:57:53 -0500 |
commit | 98c1ec7cefaadbf65680d116c3d8612b93a841a0 (patch) | |
tree | e55a7ad431b5c7134d5c71c55e026e0b07bfbca1 | |
parent | adf90eb49055636fc35aede54174456ac3520f27 (diff) |
drivers/dma/ioat: Remove now-redundant smp_read_barrier_depends()
Now that READ_ONCE() implies smp_read_barrier_depends(), the
__cleanup() and ioat_abort_descs() functions no longer need their
smp_read_barrier_depends() calls, which this commit removes.
It is actually not entirely clear why this driver ever included
smp_read_barrier_depends() given that it appears to be x86-only and
given that smp_read_barrier_depends() has no effect whatsoever except
on DEC Alpha.
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Vinod Koul <vinod.koul@intel.com>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: <dmaengine@vger.kernel.org>
-rw-r--r-- | drivers/dma/ioat/dma.c | 2 |
1 files changed, 0 insertions, 2 deletions
diff --git a/drivers/dma/ioat/dma.c b/drivers/dma/ioat/dma.c index 58d4ccd33672..8b5b23a8ace9 100644 --- a/drivers/dma/ioat/dma.c +++ b/drivers/dma/ioat/dma.c | |||
@@ -597,7 +597,6 @@ static void __cleanup(struct ioatdma_chan *ioat_chan, dma_addr_t phys_complete) | |||
597 | for (i = 0; i < active && !seen_current; i++) { | 597 | for (i = 0; i < active && !seen_current; i++) { |
598 | struct dma_async_tx_descriptor *tx; | 598 | struct dma_async_tx_descriptor *tx; |
599 | 599 | ||
600 | smp_read_barrier_depends(); | ||
601 | prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); | 600 | prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); |
602 | desc = ioat_get_ring_ent(ioat_chan, idx + i); | 601 | desc = ioat_get_ring_ent(ioat_chan, idx + i); |
603 | dump_desc_dbg(ioat_chan, desc); | 602 | dump_desc_dbg(ioat_chan, desc); |
@@ -715,7 +714,6 @@ static void ioat_abort_descs(struct ioatdma_chan *ioat_chan) | |||
715 | for (i = 1; i < active; i++) { | 714 | for (i = 1; i < active; i++) { |
716 | struct dma_async_tx_descriptor *tx; | 715 | struct dma_async_tx_descriptor *tx; |
717 | 716 | ||
718 | smp_read_barrier_depends(); | ||
719 | prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); | 717 | prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); |
720 | desc = ioat_get_ring_ent(ioat_chan, idx + i); | 718 | desc = ioat_get_ring_ent(ioat_chan, idx + i); |
721 | 719 | ||