aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/gpu/drm/i915/intel_ringbuffer.c
diff options
context:
space:
mode:
authorChris Wilson <chris@chris-wilson.co.uk>2017-04-25 09:00:49 -0400
committerChris Wilson <chris@chris-wilson.co.uk>2017-04-25 10:33:22 -0400
commite6ba9992de6c63fe86c028b4876338e1cb7dac34 (patch)
tree21a3a6a40107f920bf7c451f0aa742f07d317924 /drivers/gpu/drm/i915/intel_ringbuffer.c
parent6b764a594fc4e9f1caa537728a5a8e2192e18fc6 (diff)
drm/i915: Differentiate between sw write location into ring and last hw read
We need to keep track of the last location we ask the hw to read up to (RING_TAIL) separately from our last write location into the ring, so that in the event of a GPU reset we do not tell the HW to proceed into a partially written request (which can happen if that request is waiting for an external signal before being executed). v2: Refactor intel_ring_reset() (Mika) Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=100144 Testcase: igt/gem_exec_fence/await-hang Fixes: 821ed7df6e2a ("drm/i915: Update reset path to fix incomplete requests") Fixes: d55ac5bf97c6 ("drm/i915: Defer transfer onto execution timeline to actual hw submission") Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk> Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Cc: Mika Kuoppala <mika.kuoppala@intel.com> Link: http://patchwork.freedesktop.org/patch/msgid/20170425130049.26147-1-chris@chris-wilson.co.uk Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Diffstat (limited to 'drivers/gpu/drm/i915/intel_ringbuffer.c')
-rw-r--r--drivers/gpu/drm/i915/intel_ringbuffer.c41
1 files changed, 27 insertions, 14 deletions
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c
index 32afac6c754f..227dfcf1764e 100644
--- a/drivers/gpu/drm/i915/intel_ringbuffer.c
+++ b/drivers/gpu/drm/i915/intel_ringbuffer.c
@@ -49,7 +49,7 @@ static int __intel_ring_space(int head, int tail, int size)
49 49
50void intel_ring_update_space(struct intel_ring *ring) 50void intel_ring_update_space(struct intel_ring *ring)
51{ 51{
52 ring->space = __intel_ring_space(ring->head, ring->tail, ring->size); 52 ring->space = __intel_ring_space(ring->head, ring->emit, ring->size);
53} 53}
54 54
55static int 55static int
@@ -774,8 +774,8 @@ static void i9xx_submit_request(struct drm_i915_gem_request *request)
774 774
775 i915_gem_request_submit(request); 775 i915_gem_request_submit(request);
776 776
777 assert_ring_tail_valid(request->ring, request->tail); 777 I915_WRITE_TAIL(request->engine,
778 I915_WRITE_TAIL(request->engine, request->tail); 778 intel_ring_set_tail(request->ring, request->tail));
779} 779}
780 780
781static void i9xx_emit_breadcrumb(struct drm_i915_gem_request *req, u32 *cs) 781static void i9xx_emit_breadcrumb(struct drm_i915_gem_request *req, u32 *cs)
@@ -1319,11 +1319,23 @@ err:
1319 return PTR_ERR(addr); 1319 return PTR_ERR(addr);
1320} 1320}
1321 1321
1322void intel_ring_reset(struct intel_ring *ring, u32 tail)
1323{
1324 GEM_BUG_ON(!list_empty(&ring->request_list));
1325 ring->tail = tail;
1326 ring->head = tail;
1327 ring->emit = tail;
1328 intel_ring_update_space(ring);
1329}
1330
1322void intel_ring_unpin(struct intel_ring *ring) 1331void intel_ring_unpin(struct intel_ring *ring)
1323{ 1332{
1324 GEM_BUG_ON(!ring->vma); 1333 GEM_BUG_ON(!ring->vma);
1325 GEM_BUG_ON(!ring->vaddr); 1334 GEM_BUG_ON(!ring->vaddr);
1326 1335
1336 /* Discard any unused bytes beyond that submitted to hw. */
1337 intel_ring_reset(ring, ring->tail);
1338
1327 if (i915_vma_is_map_and_fenceable(ring->vma)) 1339 if (i915_vma_is_map_and_fenceable(ring->vma))
1328 i915_vma_unpin_iomap(ring->vma); 1340 i915_vma_unpin_iomap(ring->vma);
1329 else 1341 else
@@ -1555,8 +1567,9 @@ void intel_legacy_submission_resume(struct drm_i915_private *dev_priv)
1555 struct intel_engine_cs *engine; 1567 struct intel_engine_cs *engine;
1556 enum intel_engine_id id; 1568 enum intel_engine_id id;
1557 1569
1570 /* Restart from the beginning of the rings for convenience */
1558 for_each_engine(engine, dev_priv, id) 1571 for_each_engine(engine, dev_priv, id)
1559 engine->buffer->head = engine->buffer->tail; 1572 intel_ring_reset(engine->buffer, 0);
1560} 1573}
1561 1574
1562static int ring_request_alloc(struct drm_i915_gem_request *request) 1575static int ring_request_alloc(struct drm_i915_gem_request *request)
@@ -1609,7 +1622,7 @@ static int wait_for_space(struct drm_i915_gem_request *req, int bytes)
1609 unsigned space; 1622 unsigned space;
1610 1623
1611 /* Would completion of this request free enough space? */ 1624 /* Would completion of this request free enough space? */
1612 space = __intel_ring_space(target->postfix, ring->tail, 1625 space = __intel_ring_space(target->postfix, ring->emit,
1613 ring->size); 1626 ring->size);
1614 if (space >= bytes) 1627 if (space >= bytes)
1615 break; 1628 break;
@@ -1634,8 +1647,8 @@ static int wait_for_space(struct drm_i915_gem_request *req, int bytes)
1634u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords) 1647u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords)
1635{ 1648{
1636 struct intel_ring *ring = req->ring; 1649 struct intel_ring *ring = req->ring;
1637 int remain_actual = ring->size - ring->tail; 1650 int remain_actual = ring->size - ring->emit;
1638 int remain_usable = ring->effective_size - ring->tail; 1651 int remain_usable = ring->effective_size - ring->emit;
1639 int bytes = num_dwords * sizeof(u32); 1652 int bytes = num_dwords * sizeof(u32);
1640 int total_bytes, wait_bytes; 1653 int total_bytes, wait_bytes;
1641 bool need_wrap = false; 1654 bool need_wrap = false;
@@ -1671,17 +1684,17 @@ u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords)
1671 1684
1672 if (unlikely(need_wrap)) { 1685 if (unlikely(need_wrap)) {
1673 GEM_BUG_ON(remain_actual > ring->space); 1686 GEM_BUG_ON(remain_actual > ring->space);
1674 GEM_BUG_ON(ring->tail + remain_actual > ring->size); 1687 GEM_BUG_ON(ring->emit + remain_actual > ring->size);
1675 1688
1676 /* Fill the tail with MI_NOOP */ 1689 /* Fill the tail with MI_NOOP */
1677 memset(ring->vaddr + ring->tail, 0, remain_actual); 1690 memset(ring->vaddr + ring->emit, 0, remain_actual);
1678 ring->tail = 0; 1691 ring->emit = 0;
1679 ring->space -= remain_actual; 1692 ring->space -= remain_actual;
1680 } 1693 }
1681 1694
1682 GEM_BUG_ON(ring->tail > ring->size - bytes); 1695 GEM_BUG_ON(ring->emit > ring->size - bytes);
1683 cs = ring->vaddr + ring->tail; 1696 cs = ring->vaddr + ring->emit;
1684 ring->tail += bytes; 1697 ring->emit += bytes;
1685 ring->space -= bytes; 1698 ring->space -= bytes;
1686 GEM_BUG_ON(ring->space < 0); 1699 GEM_BUG_ON(ring->space < 0);
1687 1700
@@ -1692,7 +1705,7 @@ u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords)
1692int intel_ring_cacheline_align(struct drm_i915_gem_request *req) 1705int intel_ring_cacheline_align(struct drm_i915_gem_request *req)
1693{ 1706{
1694 int num_dwords = 1707 int num_dwords =
1695 (req->ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t); 1708 (req->ring->emit & (CACHELINE_BYTES - 1)) / sizeof(uint32_t);
1696 u32 *cs; 1709 u32 *cs;
1697 1710
1698 if (num_dwords == 0) 1711 if (num_dwords == 0)