diff options
author | Chris Wilson <chris@chris-wilson.co.uk> | 2017-04-25 09:00:49 -0400 |
---|---|---|
committer | Chris Wilson <chris@chris-wilson.co.uk> | 2017-04-25 10:33:22 -0400 |
commit | e6ba9992de6c63fe86c028b4876338e1cb7dac34 (patch) | |
tree | 21a3a6a40107f920bf7c451f0aa742f07d317924 /drivers/gpu/drm/i915/intel_ringbuffer.c | |
parent | 6b764a594fc4e9f1caa537728a5a8e2192e18fc6 (diff) |
drm/i915: Differentiate between sw write location into ring and last hw read
We need to keep track of the last location we ask the hw to read up to
(RING_TAIL) separately from our last write location into the ring, so
that in the event of a GPU reset we do not tell the HW to proceed into
a partially written request (which can happen if that request is waiting
for an external signal before being executed).
v2: Refactor intel_ring_reset() (Mika)
Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=100144
Testcase: igt/gem_exec_fence/await-hang
Fixes: 821ed7df6e2a ("drm/i915: Update reset path to fix incomplete requests")
Fixes: d55ac5bf97c6 ("drm/i915: Defer transfer onto execution timeline to actual hw submission")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170425130049.26147-1-chris@chris-wilson.co.uk
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Diffstat (limited to 'drivers/gpu/drm/i915/intel_ringbuffer.c')
-rw-r--r-- | drivers/gpu/drm/i915/intel_ringbuffer.c | 41 |
1 files changed, 27 insertions, 14 deletions
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c index 32afac6c754f..227dfcf1764e 100644 --- a/drivers/gpu/drm/i915/intel_ringbuffer.c +++ b/drivers/gpu/drm/i915/intel_ringbuffer.c | |||
@@ -49,7 +49,7 @@ static int __intel_ring_space(int head, int tail, int size) | |||
49 | 49 | ||
50 | void intel_ring_update_space(struct intel_ring *ring) | 50 | void intel_ring_update_space(struct intel_ring *ring) |
51 | { | 51 | { |
52 | ring->space = __intel_ring_space(ring->head, ring->tail, ring->size); | 52 | ring->space = __intel_ring_space(ring->head, ring->emit, ring->size); |
53 | } | 53 | } |
54 | 54 | ||
55 | static int | 55 | static int |
@@ -774,8 +774,8 @@ static void i9xx_submit_request(struct drm_i915_gem_request *request) | |||
774 | 774 | ||
775 | i915_gem_request_submit(request); | 775 | i915_gem_request_submit(request); |
776 | 776 | ||
777 | assert_ring_tail_valid(request->ring, request->tail); | 777 | I915_WRITE_TAIL(request->engine, |
778 | I915_WRITE_TAIL(request->engine, request->tail); | 778 | intel_ring_set_tail(request->ring, request->tail)); |
779 | } | 779 | } |
780 | 780 | ||
781 | static void i9xx_emit_breadcrumb(struct drm_i915_gem_request *req, u32 *cs) | 781 | static void i9xx_emit_breadcrumb(struct drm_i915_gem_request *req, u32 *cs) |
@@ -1319,11 +1319,23 @@ err: | |||
1319 | return PTR_ERR(addr); | 1319 | return PTR_ERR(addr); |
1320 | } | 1320 | } |
1321 | 1321 | ||
1322 | void intel_ring_reset(struct intel_ring *ring, u32 tail) | ||
1323 | { | ||
1324 | GEM_BUG_ON(!list_empty(&ring->request_list)); | ||
1325 | ring->tail = tail; | ||
1326 | ring->head = tail; | ||
1327 | ring->emit = tail; | ||
1328 | intel_ring_update_space(ring); | ||
1329 | } | ||
1330 | |||
1322 | void intel_ring_unpin(struct intel_ring *ring) | 1331 | void intel_ring_unpin(struct intel_ring *ring) |
1323 | { | 1332 | { |
1324 | GEM_BUG_ON(!ring->vma); | 1333 | GEM_BUG_ON(!ring->vma); |
1325 | GEM_BUG_ON(!ring->vaddr); | 1334 | GEM_BUG_ON(!ring->vaddr); |
1326 | 1335 | ||
1336 | /* Discard any unused bytes beyond that submitted to hw. */ | ||
1337 | intel_ring_reset(ring, ring->tail); | ||
1338 | |||
1327 | if (i915_vma_is_map_and_fenceable(ring->vma)) | 1339 | if (i915_vma_is_map_and_fenceable(ring->vma)) |
1328 | i915_vma_unpin_iomap(ring->vma); | 1340 | i915_vma_unpin_iomap(ring->vma); |
1329 | else | 1341 | else |
@@ -1555,8 +1567,9 @@ void intel_legacy_submission_resume(struct drm_i915_private *dev_priv) | |||
1555 | struct intel_engine_cs *engine; | 1567 | struct intel_engine_cs *engine; |
1556 | enum intel_engine_id id; | 1568 | enum intel_engine_id id; |
1557 | 1569 | ||
1570 | /* Restart from the beginning of the rings for convenience */ | ||
1558 | for_each_engine(engine, dev_priv, id) | 1571 | for_each_engine(engine, dev_priv, id) |
1559 | engine->buffer->head = engine->buffer->tail; | 1572 | intel_ring_reset(engine->buffer, 0); |
1560 | } | 1573 | } |
1561 | 1574 | ||
1562 | static int ring_request_alloc(struct drm_i915_gem_request *request) | 1575 | static int ring_request_alloc(struct drm_i915_gem_request *request) |
@@ -1609,7 +1622,7 @@ static int wait_for_space(struct drm_i915_gem_request *req, int bytes) | |||
1609 | unsigned space; | 1622 | unsigned space; |
1610 | 1623 | ||
1611 | /* Would completion of this request free enough space? */ | 1624 | /* Would completion of this request free enough space? */ |
1612 | space = __intel_ring_space(target->postfix, ring->tail, | 1625 | space = __intel_ring_space(target->postfix, ring->emit, |
1613 | ring->size); | 1626 | ring->size); |
1614 | if (space >= bytes) | 1627 | if (space >= bytes) |
1615 | break; | 1628 | break; |
@@ -1634,8 +1647,8 @@ static int wait_for_space(struct drm_i915_gem_request *req, int bytes) | |||
1634 | u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords) | 1647 | u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords) |
1635 | { | 1648 | { |
1636 | struct intel_ring *ring = req->ring; | 1649 | struct intel_ring *ring = req->ring; |
1637 | int remain_actual = ring->size - ring->tail; | 1650 | int remain_actual = ring->size - ring->emit; |
1638 | int remain_usable = ring->effective_size - ring->tail; | 1651 | int remain_usable = ring->effective_size - ring->emit; |
1639 | int bytes = num_dwords * sizeof(u32); | 1652 | int bytes = num_dwords * sizeof(u32); |
1640 | int total_bytes, wait_bytes; | 1653 | int total_bytes, wait_bytes; |
1641 | bool need_wrap = false; | 1654 | bool need_wrap = false; |
@@ -1671,17 +1684,17 @@ u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords) | |||
1671 | 1684 | ||
1672 | if (unlikely(need_wrap)) { | 1685 | if (unlikely(need_wrap)) { |
1673 | GEM_BUG_ON(remain_actual > ring->space); | 1686 | GEM_BUG_ON(remain_actual > ring->space); |
1674 | GEM_BUG_ON(ring->tail + remain_actual > ring->size); | 1687 | GEM_BUG_ON(ring->emit + remain_actual > ring->size); |
1675 | 1688 | ||
1676 | /* Fill the tail with MI_NOOP */ | 1689 | /* Fill the tail with MI_NOOP */ |
1677 | memset(ring->vaddr + ring->tail, 0, remain_actual); | 1690 | memset(ring->vaddr + ring->emit, 0, remain_actual); |
1678 | ring->tail = 0; | 1691 | ring->emit = 0; |
1679 | ring->space -= remain_actual; | 1692 | ring->space -= remain_actual; |
1680 | } | 1693 | } |
1681 | 1694 | ||
1682 | GEM_BUG_ON(ring->tail > ring->size - bytes); | 1695 | GEM_BUG_ON(ring->emit > ring->size - bytes); |
1683 | cs = ring->vaddr + ring->tail; | 1696 | cs = ring->vaddr + ring->emit; |
1684 | ring->tail += bytes; | 1697 | ring->emit += bytes; |
1685 | ring->space -= bytes; | 1698 | ring->space -= bytes; |
1686 | GEM_BUG_ON(ring->space < 0); | 1699 | GEM_BUG_ON(ring->space < 0); |
1687 | 1700 | ||
@@ -1692,7 +1705,7 @@ u32 *intel_ring_begin(struct drm_i915_gem_request *req, int num_dwords) | |||
1692 | int intel_ring_cacheline_align(struct drm_i915_gem_request *req) | 1705 | int intel_ring_cacheline_align(struct drm_i915_gem_request *req) |
1693 | { | 1706 | { |
1694 | int num_dwords = | 1707 | int num_dwords = |
1695 | (req->ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t); | 1708 | (req->ring->emit & (CACHELINE_BYTES - 1)) / sizeof(uint32_t); |
1696 | u32 *cs; | 1709 | u32 *cs; |
1697 | 1710 | ||
1698 | if (num_dwords == 0) | 1711 | if (num_dwords == 0) |