aboutsummaryrefslogtreecommitdiffstats
path: root/drivers
diff options
context:
space:
mode:
authorOscar Mateo <oscar.mateo@intel.com>2014-05-22 09:13:33 -0400
committerDaniel Vetter <daniel.vetter@ffwll.ch>2014-05-22 17:01:05 -0400
commita4872ba6d01454dfeb251d96f623ab5d1b0666a4 (patch)
tree5358d393f0af1c8a422539c2b23a254794bc7583 /drivers
parentfe5b1886a78d92e0e4681e449725714e947dfc58 (diff)
drm/i915: s/intel_ring_buffer/intel_engine_cs
In the upcoming patches we plan to break the correlation between engine command streamers (a.k.a. rings) and ringbuffers, so it makes sense to refactor the code and make the change obvious. No functional changes. Signed-off-by: Oscar Mateo <oscar.mateo@intel.com> Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
Diffstat (limited to 'drivers')
-rw-r--r--drivers/gpu/drm/i915/i915_cmd_parser.c24
-rw-r--r--drivers/gpu/drm/i915/i915_debugfs.c16
-rw-r--r--drivers/gpu/drm/i915/i915_dma.c10
-rw-r--r--drivers/gpu/drm/i915/i915_drv.h40
-rw-r--r--drivers/gpu/drm/i915/i915_gem.c56
-rw-r--r--drivers/gpu/drm/i915/i915_gem_context.c12
-rw-r--r--drivers/gpu/drm/i915/i915_gem_execbuffer.c18
-rw-r--r--drivers/gpu/drm/i915/i915_gem_gtt.c18
-rw-r--r--drivers/gpu/drm/i915/i915_gem_gtt.h2
-rw-r--r--drivers/gpu/drm/i915/i915_gem_render_state.c2
-rw-r--r--drivers/gpu/drm/i915/i915_gpu_error.c6
-rw-r--r--drivers/gpu/drm/i915/i915_irq.c28
-rw-r--r--drivers/gpu/drm/i915/i915_trace.h26
-rw-r--r--drivers/gpu/drm/i915/intel_display.c18
-rw-r--r--drivers/gpu/drm/i915/intel_drv.h4
-rw-r--r--drivers/gpu/drm/i915/intel_overlay.c12
-rw-r--r--drivers/gpu/drm/i915/intel_pm.c10
-rw-r--r--drivers/gpu/drm/i915/intel_ringbuffer.c156
-rw-r--r--drivers/gpu/drm/i915/intel_ringbuffer.h72
19 files changed, 265 insertions, 265 deletions
diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c
index d3a5b74746d1..9d7954366bd2 100644
--- a/drivers/gpu/drm/i915/i915_cmd_parser.c
+++ b/drivers/gpu/drm/i915/i915_cmd_parser.c
@@ -498,7 +498,7 @@ static u32 gen7_blt_get_cmd_length_mask(u32 cmd_header)
498 return 0; 498 return 0;
499} 499}
500 500
501static bool validate_cmds_sorted(struct intel_ring_buffer *ring, 501static bool validate_cmds_sorted(struct intel_engine_cs *ring,
502 const struct drm_i915_cmd_table *cmd_tables, 502 const struct drm_i915_cmd_table *cmd_tables,
503 int cmd_table_count) 503 int cmd_table_count)
504{ 504{
@@ -552,7 +552,7 @@ static bool check_sorted(int ring_id, const u32 *reg_table, int reg_count)
552 return ret; 552 return ret;
553} 553}
554 554
555static bool validate_regs_sorted(struct intel_ring_buffer *ring) 555static bool validate_regs_sorted(struct intel_engine_cs *ring)
556{ 556{
557 return check_sorted(ring->id, ring->reg_table, ring->reg_count) && 557 return check_sorted(ring->id, ring->reg_table, ring->reg_count) &&
558 check_sorted(ring->id, ring->master_reg_table, 558 check_sorted(ring->id, ring->master_reg_table,
@@ -580,7 +580,7 @@ struct cmd_node {
580 */ 580 */
581#define CMD_HASH_MASK STD_MI_OPCODE_MASK 581#define CMD_HASH_MASK STD_MI_OPCODE_MASK
582 582
583static int init_hash_table(struct intel_ring_buffer *ring, 583static int init_hash_table(struct intel_engine_cs *ring,
584 const struct drm_i915_cmd_table *cmd_tables, 584 const struct drm_i915_cmd_table *cmd_tables,
585 int cmd_table_count) 585 int cmd_table_count)
586{ 586{
@@ -609,7 +609,7 @@ static int init_hash_table(struct intel_ring_buffer *ring,
609 return 0; 609 return 0;
610} 610}
611 611
612static void fini_hash_table(struct intel_ring_buffer *ring) 612static void fini_hash_table(struct intel_engine_cs *ring)
613{ 613{
614 struct hlist_node *tmp; 614 struct hlist_node *tmp;
615 struct cmd_node *desc_node; 615 struct cmd_node *desc_node;
@@ -626,12 +626,12 @@ static void fini_hash_table(struct intel_ring_buffer *ring)
626 * @ring: the ringbuffer to initialize 626 * @ring: the ringbuffer to initialize
627 * 627 *
628 * Optionally initializes fields related to batch buffer command parsing in the 628 * Optionally initializes fields related to batch buffer command parsing in the
629 * struct intel_ring_buffer based on whether the platform requires software 629 * struct intel_engine_cs based on whether the platform requires software
630 * command parsing. 630 * command parsing.
631 * 631 *
632 * Return: non-zero if initialization fails 632 * Return: non-zero if initialization fails
633 */ 633 */
634int i915_cmd_parser_init_ring(struct intel_ring_buffer *ring) 634int i915_cmd_parser_init_ring(struct intel_engine_cs *ring)
635{ 635{
636 const struct drm_i915_cmd_table *cmd_tables; 636 const struct drm_i915_cmd_table *cmd_tables;
637 int cmd_table_count; 637 int cmd_table_count;
@@ -725,7 +725,7 @@ int i915_cmd_parser_init_ring(struct intel_ring_buffer *ring)
725 * Releases any resources related to command parsing that may have been 725 * Releases any resources related to command parsing that may have been
726 * initialized for the specified ring. 726 * initialized for the specified ring.
727 */ 727 */
728void i915_cmd_parser_fini_ring(struct intel_ring_buffer *ring) 728void i915_cmd_parser_fini_ring(struct intel_engine_cs *ring)
729{ 729{
730 if (!ring->needs_cmd_parser) 730 if (!ring->needs_cmd_parser)
731 return; 731 return;
@@ -734,7 +734,7 @@ void i915_cmd_parser_fini_ring(struct intel_ring_buffer *ring)
734} 734}
735 735
736static const struct drm_i915_cmd_descriptor* 736static const struct drm_i915_cmd_descriptor*
737find_cmd_in_table(struct intel_ring_buffer *ring, 737find_cmd_in_table(struct intel_engine_cs *ring,
738 u32 cmd_header) 738 u32 cmd_header)
739{ 739{
740 struct cmd_node *desc_node; 740 struct cmd_node *desc_node;
@@ -761,7 +761,7 @@ find_cmd_in_table(struct intel_ring_buffer *ring,
761 * ring's default length encoding and returns default_desc. 761 * ring's default length encoding and returns default_desc.
762 */ 762 */
763static const struct drm_i915_cmd_descriptor* 763static const struct drm_i915_cmd_descriptor*
764find_cmd(struct intel_ring_buffer *ring, 764find_cmd(struct intel_engine_cs *ring,
765 u32 cmd_header, 765 u32 cmd_header,
766 struct drm_i915_cmd_descriptor *default_desc) 766 struct drm_i915_cmd_descriptor *default_desc)
767{ 767{
@@ -837,7 +837,7 @@ finish:
837 * 837 *
838 * Return: true if the ring requires software command parsing 838 * Return: true if the ring requires software command parsing
839 */ 839 */
840bool i915_needs_cmd_parser(struct intel_ring_buffer *ring) 840bool i915_needs_cmd_parser(struct intel_engine_cs *ring)
841{ 841{
842 struct drm_i915_private *dev_priv = ring->dev->dev_private; 842 struct drm_i915_private *dev_priv = ring->dev->dev_private;
843 843
@@ -855,7 +855,7 @@ bool i915_needs_cmd_parser(struct intel_ring_buffer *ring)
855 return (i915.enable_cmd_parser == 1); 855 return (i915.enable_cmd_parser == 1);
856} 856}
857 857
858static bool check_cmd(const struct intel_ring_buffer *ring, 858static bool check_cmd(const struct intel_engine_cs *ring,
859 const struct drm_i915_cmd_descriptor *desc, 859 const struct drm_i915_cmd_descriptor *desc,
860 const u32 *cmd, 860 const u32 *cmd,
861 const bool is_master, 861 const bool is_master,
@@ -957,7 +957,7 @@ static bool check_cmd(const struct intel_ring_buffer *ring,
957 * 957 *
958 * Return: non-zero if the parser finds violations or otherwise fails 958 * Return: non-zero if the parser finds violations or otherwise fails
959 */ 959 */
960int i915_parse_cmds(struct intel_ring_buffer *ring, 960int i915_parse_cmds(struct intel_engine_cs *ring,
961 struct drm_i915_gem_object *batch_obj, 961 struct drm_i915_gem_object *batch_obj,
962 u32 batch_start_offset, 962 u32 batch_start_offset,
963 bool is_master) 963 bool is_master)
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c
index 16bbdc7243df..833fdd2d1071 100644
--- a/drivers/gpu/drm/i915/i915_debugfs.c
+++ b/drivers/gpu/drm/i915/i915_debugfs.c
@@ -562,7 +562,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data)
562 struct drm_info_node *node = m->private; 562 struct drm_info_node *node = m->private;
563 struct drm_device *dev = node->minor->dev; 563 struct drm_device *dev = node->minor->dev;
564 struct drm_i915_private *dev_priv = dev->dev_private; 564 struct drm_i915_private *dev_priv = dev->dev_private;
565 struct intel_ring_buffer *ring; 565 struct intel_engine_cs *ring;
566 struct drm_i915_gem_request *gem_request; 566 struct drm_i915_gem_request *gem_request;
567 int ret, count, i; 567 int ret, count, i;
568 568
@@ -594,7 +594,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data)
594} 594}
595 595
596static void i915_ring_seqno_info(struct seq_file *m, 596static void i915_ring_seqno_info(struct seq_file *m,
597 struct intel_ring_buffer *ring) 597 struct intel_engine_cs *ring)
598{ 598{
599 if (ring->get_seqno) { 599 if (ring->get_seqno) {
600 seq_printf(m, "Current sequence (%s): %u\n", 600 seq_printf(m, "Current sequence (%s): %u\n",
@@ -607,7 +607,7 @@ static int i915_gem_seqno_info(struct seq_file *m, void *data)
607 struct drm_info_node *node = m->private; 607 struct drm_info_node *node = m->private;
608 struct drm_device *dev = node->minor->dev; 608 struct drm_device *dev = node->minor->dev;
609 struct drm_i915_private *dev_priv = dev->dev_private; 609 struct drm_i915_private *dev_priv = dev->dev_private;
610 struct intel_ring_buffer *ring; 610 struct intel_engine_cs *ring;
611 int ret, i; 611 int ret, i;
612 612
613 ret = mutex_lock_interruptible(&dev->struct_mutex); 613 ret = mutex_lock_interruptible(&dev->struct_mutex);
@@ -630,7 +630,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data)
630 struct drm_info_node *node = m->private; 630 struct drm_info_node *node = m->private;
631 struct drm_device *dev = node->minor->dev; 631 struct drm_device *dev = node->minor->dev;
632 struct drm_i915_private *dev_priv = dev->dev_private; 632 struct drm_i915_private *dev_priv = dev->dev_private;
633 struct intel_ring_buffer *ring; 633 struct intel_engine_cs *ring;
634 int ret, i, pipe; 634 int ret, i, pipe;
635 635
636 ret = mutex_lock_interruptible(&dev->struct_mutex); 636 ret = mutex_lock_interruptible(&dev->struct_mutex);
@@ -840,7 +840,7 @@ static int i915_hws_info(struct seq_file *m, void *data)
840 struct drm_info_node *node = m->private; 840 struct drm_info_node *node = m->private;
841 struct drm_device *dev = node->minor->dev; 841 struct drm_device *dev = node->minor->dev;
842 struct drm_i915_private *dev_priv = dev->dev_private; 842 struct drm_i915_private *dev_priv = dev->dev_private;
843 struct intel_ring_buffer *ring; 843 struct intel_engine_cs *ring;
844 const u32 *hws; 844 const u32 *hws;
845 int i; 845 int i;
846 846
@@ -1717,7 +1717,7 @@ static int i915_context_status(struct seq_file *m, void *unused)
1717 struct drm_info_node *node = m->private; 1717 struct drm_info_node *node = m->private;
1718 struct drm_device *dev = node->minor->dev; 1718 struct drm_device *dev = node->minor->dev;
1719 struct drm_i915_private *dev_priv = dev->dev_private; 1719 struct drm_i915_private *dev_priv = dev->dev_private;
1720 struct intel_ring_buffer *ring; 1720 struct intel_engine_cs *ring;
1721 struct i915_hw_context *ctx; 1721 struct i915_hw_context *ctx;
1722 int ret, i; 1722 int ret, i;
1723 1723
@@ -1866,7 +1866,7 @@ static int per_file_ctx(int id, void *ptr, void *data)
1866static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev) 1866static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev)
1867{ 1867{
1868 struct drm_i915_private *dev_priv = dev->dev_private; 1868 struct drm_i915_private *dev_priv = dev->dev_private;
1869 struct intel_ring_buffer *ring; 1869 struct intel_engine_cs *ring;
1870 struct i915_hw_ppgtt *ppgtt = dev_priv->mm.aliasing_ppgtt; 1870 struct i915_hw_ppgtt *ppgtt = dev_priv->mm.aliasing_ppgtt;
1871 int unused, i; 1871 int unused, i;
1872 1872
@@ -1890,7 +1890,7 @@ static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev)
1890static void gen6_ppgtt_info(struct seq_file *m, struct drm_device *dev) 1890static void gen6_ppgtt_info(struct seq_file *m, struct drm_device *dev)
1891{ 1891{
1892 struct drm_i915_private *dev_priv = dev->dev_private; 1892 struct drm_i915_private *dev_priv = dev->dev_private;
1893 struct intel_ring_buffer *ring; 1893 struct intel_engine_cs *ring;
1894 struct drm_file *file; 1894 struct drm_file *file;
1895 int i; 1895 int i;
1896 1896
diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c
index ead67c0c4109..13784fefa67d 100644
--- a/drivers/gpu/drm/i915/i915_dma.c
+++ b/drivers/gpu/drm/i915/i915_dma.c
@@ -120,7 +120,7 @@ static void i915_write_hws_pga(struct drm_device *dev)
120static void i915_free_hws(struct drm_device *dev) 120static void i915_free_hws(struct drm_device *dev)
121{ 121{
122 struct drm_i915_private *dev_priv = dev->dev_private; 122 struct drm_i915_private *dev_priv = dev->dev_private;
123 struct intel_ring_buffer *ring = LP_RING(dev_priv); 123 struct intel_engine_cs *ring = LP_RING(dev_priv);
124 124
125 if (dev_priv->status_page_dmah) { 125 if (dev_priv->status_page_dmah) {
126 drm_pci_free(dev, dev_priv->status_page_dmah); 126 drm_pci_free(dev, dev_priv->status_page_dmah);
@@ -140,7 +140,7 @@ void i915_kernel_lost_context(struct drm_device * dev)
140{ 140{
141 struct drm_i915_private *dev_priv = dev->dev_private; 141 struct drm_i915_private *dev_priv = dev->dev_private;
142 struct drm_i915_master_private *master_priv; 142 struct drm_i915_master_private *master_priv;
143 struct intel_ring_buffer *ring = LP_RING(dev_priv); 143 struct intel_engine_cs *ring = LP_RING(dev_priv);
144 144
145 /* 145 /*
146 * We should never lose context on the ring with modesetting 146 * We should never lose context on the ring with modesetting
@@ -235,7 +235,7 @@ static int i915_initialize(struct drm_device * dev, drm_i915_init_t * init)
235static int i915_dma_resume(struct drm_device * dev) 235static int i915_dma_resume(struct drm_device * dev)
236{ 236{
237 struct drm_i915_private *dev_priv = dev->dev_private; 237 struct drm_i915_private *dev_priv = dev->dev_private;
238 struct intel_ring_buffer *ring = LP_RING(dev_priv); 238 struct intel_engine_cs *ring = LP_RING(dev_priv);
239 239
240 DRM_DEBUG_DRIVER("%s\n", __func__); 240 DRM_DEBUG_DRIVER("%s\n", __func__);
241 241
@@ -783,7 +783,7 @@ static int i915_wait_irq(struct drm_device * dev, int irq_nr)
783 struct drm_i915_private *dev_priv = dev->dev_private; 783 struct drm_i915_private *dev_priv = dev->dev_private;
784 struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv; 784 struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
785 int ret = 0; 785 int ret = 0;
786 struct intel_ring_buffer *ring = LP_RING(dev_priv); 786 struct intel_engine_cs *ring = LP_RING(dev_priv);
787 787
788 DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr, 788 DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr,
789 READ_BREADCRUMB(dev_priv)); 789 READ_BREADCRUMB(dev_priv));
@@ -1074,7 +1074,7 @@ static int i915_set_status_page(struct drm_device *dev, void *data,
1074{ 1074{
1075 struct drm_i915_private *dev_priv = dev->dev_private; 1075 struct drm_i915_private *dev_priv = dev->dev_private;
1076 drm_i915_hws_addr_t *hws = data; 1076 drm_i915_hws_addr_t *hws = data;
1077 struct intel_ring_buffer *ring; 1077 struct intel_engine_cs *ring;
1078 1078
1079 if (drm_core_check_feature(dev, DRIVER_MODESET)) 1079 if (drm_core_check_feature(dev, DRIVER_MODESET))
1080 return -ENODEV; 1080 return -ENODEV;
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h
index de899b30ae0c..ca5bd572a34d 100644
--- a/drivers/gpu/drm/i915/i915_drv.h
+++ b/drivers/gpu/drm/i915/i915_drv.h
@@ -470,7 +470,7 @@ struct drm_i915_display_funcs {
470 int (*queue_flip)(struct drm_device *dev, struct drm_crtc *crtc, 470 int (*queue_flip)(struct drm_device *dev, struct drm_crtc *crtc,
471 struct drm_framebuffer *fb, 471 struct drm_framebuffer *fb,
472 struct drm_i915_gem_object *obj, 472 struct drm_i915_gem_object *obj,
473 struct intel_ring_buffer *ring, 473 struct intel_engine_cs *ring,
474 uint32_t flags); 474 uint32_t flags);
475 void (*update_primary_plane)(struct drm_crtc *crtc, 475 void (*update_primary_plane)(struct drm_crtc *crtc,
476 struct drm_framebuffer *fb, 476 struct drm_framebuffer *fb,
@@ -605,7 +605,7 @@ struct i915_hw_context {
605 bool is_initialized; 605 bool is_initialized;
606 uint8_t remap_slice; 606 uint8_t remap_slice;
607 struct drm_i915_file_private *file_priv; 607 struct drm_i915_file_private *file_priv;
608 struct intel_ring_buffer *last_ring; 608 struct intel_engine_cs *last_ring;
609 struct drm_i915_gem_object *obj; 609 struct drm_i915_gem_object *obj;
610 struct i915_ctx_hang_stats hang_stats; 610 struct i915_ctx_hang_stats hang_stats;
611 struct i915_address_space *vm; 611 struct i915_address_space *vm;
@@ -1372,7 +1372,7 @@ struct drm_i915_private {
1372 wait_queue_head_t gmbus_wait_queue; 1372 wait_queue_head_t gmbus_wait_queue;
1373 1373
1374 struct pci_dev *bridge_dev; 1374 struct pci_dev *bridge_dev;
1375 struct intel_ring_buffer ring[I915_NUM_RINGS]; 1375 struct intel_engine_cs ring[I915_NUM_RINGS];
1376 uint32_t last_seqno, next_seqno; 1376 uint32_t last_seqno, next_seqno;
1377 1377
1378 drm_dma_handle_t *status_page_dmah; 1378 drm_dma_handle_t *status_page_dmah;
@@ -1690,7 +1690,7 @@ struct drm_i915_gem_object {
1690 void *dma_buf_vmapping; 1690 void *dma_buf_vmapping;
1691 int vmapping_count; 1691 int vmapping_count;
1692 1692
1693 struct intel_ring_buffer *ring; 1693 struct intel_engine_cs *ring;
1694 1694
1695 /** Breadcrumb of last rendering to the buffer. */ 1695 /** Breadcrumb of last rendering to the buffer. */
1696 uint32_t last_read_seqno; 1696 uint32_t last_read_seqno;
@@ -1741,7 +1741,7 @@ struct drm_i915_gem_object {
1741 */ 1741 */
1742struct drm_i915_gem_request { 1742struct drm_i915_gem_request {
1743 /** On Which ring this request was generated */ 1743 /** On Which ring this request was generated */
1744 struct intel_ring_buffer *ring; 1744 struct intel_engine_cs *ring;
1745 1745
1746 /** GEM sequence number associated with this request. */ 1746 /** GEM sequence number associated with this request. */
1747 uint32_t seqno; 1747 uint32_t seqno;
@@ -1782,7 +1782,7 @@ struct drm_i915_file_private {
1782 1782
1783 struct i915_hw_context *private_default_ctx; 1783 struct i915_hw_context *private_default_ctx;
1784 atomic_t rps_wait_boost; 1784 atomic_t rps_wait_boost;
1785 struct intel_ring_buffer *bsd_ring; 1785 struct intel_engine_cs *bsd_ring;
1786}; 1786};
1787 1787
1788/* 1788/*
@@ -2209,9 +2209,9 @@ static inline void i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj)
2209 2209
2210int __must_check i915_mutex_lock_interruptible(struct drm_device *dev); 2210int __must_check i915_mutex_lock_interruptible(struct drm_device *dev);
2211int i915_gem_object_sync(struct drm_i915_gem_object *obj, 2211int i915_gem_object_sync(struct drm_i915_gem_object *obj,
2212 struct intel_ring_buffer *to); 2212 struct intel_engine_cs *to);
2213void i915_vma_move_to_active(struct i915_vma *vma, 2213void i915_vma_move_to_active(struct i915_vma *vma,
2214 struct intel_ring_buffer *ring); 2214 struct intel_engine_cs *ring);
2215int i915_gem_dumb_create(struct drm_file *file_priv, 2215int i915_gem_dumb_create(struct drm_file *file_priv,
2216 struct drm_device *dev, 2216 struct drm_device *dev,
2217 struct drm_mode_create_dumb *args); 2217 struct drm_mode_create_dumb *args);
@@ -2235,10 +2235,10 @@ bool i915_gem_object_pin_fence(struct drm_i915_gem_object *obj);
2235void i915_gem_object_unpin_fence(struct drm_i915_gem_object *obj); 2235void i915_gem_object_unpin_fence(struct drm_i915_gem_object *obj);
2236 2236
2237struct drm_i915_gem_request * 2237struct drm_i915_gem_request *
2238i915_gem_find_active_request(struct intel_ring_buffer *ring); 2238i915_gem_find_active_request(struct intel_engine_cs *ring);
2239 2239
2240bool i915_gem_retire_requests(struct drm_device *dev); 2240bool i915_gem_retire_requests(struct drm_device *dev);
2241void i915_gem_retire_requests_ring(struct intel_ring_buffer *ring); 2241void i915_gem_retire_requests_ring(struct intel_engine_cs *ring);
2242int __must_check i915_gem_check_wedge(struct i915_gpu_error *error, 2242int __must_check i915_gem_check_wedge(struct i915_gpu_error *error,
2243 bool interruptible); 2243 bool interruptible);
2244static inline bool i915_reset_in_progress(struct i915_gpu_error *error) 2244static inline bool i915_reset_in_progress(struct i915_gpu_error *error)
@@ -2274,18 +2274,18 @@ bool i915_gem_clflush_object(struct drm_i915_gem_object *obj, bool force);
2274int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj); 2274int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj);
2275int __must_check i915_gem_init(struct drm_device *dev); 2275int __must_check i915_gem_init(struct drm_device *dev);
2276int __must_check i915_gem_init_hw(struct drm_device *dev); 2276int __must_check i915_gem_init_hw(struct drm_device *dev);
2277int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice); 2277int i915_gem_l3_remap(struct intel_engine_cs *ring, int slice);
2278void i915_gem_init_swizzling(struct drm_device *dev); 2278void i915_gem_init_swizzling(struct drm_device *dev);
2279void i915_gem_cleanup_ringbuffer(struct drm_device *dev); 2279void i915_gem_cleanup_ringbuffer(struct drm_device *dev);
2280int __must_check i915_gpu_idle(struct drm_device *dev); 2280int __must_check i915_gpu_idle(struct drm_device *dev);
2281int __must_check i915_gem_suspend(struct drm_device *dev); 2281int __must_check i915_gem_suspend(struct drm_device *dev);
2282int __i915_add_request(struct intel_ring_buffer *ring, 2282int __i915_add_request(struct intel_engine_cs *ring,
2283 struct drm_file *file, 2283 struct drm_file *file,
2284 struct drm_i915_gem_object *batch_obj, 2284 struct drm_i915_gem_object *batch_obj,
2285 u32 *seqno); 2285 u32 *seqno);
2286#define i915_add_request(ring, seqno) \ 2286#define i915_add_request(ring, seqno) \
2287 __i915_add_request(ring, NULL, NULL, seqno) 2287 __i915_add_request(ring, NULL, NULL, seqno)
2288int __must_check i915_wait_seqno(struct intel_ring_buffer *ring, 2288int __must_check i915_wait_seqno(struct intel_engine_cs *ring,
2289 uint32_t seqno); 2289 uint32_t seqno);
2290int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); 2290int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
2291int __must_check 2291int __must_check
@@ -2296,7 +2296,7 @@ i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write);
2296int __must_check 2296int __must_check
2297i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, 2297i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
2298 u32 alignment, 2298 u32 alignment,
2299 struct intel_ring_buffer *pipelined); 2299 struct intel_engine_cs *pipelined);
2300void i915_gem_object_unpin_from_display_plane(struct drm_i915_gem_object *obj); 2300void i915_gem_object_unpin_from_display_plane(struct drm_i915_gem_object *obj);
2301int i915_gem_attach_phys_object(struct drm_device *dev, 2301int i915_gem_attach_phys_object(struct drm_device *dev,
2302 struct drm_i915_gem_object *obj, 2302 struct drm_i915_gem_object *obj,
@@ -2398,7 +2398,7 @@ void i915_gem_context_reset(struct drm_device *dev);
2398int i915_gem_context_open(struct drm_device *dev, struct drm_file *file); 2398int i915_gem_context_open(struct drm_device *dev, struct drm_file *file);
2399int i915_gem_context_enable(struct drm_i915_private *dev_priv); 2399int i915_gem_context_enable(struct drm_i915_private *dev_priv);
2400void i915_gem_context_close(struct drm_device *dev, struct drm_file *file); 2400void i915_gem_context_close(struct drm_device *dev, struct drm_file *file);
2401int i915_switch_context(struct intel_ring_buffer *ring, 2401int i915_switch_context(struct intel_engine_cs *ring,
2402 struct i915_hw_context *to); 2402 struct i915_hw_context *to);
2403struct i915_hw_context * 2403struct i915_hw_context *
2404i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id); 2404i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id);
@@ -2424,7 +2424,7 @@ int i915_gem_context_destroy_ioctl(struct drm_device *dev, void *data,
2424 struct drm_file *file); 2424 struct drm_file *file);
2425 2425
2426/* i915_gem_render_state.c */ 2426/* i915_gem_render_state.c */
2427int i915_gem_render_state_init(struct intel_ring_buffer *ring); 2427int i915_gem_render_state_init(struct intel_engine_cs *ring);
2428/* i915_gem_evict.c */ 2428/* i915_gem_evict.c */
2429int __must_check i915_gem_evict_something(struct drm_device *dev, 2429int __must_check i915_gem_evict_something(struct drm_device *dev,
2430 struct i915_address_space *vm, 2430 struct i915_address_space *vm,
@@ -2509,10 +2509,10 @@ const char *i915_cache_level_str(int type);
2509 2509
2510/* i915_cmd_parser.c */ 2510/* i915_cmd_parser.c */
2511int i915_cmd_parser_get_version(void); 2511int i915_cmd_parser_get_version(void);
2512int i915_cmd_parser_init_ring(struct intel_ring_buffer *ring); 2512int i915_cmd_parser_init_ring(struct intel_engine_cs *ring);
2513void i915_cmd_parser_fini_ring(struct intel_ring_buffer *ring); 2513void i915_cmd_parser_fini_ring(struct intel_engine_cs *ring);
2514bool i915_needs_cmd_parser(struct intel_ring_buffer *ring); 2514bool i915_needs_cmd_parser(struct intel_engine_cs *ring);
2515int i915_parse_cmds(struct intel_ring_buffer *ring, 2515int i915_parse_cmds(struct intel_engine_cs *ring,
2516 struct drm_i915_gem_object *batch_obj, 2516 struct drm_i915_gem_object *batch_obj,
2517 u32 batch_start_offset, 2517 u32 batch_start_offset,
2518 bool is_master); 2518 bool is_master);
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c
index 6d64cff37fe2..382eae74eebf 100644
--- a/drivers/gpu/drm/i915/i915_gem.c
+++ b/drivers/gpu/drm/i915/i915_gem.c
@@ -979,7 +979,7 @@ i915_gem_check_wedge(struct i915_gpu_error *error,
979 * equal. 979 * equal.
980 */ 980 */
981static int 981static int
982i915_gem_check_olr(struct intel_ring_buffer *ring, u32 seqno) 982i915_gem_check_olr(struct intel_engine_cs *ring, u32 seqno)
983{ 983{
984 int ret; 984 int ret;
985 985
@@ -998,7 +998,7 @@ static void fake_irq(unsigned long data)
998} 998}
999 999
1000static bool missed_irq(struct drm_i915_private *dev_priv, 1000static bool missed_irq(struct drm_i915_private *dev_priv,
1001 struct intel_ring_buffer *ring) 1001 struct intel_engine_cs *ring)
1002{ 1002{
1003 return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings); 1003 return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings);
1004} 1004}
@@ -1029,7 +1029,7 @@ static bool can_wait_boost(struct drm_i915_file_private *file_priv)
1029 * Returns 0 if the seqno was found within the alloted time. Else returns the 1029 * Returns 0 if the seqno was found within the alloted time. Else returns the
1030 * errno with remaining time filled in timeout argument. 1030 * errno with remaining time filled in timeout argument.
1031 */ 1031 */
1032static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno, 1032static int __wait_seqno(struct intel_engine_cs *ring, u32 seqno,
1033 unsigned reset_counter, 1033 unsigned reset_counter,
1034 bool interruptible, 1034 bool interruptible,
1035 struct timespec *timeout, 1035 struct timespec *timeout,
@@ -1136,7 +1136,7 @@ static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno,
1136 * request and object lists appropriately for that event. 1136 * request and object lists appropriately for that event.
1137 */ 1137 */
1138int 1138int
1139i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno) 1139i915_wait_seqno(struct intel_engine_cs *ring, uint32_t seqno)
1140{ 1140{
1141 struct drm_device *dev = ring->dev; 1141 struct drm_device *dev = ring->dev;
1142 struct drm_i915_private *dev_priv = dev->dev_private; 1142 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1161,7 +1161,7 @@ i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno)
1161 1161
1162static int 1162static int
1163i915_gem_object_wait_rendering__tail(struct drm_i915_gem_object *obj, 1163i915_gem_object_wait_rendering__tail(struct drm_i915_gem_object *obj,
1164 struct intel_ring_buffer *ring) 1164 struct intel_engine_cs *ring)
1165{ 1165{
1166 if (!obj->active) 1166 if (!obj->active)
1167 return 0; 1167 return 0;
@@ -1186,7 +1186,7 @@ static __must_check int
1186i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj, 1186i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj,
1187 bool readonly) 1187 bool readonly)
1188{ 1188{
1189 struct intel_ring_buffer *ring = obj->ring; 1189 struct intel_engine_cs *ring = obj->ring;
1190 u32 seqno; 1190 u32 seqno;
1191 int ret; 1191 int ret;
1192 1192
@@ -1211,7 +1211,7 @@ i915_gem_object_wait_rendering__nonblocking(struct drm_i915_gem_object *obj,
1211{ 1211{
1212 struct drm_device *dev = obj->base.dev; 1212 struct drm_device *dev = obj->base.dev;
1213 struct drm_i915_private *dev_priv = dev->dev_private; 1213 struct drm_i915_private *dev_priv = dev->dev_private;
1214 struct intel_ring_buffer *ring = obj->ring; 1214 struct intel_engine_cs *ring = obj->ring;
1215 unsigned reset_counter; 1215 unsigned reset_counter;
1216 u32 seqno; 1216 u32 seqno;
1217 int ret; 1217 int ret;
@@ -2040,7 +2040,7 @@ i915_gem_object_get_pages(struct drm_i915_gem_object *obj)
2040 2040
2041static void 2041static void
2042i915_gem_object_move_to_active(struct drm_i915_gem_object *obj, 2042i915_gem_object_move_to_active(struct drm_i915_gem_object *obj,
2043 struct intel_ring_buffer *ring) 2043 struct intel_engine_cs *ring)
2044{ 2044{
2045 struct drm_device *dev = obj->base.dev; 2045 struct drm_device *dev = obj->base.dev;
2046 struct drm_i915_private *dev_priv = dev->dev_private; 2046 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -2078,7 +2078,7 @@ i915_gem_object_move_to_active(struct drm_i915_gem_object *obj,
2078} 2078}
2079 2079
2080void i915_vma_move_to_active(struct i915_vma *vma, 2080void i915_vma_move_to_active(struct i915_vma *vma,
2081 struct intel_ring_buffer *ring) 2081 struct intel_engine_cs *ring)
2082{ 2082{
2083 list_move_tail(&vma->mm_list, &vma->vm->active_list); 2083 list_move_tail(&vma->mm_list, &vma->vm->active_list);
2084 return i915_gem_object_move_to_active(vma->obj, ring); 2084 return i915_gem_object_move_to_active(vma->obj, ring);
@@ -2119,7 +2119,7 @@ i915_gem_object_move_to_inactive(struct drm_i915_gem_object *obj)
2119static void 2119static void
2120i915_gem_object_retire(struct drm_i915_gem_object *obj) 2120i915_gem_object_retire(struct drm_i915_gem_object *obj)
2121{ 2121{
2122 struct intel_ring_buffer *ring = obj->ring; 2122 struct intel_engine_cs *ring = obj->ring;
2123 2123
2124 if (ring == NULL) 2124 if (ring == NULL)
2125 return; 2125 return;
@@ -2133,7 +2133,7 @@ static int
2133i915_gem_init_seqno(struct drm_device *dev, u32 seqno) 2133i915_gem_init_seqno(struct drm_device *dev, u32 seqno)
2134{ 2134{
2135 struct drm_i915_private *dev_priv = dev->dev_private; 2135 struct drm_i915_private *dev_priv = dev->dev_private;
2136 struct intel_ring_buffer *ring; 2136 struct intel_engine_cs *ring;
2137 int ret, i, j; 2137 int ret, i, j;
2138 2138
2139 /* Carefully retire all requests without writing to the rings */ 2139 /* Carefully retire all requests without writing to the rings */
@@ -2199,7 +2199,7 @@ i915_gem_get_seqno(struct drm_device *dev, u32 *seqno)
2199 return 0; 2199 return 0;
2200} 2200}
2201 2201
2202int __i915_add_request(struct intel_ring_buffer *ring, 2202int __i915_add_request(struct intel_engine_cs *ring,
2203 struct drm_file *file, 2203 struct drm_file *file,
2204 struct drm_i915_gem_object *obj, 2204 struct drm_i915_gem_object *obj,
2205 u32 *out_seqno) 2205 u32 *out_seqno)
@@ -2359,7 +2359,7 @@ static void i915_gem_free_request(struct drm_i915_gem_request *request)
2359} 2359}
2360 2360
2361struct drm_i915_gem_request * 2361struct drm_i915_gem_request *
2362i915_gem_find_active_request(struct intel_ring_buffer *ring) 2362i915_gem_find_active_request(struct intel_engine_cs *ring)
2363{ 2363{
2364 struct drm_i915_gem_request *request; 2364 struct drm_i915_gem_request *request;
2365 u32 completed_seqno; 2365 u32 completed_seqno;
@@ -2377,7 +2377,7 @@ i915_gem_find_active_request(struct intel_ring_buffer *ring)
2377} 2377}
2378 2378
2379static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv, 2379static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
2380 struct intel_ring_buffer *ring) 2380 struct intel_engine_cs *ring)
2381{ 2381{
2382 struct drm_i915_gem_request *request; 2382 struct drm_i915_gem_request *request;
2383 bool ring_hung; 2383 bool ring_hung;
@@ -2396,7 +2396,7 @@ static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv,
2396} 2396}
2397 2397
2398static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv, 2398static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv,
2399 struct intel_ring_buffer *ring) 2399 struct intel_engine_cs *ring)
2400{ 2400{
2401 while (!list_empty(&ring->active_list)) { 2401 while (!list_empty(&ring->active_list)) {
2402 struct drm_i915_gem_object *obj; 2402 struct drm_i915_gem_object *obj;
@@ -2455,7 +2455,7 @@ void i915_gem_restore_fences(struct drm_device *dev)
2455void i915_gem_reset(struct drm_device *dev) 2455void i915_gem_reset(struct drm_device *dev)
2456{ 2456{
2457 struct drm_i915_private *dev_priv = dev->dev_private; 2457 struct drm_i915_private *dev_priv = dev->dev_private;
2458 struct intel_ring_buffer *ring; 2458 struct intel_engine_cs *ring;
2459 int i; 2459 int i;
2460 2460
2461 /* 2461 /*
@@ -2478,7 +2478,7 @@ void i915_gem_reset(struct drm_device *dev)
2478 * This function clears the request list as sequence numbers are passed. 2478 * This function clears the request list as sequence numbers are passed.
2479 */ 2479 */
2480void 2480void
2481i915_gem_retire_requests_ring(struct intel_ring_buffer *ring) 2481i915_gem_retire_requests_ring(struct intel_engine_cs *ring)
2482{ 2482{
2483 uint32_t seqno; 2483 uint32_t seqno;
2484 2484
@@ -2541,7 +2541,7 @@ bool
2541i915_gem_retire_requests(struct drm_device *dev) 2541i915_gem_retire_requests(struct drm_device *dev)
2542{ 2542{
2543 struct drm_i915_private *dev_priv = dev->dev_private; 2543 struct drm_i915_private *dev_priv = dev->dev_private;
2544 struct intel_ring_buffer *ring; 2544 struct intel_engine_cs *ring;
2545 bool idle = true; 2545 bool idle = true;
2546 int i; 2546 int i;
2547 2547
@@ -2635,7 +2635,7 @@ i915_gem_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
2635 struct drm_i915_private *dev_priv = dev->dev_private; 2635 struct drm_i915_private *dev_priv = dev->dev_private;
2636 struct drm_i915_gem_wait *args = data; 2636 struct drm_i915_gem_wait *args = data;
2637 struct drm_i915_gem_object *obj; 2637 struct drm_i915_gem_object *obj;
2638 struct intel_ring_buffer *ring = NULL; 2638 struct intel_engine_cs *ring = NULL;
2639 struct timespec timeout_stack, *timeout = NULL; 2639 struct timespec timeout_stack, *timeout = NULL;
2640 unsigned reset_counter; 2640 unsigned reset_counter;
2641 u32 seqno = 0; 2641 u32 seqno = 0;
@@ -2706,9 +2706,9 @@ out:
2706 */ 2706 */
2707int 2707int
2708i915_gem_object_sync(struct drm_i915_gem_object *obj, 2708i915_gem_object_sync(struct drm_i915_gem_object *obj,
2709 struct intel_ring_buffer *to) 2709 struct intel_engine_cs *to)
2710{ 2710{
2711 struct intel_ring_buffer *from = obj->ring; 2711 struct intel_engine_cs *from = obj->ring;
2712 u32 seqno; 2712 u32 seqno;
2713 int ret, idx; 2713 int ret, idx;
2714 2714
@@ -2831,7 +2831,7 @@ int i915_vma_unbind(struct i915_vma *vma)
2831int i915_gpu_idle(struct drm_device *dev) 2831int i915_gpu_idle(struct drm_device *dev)
2832{ 2832{
2833 struct drm_i915_private *dev_priv = dev->dev_private; 2833 struct drm_i915_private *dev_priv = dev->dev_private;
2834 struct intel_ring_buffer *ring; 2834 struct intel_engine_cs *ring;
2835 int ret, i; 2835 int ret, i;
2836 2836
2837 /* Flush everything onto the inactive list. */ 2837 /* Flush everything onto the inactive list. */
@@ -3702,7 +3702,7 @@ static bool is_pin_display(struct drm_i915_gem_object *obj)
3702int 3702int
3703i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, 3703i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
3704 u32 alignment, 3704 u32 alignment,
3705 struct intel_ring_buffer *pipelined) 3705 struct intel_engine_cs *pipelined)
3706{ 3706{
3707 u32 old_read_domains, old_write_domain; 3707 u32 old_read_domains, old_write_domain;
3708 bool was_pin_display; 3708 bool was_pin_display;
@@ -3858,7 +3858,7 @@ i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file)
3858 struct drm_i915_file_private *file_priv = file->driver_priv; 3858 struct drm_i915_file_private *file_priv = file->driver_priv;
3859 unsigned long recent_enough = jiffies - msecs_to_jiffies(20); 3859 unsigned long recent_enough = jiffies - msecs_to_jiffies(20);
3860 struct drm_i915_gem_request *request; 3860 struct drm_i915_gem_request *request;
3861 struct intel_ring_buffer *ring = NULL; 3861 struct intel_engine_cs *ring = NULL;
3862 unsigned reset_counter; 3862 unsigned reset_counter;
3863 u32 seqno = 0; 3863 u32 seqno = 0;
3864 int ret; 3864 int ret;
@@ -4359,7 +4359,7 @@ static void
4359i915_gem_stop_ringbuffers(struct drm_device *dev) 4359i915_gem_stop_ringbuffers(struct drm_device *dev)
4360{ 4360{
4361 struct drm_i915_private *dev_priv = dev->dev_private; 4361 struct drm_i915_private *dev_priv = dev->dev_private;
4362 struct intel_ring_buffer *ring; 4362 struct intel_engine_cs *ring;
4363 int i; 4363 int i;
4364 4364
4365 for_each_ring(ring, dev_priv, i) 4365 for_each_ring(ring, dev_priv, i)
@@ -4408,7 +4408,7 @@ err:
4408 return ret; 4408 return ret;
4409} 4409}
4410 4410
4411int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice) 4411int i915_gem_l3_remap(struct intel_engine_cs *ring, int slice)
4412{ 4412{
4413 struct drm_device *dev = ring->dev; 4413 struct drm_device *dev = ring->dev;
4414 struct drm_i915_private *dev_priv = dev->dev_private; 4414 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -4632,7 +4632,7 @@ void
4632i915_gem_cleanup_ringbuffer(struct drm_device *dev) 4632i915_gem_cleanup_ringbuffer(struct drm_device *dev)
4633{ 4633{
4634 struct drm_i915_private *dev_priv = dev->dev_private; 4634 struct drm_i915_private *dev_priv = dev->dev_private;
4635 struct intel_ring_buffer *ring; 4635 struct intel_engine_cs *ring;
4636 int i; 4636 int i;
4637 4637
4638 for_each_ring(ring, dev_priv, i) 4638 for_each_ring(ring, dev_priv, i)
@@ -4708,7 +4708,7 @@ i915_gem_lastclose(struct drm_device *dev)
4708} 4708}
4709 4709
4710static void 4710static void
4711init_ring_lists(struct intel_ring_buffer *ring) 4711init_ring_lists(struct intel_engine_cs *ring)
4712{ 4712{
4713 INIT_LIST_HEAD(&ring->active_list); 4713 INIT_LIST_HEAD(&ring->active_list);
4714 INIT_LIST_HEAD(&ring->request_list); 4714 INIT_LIST_HEAD(&ring->request_list);
diff --git a/drivers/gpu/drm/i915/i915_gem_context.c b/drivers/gpu/drm/i915/i915_gem_context.c
index 6a2d847a637a..8e57e1d63fb1 100644
--- a/drivers/gpu/drm/i915/i915_gem_context.c
+++ b/drivers/gpu/drm/i915/i915_gem_context.c
@@ -364,7 +364,7 @@ void i915_gem_context_reset(struct drm_device *dev)
364 /* Prevent the hardware from restoring the last context (which hung) on 364 /* Prevent the hardware from restoring the last context (which hung) on
365 * the next switch */ 365 * the next switch */
366 for (i = 0; i < I915_NUM_RINGS; i++) { 366 for (i = 0; i < I915_NUM_RINGS; i++) {
367 struct intel_ring_buffer *ring = &dev_priv->ring[i]; 367 struct intel_engine_cs *ring = &dev_priv->ring[i];
368 struct i915_hw_context *dctx = ring->default_context; 368 struct i915_hw_context *dctx = ring->default_context;
369 369
370 /* Do a fake switch to the default context */ 370 /* Do a fake switch to the default context */
@@ -454,7 +454,7 @@ void i915_gem_context_fini(struct drm_device *dev)
454 } 454 }
455 455
456 for (i = 0; i < I915_NUM_RINGS; i++) { 456 for (i = 0; i < I915_NUM_RINGS; i++) {
457 struct intel_ring_buffer *ring = &dev_priv->ring[i]; 457 struct intel_engine_cs *ring = &dev_priv->ring[i];
458 458
459 if (ring->last_context) 459 if (ring->last_context)
460 i915_gem_context_unreference(ring->last_context); 460 i915_gem_context_unreference(ring->last_context);
@@ -468,7 +468,7 @@ void i915_gem_context_fini(struct drm_device *dev)
468 468
469int i915_gem_context_enable(struct drm_i915_private *dev_priv) 469int i915_gem_context_enable(struct drm_i915_private *dev_priv)
470{ 470{
471 struct intel_ring_buffer *ring; 471 struct intel_engine_cs *ring;
472 int ret, i; 472 int ret, i;
473 473
474 /* This is the only place the aliasing PPGTT gets enabled, which means 474 /* This is the only place the aliasing PPGTT gets enabled, which means
@@ -547,7 +547,7 @@ i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id)
547} 547}
548 548
549static inline int 549static inline int
550mi_set_context(struct intel_ring_buffer *ring, 550mi_set_context(struct intel_engine_cs *ring,
551 struct i915_hw_context *new_context, 551 struct i915_hw_context *new_context,
552 u32 hw_flags) 552 u32 hw_flags)
553{ 553{
@@ -597,7 +597,7 @@ mi_set_context(struct intel_ring_buffer *ring,
597 return ret; 597 return ret;
598} 598}
599 599
600static int do_switch(struct intel_ring_buffer *ring, 600static int do_switch(struct intel_engine_cs *ring,
601 struct i915_hw_context *to) 601 struct i915_hw_context *to)
602{ 602{
603 struct drm_i915_private *dev_priv = ring->dev->dev_private; 603 struct drm_i915_private *dev_priv = ring->dev->dev_private;
@@ -733,7 +733,7 @@ unpin_out:
733 * it will have a refoucnt > 1. This allows us to destroy the context abstract 733 * it will have a refoucnt > 1. This allows us to destroy the context abstract
734 * object while letting the normal object tracking destroy the backing BO. 734 * object while letting the normal object tracking destroy the backing BO.
735 */ 735 */
736int i915_switch_context(struct intel_ring_buffer *ring, 736int i915_switch_context(struct intel_engine_cs *ring,
737 struct i915_hw_context *to) 737 struct i915_hw_context *to)
738{ 738{
739 struct drm_i915_private *dev_priv = ring->dev->dev_private; 739 struct drm_i915_private *dev_priv = ring->dev->dev_private;
diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
index de2fd90bdd0f..7829b9074ee4 100644
--- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
@@ -541,7 +541,7 @@ need_reloc_mappable(struct i915_vma *vma)
541 541
542static int 542static int
543i915_gem_execbuffer_reserve_vma(struct i915_vma *vma, 543i915_gem_execbuffer_reserve_vma(struct i915_vma *vma,
544 struct intel_ring_buffer *ring, 544 struct intel_engine_cs *ring,
545 bool *need_reloc) 545 bool *need_reloc)
546{ 546{
547 struct drm_i915_gem_object *obj = vma->obj; 547 struct drm_i915_gem_object *obj = vma->obj;
@@ -596,7 +596,7 @@ i915_gem_execbuffer_reserve_vma(struct i915_vma *vma,
596} 596}
597 597
598static int 598static int
599i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, 599i915_gem_execbuffer_reserve(struct intel_engine_cs *ring,
600 struct list_head *vmas, 600 struct list_head *vmas,
601 bool *need_relocs) 601 bool *need_relocs)
602{ 602{
@@ -713,7 +713,7 @@ static int
713i915_gem_execbuffer_relocate_slow(struct drm_device *dev, 713i915_gem_execbuffer_relocate_slow(struct drm_device *dev,
714 struct drm_i915_gem_execbuffer2 *args, 714 struct drm_i915_gem_execbuffer2 *args,
715 struct drm_file *file, 715 struct drm_file *file,
716 struct intel_ring_buffer *ring, 716 struct intel_engine_cs *ring,
717 struct eb_vmas *eb, 717 struct eb_vmas *eb,
718 struct drm_i915_gem_exec_object2 *exec) 718 struct drm_i915_gem_exec_object2 *exec)
719{ 719{
@@ -829,7 +829,7 @@ err:
829} 829}
830 830
831static int 831static int
832i915_gem_execbuffer_move_to_gpu(struct intel_ring_buffer *ring, 832i915_gem_execbuffer_move_to_gpu(struct intel_engine_cs *ring,
833 struct list_head *vmas) 833 struct list_head *vmas)
834{ 834{
835 struct i915_vma *vma; 835 struct i915_vma *vma;
@@ -914,7 +914,7 @@ validate_exec_list(struct drm_i915_gem_exec_object2 *exec,
914 914
915static struct i915_hw_context * 915static struct i915_hw_context *
916i915_gem_validate_context(struct drm_device *dev, struct drm_file *file, 916i915_gem_validate_context(struct drm_device *dev, struct drm_file *file,
917 struct intel_ring_buffer *ring, const u32 ctx_id) 917 struct intel_engine_cs *ring, const u32 ctx_id)
918{ 918{
919 struct i915_hw_context *ctx = NULL; 919 struct i915_hw_context *ctx = NULL;
920 struct i915_ctx_hang_stats *hs; 920 struct i915_ctx_hang_stats *hs;
@@ -937,7 +937,7 @@ i915_gem_validate_context(struct drm_device *dev, struct drm_file *file,
937 937
938static void 938static void
939i915_gem_execbuffer_move_to_active(struct list_head *vmas, 939i915_gem_execbuffer_move_to_active(struct list_head *vmas,
940 struct intel_ring_buffer *ring) 940 struct intel_engine_cs *ring)
941{ 941{
942 struct i915_vma *vma; 942 struct i915_vma *vma;
943 943
@@ -972,7 +972,7 @@ i915_gem_execbuffer_move_to_active(struct list_head *vmas,
972static void 972static void
973i915_gem_execbuffer_retire_commands(struct drm_device *dev, 973i915_gem_execbuffer_retire_commands(struct drm_device *dev,
974 struct drm_file *file, 974 struct drm_file *file,
975 struct intel_ring_buffer *ring, 975 struct intel_engine_cs *ring,
976 struct drm_i915_gem_object *obj) 976 struct drm_i915_gem_object *obj)
977{ 977{
978 /* Unconditionally force add_request to emit a full flush. */ 978 /* Unconditionally force add_request to emit a full flush. */
@@ -984,7 +984,7 @@ i915_gem_execbuffer_retire_commands(struct drm_device *dev,
984 984
985static int 985static int
986i915_reset_gen7_sol_offsets(struct drm_device *dev, 986i915_reset_gen7_sol_offsets(struct drm_device *dev,
987 struct intel_ring_buffer *ring) 987 struct intel_engine_cs *ring)
988{ 988{
989 struct drm_i915_private *dev_priv = dev->dev_private; 989 struct drm_i915_private *dev_priv = dev->dev_private;
990 int ret, i; 990 int ret, i;
@@ -1050,7 +1050,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data,
1050 struct eb_vmas *eb; 1050 struct eb_vmas *eb;
1051 struct drm_i915_gem_object *batch_obj; 1051 struct drm_i915_gem_object *batch_obj;
1052 struct drm_clip_rect *cliprects = NULL; 1052 struct drm_clip_rect *cliprects = NULL;
1053 struct intel_ring_buffer *ring; 1053 struct intel_engine_cs *ring;
1054 struct i915_hw_context *ctx; 1054 struct i915_hw_context *ctx;
1055 struct i915_address_space *vm; 1055 struct i915_address_space *vm;
1056 const u32 ctx_id = i915_execbuffer2_get_context_id(*args); 1056 const u32 ctx_id = i915_execbuffer2_get_context_id(*args);
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
index ca213c59bf27..94916362b61c 100644
--- a/drivers/gpu/drm/i915/i915_gem_gtt.c
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
@@ -197,7 +197,7 @@ static gen6_gtt_pte_t iris_pte_encode(dma_addr_t addr,
197} 197}
198 198
199/* Broadwell Page Directory Pointer Descriptors */ 199/* Broadwell Page Directory Pointer Descriptors */
200static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry, 200static int gen8_write_pdp(struct intel_engine_cs *ring, unsigned entry,
201 uint64_t val, bool synchronous) 201 uint64_t val, bool synchronous)
202{ 202{
203 struct drm_i915_private *dev_priv = ring->dev->dev_private; 203 struct drm_i915_private *dev_priv = ring->dev->dev_private;
@@ -227,7 +227,7 @@ static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry,
227} 227}
228 228
229static int gen8_mm_switch(struct i915_hw_ppgtt *ppgtt, 229static int gen8_mm_switch(struct i915_hw_ppgtt *ppgtt,
230 struct intel_ring_buffer *ring, 230 struct intel_engine_cs *ring,
231 bool synchronous) 231 bool synchronous)
232{ 232{
233 int i, ret; 233 int i, ret;
@@ -706,7 +706,7 @@ static uint32_t get_pd_offset(struct i915_hw_ppgtt *ppgtt)
706} 706}
707 707
708static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt, 708static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt,
709 struct intel_ring_buffer *ring, 709 struct intel_engine_cs *ring,
710 bool synchronous) 710 bool synchronous)
711{ 711{
712 struct drm_device *dev = ppgtt->base.dev; 712 struct drm_device *dev = ppgtt->base.dev;
@@ -750,7 +750,7 @@ static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt,
750} 750}
751 751
752static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt, 752static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt,
753 struct intel_ring_buffer *ring, 753 struct intel_engine_cs *ring,
754 bool synchronous) 754 bool synchronous)
755{ 755{
756 struct drm_device *dev = ppgtt->base.dev; 756 struct drm_device *dev = ppgtt->base.dev;
@@ -801,7 +801,7 @@ static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt,
801} 801}
802 802
803static int gen6_mm_switch(struct i915_hw_ppgtt *ppgtt, 803static int gen6_mm_switch(struct i915_hw_ppgtt *ppgtt,
804 struct intel_ring_buffer *ring, 804 struct intel_engine_cs *ring,
805 bool synchronous) 805 bool synchronous)
806{ 806{
807 struct drm_device *dev = ppgtt->base.dev; 807 struct drm_device *dev = ppgtt->base.dev;
@@ -822,7 +822,7 @@ static int gen8_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
822{ 822{
823 struct drm_device *dev = ppgtt->base.dev; 823 struct drm_device *dev = ppgtt->base.dev;
824 struct drm_i915_private *dev_priv = dev->dev_private; 824 struct drm_i915_private *dev_priv = dev->dev_private;
825 struct intel_ring_buffer *ring; 825 struct intel_engine_cs *ring;
826 int j, ret; 826 int j, ret;
827 827
828 for_each_ring(ring, dev_priv, j) { 828 for_each_ring(ring, dev_priv, j) {
@@ -852,7 +852,7 @@ static int gen7_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
852{ 852{
853 struct drm_device *dev = ppgtt->base.dev; 853 struct drm_device *dev = ppgtt->base.dev;
854 struct drm_i915_private *dev_priv = dev->dev_private; 854 struct drm_i915_private *dev_priv = dev->dev_private;
855 struct intel_ring_buffer *ring; 855 struct intel_engine_cs *ring;
856 uint32_t ecochk, ecobits; 856 uint32_t ecochk, ecobits;
857 int i; 857 int i;
858 858
@@ -891,7 +891,7 @@ static int gen6_ppgtt_enable(struct i915_hw_ppgtt *ppgtt)
891{ 891{
892 struct drm_device *dev = ppgtt->base.dev; 892 struct drm_device *dev = ppgtt->base.dev;
893 struct drm_i915_private *dev_priv = dev->dev_private; 893 struct drm_i915_private *dev_priv = dev->dev_private;
894 struct intel_ring_buffer *ring; 894 struct intel_engine_cs *ring;
895 uint32_t ecochk, gab_ctl, ecobits; 895 uint32_t ecochk, gab_ctl, ecobits;
896 int i; 896 int i;
897 897
@@ -1259,7 +1259,7 @@ static void undo_idling(struct drm_i915_private *dev_priv, bool interruptible)
1259void i915_check_and_clear_faults(struct drm_device *dev) 1259void i915_check_and_clear_faults(struct drm_device *dev)
1260{ 1260{
1261 struct drm_i915_private *dev_priv = dev->dev_private; 1261 struct drm_i915_private *dev_priv = dev->dev_private;
1262 struct intel_ring_buffer *ring; 1262 struct intel_engine_cs *ring;
1263 int i; 1263 int i;
1264 1264
1265 if (INTEL_INFO(dev)->gen < 6) 1265 if (INTEL_INFO(dev)->gen < 6)
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h
index cfca023c4076..d187c0282860 100644
--- a/drivers/gpu/drm/i915/i915_gem_gtt.h
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.h
@@ -261,7 +261,7 @@ struct i915_hw_ppgtt {
261 261
262 int (*enable)(struct i915_hw_ppgtt *ppgtt); 262 int (*enable)(struct i915_hw_ppgtt *ppgtt);
263 int (*switch_mm)(struct i915_hw_ppgtt *ppgtt, 263 int (*switch_mm)(struct i915_hw_ppgtt *ppgtt,
264 struct intel_ring_buffer *ring, 264 struct intel_engine_cs *ring,
265 bool synchronous); 265 bool synchronous);
266 void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m); 266 void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m);
267}; 267};
diff --git a/drivers/gpu/drm/i915/i915_gem_render_state.c b/drivers/gpu/drm/i915/i915_gem_render_state.c
index cfbf6fc02a9e..3521f998a178 100644
--- a/drivers/gpu/drm/i915/i915_gem_render_state.c
+++ b/drivers/gpu/drm/i915/i915_gem_render_state.c
@@ -159,7 +159,7 @@ static int render_state_setup(const int gen,
159 return 0; 159 return 0;
160} 160}
161 161
162int i915_gem_render_state_init(struct intel_ring_buffer *ring) 162int i915_gem_render_state_init(struct intel_engine_cs *ring)
163{ 163{
164 const int gen = INTEL_INFO(ring->dev)->gen; 164 const int gen = INTEL_INFO(ring->dev)->gen;
165 struct i915_render_state *so; 165 struct i915_render_state *so;
diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c
index dcbec692f60f..632db42b3f89 100644
--- a/drivers/gpu/drm/i915/i915_gpu_error.c
+++ b/drivers/gpu/drm/i915/i915_gpu_error.c
@@ -747,7 +747,7 @@ static void i915_gem_record_fences(struct drm_device *dev,
747} 747}
748 748
749static void i915_record_ring_state(struct drm_device *dev, 749static void i915_record_ring_state(struct drm_device *dev,
750 struct intel_ring_buffer *ring, 750 struct intel_engine_cs *ring,
751 struct drm_i915_error_ring *ering) 751 struct drm_i915_error_ring *ering)
752{ 752{
753 struct drm_i915_private *dev_priv = dev->dev_private; 753 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -859,7 +859,7 @@ static void i915_record_ring_state(struct drm_device *dev,
859} 859}
860 860
861 861
862static void i915_gem_record_active_context(struct intel_ring_buffer *ring, 862static void i915_gem_record_active_context(struct intel_engine_cs *ring,
863 struct drm_i915_error_state *error, 863 struct drm_i915_error_state *error,
864 struct drm_i915_error_ring *ering) 864 struct drm_i915_error_ring *ering)
865{ 865{
@@ -886,7 +886,7 @@ static void i915_gem_record_rings(struct drm_device *dev,
886 int i, count; 886 int i, count;
887 887
888 for (i = 0; i < I915_NUM_RINGS; i++) { 888 for (i = 0; i < I915_NUM_RINGS; i++) {
889 struct intel_ring_buffer *ring = &dev_priv->ring[i]; 889 struct intel_engine_cs *ring = &dev_priv->ring[i];
890 890
891 if (ring->dev == NULL) 891 if (ring->dev == NULL)
892 continue; 892 continue;
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
index 304f86a3162c..c28e0dae38f7 100644
--- a/drivers/gpu/drm/i915/i915_irq.c
+++ b/drivers/gpu/drm/i915/i915_irq.c
@@ -1217,7 +1217,7 @@ static void ironlake_rps_change_irq_handler(struct drm_device *dev)
1217} 1217}
1218 1218
1219static void notify_ring(struct drm_device *dev, 1219static void notify_ring(struct drm_device *dev,
1220 struct intel_ring_buffer *ring) 1220 struct intel_engine_cs *ring)
1221{ 1221{
1222 if (ring->obj == NULL) 1222 if (ring->obj == NULL)
1223 return; 1223 return;
@@ -2316,7 +2316,7 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg)
2316static void i915_error_wake_up(struct drm_i915_private *dev_priv, 2316static void i915_error_wake_up(struct drm_i915_private *dev_priv,
2317 bool reset_completed) 2317 bool reset_completed)
2318{ 2318{
2319 struct intel_ring_buffer *ring; 2319 struct intel_engine_cs *ring;
2320 int i; 2320 int i;
2321 2321
2322 /* 2322 /*
@@ -2749,14 +2749,14 @@ static void gen8_disable_vblank(struct drm_device *dev, int pipe)
2749} 2749}
2750 2750
2751static u32 2751static u32
2752ring_last_seqno(struct intel_ring_buffer *ring) 2752ring_last_seqno(struct intel_engine_cs *ring)
2753{ 2753{
2754 return list_entry(ring->request_list.prev, 2754 return list_entry(ring->request_list.prev,
2755 struct drm_i915_gem_request, list)->seqno; 2755 struct drm_i915_gem_request, list)->seqno;
2756} 2756}
2757 2757
2758static bool 2758static bool
2759ring_idle(struct intel_ring_buffer *ring, u32 seqno) 2759ring_idle(struct intel_engine_cs *ring, u32 seqno)
2760{ 2760{
2761 return (list_empty(&ring->request_list) || 2761 return (list_empty(&ring->request_list) ||
2762 i915_seqno_passed(seqno, ring_last_seqno(ring))); 2762 i915_seqno_passed(seqno, ring_last_seqno(ring)));
@@ -2779,11 +2779,11 @@ ipehr_is_semaphore_wait(struct drm_device *dev, u32 ipehr)
2779 } 2779 }
2780} 2780}
2781 2781
2782static struct intel_ring_buffer * 2782static struct intel_engine_cs *
2783semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr) 2783semaphore_wait_to_signaller_ring(struct intel_engine_cs *ring, u32 ipehr)
2784{ 2784{
2785 struct drm_i915_private *dev_priv = ring->dev->dev_private; 2785 struct drm_i915_private *dev_priv = ring->dev->dev_private;
2786 struct intel_ring_buffer *signaller; 2786 struct intel_engine_cs *signaller;
2787 int i; 2787 int i;
2788 2788
2789 if (INTEL_INFO(dev_priv->dev)->gen >= 8) { 2789 if (INTEL_INFO(dev_priv->dev)->gen >= 8) {
@@ -2811,8 +2811,8 @@ semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr)
2811 return NULL; 2811 return NULL;
2812} 2812}
2813 2813
2814static struct intel_ring_buffer * 2814static struct intel_engine_cs *
2815semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno) 2815semaphore_waits_for(struct intel_engine_cs *ring, u32 *seqno)
2816{ 2816{
2817 struct drm_i915_private *dev_priv = ring->dev->dev_private; 2817 struct drm_i915_private *dev_priv = ring->dev->dev_private;
2818 u32 cmd, ipehr, head; 2818 u32 cmd, ipehr, head;
@@ -2854,10 +2854,10 @@ semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno)
2854 return semaphore_wait_to_signaller_ring(ring, ipehr); 2854 return semaphore_wait_to_signaller_ring(ring, ipehr);
2855} 2855}
2856 2856
2857static int semaphore_passed(struct intel_ring_buffer *ring) 2857static int semaphore_passed(struct intel_engine_cs *ring)
2858{ 2858{
2859 struct drm_i915_private *dev_priv = ring->dev->dev_private; 2859 struct drm_i915_private *dev_priv = ring->dev->dev_private;
2860 struct intel_ring_buffer *signaller; 2860 struct intel_engine_cs *signaller;
2861 u32 seqno, ctl; 2861 u32 seqno, ctl;
2862 2862
2863 ring->hangcheck.deadlock = true; 2863 ring->hangcheck.deadlock = true;
@@ -2876,7 +2876,7 @@ static int semaphore_passed(struct intel_ring_buffer *ring)
2876 2876
2877static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv) 2877static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv)
2878{ 2878{
2879 struct intel_ring_buffer *ring; 2879 struct intel_engine_cs *ring;
2880 int i; 2880 int i;
2881 2881
2882 for_each_ring(ring, dev_priv, i) 2882 for_each_ring(ring, dev_priv, i)
@@ -2884,7 +2884,7 @@ static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv)
2884} 2884}
2885 2885
2886static enum intel_ring_hangcheck_action 2886static enum intel_ring_hangcheck_action
2887ring_stuck(struct intel_ring_buffer *ring, u64 acthd) 2887ring_stuck(struct intel_engine_cs *ring, u64 acthd)
2888{ 2888{
2889 struct drm_device *dev = ring->dev; 2889 struct drm_device *dev = ring->dev;
2890 struct drm_i915_private *dev_priv = dev->dev_private; 2890 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -2940,7 +2940,7 @@ static void i915_hangcheck_elapsed(unsigned long data)
2940{ 2940{
2941 struct drm_device *dev = (struct drm_device *)data; 2941 struct drm_device *dev = (struct drm_device *)data;
2942 struct drm_i915_private *dev_priv = dev->dev_private; 2942 struct drm_i915_private *dev_priv = dev->dev_private;
2943 struct intel_ring_buffer *ring; 2943 struct intel_engine_cs *ring;
2944 int i; 2944 int i;
2945 int busy_count = 0, rings_hung = 0; 2945 int busy_count = 0, rings_hung = 0;
2946 bool stuck[I915_NUM_RINGS] = { 0 }; 2946 bool stuck[I915_NUM_RINGS] = { 0 };
diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
index b29d7b1828f1..f5aa0067755a 100644
--- a/drivers/gpu/drm/i915/i915_trace.h
+++ b/drivers/gpu/drm/i915/i915_trace.h
@@ -326,8 +326,8 @@ TRACE_EVENT(i915_gem_evict_vm,
326); 326);
327 327
328TRACE_EVENT(i915_gem_ring_sync_to, 328TRACE_EVENT(i915_gem_ring_sync_to,
329 TP_PROTO(struct intel_ring_buffer *from, 329 TP_PROTO(struct intel_engine_cs *from,
330 struct intel_ring_buffer *to, 330 struct intel_engine_cs *to,
331 u32 seqno), 331 u32 seqno),
332 TP_ARGS(from, to, seqno), 332 TP_ARGS(from, to, seqno),
333 333
@@ -352,7 +352,7 @@ TRACE_EVENT(i915_gem_ring_sync_to,
352); 352);
353 353
354TRACE_EVENT(i915_gem_ring_dispatch, 354TRACE_EVENT(i915_gem_ring_dispatch,
355 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno, u32 flags), 355 TP_PROTO(struct intel_engine_cs *ring, u32 seqno, u32 flags),
356 TP_ARGS(ring, seqno, flags), 356 TP_ARGS(ring, seqno, flags),
357 357
358 TP_STRUCT__entry( 358 TP_STRUCT__entry(
@@ -375,7 +375,7 @@ TRACE_EVENT(i915_gem_ring_dispatch,
375); 375);
376 376
377TRACE_EVENT(i915_gem_ring_flush, 377TRACE_EVENT(i915_gem_ring_flush,
378 TP_PROTO(struct intel_ring_buffer *ring, u32 invalidate, u32 flush), 378 TP_PROTO(struct intel_engine_cs *ring, u32 invalidate, u32 flush),
379 TP_ARGS(ring, invalidate, flush), 379 TP_ARGS(ring, invalidate, flush),
380 380
381 TP_STRUCT__entry( 381 TP_STRUCT__entry(
@@ -398,7 +398,7 @@ TRACE_EVENT(i915_gem_ring_flush,
398); 398);
399 399
400DECLARE_EVENT_CLASS(i915_gem_request, 400DECLARE_EVENT_CLASS(i915_gem_request,
401 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 401 TP_PROTO(struct intel_engine_cs *ring, u32 seqno),
402 TP_ARGS(ring, seqno), 402 TP_ARGS(ring, seqno),
403 403
404 TP_STRUCT__entry( 404 TP_STRUCT__entry(
@@ -418,12 +418,12 @@ DECLARE_EVENT_CLASS(i915_gem_request,
418); 418);
419 419
420DEFINE_EVENT(i915_gem_request, i915_gem_request_add, 420DEFINE_EVENT(i915_gem_request, i915_gem_request_add,
421 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 421 TP_PROTO(struct intel_engine_cs *ring, u32 seqno),
422 TP_ARGS(ring, seqno) 422 TP_ARGS(ring, seqno)
423); 423);
424 424
425TRACE_EVENT(i915_gem_request_complete, 425TRACE_EVENT(i915_gem_request_complete,
426 TP_PROTO(struct intel_ring_buffer *ring), 426 TP_PROTO(struct intel_engine_cs *ring),
427 TP_ARGS(ring), 427 TP_ARGS(ring),
428 428
429 TP_STRUCT__entry( 429 TP_STRUCT__entry(
@@ -443,12 +443,12 @@ TRACE_EVENT(i915_gem_request_complete,
443); 443);
444 444
445DEFINE_EVENT(i915_gem_request, i915_gem_request_retire, 445DEFINE_EVENT(i915_gem_request, i915_gem_request_retire,
446 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 446 TP_PROTO(struct intel_engine_cs *ring, u32 seqno),
447 TP_ARGS(ring, seqno) 447 TP_ARGS(ring, seqno)
448); 448);
449 449
450TRACE_EVENT(i915_gem_request_wait_begin, 450TRACE_EVENT(i915_gem_request_wait_begin,
451 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 451 TP_PROTO(struct intel_engine_cs *ring, u32 seqno),
452 TP_ARGS(ring, seqno), 452 TP_ARGS(ring, seqno),
453 453
454 TP_STRUCT__entry( 454 TP_STRUCT__entry(
@@ -477,12 +477,12 @@ TRACE_EVENT(i915_gem_request_wait_begin,
477); 477);
478 478
479DEFINE_EVENT(i915_gem_request, i915_gem_request_wait_end, 479DEFINE_EVENT(i915_gem_request, i915_gem_request_wait_end,
480 TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), 480 TP_PROTO(struct intel_engine_cs *ring, u32 seqno),
481 TP_ARGS(ring, seqno) 481 TP_ARGS(ring, seqno)
482); 482);
483 483
484DECLARE_EVENT_CLASS(i915_ring, 484DECLARE_EVENT_CLASS(i915_ring,
485 TP_PROTO(struct intel_ring_buffer *ring), 485 TP_PROTO(struct intel_engine_cs *ring),
486 TP_ARGS(ring), 486 TP_ARGS(ring),
487 487
488 TP_STRUCT__entry( 488 TP_STRUCT__entry(
@@ -499,12 +499,12 @@ DECLARE_EVENT_CLASS(i915_ring,
499); 499);
500 500
501DEFINE_EVENT(i915_ring, i915_ring_wait_begin, 501DEFINE_EVENT(i915_ring, i915_ring_wait_begin,
502 TP_PROTO(struct intel_ring_buffer *ring), 502 TP_PROTO(struct intel_engine_cs *ring),
503 TP_ARGS(ring) 503 TP_ARGS(ring)
504); 504);
505 505
506DEFINE_EVENT(i915_ring, i915_ring_wait_end, 506DEFINE_EVENT(i915_ring, i915_ring_wait_end,
507 TP_PROTO(struct intel_ring_buffer *ring), 507 TP_PROTO(struct intel_engine_cs *ring),
508 TP_ARGS(ring) 508 TP_ARGS(ring)
509); 509);
510 510
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
index 6db85e28cb66..90f4f05c3000 100644
--- a/drivers/gpu/drm/i915/intel_display.c
+++ b/drivers/gpu/drm/i915/intel_display.c
@@ -2159,7 +2159,7 @@ static int intel_align_height(struct drm_device *dev, int height, bool tiled)
2159int 2159int
2160intel_pin_and_fence_fb_obj(struct drm_device *dev, 2160intel_pin_and_fence_fb_obj(struct drm_device *dev,
2161 struct drm_i915_gem_object *obj, 2161 struct drm_i915_gem_object *obj,
2162 struct intel_ring_buffer *pipelined) 2162 struct intel_engine_cs *pipelined)
2163{ 2163{
2164 struct drm_i915_private *dev_priv = dev->dev_private; 2164 struct drm_i915_private *dev_priv = dev->dev_private;
2165 u32 alignment; 2165 u32 alignment;
@@ -8756,7 +8756,7 @@ out:
8756} 8756}
8757 8757
8758void intel_mark_fb_busy(struct drm_i915_gem_object *obj, 8758void intel_mark_fb_busy(struct drm_i915_gem_object *obj,
8759 struct intel_ring_buffer *ring) 8759 struct intel_engine_cs *ring)
8760{ 8760{
8761 struct drm_device *dev = obj->base.dev; 8761 struct drm_device *dev = obj->base.dev;
8762 struct drm_crtc *crtc; 8762 struct drm_crtc *crtc;
@@ -8951,7 +8951,7 @@ static int intel_gen2_queue_flip(struct drm_device *dev,
8951 struct drm_crtc *crtc, 8951 struct drm_crtc *crtc,
8952 struct drm_framebuffer *fb, 8952 struct drm_framebuffer *fb,
8953 struct drm_i915_gem_object *obj, 8953 struct drm_i915_gem_object *obj,
8954 struct intel_ring_buffer *ring, 8954 struct intel_engine_cs *ring,
8955 uint32_t flags) 8955 uint32_t flags)
8956{ 8956{
8957 struct intel_crtc *intel_crtc = to_intel_crtc(crtc); 8957 struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
@@ -8986,7 +8986,7 @@ static int intel_gen3_queue_flip(struct drm_device *dev,
8986 struct drm_crtc *crtc, 8986 struct drm_crtc *crtc,
8987 struct drm_framebuffer *fb, 8987 struct drm_framebuffer *fb,
8988 struct drm_i915_gem_object *obj, 8988 struct drm_i915_gem_object *obj,
8989 struct intel_ring_buffer *ring, 8989 struct intel_engine_cs *ring,
8990 uint32_t flags) 8990 uint32_t flags)
8991{ 8991{
8992 struct intel_crtc *intel_crtc = to_intel_crtc(crtc); 8992 struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
@@ -9018,7 +9018,7 @@ static int intel_gen4_queue_flip(struct drm_device *dev,
9018 struct drm_crtc *crtc, 9018 struct drm_crtc *crtc,
9019 struct drm_framebuffer *fb, 9019 struct drm_framebuffer *fb,
9020 struct drm_i915_gem_object *obj, 9020 struct drm_i915_gem_object *obj,
9021 struct intel_ring_buffer *ring, 9021 struct intel_engine_cs *ring,
9022 uint32_t flags) 9022 uint32_t flags)
9023{ 9023{
9024 struct drm_i915_private *dev_priv = dev->dev_private; 9024 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -9057,7 +9057,7 @@ static int intel_gen6_queue_flip(struct drm_device *dev,
9057 struct drm_crtc *crtc, 9057 struct drm_crtc *crtc,
9058 struct drm_framebuffer *fb, 9058 struct drm_framebuffer *fb,
9059 struct drm_i915_gem_object *obj, 9059 struct drm_i915_gem_object *obj,
9060 struct intel_ring_buffer *ring, 9060 struct intel_engine_cs *ring,
9061 uint32_t flags) 9061 uint32_t flags)
9062{ 9062{
9063 struct drm_i915_private *dev_priv = dev->dev_private; 9063 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -9093,7 +9093,7 @@ static int intel_gen7_queue_flip(struct drm_device *dev,
9093 struct drm_crtc *crtc, 9093 struct drm_crtc *crtc,
9094 struct drm_framebuffer *fb, 9094 struct drm_framebuffer *fb,
9095 struct drm_i915_gem_object *obj, 9095 struct drm_i915_gem_object *obj,
9096 struct intel_ring_buffer *ring, 9096 struct intel_engine_cs *ring,
9097 uint32_t flags) 9097 uint32_t flags)
9098{ 9098{
9099 struct intel_crtc *intel_crtc = to_intel_crtc(crtc); 9099 struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
@@ -9188,7 +9188,7 @@ static int intel_default_queue_flip(struct drm_device *dev,
9188 struct drm_crtc *crtc, 9188 struct drm_crtc *crtc,
9189 struct drm_framebuffer *fb, 9189 struct drm_framebuffer *fb,
9190 struct drm_i915_gem_object *obj, 9190 struct drm_i915_gem_object *obj,
9191 struct intel_ring_buffer *ring, 9191 struct intel_engine_cs *ring,
9192 uint32_t flags) 9192 uint32_t flags)
9193{ 9193{
9194 return -ENODEV; 9194 return -ENODEV;
@@ -9205,7 +9205,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc,
9205 struct drm_i915_gem_object *obj = to_intel_framebuffer(fb)->obj; 9205 struct drm_i915_gem_object *obj = to_intel_framebuffer(fb)->obj;
9206 struct intel_crtc *intel_crtc = to_intel_crtc(crtc); 9206 struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
9207 struct intel_unpin_work *work; 9207 struct intel_unpin_work *work;
9208 struct intel_ring_buffer *ring; 9208 struct intel_engine_cs *ring;
9209 unsigned long flags; 9209 unsigned long flags;
9210 int ret; 9210 int ret;
9211 9211
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
index d7c52b2c546a..f1d5897c96cd 100644
--- a/drivers/gpu/drm/i915/intel_drv.h
+++ b/drivers/gpu/drm/i915/intel_drv.h
@@ -721,7 +721,7 @@ int intel_pch_rawclk(struct drm_device *dev);
721int valleyview_cur_cdclk(struct drm_i915_private *dev_priv); 721int valleyview_cur_cdclk(struct drm_i915_private *dev_priv);
722void intel_mark_busy(struct drm_device *dev); 722void intel_mark_busy(struct drm_device *dev);
723void intel_mark_fb_busy(struct drm_i915_gem_object *obj, 723void intel_mark_fb_busy(struct drm_i915_gem_object *obj,
724 struct intel_ring_buffer *ring); 724 struct intel_engine_cs *ring);
725void intel_mark_idle(struct drm_device *dev); 725void intel_mark_idle(struct drm_device *dev);
726void intel_crtc_restore_mode(struct drm_crtc *crtc); 726void intel_crtc_restore_mode(struct drm_crtc *crtc);
727void intel_crtc_update_dpms(struct drm_crtc *crtc); 727void intel_crtc_update_dpms(struct drm_crtc *crtc);
@@ -753,7 +753,7 @@ void intel_release_load_detect_pipe(struct drm_connector *connector,
753 struct intel_load_detect_pipe *old); 753 struct intel_load_detect_pipe *old);
754int intel_pin_and_fence_fb_obj(struct drm_device *dev, 754int intel_pin_and_fence_fb_obj(struct drm_device *dev,
755 struct drm_i915_gem_object *obj, 755 struct drm_i915_gem_object *obj,
756 struct intel_ring_buffer *pipelined); 756 struct intel_engine_cs *pipelined);
757void intel_unpin_fb_obj(struct drm_i915_gem_object *obj); 757void intel_unpin_fb_obj(struct drm_i915_gem_object *obj);
758struct drm_framebuffer * 758struct drm_framebuffer *
759__intel_framebuffer_create(struct drm_device *dev, 759__intel_framebuffer_create(struct drm_device *dev,
diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c
index d8adc9104dca..27eb820d1b19 100644
--- a/drivers/gpu/drm/i915/intel_overlay.c
+++ b/drivers/gpu/drm/i915/intel_overlay.c
@@ -213,7 +213,7 @@ static int intel_overlay_do_wait_request(struct intel_overlay *overlay,
213{ 213{
214 struct drm_device *dev = overlay->dev; 214 struct drm_device *dev = overlay->dev;
215 struct drm_i915_private *dev_priv = dev->dev_private; 215 struct drm_i915_private *dev_priv = dev->dev_private;
216 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 216 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
217 int ret; 217 int ret;
218 218
219 BUG_ON(overlay->last_flip_req); 219 BUG_ON(overlay->last_flip_req);
@@ -236,7 +236,7 @@ static int intel_overlay_on(struct intel_overlay *overlay)
236{ 236{
237 struct drm_device *dev = overlay->dev; 237 struct drm_device *dev = overlay->dev;
238 struct drm_i915_private *dev_priv = dev->dev_private; 238 struct drm_i915_private *dev_priv = dev->dev_private;
239 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 239 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
240 int ret; 240 int ret;
241 241
242 BUG_ON(overlay->active); 242 BUG_ON(overlay->active);
@@ -263,7 +263,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay,
263{ 263{
264 struct drm_device *dev = overlay->dev; 264 struct drm_device *dev = overlay->dev;
265 struct drm_i915_private *dev_priv = dev->dev_private; 265 struct drm_i915_private *dev_priv = dev->dev_private;
266 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 266 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
267 u32 flip_addr = overlay->flip_addr; 267 u32 flip_addr = overlay->flip_addr;
268 u32 tmp; 268 u32 tmp;
269 int ret; 269 int ret;
@@ -320,7 +320,7 @@ static int intel_overlay_off(struct intel_overlay *overlay)
320{ 320{
321 struct drm_device *dev = overlay->dev; 321 struct drm_device *dev = overlay->dev;
322 struct drm_i915_private *dev_priv = dev->dev_private; 322 struct drm_i915_private *dev_priv = dev->dev_private;
323 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 323 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
324 u32 flip_addr = overlay->flip_addr; 324 u32 flip_addr = overlay->flip_addr;
325 int ret; 325 int ret;
326 326
@@ -363,7 +363,7 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay)
363{ 363{
364 struct drm_device *dev = overlay->dev; 364 struct drm_device *dev = overlay->dev;
365 struct drm_i915_private *dev_priv = dev->dev_private; 365 struct drm_i915_private *dev_priv = dev->dev_private;
366 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 366 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
367 int ret; 367 int ret;
368 368
369 if (overlay->last_flip_req == 0) 369 if (overlay->last_flip_req == 0)
@@ -389,7 +389,7 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay)
389{ 389{
390 struct drm_device *dev = overlay->dev; 390 struct drm_device *dev = overlay->dev;
391 struct drm_i915_private *dev_priv = dev->dev_private; 391 struct drm_i915_private *dev_priv = dev->dev_private;
392 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 392 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
393 int ret; 393 int ret;
394 394
395 /* Only wait if there is actually an old frame to release to 395 /* Only wait if there is actually an old frame to release to
diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c
index 124e82e6cc38..b86b58c44228 100644
--- a/drivers/gpu/drm/i915/intel_pm.c
+++ b/drivers/gpu/drm/i915/intel_pm.c
@@ -3412,7 +3412,7 @@ static void parse_rp_state_cap(struct drm_i915_private *dev_priv, u32 rp_state_c
3412static void gen8_enable_rps(struct drm_device *dev) 3412static void gen8_enable_rps(struct drm_device *dev)
3413{ 3413{
3414 struct drm_i915_private *dev_priv = dev->dev_private; 3414 struct drm_i915_private *dev_priv = dev->dev_private;
3415 struct intel_ring_buffer *ring; 3415 struct intel_engine_cs *ring;
3416 uint32_t rc6_mask = 0, rp_state_cap; 3416 uint32_t rc6_mask = 0, rp_state_cap;
3417 int unused; 3417 int unused;
3418 3418
@@ -3491,7 +3491,7 @@ static void gen8_enable_rps(struct drm_device *dev)
3491static void gen6_enable_rps(struct drm_device *dev) 3491static void gen6_enable_rps(struct drm_device *dev)
3492{ 3492{
3493 struct drm_i915_private *dev_priv = dev->dev_private; 3493 struct drm_i915_private *dev_priv = dev->dev_private;
3494 struct intel_ring_buffer *ring; 3494 struct intel_engine_cs *ring;
3495 u32 rp_state_cap; 3495 u32 rp_state_cap;
3496 u32 gt_perf_status; 3496 u32 gt_perf_status;
3497 u32 rc6vids, pcu_mbox = 0, rc6_mask = 0; 3497 u32 rc6vids, pcu_mbox = 0, rc6_mask = 0;
@@ -3820,7 +3820,7 @@ static void valleyview_cleanup_gt_powersave(struct drm_device *dev)
3820static void valleyview_enable_rps(struct drm_device *dev) 3820static void valleyview_enable_rps(struct drm_device *dev)
3821{ 3821{
3822 struct drm_i915_private *dev_priv = dev->dev_private; 3822 struct drm_i915_private *dev_priv = dev->dev_private;
3823 struct intel_ring_buffer *ring; 3823 struct intel_engine_cs *ring;
3824 u32 gtfifodbg, val, rc6_mode = 0; 3824 u32 gtfifodbg, val, rc6_mode = 0;
3825 int i; 3825 int i;
3826 3826
@@ -3951,7 +3951,7 @@ static int ironlake_setup_rc6(struct drm_device *dev)
3951static void ironlake_enable_rc6(struct drm_device *dev) 3951static void ironlake_enable_rc6(struct drm_device *dev)
3952{ 3952{
3953 struct drm_i915_private *dev_priv = dev->dev_private; 3953 struct drm_i915_private *dev_priv = dev->dev_private;
3954 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 3954 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
3955 bool was_interruptible; 3955 bool was_interruptible;
3956 int ret; 3956 int ret;
3957 3957
@@ -4463,7 +4463,7 @@ EXPORT_SYMBOL_GPL(i915_gpu_lower);
4463bool i915_gpu_busy(void) 4463bool i915_gpu_busy(void)
4464{ 4464{
4465 struct drm_i915_private *dev_priv; 4465 struct drm_i915_private *dev_priv;
4466 struct intel_ring_buffer *ring; 4466 struct intel_engine_cs *ring;
4467 bool ret = false; 4467 bool ret = false;
4468 int i; 4468 int i;
4469 4469
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c
index 72bf57fe3e54..748041aa84df 100644
--- a/drivers/gpu/drm/i915/intel_ringbuffer.c
+++ b/drivers/gpu/drm/i915/intel_ringbuffer.c
@@ -48,18 +48,18 @@ static inline int __ring_space(int head, int tail, int size)
48 return space; 48 return space;
49} 49}
50 50
51static inline int ring_space(struct intel_ring_buffer *ring) 51static inline int ring_space(struct intel_engine_cs *ring)
52{ 52{
53 return __ring_space(ring->head & HEAD_ADDR, ring->tail, ring->size); 53 return __ring_space(ring->head & HEAD_ADDR, ring->tail, ring->size);
54} 54}
55 55
56static bool intel_ring_stopped(struct intel_ring_buffer *ring) 56static bool intel_ring_stopped(struct intel_engine_cs *ring)
57{ 57{
58 struct drm_i915_private *dev_priv = ring->dev->dev_private; 58 struct drm_i915_private *dev_priv = ring->dev->dev_private;
59 return dev_priv->gpu_error.stop_rings & intel_ring_flag(ring); 59 return dev_priv->gpu_error.stop_rings & intel_ring_flag(ring);
60} 60}
61 61
62void __intel_ring_advance(struct intel_ring_buffer *ring) 62void __intel_ring_advance(struct intel_engine_cs *ring)
63{ 63{
64 ring->tail &= ring->size - 1; 64 ring->tail &= ring->size - 1;
65 if (intel_ring_stopped(ring)) 65 if (intel_ring_stopped(ring))
@@ -68,7 +68,7 @@ void __intel_ring_advance(struct intel_ring_buffer *ring)
68} 68}
69 69
70static int 70static int
71gen2_render_ring_flush(struct intel_ring_buffer *ring, 71gen2_render_ring_flush(struct intel_engine_cs *ring,
72 u32 invalidate_domains, 72 u32 invalidate_domains,
73 u32 flush_domains) 73 u32 flush_domains)
74{ 74{
@@ -94,7 +94,7 @@ gen2_render_ring_flush(struct intel_ring_buffer *ring,
94} 94}
95 95
96static int 96static int
97gen4_render_ring_flush(struct intel_ring_buffer *ring, 97gen4_render_ring_flush(struct intel_engine_cs *ring,
98 u32 invalidate_domains, 98 u32 invalidate_domains,
99 u32 flush_domains) 99 u32 flush_domains)
100{ 100{
@@ -189,7 +189,7 @@ gen4_render_ring_flush(struct intel_ring_buffer *ring,
189 * really our business. That leaves only stall at scoreboard. 189 * really our business. That leaves only stall at scoreboard.
190 */ 190 */
191static int 191static int
192intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring) 192intel_emit_post_sync_nonzero_flush(struct intel_engine_cs *ring)
193{ 193{
194 u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES; 194 u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES;
195 int ret; 195 int ret;
@@ -224,7 +224,7 @@ intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring)
224} 224}
225 225
226static int 226static int
227gen6_render_ring_flush(struct intel_ring_buffer *ring, 227gen6_render_ring_flush(struct intel_engine_cs *ring,
228 u32 invalidate_domains, u32 flush_domains) 228 u32 invalidate_domains, u32 flush_domains)
229{ 229{
230 u32 flags = 0; 230 u32 flags = 0;
@@ -276,7 +276,7 @@ gen6_render_ring_flush(struct intel_ring_buffer *ring,
276} 276}
277 277
278static int 278static int
279gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring) 279gen7_render_ring_cs_stall_wa(struct intel_engine_cs *ring)
280{ 280{
281 int ret; 281 int ret;
282 282
@@ -294,7 +294,7 @@ gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring)
294 return 0; 294 return 0;
295} 295}
296 296
297static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value) 297static int gen7_ring_fbc_flush(struct intel_engine_cs *ring, u32 value)
298{ 298{
299 int ret; 299 int ret;
300 300
@@ -318,7 +318,7 @@ static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value)
318} 318}
319 319
320static int 320static int
321gen7_render_ring_flush(struct intel_ring_buffer *ring, 321gen7_render_ring_flush(struct intel_engine_cs *ring,
322 u32 invalidate_domains, u32 flush_domains) 322 u32 invalidate_domains, u32 flush_domains)
323{ 323{
324 u32 flags = 0; 324 u32 flags = 0;
@@ -379,7 +379,7 @@ gen7_render_ring_flush(struct intel_ring_buffer *ring,
379} 379}
380 380
381static int 381static int
382gen8_render_ring_flush(struct intel_ring_buffer *ring, 382gen8_render_ring_flush(struct intel_engine_cs *ring,
383 u32 invalidate_domains, u32 flush_domains) 383 u32 invalidate_domains, u32 flush_domains)
384{ 384{
385 u32 flags = 0; 385 u32 flags = 0;
@@ -419,14 +419,14 @@ gen8_render_ring_flush(struct intel_ring_buffer *ring,
419 419
420} 420}
421 421
422static void ring_write_tail(struct intel_ring_buffer *ring, 422static void ring_write_tail(struct intel_engine_cs *ring,
423 u32 value) 423 u32 value)
424{ 424{
425 struct drm_i915_private *dev_priv = ring->dev->dev_private; 425 struct drm_i915_private *dev_priv = ring->dev->dev_private;
426 I915_WRITE_TAIL(ring, value); 426 I915_WRITE_TAIL(ring, value);
427} 427}
428 428
429u64 intel_ring_get_active_head(struct intel_ring_buffer *ring) 429u64 intel_ring_get_active_head(struct intel_engine_cs *ring)
430{ 430{
431 struct drm_i915_private *dev_priv = ring->dev->dev_private; 431 struct drm_i915_private *dev_priv = ring->dev->dev_private;
432 u64 acthd; 432 u64 acthd;
@@ -442,7 +442,7 @@ u64 intel_ring_get_active_head(struct intel_ring_buffer *ring)
442 return acthd; 442 return acthd;
443} 443}
444 444
445static void ring_setup_phys_status_page(struct intel_ring_buffer *ring) 445static void ring_setup_phys_status_page(struct intel_engine_cs *ring)
446{ 446{
447 struct drm_i915_private *dev_priv = ring->dev->dev_private; 447 struct drm_i915_private *dev_priv = ring->dev->dev_private;
448 u32 addr; 448 u32 addr;
@@ -453,7 +453,7 @@ static void ring_setup_phys_status_page(struct intel_ring_buffer *ring)
453 I915_WRITE(HWS_PGA, addr); 453 I915_WRITE(HWS_PGA, addr);
454} 454}
455 455
456static bool stop_ring(struct intel_ring_buffer *ring) 456static bool stop_ring(struct intel_engine_cs *ring)
457{ 457{
458 struct drm_i915_private *dev_priv = to_i915(ring->dev); 458 struct drm_i915_private *dev_priv = to_i915(ring->dev);
459 459
@@ -477,7 +477,7 @@ static bool stop_ring(struct intel_ring_buffer *ring)
477 return (I915_READ_HEAD(ring) & HEAD_ADDR) == 0; 477 return (I915_READ_HEAD(ring) & HEAD_ADDR) == 0;
478} 478}
479 479
480static int init_ring_common(struct intel_ring_buffer *ring) 480static int init_ring_common(struct intel_engine_cs *ring)
481{ 481{
482 struct drm_device *dev = ring->dev; 482 struct drm_device *dev = ring->dev;
483 struct drm_i915_private *dev_priv = dev->dev_private; 483 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -555,7 +555,7 @@ out:
555} 555}
556 556
557static int 557static int
558init_pipe_control(struct intel_ring_buffer *ring) 558init_pipe_control(struct intel_engine_cs *ring)
559{ 559{
560 int ret; 560 int ret;
561 561
@@ -596,7 +596,7 @@ err:
596 return ret; 596 return ret;
597} 597}
598 598
599static int init_render_ring(struct intel_ring_buffer *ring) 599static int init_render_ring(struct intel_engine_cs *ring)
600{ 600{
601 struct drm_device *dev = ring->dev; 601 struct drm_device *dev = ring->dev;
602 struct drm_i915_private *dev_priv = dev->dev_private; 602 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -652,7 +652,7 @@ static int init_render_ring(struct intel_ring_buffer *ring)
652 return ret; 652 return ret;
653} 653}
654 654
655static void render_ring_cleanup(struct intel_ring_buffer *ring) 655static void render_ring_cleanup(struct intel_engine_cs *ring)
656{ 656{
657 struct drm_device *dev = ring->dev; 657 struct drm_device *dev = ring->dev;
658 658
@@ -668,12 +668,12 @@ static void render_ring_cleanup(struct intel_ring_buffer *ring)
668 ring->scratch.obj = NULL; 668 ring->scratch.obj = NULL;
669} 669}
670 670
671static int gen6_signal(struct intel_ring_buffer *signaller, 671static int gen6_signal(struct intel_engine_cs *signaller,
672 unsigned int num_dwords) 672 unsigned int num_dwords)
673{ 673{
674 struct drm_device *dev = signaller->dev; 674 struct drm_device *dev = signaller->dev;
675 struct drm_i915_private *dev_priv = dev->dev_private; 675 struct drm_i915_private *dev_priv = dev->dev_private;
676 struct intel_ring_buffer *useless; 676 struct intel_engine_cs *useless;
677 int i, ret; 677 int i, ret;
678 678
679 /* NB: In order to be able to do semaphore MBOX updates for varying 679 /* NB: In order to be able to do semaphore MBOX updates for varying
@@ -720,7 +720,7 @@ static int gen6_signal(struct intel_ring_buffer *signaller,
720 * This acts like a signal in the canonical semaphore. 720 * This acts like a signal in the canonical semaphore.
721 */ 721 */
722static int 722static int
723gen6_add_request(struct intel_ring_buffer *ring) 723gen6_add_request(struct intel_engine_cs *ring)
724{ 724{
725 int ret; 725 int ret;
726 726
@@ -752,8 +752,8 @@ static inline bool i915_gem_has_seqno_wrapped(struct drm_device *dev,
752 * @seqno - seqno which the waiter will block on 752 * @seqno - seqno which the waiter will block on
753 */ 753 */
754static int 754static int
755gen6_ring_sync(struct intel_ring_buffer *waiter, 755gen6_ring_sync(struct intel_engine_cs *waiter,
756 struct intel_ring_buffer *signaller, 756 struct intel_engine_cs *signaller,
757 u32 seqno) 757 u32 seqno)
758{ 758{
759 u32 dw1 = MI_SEMAPHORE_MBOX | 759 u32 dw1 = MI_SEMAPHORE_MBOX |
@@ -801,7 +801,7 @@ do { \
801} while (0) 801} while (0)
802 802
803static int 803static int
804pc_render_add_request(struct intel_ring_buffer *ring) 804pc_render_add_request(struct intel_engine_cs *ring)
805{ 805{
806 u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES; 806 u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES;
807 int ret; 807 int ret;
@@ -849,7 +849,7 @@ pc_render_add_request(struct intel_ring_buffer *ring)
849} 849}
850 850
851static u32 851static u32
852gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) 852gen6_ring_get_seqno(struct intel_engine_cs *ring, bool lazy_coherency)
853{ 853{
854 /* Workaround to force correct ordering between irq and seqno writes on 854 /* Workaround to force correct ordering between irq and seqno writes on
855 * ivb (and maybe also on snb) by reading from a CS register (like 855 * ivb (and maybe also on snb) by reading from a CS register (like
@@ -863,31 +863,31 @@ gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency)
863} 863}
864 864
865static u32 865static u32
866ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) 866ring_get_seqno(struct intel_engine_cs *ring, bool lazy_coherency)
867{ 867{
868 return intel_read_status_page(ring, I915_GEM_HWS_INDEX); 868 return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
869} 869}
870 870
871static void 871static void
872ring_set_seqno(struct intel_ring_buffer *ring, u32 seqno) 872ring_set_seqno(struct intel_engine_cs *ring, u32 seqno)
873{ 873{
874 intel_write_status_page(ring, I915_GEM_HWS_INDEX, seqno); 874 intel_write_status_page(ring, I915_GEM_HWS_INDEX, seqno);
875} 875}
876 876
877static u32 877static u32
878pc_render_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) 878pc_render_get_seqno(struct intel_engine_cs *ring, bool lazy_coherency)
879{ 879{
880 return ring->scratch.cpu_page[0]; 880 return ring->scratch.cpu_page[0];
881} 881}
882 882
883static void 883static void
884pc_render_set_seqno(struct intel_ring_buffer *ring, u32 seqno) 884pc_render_set_seqno(struct intel_engine_cs *ring, u32 seqno)
885{ 885{
886 ring->scratch.cpu_page[0] = seqno; 886 ring->scratch.cpu_page[0] = seqno;
887} 887}
888 888
889static bool 889static bool
890gen5_ring_get_irq(struct intel_ring_buffer *ring) 890gen5_ring_get_irq(struct intel_engine_cs *ring)
891{ 891{
892 struct drm_device *dev = ring->dev; 892 struct drm_device *dev = ring->dev;
893 struct drm_i915_private *dev_priv = dev->dev_private; 893 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -905,7 +905,7 @@ gen5_ring_get_irq(struct intel_ring_buffer *ring)
905} 905}
906 906
907static void 907static void
908gen5_ring_put_irq(struct intel_ring_buffer *ring) 908gen5_ring_put_irq(struct intel_engine_cs *ring)
909{ 909{
910 struct drm_device *dev = ring->dev; 910 struct drm_device *dev = ring->dev;
911 struct drm_i915_private *dev_priv = dev->dev_private; 911 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -918,7 +918,7 @@ gen5_ring_put_irq(struct intel_ring_buffer *ring)
918} 918}
919 919
920static bool 920static bool
921i9xx_ring_get_irq(struct intel_ring_buffer *ring) 921i9xx_ring_get_irq(struct intel_engine_cs *ring)
922{ 922{
923 struct drm_device *dev = ring->dev; 923 struct drm_device *dev = ring->dev;
924 struct drm_i915_private *dev_priv = dev->dev_private; 924 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -939,7 +939,7 @@ i9xx_ring_get_irq(struct intel_ring_buffer *ring)
939} 939}
940 940
941static void 941static void
942i9xx_ring_put_irq(struct intel_ring_buffer *ring) 942i9xx_ring_put_irq(struct intel_engine_cs *ring)
943{ 943{
944 struct drm_device *dev = ring->dev; 944 struct drm_device *dev = ring->dev;
945 struct drm_i915_private *dev_priv = dev->dev_private; 945 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -955,7 +955,7 @@ i9xx_ring_put_irq(struct intel_ring_buffer *ring)
955} 955}
956 956
957static bool 957static bool
958i8xx_ring_get_irq(struct intel_ring_buffer *ring) 958i8xx_ring_get_irq(struct intel_engine_cs *ring)
959{ 959{
960 struct drm_device *dev = ring->dev; 960 struct drm_device *dev = ring->dev;
961 struct drm_i915_private *dev_priv = dev->dev_private; 961 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -976,7 +976,7 @@ i8xx_ring_get_irq(struct intel_ring_buffer *ring)
976} 976}
977 977
978static void 978static void
979i8xx_ring_put_irq(struct intel_ring_buffer *ring) 979i8xx_ring_put_irq(struct intel_engine_cs *ring)
980{ 980{
981 struct drm_device *dev = ring->dev; 981 struct drm_device *dev = ring->dev;
982 struct drm_i915_private *dev_priv = dev->dev_private; 982 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -991,7 +991,7 @@ i8xx_ring_put_irq(struct intel_ring_buffer *ring)
991 spin_unlock_irqrestore(&dev_priv->irq_lock, flags); 991 spin_unlock_irqrestore(&dev_priv->irq_lock, flags);
992} 992}
993 993
994void intel_ring_setup_status_page(struct intel_ring_buffer *ring) 994void intel_ring_setup_status_page(struct intel_engine_cs *ring)
995{ 995{
996 struct drm_device *dev = ring->dev; 996 struct drm_device *dev = ring->dev;
997 struct drm_i915_private *dev_priv = ring->dev->dev_private; 997 struct drm_i915_private *dev_priv = ring->dev->dev_private;
@@ -1054,7 +1054,7 @@ void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
1054} 1054}
1055 1055
1056static int 1056static int
1057bsd_ring_flush(struct intel_ring_buffer *ring, 1057bsd_ring_flush(struct intel_engine_cs *ring,
1058 u32 invalidate_domains, 1058 u32 invalidate_domains,
1059 u32 flush_domains) 1059 u32 flush_domains)
1060{ 1060{
@@ -1071,7 +1071,7 @@ bsd_ring_flush(struct intel_ring_buffer *ring,
1071} 1071}
1072 1072
1073static int 1073static int
1074i9xx_add_request(struct intel_ring_buffer *ring) 1074i9xx_add_request(struct intel_engine_cs *ring)
1075{ 1075{
1076 int ret; 1076 int ret;
1077 1077
@@ -1089,7 +1089,7 @@ i9xx_add_request(struct intel_ring_buffer *ring)
1089} 1089}
1090 1090
1091static bool 1091static bool
1092gen6_ring_get_irq(struct intel_ring_buffer *ring) 1092gen6_ring_get_irq(struct intel_engine_cs *ring)
1093{ 1093{
1094 struct drm_device *dev = ring->dev; 1094 struct drm_device *dev = ring->dev;
1095 struct drm_i915_private *dev_priv = dev->dev_private; 1095 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1114,7 +1114,7 @@ gen6_ring_get_irq(struct intel_ring_buffer *ring)
1114} 1114}
1115 1115
1116static void 1116static void
1117gen6_ring_put_irq(struct intel_ring_buffer *ring) 1117gen6_ring_put_irq(struct intel_engine_cs *ring)
1118{ 1118{
1119 struct drm_device *dev = ring->dev; 1119 struct drm_device *dev = ring->dev;
1120 struct drm_i915_private *dev_priv = dev->dev_private; 1120 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1132,7 +1132,7 @@ gen6_ring_put_irq(struct intel_ring_buffer *ring)
1132} 1132}
1133 1133
1134static bool 1134static bool
1135hsw_vebox_get_irq(struct intel_ring_buffer *ring) 1135hsw_vebox_get_irq(struct intel_engine_cs *ring)
1136{ 1136{
1137 struct drm_device *dev = ring->dev; 1137 struct drm_device *dev = ring->dev;
1138 struct drm_i915_private *dev_priv = dev->dev_private; 1138 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1152,7 +1152,7 @@ hsw_vebox_get_irq(struct intel_ring_buffer *ring)
1152} 1152}
1153 1153
1154static void 1154static void
1155hsw_vebox_put_irq(struct intel_ring_buffer *ring) 1155hsw_vebox_put_irq(struct intel_engine_cs *ring)
1156{ 1156{
1157 struct drm_device *dev = ring->dev; 1157 struct drm_device *dev = ring->dev;
1158 struct drm_i915_private *dev_priv = dev->dev_private; 1158 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1170,7 +1170,7 @@ hsw_vebox_put_irq(struct intel_ring_buffer *ring)
1170} 1170}
1171 1171
1172static bool 1172static bool
1173gen8_ring_get_irq(struct intel_ring_buffer *ring) 1173gen8_ring_get_irq(struct intel_engine_cs *ring)
1174{ 1174{
1175 struct drm_device *dev = ring->dev; 1175 struct drm_device *dev = ring->dev;
1176 struct drm_i915_private *dev_priv = dev->dev_private; 1176 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1196,7 +1196,7 @@ gen8_ring_get_irq(struct intel_ring_buffer *ring)
1196} 1196}
1197 1197
1198static void 1198static void
1199gen8_ring_put_irq(struct intel_ring_buffer *ring) 1199gen8_ring_put_irq(struct intel_engine_cs *ring)
1200{ 1200{
1201 struct drm_device *dev = ring->dev; 1201 struct drm_device *dev = ring->dev;
1202 struct drm_i915_private *dev_priv = dev->dev_private; 1202 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1216,7 +1216,7 @@ gen8_ring_put_irq(struct intel_ring_buffer *ring)
1216} 1216}
1217 1217
1218static int 1218static int
1219i965_dispatch_execbuffer(struct intel_ring_buffer *ring, 1219i965_dispatch_execbuffer(struct intel_engine_cs *ring,
1220 u64 offset, u32 length, 1220 u64 offset, u32 length,
1221 unsigned flags) 1221 unsigned flags)
1222{ 1222{
@@ -1239,7 +1239,7 @@ i965_dispatch_execbuffer(struct intel_ring_buffer *ring,
1239/* Just userspace ABI convention to limit the wa batch bo to a resonable size */ 1239/* Just userspace ABI convention to limit the wa batch bo to a resonable size */
1240#define I830_BATCH_LIMIT (256*1024) 1240#define I830_BATCH_LIMIT (256*1024)
1241static int 1241static int
1242i830_dispatch_execbuffer(struct intel_ring_buffer *ring, 1242i830_dispatch_execbuffer(struct intel_engine_cs *ring,
1243 u64 offset, u32 len, 1243 u64 offset, u32 len,
1244 unsigned flags) 1244 unsigned flags)
1245{ 1245{
@@ -1290,7 +1290,7 @@ i830_dispatch_execbuffer(struct intel_ring_buffer *ring,
1290} 1290}
1291 1291
1292static int 1292static int
1293i915_dispatch_execbuffer(struct intel_ring_buffer *ring, 1293i915_dispatch_execbuffer(struct intel_engine_cs *ring,
1294 u64 offset, u32 len, 1294 u64 offset, u32 len,
1295 unsigned flags) 1295 unsigned flags)
1296{ 1296{
@@ -1307,7 +1307,7 @@ i915_dispatch_execbuffer(struct intel_ring_buffer *ring,
1307 return 0; 1307 return 0;
1308} 1308}
1309 1309
1310static void cleanup_status_page(struct intel_ring_buffer *ring) 1310static void cleanup_status_page(struct intel_engine_cs *ring)
1311{ 1311{
1312 struct drm_i915_gem_object *obj; 1312 struct drm_i915_gem_object *obj;
1313 1313
@@ -1321,7 +1321,7 @@ static void cleanup_status_page(struct intel_ring_buffer *ring)
1321 ring->status_page.obj = NULL; 1321 ring->status_page.obj = NULL;
1322} 1322}
1323 1323
1324static int init_status_page(struct intel_ring_buffer *ring) 1324static int init_status_page(struct intel_engine_cs *ring)
1325{ 1325{
1326 struct drm_i915_gem_object *obj; 1326 struct drm_i915_gem_object *obj;
1327 1327
@@ -1358,7 +1358,7 @@ err_unref:
1358 return 0; 1358 return 0;
1359} 1359}
1360 1360
1361static int init_phys_status_page(struct intel_ring_buffer *ring) 1361static int init_phys_status_page(struct intel_engine_cs *ring)
1362{ 1362{
1363 struct drm_i915_private *dev_priv = ring->dev->dev_private; 1363 struct drm_i915_private *dev_priv = ring->dev->dev_private;
1364 1364
@@ -1375,7 +1375,7 @@ static int init_phys_status_page(struct intel_ring_buffer *ring)
1375 return 0; 1375 return 0;
1376} 1376}
1377 1377
1378static int allocate_ring_buffer(struct intel_ring_buffer *ring) 1378static int allocate_ring_buffer(struct intel_engine_cs *ring)
1379{ 1379{
1380 struct drm_device *dev = ring->dev; 1380 struct drm_device *dev = ring->dev;
1381 struct drm_i915_private *dev_priv = to_i915(dev); 1381 struct drm_i915_private *dev_priv = to_i915(dev);
@@ -1420,7 +1420,7 @@ err_unref:
1420} 1420}
1421 1421
1422static int intel_init_ring_buffer(struct drm_device *dev, 1422static int intel_init_ring_buffer(struct drm_device *dev,
1423 struct intel_ring_buffer *ring) 1423 struct intel_engine_cs *ring)
1424{ 1424{
1425 int ret; 1425 int ret;
1426 1426
@@ -1464,7 +1464,7 @@ static int intel_init_ring_buffer(struct drm_device *dev,
1464 return ring->init(ring); 1464 return ring->init(ring);
1465} 1465}
1466 1466
1467void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) 1467void intel_cleanup_ring_buffer(struct intel_engine_cs *ring)
1468{ 1468{
1469 struct drm_i915_private *dev_priv = to_i915(ring->dev); 1469 struct drm_i915_private *dev_priv = to_i915(ring->dev);
1470 1470
@@ -1490,7 +1490,7 @@ void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
1490 i915_cmd_parser_fini_ring(ring); 1490 i915_cmd_parser_fini_ring(ring);
1491} 1491}
1492 1492
1493static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n) 1493static int intel_ring_wait_request(struct intel_engine_cs *ring, int n)
1494{ 1494{
1495 struct drm_i915_gem_request *request; 1495 struct drm_i915_gem_request *request;
1496 u32 seqno = 0; 1496 u32 seqno = 0;
@@ -1527,7 +1527,7 @@ static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n)
1527 return 0; 1527 return 0;
1528} 1528}
1529 1529
1530static int ring_wait_for_space(struct intel_ring_buffer *ring, int n) 1530static int ring_wait_for_space(struct intel_engine_cs *ring, int n)
1531{ 1531{
1532 struct drm_device *dev = ring->dev; 1532 struct drm_device *dev = ring->dev;
1533 struct drm_i915_private *dev_priv = dev->dev_private; 1533 struct drm_i915_private *dev_priv = dev->dev_private;
@@ -1585,7 +1585,7 @@ static int ring_wait_for_space(struct intel_ring_buffer *ring, int n)
1585 return ret; 1585 return ret;
1586} 1586}
1587 1587
1588static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring) 1588static int intel_wrap_ring_buffer(struct intel_engine_cs *ring)
1589{ 1589{
1590 uint32_t __iomem *virt; 1590 uint32_t __iomem *virt;
1591 int rem = ring->size - ring->tail; 1591 int rem = ring->size - ring->tail;
@@ -1607,7 +1607,7 @@ static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
1607 return 0; 1607 return 0;
1608} 1608}
1609 1609
1610int intel_ring_idle(struct intel_ring_buffer *ring) 1610int intel_ring_idle(struct intel_engine_cs *ring)
1611{ 1611{
1612 u32 seqno; 1612 u32 seqno;
1613 int ret; 1613 int ret;
@@ -1631,7 +1631,7 @@ int intel_ring_idle(struct intel_ring_buffer *ring)
1631} 1631}
1632 1632
1633static int 1633static int
1634intel_ring_alloc_seqno(struct intel_ring_buffer *ring) 1634intel_ring_alloc_seqno(struct intel_engine_cs *ring)
1635{ 1635{
1636 if (ring->outstanding_lazy_seqno) 1636 if (ring->outstanding_lazy_seqno)
1637 return 0; 1637 return 0;
@@ -1649,7 +1649,7 @@ intel_ring_alloc_seqno(struct intel_ring_buffer *ring)
1649 return i915_gem_get_seqno(ring->dev, &ring->outstanding_lazy_seqno); 1649 return i915_gem_get_seqno(ring->dev, &ring->outstanding_lazy_seqno);
1650} 1650}
1651 1651
1652static int __intel_ring_prepare(struct intel_ring_buffer *ring, 1652static int __intel_ring_prepare(struct intel_engine_cs *ring,
1653 int bytes) 1653 int bytes)
1654{ 1654{
1655 int ret; 1655 int ret;
@@ -1669,7 +1669,7 @@ static int __intel_ring_prepare(struct intel_ring_buffer *ring,
1669 return 0; 1669 return 0;
1670} 1670}
1671 1671
1672int intel_ring_begin(struct intel_ring_buffer *ring, 1672int intel_ring_begin(struct intel_engine_cs *ring,
1673 int num_dwords) 1673 int num_dwords)
1674{ 1674{
1675 struct drm_i915_private *dev_priv = ring->dev->dev_private; 1675 struct drm_i915_private *dev_priv = ring->dev->dev_private;
@@ -1694,7 +1694,7 @@ int intel_ring_begin(struct intel_ring_buffer *ring,
1694} 1694}
1695 1695
1696/* Align the ring tail to a cacheline boundary */ 1696/* Align the ring tail to a cacheline boundary */
1697int intel_ring_cacheline_align(struct intel_ring_buffer *ring) 1697int intel_ring_cacheline_align(struct intel_engine_cs *ring)
1698{ 1698{
1699 int num_dwords = (ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t); 1699 int num_dwords = (ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t);
1700 int ret; 1700 int ret;
@@ -1715,7 +1715,7 @@ int intel_ring_cacheline_align(struct intel_ring_buffer *ring)
1715 return 0; 1715 return 0;
1716} 1716}
1717 1717
1718void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno) 1718void intel_ring_init_seqno(struct intel_engine_cs *ring, u32 seqno)
1719{ 1719{
1720 struct drm_i915_private *dev_priv = ring->dev->dev_private; 1720 struct drm_i915_private *dev_priv = ring->dev->dev_private;
1721 1721
@@ -1732,7 +1732,7 @@ void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno)
1732 ring->hangcheck.seqno = seqno; 1732 ring->hangcheck.seqno = seqno;
1733} 1733}
1734 1734
1735static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, 1735static void gen6_bsd_ring_write_tail(struct intel_engine_cs *ring,
1736 u32 value) 1736 u32 value)
1737{ 1737{
1738 struct drm_i915_private *dev_priv = ring->dev->dev_private; 1738 struct drm_i915_private *dev_priv = ring->dev->dev_private;
@@ -1765,7 +1765,7 @@ static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
1765 _MASKED_BIT_DISABLE(GEN6_BSD_SLEEP_MSG_DISABLE)); 1765 _MASKED_BIT_DISABLE(GEN6_BSD_SLEEP_MSG_DISABLE));
1766} 1766}
1767 1767
1768static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring, 1768static int gen6_bsd_ring_flush(struct intel_engine_cs *ring,
1769 u32 invalidate, u32 flush) 1769 u32 invalidate, u32 flush)
1770{ 1770{
1771 uint32_t cmd; 1771 uint32_t cmd;
@@ -1801,7 +1801,7 @@ static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring,
1801} 1801}
1802 1802
1803static int 1803static int
1804gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, 1804gen8_ring_dispatch_execbuffer(struct intel_engine_cs *ring,
1805 u64 offset, u32 len, 1805 u64 offset, u32 len,
1806 unsigned flags) 1806 unsigned flags)
1807{ 1807{
@@ -1825,7 +1825,7 @@ gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
1825} 1825}
1826 1826
1827static int 1827static int
1828hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, 1828hsw_ring_dispatch_execbuffer(struct intel_engine_cs *ring,
1829 u64 offset, u32 len, 1829 u64 offset, u32 len,
1830 unsigned flags) 1830 unsigned flags)
1831{ 1831{
@@ -1846,7 +1846,7 @@ hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
1846} 1846}
1847 1847
1848static int 1848static int
1849gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, 1849gen6_ring_dispatch_execbuffer(struct intel_engine_cs *ring,
1850 u64 offset, u32 len, 1850 u64 offset, u32 len,
1851 unsigned flags) 1851 unsigned flags)
1852{ 1852{
@@ -1868,7 +1868,7 @@ gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
1868 1868
1869/* Blitter support (SandyBridge+) */ 1869/* Blitter support (SandyBridge+) */
1870 1870
1871static int gen6_ring_flush(struct intel_ring_buffer *ring, 1871static int gen6_ring_flush(struct intel_engine_cs *ring,
1872 u32 invalidate, u32 flush) 1872 u32 invalidate, u32 flush)
1873{ 1873{
1874 struct drm_device *dev = ring->dev; 1874 struct drm_device *dev = ring->dev;
@@ -1911,7 +1911,7 @@ static int gen6_ring_flush(struct intel_ring_buffer *ring,
1911int intel_init_render_ring_buffer(struct drm_device *dev) 1911int intel_init_render_ring_buffer(struct drm_device *dev)
1912{ 1912{
1913 struct drm_i915_private *dev_priv = dev->dev_private; 1913 struct drm_i915_private *dev_priv = dev->dev_private;
1914 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 1914 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
1915 1915
1916 ring->name = "render ring"; 1916 ring->name = "render ring";
1917 ring->id = RCS; 1917 ring->id = RCS;
@@ -2021,7 +2021,7 @@ int intel_init_render_ring_buffer(struct drm_device *dev)
2021int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size) 2021int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
2022{ 2022{
2023 struct drm_i915_private *dev_priv = dev->dev_private; 2023 struct drm_i915_private *dev_priv = dev->dev_private;
2024 struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; 2024 struct intel_engine_cs *ring = &dev_priv->ring[RCS];
2025 int ret; 2025 int ret;
2026 2026
2027 ring->name = "render ring"; 2027 ring->name = "render ring";
@@ -2089,7 +2089,7 @@ int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
2089int intel_init_bsd_ring_buffer(struct drm_device *dev) 2089int intel_init_bsd_ring_buffer(struct drm_device *dev)
2090{ 2090{
2091 struct drm_i915_private *dev_priv = dev->dev_private; 2091 struct drm_i915_private *dev_priv = dev->dev_private;
2092 struct intel_ring_buffer *ring = &dev_priv->ring[VCS]; 2092 struct intel_engine_cs *ring = &dev_priv->ring[VCS];
2093 2093
2094 ring->name = "bsd ring"; 2094 ring->name = "bsd ring";
2095 ring->id = VCS; 2095 ring->id = VCS;
@@ -2165,7 +2165,7 @@ int intel_init_bsd_ring_buffer(struct drm_device *dev)
2165int intel_init_bsd2_ring_buffer(struct drm_device *dev) 2165int intel_init_bsd2_ring_buffer(struct drm_device *dev)
2166{ 2166{
2167 struct drm_i915_private *dev_priv = dev->dev_private; 2167 struct drm_i915_private *dev_priv = dev->dev_private;
2168 struct intel_ring_buffer *ring = &dev_priv->ring[VCS2]; 2168 struct intel_engine_cs *ring = &dev_priv->ring[VCS2];
2169 2169
2170 if ((INTEL_INFO(dev)->gen != 8)) { 2170 if ((INTEL_INFO(dev)->gen != 8)) {
2171 DRM_ERROR("No dual-BSD ring on non-BDW machine\n"); 2171 DRM_ERROR("No dual-BSD ring on non-BDW machine\n");
@@ -2214,7 +2214,7 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev)
2214int intel_init_blt_ring_buffer(struct drm_device *dev) 2214int intel_init_blt_ring_buffer(struct drm_device *dev)
2215{ 2215{
2216 struct drm_i915_private *dev_priv = dev->dev_private; 2216 struct drm_i915_private *dev_priv = dev->dev_private;
2217 struct intel_ring_buffer *ring = &dev_priv->ring[BCS]; 2217 struct intel_engine_cs *ring = &dev_priv->ring[BCS];
2218 2218
2219 ring->name = "blitter ring"; 2219 ring->name = "blitter ring";
2220 ring->id = BCS; 2220 ring->id = BCS;
@@ -2263,7 +2263,7 @@ int intel_init_blt_ring_buffer(struct drm_device *dev)
2263int intel_init_vebox_ring_buffer(struct drm_device *dev) 2263int intel_init_vebox_ring_buffer(struct drm_device *dev)
2264{ 2264{
2265 struct drm_i915_private *dev_priv = dev->dev_private; 2265 struct drm_i915_private *dev_priv = dev->dev_private;
2266 struct intel_ring_buffer *ring = &dev_priv->ring[VECS]; 2266 struct intel_engine_cs *ring = &dev_priv->ring[VECS];
2267 2267
2268 ring->name = "video enhancement ring"; 2268 ring->name = "video enhancement ring";
2269 ring->id = VECS; 2269 ring->id = VECS;
@@ -2305,7 +2305,7 @@ int intel_init_vebox_ring_buffer(struct drm_device *dev)
2305} 2305}
2306 2306
2307int 2307int
2308intel_ring_flush_all_caches(struct intel_ring_buffer *ring) 2308intel_ring_flush_all_caches(struct intel_engine_cs *ring)
2309{ 2309{
2310 int ret; 2310 int ret;
2311 2311
@@ -2323,7 +2323,7 @@ intel_ring_flush_all_caches(struct intel_ring_buffer *ring)
2323} 2323}
2324 2324
2325int 2325int
2326intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring) 2326intel_ring_invalidate_all_caches(struct intel_engine_cs *ring)
2327{ 2327{
2328 uint32_t flush_domains; 2328 uint32_t flush_domains;
2329 int ret; 2329 int ret;
@@ -2343,7 +2343,7 @@ intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring)
2343} 2343}
2344 2344
2345void 2345void
2346intel_stop_ring_buffer(struct intel_ring_buffer *ring) 2346intel_stop_ring_buffer(struct intel_engine_cs *ring)
2347{ 2347{
2348 int ret; 2348 int ret;
2349 2349
diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h
index a505a71ee639..7027473c59d8 100644
--- a/drivers/gpu/drm/i915/intel_ringbuffer.h
+++ b/drivers/gpu/drm/i915/intel_ringbuffer.h
@@ -58,7 +58,7 @@ struct intel_ring_hangcheck {
58 bool deadlock; 58 bool deadlock;
59}; 59};
60 60
61struct intel_ring_buffer { 61struct intel_engine_cs {
62 const char *name; 62 const char *name;
63 enum intel_ring_id { 63 enum intel_ring_id {
64 RCS = 0x0, 64 RCS = 0x0,
@@ -94,33 +94,33 @@ struct intel_ring_buffer {
94 unsigned irq_refcount; /* protected by dev_priv->irq_lock */ 94 unsigned irq_refcount; /* protected by dev_priv->irq_lock */
95 u32 irq_enable_mask; /* bitmask to enable ring interrupt */ 95 u32 irq_enable_mask; /* bitmask to enable ring interrupt */
96 u32 trace_irq_seqno; 96 u32 trace_irq_seqno;
97 bool __must_check (*irq_get)(struct intel_ring_buffer *ring); 97 bool __must_check (*irq_get)(struct intel_engine_cs *ring);
98 void (*irq_put)(struct intel_ring_buffer *ring); 98 void (*irq_put)(struct intel_engine_cs *ring);
99 99
100 int (*init)(struct intel_ring_buffer *ring); 100 int (*init)(struct intel_engine_cs *ring);
101 101
102 void (*write_tail)(struct intel_ring_buffer *ring, 102 void (*write_tail)(struct intel_engine_cs *ring,
103 u32 value); 103 u32 value);
104 int __must_check (*flush)(struct intel_ring_buffer *ring, 104 int __must_check (*flush)(struct intel_engine_cs *ring,
105 u32 invalidate_domains, 105 u32 invalidate_domains,
106 u32 flush_domains); 106 u32 flush_domains);
107 int (*add_request)(struct intel_ring_buffer *ring); 107 int (*add_request)(struct intel_engine_cs *ring);
108 /* Some chipsets are not quite as coherent as advertised and need 108 /* Some chipsets are not quite as coherent as advertised and need
109 * an expensive kick to force a true read of the up-to-date seqno. 109 * an expensive kick to force a true read of the up-to-date seqno.
110 * However, the up-to-date seqno is not always required and the last 110 * However, the up-to-date seqno is not always required and the last
111 * seen value is good enough. Note that the seqno will always be 111 * seen value is good enough. Note that the seqno will always be
112 * monotonic, even if not coherent. 112 * monotonic, even if not coherent.
113 */ 113 */
114 u32 (*get_seqno)(struct intel_ring_buffer *ring, 114 u32 (*get_seqno)(struct intel_engine_cs *ring,
115 bool lazy_coherency); 115 bool lazy_coherency);
116 void (*set_seqno)(struct intel_ring_buffer *ring, 116 void (*set_seqno)(struct intel_engine_cs *ring,
117 u32 seqno); 117 u32 seqno);
118 int (*dispatch_execbuffer)(struct intel_ring_buffer *ring, 118 int (*dispatch_execbuffer)(struct intel_engine_cs *ring,
119 u64 offset, u32 length, 119 u64 offset, u32 length,
120 unsigned flags); 120 unsigned flags);
121#define I915_DISPATCH_SECURE 0x1 121#define I915_DISPATCH_SECURE 0x1
122#define I915_DISPATCH_PINNED 0x2 122#define I915_DISPATCH_PINNED 0x2
123 void (*cleanup)(struct intel_ring_buffer *ring); 123 void (*cleanup)(struct intel_engine_cs *ring);
124 124
125 struct { 125 struct {
126 u32 sync_seqno[I915_NUM_RINGS-1]; 126 u32 sync_seqno[I915_NUM_RINGS-1];
@@ -133,10 +133,10 @@ struct intel_ring_buffer {
133 } mbox; 133 } mbox;
134 134
135 /* AKA wait() */ 135 /* AKA wait() */
136 int (*sync_to)(struct intel_ring_buffer *ring, 136 int (*sync_to)(struct intel_engine_cs *ring,
137 struct intel_ring_buffer *to, 137 struct intel_engine_cs *to,
138 u32 seqno); 138 u32 seqno);
139 int (*signal)(struct intel_ring_buffer *signaller, 139 int (*signal)(struct intel_engine_cs *signaller,
140 /* num_dwords needed by caller */ 140 /* num_dwords needed by caller */
141 unsigned int num_dwords); 141 unsigned int num_dwords);
142 } semaphore; 142 } semaphore;
@@ -215,20 +215,20 @@ struct intel_ring_buffer {
215}; 215};
216 216
217static inline bool 217static inline bool
218intel_ring_initialized(struct intel_ring_buffer *ring) 218intel_ring_initialized(struct intel_engine_cs *ring)
219{ 219{
220 return ring->obj != NULL; 220 return ring->obj != NULL;
221} 221}
222 222
223static inline unsigned 223static inline unsigned
224intel_ring_flag(struct intel_ring_buffer *ring) 224intel_ring_flag(struct intel_engine_cs *ring)
225{ 225{
226 return 1 << ring->id; 226 return 1 << ring->id;
227} 227}
228 228
229static inline u32 229static inline u32
230intel_ring_sync_index(struct intel_ring_buffer *ring, 230intel_ring_sync_index(struct intel_engine_cs *ring,
231 struct intel_ring_buffer *other) 231 struct intel_engine_cs *other)
232{ 232{
233 int idx; 233 int idx;
234 234
@@ -246,7 +246,7 @@ intel_ring_sync_index(struct intel_ring_buffer *ring,
246} 246}
247 247
248static inline u32 248static inline u32
249intel_read_status_page(struct intel_ring_buffer *ring, 249intel_read_status_page(struct intel_engine_cs *ring,
250 int reg) 250 int reg)
251{ 251{
252 /* Ensure that the compiler doesn't optimize away the load. */ 252 /* Ensure that the compiler doesn't optimize away the load. */
@@ -255,7 +255,7 @@ intel_read_status_page(struct intel_ring_buffer *ring,
255} 255}
256 256
257static inline void 257static inline void
258intel_write_status_page(struct intel_ring_buffer *ring, 258intel_write_status_page(struct intel_engine_cs *ring,
259 int reg, u32 value) 259 int reg, u32 value)
260{ 260{
261 ring->status_page.page_addr[reg] = value; 261 ring->status_page.page_addr[reg] = value;
@@ -280,27 +280,27 @@ intel_write_status_page(struct intel_ring_buffer *ring,
280#define I915_GEM_HWS_SCRATCH_INDEX 0x30 280#define I915_GEM_HWS_SCRATCH_INDEX 0x30
281#define I915_GEM_HWS_SCRATCH_ADDR (I915_GEM_HWS_SCRATCH_INDEX << MI_STORE_DWORD_INDEX_SHIFT) 281#define I915_GEM_HWS_SCRATCH_ADDR (I915_GEM_HWS_SCRATCH_INDEX << MI_STORE_DWORD_INDEX_SHIFT)
282 282
283void intel_stop_ring_buffer(struct intel_ring_buffer *ring); 283void intel_stop_ring_buffer(struct intel_engine_cs *ring);
284void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring); 284void intel_cleanup_ring_buffer(struct intel_engine_cs *ring);
285 285
286int __must_check intel_ring_begin(struct intel_ring_buffer *ring, int n); 286int __must_check intel_ring_begin(struct intel_engine_cs *ring, int n);
287int __must_check intel_ring_cacheline_align(struct intel_ring_buffer *ring); 287int __must_check intel_ring_cacheline_align(struct intel_engine_cs *ring);
288static inline void intel_ring_emit(struct intel_ring_buffer *ring, 288static inline void intel_ring_emit(struct intel_engine_cs *ring,
289 u32 data) 289 u32 data)
290{ 290{
291 iowrite32(data, ring->virtual_start + ring->tail); 291 iowrite32(data, ring->virtual_start + ring->tail);
292 ring->tail += 4; 292 ring->tail += 4;
293} 293}
294static inline void intel_ring_advance(struct intel_ring_buffer *ring) 294static inline void intel_ring_advance(struct intel_engine_cs *ring)
295{ 295{
296 ring->tail &= ring->size - 1; 296 ring->tail &= ring->size - 1;
297} 297}
298void __intel_ring_advance(struct intel_ring_buffer *ring); 298void __intel_ring_advance(struct intel_engine_cs *ring);
299 299
300int __must_check intel_ring_idle(struct intel_ring_buffer *ring); 300int __must_check intel_ring_idle(struct intel_engine_cs *ring);
301void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno); 301void intel_ring_init_seqno(struct intel_engine_cs *ring, u32 seqno);
302int intel_ring_flush_all_caches(struct intel_ring_buffer *ring); 302int intel_ring_flush_all_caches(struct intel_engine_cs *ring);
303int intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring); 303int intel_ring_invalidate_all_caches(struct intel_engine_cs *ring);
304 304
305int intel_init_render_ring_buffer(struct drm_device *dev); 305int intel_init_render_ring_buffer(struct drm_device *dev);
306int intel_init_bsd_ring_buffer(struct drm_device *dev); 306int intel_init_bsd_ring_buffer(struct drm_device *dev);
@@ -308,21 +308,21 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev);
308int intel_init_blt_ring_buffer(struct drm_device *dev); 308int intel_init_blt_ring_buffer(struct drm_device *dev);
309int intel_init_vebox_ring_buffer(struct drm_device *dev); 309int intel_init_vebox_ring_buffer(struct drm_device *dev);
310 310
311u64 intel_ring_get_active_head(struct intel_ring_buffer *ring); 311u64 intel_ring_get_active_head(struct intel_engine_cs *ring);
312void intel_ring_setup_status_page(struct intel_ring_buffer *ring); 312void intel_ring_setup_status_page(struct intel_engine_cs *ring);
313 313
314static inline u32 intel_ring_get_tail(struct intel_ring_buffer *ring) 314static inline u32 intel_ring_get_tail(struct intel_engine_cs *ring)
315{ 315{
316 return ring->tail; 316 return ring->tail;
317} 317}
318 318
319static inline u32 intel_ring_get_seqno(struct intel_ring_buffer *ring) 319static inline u32 intel_ring_get_seqno(struct intel_engine_cs *ring)
320{ 320{
321 BUG_ON(ring->outstanding_lazy_seqno == 0); 321 BUG_ON(ring->outstanding_lazy_seqno == 0);
322 return ring->outstanding_lazy_seqno; 322 return ring->outstanding_lazy_seqno;
323} 323}
324 324
325static inline void i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno) 325static inline void i915_trace_irq_get(struct intel_engine_cs *ring, u32 seqno)
326{ 326{
327 if (ring->trace_irq_seqno == 0 && ring->irq_get(ring)) 327 if (ring->trace_irq_seqno == 0 && ring->irq_get(ring))
328 ring->trace_irq_seqno = seqno; 328 ring->trace_irq_seqno = seqno;