diff options
author | Francisco Jerez <currojerez@riseup.net> | 2010-09-21 12:57:11 -0400 |
---|---|---|
committer | Ben Skeggs <bskeggs@redhat.com> | 2010-10-04 19:58:59 -0400 |
commit | 2730723bbc4a8b289fa536fc3555e15947da09c1 (patch) | |
tree | 0d90e13273f7e3304532167ff91ee668d3b92092 /drivers | |
parent | 0cba1b7644cbcd855d0a2b2ea4d8da26fd08dec4 (diff) |
drm/nouveau: Minor refactoring/cleanup of the fence code.
Mainly to make room for inter-channel sync.
Signed-off-by: Francisco Jerez <currojerez@riseup.net>
Signed-off-by: Ben Skeggs <bskeggs@redhat.com>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_bo.c | 15 | ||||
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_channel.c | 4 | ||||
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_dma.c | 17 | ||||
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_drv.h | 11 | ||||
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_fence.c | 63 | ||||
-rw-r--r-- | drivers/gpu/drm/nouveau/nouveau_gem.c | 4 |
6 files changed, 59 insertions, 55 deletions
diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c index 2148e2d73de3..03d842a06b04 100644 --- a/drivers/gpu/drm/nouveau/nouveau_bo.c +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c | |||
@@ -36,21 +36,6 @@ | |||
36 | #include <linux/log2.h> | 36 | #include <linux/log2.h> |
37 | #include <linux/slab.h> | 37 | #include <linux/slab.h> |
38 | 38 | ||
39 | int | ||
40 | nouveau_bo_sync_gpu(struct nouveau_bo *nvbo, struct nouveau_channel *chan) | ||
41 | { | ||
42 | struct nouveau_fence *prev_fence = nvbo->bo.sync_obj; | ||
43 | int ret; | ||
44 | |||
45 | if (!prev_fence || nouveau_fence_channel(prev_fence) == chan) | ||
46 | return 0; | ||
47 | |||
48 | spin_lock(&nvbo->bo.lock); | ||
49 | ret = ttm_bo_wait(&nvbo->bo, false, false, false); | ||
50 | spin_unlock(&nvbo->bo.lock); | ||
51 | return ret; | ||
52 | } | ||
53 | |||
54 | static void | 39 | static void |
55 | nouveau_bo_del_ttm(struct ttm_buffer_object *bo) | 40 | nouveau_bo_del_ttm(struct ttm_buffer_object *bo) |
56 | { | 41 | { |
diff --git a/drivers/gpu/drm/nouveau/nouveau_channel.c b/drivers/gpu/drm/nouveau/nouveau_channel.c index e01396747f6f..5eb4c966273f 100644 --- a/drivers/gpu/drm/nouveau/nouveau_channel.c +++ b/drivers/gpu/drm/nouveau/nouveau_channel.c | |||
@@ -223,7 +223,7 @@ nouveau_channel_alloc(struct drm_device *dev, struct nouveau_channel **chan_ret, | |||
223 | 223 | ||
224 | ret = nouveau_dma_init(chan); | 224 | ret = nouveau_dma_init(chan); |
225 | if (!ret) | 225 | if (!ret) |
226 | ret = nouveau_fence_init(chan); | 226 | ret = nouveau_fence_channel_init(chan); |
227 | if (ret) { | 227 | if (ret) { |
228 | nouveau_channel_free(chan); | 228 | nouveau_channel_free(chan); |
229 | return ret; | 229 | return ret; |
@@ -270,7 +270,7 @@ nouveau_channel_free(struct nouveau_channel *chan) | |||
270 | * above attempts at idling were OK, but if we failed this'll tell TTM | 270 | * above attempts at idling were OK, but if we failed this'll tell TTM |
271 | * we're done with the buffers. | 271 | * we're done with the buffers. |
272 | */ | 272 | */ |
273 | nouveau_fence_fini(chan); | 273 | nouveau_fence_channel_fini(chan); |
274 | 274 | ||
275 | /* This will prevent pfifo from switching channels. */ | 275 | /* This will prevent pfifo from switching channels. */ |
276 | pfifo->reassign(dev, false); | 276 | pfifo->reassign(dev, false); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_dma.c b/drivers/gpu/drm/nouveau/nouveau_dma.c index eb24e2b05193..82581e600dcd 100644 --- a/drivers/gpu/drm/nouveau/nouveau_dma.c +++ b/drivers/gpu/drm/nouveau/nouveau_dma.c | |||
@@ -73,16 +73,6 @@ nouveau_dma_init(struct nouveau_channel *chan) | |||
73 | if (ret) | 73 | if (ret) |
74 | return ret; | 74 | return ret; |
75 | 75 | ||
76 | /* Create an NV_SW object for various sync purposes */ | ||
77 | ret = nouveau_gpuobj_sw_new(chan, NV_SW, &obj); | ||
78 | if (ret) | ||
79 | return ret; | ||
80 | |||
81 | ret = nouveau_ramht_insert(chan, NvSw, obj); | ||
82 | nouveau_gpuobj_ref(NULL, &obj); | ||
83 | if (ret) | ||
84 | return ret; | ||
85 | |||
86 | /* NV_MEMORY_TO_MEMORY_FORMAT requires a notifier object */ | 76 | /* NV_MEMORY_TO_MEMORY_FORMAT requires a notifier object */ |
87 | ret = nouveau_notifier_alloc(chan, NvNotify0, 32, &chan->m2mf_ntfy); | 77 | ret = nouveau_notifier_alloc(chan, NvNotify0, 32, &chan->m2mf_ntfy); |
88 | if (ret) | 78 | if (ret) |
@@ -110,13 +100,6 @@ nouveau_dma_init(struct nouveau_channel *chan) | |||
110 | BEGIN_RING(chan, NvSubM2MF, NV_MEMORY_TO_MEMORY_FORMAT_DMA_NOTIFY, 1); | 100 | BEGIN_RING(chan, NvSubM2MF, NV_MEMORY_TO_MEMORY_FORMAT_DMA_NOTIFY, 1); |
111 | OUT_RING(chan, NvNotify0); | 101 | OUT_RING(chan, NvNotify0); |
112 | 102 | ||
113 | /* Initialise NV_SW */ | ||
114 | ret = RING_SPACE(chan, 2); | ||
115 | if (ret) | ||
116 | return ret; | ||
117 | BEGIN_RING(chan, NvSubSw, 0, 1); | ||
118 | OUT_RING(chan, NvSw); | ||
119 | |||
120 | /* Sit back and pray the channel works.. */ | 103 | /* Sit back and pray the channel works.. */ |
121 | FIRE_RING(chan); | 104 | FIRE_RING(chan); |
122 | 105 | ||
diff --git a/drivers/gpu/drm/nouveau/nouveau_drv.h b/drivers/gpu/drm/nouveau/nouveau_drv.h index b209a6208757..9a9066f1fbcc 100644 --- a/drivers/gpu/drm/nouveau/nouveau_drv.h +++ b/drivers/gpu/drm/nouveau/nouveau_drv.h | |||
@@ -706,6 +706,12 @@ struct drm_nouveau_private { | |||
706 | }; | 706 | }; |
707 | 707 | ||
708 | static inline struct drm_nouveau_private * | 708 | static inline struct drm_nouveau_private * |
709 | nouveau_private(struct drm_device *dev) | ||
710 | { | ||
711 | return dev->dev_private; | ||
712 | } | ||
713 | |||
714 | static inline struct drm_nouveau_private * | ||
709 | nouveau_bdev(struct ttm_bo_device *bd) | 715 | nouveau_bdev(struct ttm_bo_device *bd) |
710 | { | 716 | { |
711 | return container_of(bd, struct drm_nouveau_private, ttm.bdev); | 717 | return container_of(bd, struct drm_nouveau_private, ttm.bdev); |
@@ -1231,8 +1237,8 @@ extern int nouveau_bo_sync_gpu(struct nouveau_bo *, struct nouveau_channel *); | |||
1231 | 1237 | ||
1232 | /* nouveau_fence.c */ | 1238 | /* nouveau_fence.c */ |
1233 | struct nouveau_fence; | 1239 | struct nouveau_fence; |
1234 | extern int nouveau_fence_init(struct nouveau_channel *); | 1240 | extern int nouveau_fence_channel_init(struct nouveau_channel *); |
1235 | extern void nouveau_fence_fini(struct nouveau_channel *); | 1241 | extern void nouveau_fence_channel_fini(struct nouveau_channel *); |
1236 | extern void nouveau_fence_update(struct nouveau_channel *); | 1242 | extern void nouveau_fence_update(struct nouveau_channel *); |
1237 | extern int nouveau_fence_new(struct nouveau_channel *, struct nouveau_fence **, | 1243 | extern int nouveau_fence_new(struct nouveau_channel *, struct nouveau_fence **, |
1238 | bool emit); | 1244 | bool emit); |
@@ -1240,6 +1246,7 @@ extern int nouveau_fence_emit(struct nouveau_fence *); | |||
1240 | struct nouveau_channel *nouveau_fence_channel(struct nouveau_fence *); | 1246 | struct nouveau_channel *nouveau_fence_channel(struct nouveau_fence *); |
1241 | extern bool nouveau_fence_signalled(void *obj, void *arg); | 1247 | extern bool nouveau_fence_signalled(void *obj, void *arg); |
1242 | extern int nouveau_fence_wait(void *obj, void *arg, bool lazy, bool intr); | 1248 | extern int nouveau_fence_wait(void *obj, void *arg, bool lazy, bool intr); |
1249 | extern int nouveau_fence_sync(struct nouveau_fence *, struct nouveau_channel *); | ||
1243 | extern int nouveau_fence_flush(void *obj, void *arg); | 1250 | extern int nouveau_fence_flush(void *obj, void *arg); |
1244 | extern void nouveau_fence_unref(void **obj); | 1251 | extern void nouveau_fence_unref(void **obj); |
1245 | extern void *nouveau_fence_ref(void *obj); | 1252 | extern void *nouveau_fence_ref(void *obj); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fence.c b/drivers/gpu/drm/nouveau/nouveau_fence.c index 87ac21ec23d2..62f13189698a 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fence.c +++ b/drivers/gpu/drm/nouveau/nouveau_fence.c | |||
@@ -30,7 +30,7 @@ | |||
30 | #include "nouveau_drv.h" | 30 | #include "nouveau_drv.h" |
31 | #include "nouveau_dma.h" | 31 | #include "nouveau_dma.h" |
32 | 32 | ||
33 | #define USE_REFCNT (dev_priv->card_type >= NV_10) | 33 | #define USE_REFCNT(dev) (nouveau_private(dev)->chipset >= 0x10) |
34 | 34 | ||
35 | struct nouveau_fence { | 35 | struct nouveau_fence { |
36 | struct nouveau_channel *channel; | 36 | struct nouveau_channel *channel; |
@@ -59,14 +59,13 @@ nouveau_fence_del(struct kref *ref) | |||
59 | void | 59 | void |
60 | nouveau_fence_update(struct nouveau_channel *chan) | 60 | nouveau_fence_update(struct nouveau_channel *chan) |
61 | { | 61 | { |
62 | struct drm_nouveau_private *dev_priv = chan->dev->dev_private; | 62 | struct drm_device *dev = chan->dev; |
63 | struct list_head *entry, *tmp; | 63 | struct nouveau_fence *tmp, *fence; |
64 | struct nouveau_fence *fence; | ||
65 | uint32_t sequence; | 64 | uint32_t sequence; |
66 | 65 | ||
67 | spin_lock(&chan->fence.lock); | 66 | spin_lock(&chan->fence.lock); |
68 | 67 | ||
69 | if (USE_REFCNT) | 68 | if (USE_REFCNT(dev)) |
70 | sequence = nvchan_rd32(chan, 0x48); | 69 | sequence = nvchan_rd32(chan, 0x48); |
71 | else | 70 | else |
72 | sequence = atomic_read(&chan->fence.last_sequence_irq); | 71 | sequence = atomic_read(&chan->fence.last_sequence_irq); |
@@ -75,9 +74,7 @@ nouveau_fence_update(struct nouveau_channel *chan) | |||
75 | goto out; | 74 | goto out; |
76 | chan->fence.sequence_ack = sequence; | 75 | chan->fence.sequence_ack = sequence; |
77 | 76 | ||
78 | list_for_each_safe(entry, tmp, &chan->fence.pending) { | 77 | list_for_each_entry_safe(fence, tmp, &chan->fence.pending, entry) { |
79 | fence = list_entry(entry, struct nouveau_fence, entry); | ||
80 | |||
81 | sequence = fence->sequence; | 78 | sequence = fence->sequence; |
82 | fence->signalled = true; | 79 | fence->signalled = true; |
83 | list_del(&fence->entry); | 80 | list_del(&fence->entry); |
@@ -121,8 +118,8 @@ nouveau_fence_channel(struct nouveau_fence *fence) | |||
121 | int | 118 | int |
122 | nouveau_fence_emit(struct nouveau_fence *fence) | 119 | nouveau_fence_emit(struct nouveau_fence *fence) |
123 | { | 120 | { |
124 | struct drm_nouveau_private *dev_priv = fence->channel->dev->dev_private; | ||
125 | struct nouveau_channel *chan = fence->channel; | 121 | struct nouveau_channel *chan = fence->channel; |
122 | struct drm_device *dev = chan->dev; | ||
126 | int ret; | 123 | int ret; |
127 | 124 | ||
128 | ret = RING_SPACE(chan, 2); | 125 | ret = RING_SPACE(chan, 2); |
@@ -143,7 +140,7 @@ nouveau_fence_emit(struct nouveau_fence *fence) | |||
143 | list_add_tail(&fence->entry, &chan->fence.pending); | 140 | list_add_tail(&fence->entry, &chan->fence.pending); |
144 | spin_unlock(&chan->fence.lock); | 141 | spin_unlock(&chan->fence.lock); |
145 | 142 | ||
146 | BEGIN_RING(chan, NvSubSw, USE_REFCNT ? 0x0050 : 0x0150, 1); | 143 | BEGIN_RING(chan, NvSubSw, USE_REFCNT(dev) ? 0x0050 : 0x0150, 1); |
147 | OUT_RING(chan, fence->sequence); | 144 | OUT_RING(chan, fence->sequence); |
148 | FIRE_RING(chan); | 145 | FIRE_RING(chan); |
149 | 146 | ||
@@ -214,29 +211,61 @@ nouveau_fence_wait(void *sync_obj, void *sync_arg, bool lazy, bool intr) | |||
214 | } | 211 | } |
215 | 212 | ||
216 | int | 213 | int |
214 | nouveau_fence_sync(struct nouveau_fence *fence, | ||
215 | struct nouveau_channel *wchan) | ||
216 | { | ||
217 | struct nouveau_channel *chan = nouveau_fence_channel(fence); | ||
218 | |||
219 | if (likely(!fence || chan == wchan || | ||
220 | nouveau_fence_signalled(fence, NULL))) | ||
221 | return 0; | ||
222 | |||
223 | return nouveau_fence_wait(fence, NULL, false, false); | ||
224 | } | ||
225 | |||
226 | int | ||
217 | nouveau_fence_flush(void *sync_obj, void *sync_arg) | 227 | nouveau_fence_flush(void *sync_obj, void *sync_arg) |
218 | { | 228 | { |
219 | return 0; | 229 | return 0; |
220 | } | 230 | } |
221 | 231 | ||
222 | int | 232 | int |
223 | nouveau_fence_init(struct nouveau_channel *chan) | 233 | nouveau_fence_channel_init(struct nouveau_channel *chan) |
224 | { | 234 | { |
235 | struct nouveau_gpuobj *obj = NULL; | ||
236 | int ret; | ||
237 | |||
238 | /* Create an NV_SW object for various sync purposes */ | ||
239 | ret = nouveau_gpuobj_sw_new(chan, NV_SW, &obj); | ||
240 | if (ret) | ||
241 | return ret; | ||
242 | |||
243 | ret = nouveau_ramht_insert(chan, NvSw, obj); | ||
244 | nouveau_gpuobj_ref(NULL, &obj); | ||
245 | if (ret) | ||
246 | return ret; | ||
247 | |||
248 | ret = RING_SPACE(chan, 2); | ||
249 | if (ret) | ||
250 | return ret; | ||
251 | BEGIN_RING(chan, NvSubSw, 0, 1); | ||
252 | OUT_RING(chan, NvSw); | ||
253 | |||
254 | FIRE_RING(chan); | ||
255 | |||
225 | INIT_LIST_HEAD(&chan->fence.pending); | 256 | INIT_LIST_HEAD(&chan->fence.pending); |
226 | spin_lock_init(&chan->fence.lock); | 257 | spin_lock_init(&chan->fence.lock); |
227 | atomic_set(&chan->fence.last_sequence_irq, 0); | 258 | atomic_set(&chan->fence.last_sequence_irq, 0); |
259 | |||
228 | return 0; | 260 | return 0; |
229 | } | 261 | } |
230 | 262 | ||
231 | void | 263 | void |
232 | nouveau_fence_fini(struct nouveau_channel *chan) | 264 | nouveau_fence_channel_fini(struct nouveau_channel *chan) |
233 | { | 265 | { |
234 | struct list_head *entry, *tmp; | 266 | struct nouveau_fence *tmp, *fence; |
235 | struct nouveau_fence *fence; | ||
236 | |||
237 | list_for_each_safe(entry, tmp, &chan->fence.pending) { | ||
238 | fence = list_entry(entry, struct nouveau_fence, entry); | ||
239 | 267 | ||
268 | list_for_each_entry_safe(fence, tmp, &chan->fence.pending, entry) { | ||
240 | fence->signalled = true; | 269 | fence->signalled = true; |
241 | list_del(&fence->entry); | 270 | list_del(&fence->entry); |
242 | kref_put(&fence->refcount, nouveau_fence_del); | 271 | kref_put(&fence->refcount, nouveau_fence_del); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_gem.c b/drivers/gpu/drm/nouveau/nouveau_gem.c index ead7b8fc53fc..1e630987543e 100644 --- a/drivers/gpu/drm/nouveau/nouveau_gem.c +++ b/drivers/gpu/drm/nouveau/nouveau_gem.c | |||
@@ -364,7 +364,7 @@ validate_list(struct nouveau_channel *chan, struct list_head *list, | |||
364 | list_for_each_entry(nvbo, list, entry) { | 364 | list_for_each_entry(nvbo, list, entry) { |
365 | struct drm_nouveau_gem_pushbuf_bo *b = &pbbo[nvbo->pbbo_index]; | 365 | struct drm_nouveau_gem_pushbuf_bo *b = &pbbo[nvbo->pbbo_index]; |
366 | 366 | ||
367 | ret = nouveau_bo_sync_gpu(nvbo, chan); | 367 | ret = nouveau_fence_sync(nvbo->bo.sync_obj, chan); |
368 | if (unlikely(ret)) { | 368 | if (unlikely(ret)) { |
369 | NV_ERROR(dev, "fail pre-validate sync\n"); | 369 | NV_ERROR(dev, "fail pre-validate sync\n"); |
370 | return ret; | 370 | return ret; |
@@ -387,7 +387,7 @@ validate_list(struct nouveau_channel *chan, struct list_head *list, | |||
387 | return ret; | 387 | return ret; |
388 | } | 388 | } |
389 | 389 | ||
390 | ret = nouveau_bo_sync_gpu(nvbo, chan); | 390 | ret = nouveau_fence_sync(nvbo->bo.sync_obj, chan); |
391 | if (unlikely(ret)) { | 391 | if (unlikely(ret)) { |
392 | NV_ERROR(dev, "fail post-validate sync\n"); | 392 | NV_ERROR(dev, "fail post-validate sync\n"); |
393 | return ret; | 393 | return ret; |