diff options
Diffstat (limited to 'drivers/gpu/drm/nouveau/nv50_fifo.c')
-rw-r--r-- | drivers/gpu/drm/nouveau/nv50_fifo.c | 500 |
1 files changed, 500 insertions, 0 deletions
diff --git a/drivers/gpu/drm/nouveau/nv50_fifo.c b/drivers/gpu/drm/nouveau/nv50_fifo.c new file mode 100644 index 000000000000..e20c0e2474f3 --- /dev/null +++ b/drivers/gpu/drm/nouveau/nv50_fifo.c | |||
@@ -0,0 +1,500 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2007 Ben Skeggs. | ||
3 | * All Rights Reserved. | ||
4 | * | ||
5 | * Permission is hereby granted, free of charge, to any person obtaining | ||
6 | * a copy of this software and associated documentation files (the | ||
7 | * "Software"), to deal in the Software without restriction, including | ||
8 | * without limitation the rights to use, copy, modify, merge, publish, | ||
9 | * distribute, sublicense, and/or sell copies of the Software, and to | ||
10 | * permit persons to whom the Software is furnished to do so, subject to | ||
11 | * the following conditions: | ||
12 | * | ||
13 | * The above copyright notice and this permission notice (including the | ||
14 | * next paragraph) shall be included in all copies or substantial | ||
15 | * portions of the Software. | ||
16 | * | ||
17 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | ||
18 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | ||
19 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. | ||
20 | * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE | ||
21 | * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION | ||
22 | * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION | ||
23 | * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
24 | * | ||
25 | */ | ||
26 | |||
27 | #include "drmP.h" | ||
28 | #include "drm.h" | ||
29 | #include "nouveau_drv.h" | ||
30 | |||
31 | struct nv50_fifo_priv { | ||
32 | struct nouveau_gpuobj_ref *thingo[2]; | ||
33 | int cur_thingo; | ||
34 | }; | ||
35 | |||
36 | #define IS_G80 ((dev_priv->chipset & 0xf0) == 0x50) | ||
37 | |||
38 | static void | ||
39 | nv50_fifo_init_thingo(struct drm_device *dev) | ||
40 | { | ||
41 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
42 | struct nv50_fifo_priv *priv = dev_priv->engine.fifo.priv; | ||
43 | struct nouveau_gpuobj_ref *cur; | ||
44 | int i, nr; | ||
45 | |||
46 | NV_DEBUG(dev, "\n"); | ||
47 | |||
48 | cur = priv->thingo[priv->cur_thingo]; | ||
49 | priv->cur_thingo = !priv->cur_thingo; | ||
50 | |||
51 | /* We never schedule channel 0 or 127 */ | ||
52 | dev_priv->engine.instmem.prepare_access(dev, true); | ||
53 | for (i = 1, nr = 0; i < 127; i++) { | ||
54 | if (dev_priv->fifos[i] && dev_priv->fifos[i]->ramfc) | ||
55 | nv_wo32(dev, cur->gpuobj, nr++, i); | ||
56 | } | ||
57 | dev_priv->engine.instmem.finish_access(dev); | ||
58 | |||
59 | nv_wr32(dev, 0x32f4, cur->instance >> 12); | ||
60 | nv_wr32(dev, 0x32ec, nr); | ||
61 | nv_wr32(dev, 0x2500, 0x101); | ||
62 | } | ||
63 | |||
64 | static int | ||
65 | nv50_fifo_channel_enable(struct drm_device *dev, int channel, bool nt) | ||
66 | { | ||
67 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
68 | struct nouveau_channel *chan = dev_priv->fifos[channel]; | ||
69 | uint32_t inst; | ||
70 | |||
71 | NV_DEBUG(dev, "ch%d\n", channel); | ||
72 | |||
73 | if (!chan->ramfc) | ||
74 | return -EINVAL; | ||
75 | |||
76 | if (IS_G80) | ||
77 | inst = chan->ramfc->instance >> 12; | ||
78 | else | ||
79 | inst = chan->ramfc->instance >> 8; | ||
80 | nv_wr32(dev, NV50_PFIFO_CTX_TABLE(channel), | ||
81 | inst | NV50_PFIFO_CTX_TABLE_CHANNEL_ENABLED); | ||
82 | |||
83 | if (!nt) | ||
84 | nv50_fifo_init_thingo(dev); | ||
85 | return 0; | ||
86 | } | ||
87 | |||
88 | static void | ||
89 | nv50_fifo_channel_disable(struct drm_device *dev, int channel, bool nt) | ||
90 | { | ||
91 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
92 | uint32_t inst; | ||
93 | |||
94 | NV_DEBUG(dev, "ch%d, nt=%d\n", channel, nt); | ||
95 | |||
96 | if (IS_G80) | ||
97 | inst = NV50_PFIFO_CTX_TABLE_INSTANCE_MASK_G80; | ||
98 | else | ||
99 | inst = NV50_PFIFO_CTX_TABLE_INSTANCE_MASK_G84; | ||
100 | nv_wr32(dev, NV50_PFIFO_CTX_TABLE(channel), inst); | ||
101 | |||
102 | if (!nt) | ||
103 | nv50_fifo_init_thingo(dev); | ||
104 | } | ||
105 | |||
106 | static void | ||
107 | nv50_fifo_init_reset(struct drm_device *dev) | ||
108 | { | ||
109 | uint32_t pmc_e = NV_PMC_ENABLE_PFIFO; | ||
110 | |||
111 | NV_DEBUG(dev, "\n"); | ||
112 | |||
113 | nv_wr32(dev, NV03_PMC_ENABLE, nv_rd32(dev, NV03_PMC_ENABLE) & ~pmc_e); | ||
114 | nv_wr32(dev, NV03_PMC_ENABLE, nv_rd32(dev, NV03_PMC_ENABLE) | pmc_e); | ||
115 | } | ||
116 | |||
117 | static void | ||
118 | nv50_fifo_init_intr(struct drm_device *dev) | ||
119 | { | ||
120 | NV_DEBUG(dev, "\n"); | ||
121 | |||
122 | nv_wr32(dev, NV03_PFIFO_INTR_0, 0xFFFFFFFF); | ||
123 | nv_wr32(dev, NV03_PFIFO_INTR_EN_0, 0xFFFFFFFF); | ||
124 | } | ||
125 | |||
126 | static void | ||
127 | nv50_fifo_init_context_table(struct drm_device *dev) | ||
128 | { | ||
129 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
130 | int i; | ||
131 | |||
132 | NV_DEBUG(dev, "\n"); | ||
133 | |||
134 | for (i = 0; i < NV50_PFIFO_CTX_TABLE__SIZE; i++) { | ||
135 | if (dev_priv->fifos[i]) | ||
136 | nv50_fifo_channel_enable(dev, i, true); | ||
137 | else | ||
138 | nv50_fifo_channel_disable(dev, i, true); | ||
139 | } | ||
140 | |||
141 | nv50_fifo_init_thingo(dev); | ||
142 | } | ||
143 | |||
144 | static void | ||
145 | nv50_fifo_init_regs__nv(struct drm_device *dev) | ||
146 | { | ||
147 | NV_DEBUG(dev, "\n"); | ||
148 | |||
149 | nv_wr32(dev, 0x250c, 0x6f3cfc34); | ||
150 | } | ||
151 | |||
152 | static void | ||
153 | nv50_fifo_init_regs(struct drm_device *dev) | ||
154 | { | ||
155 | NV_DEBUG(dev, "\n"); | ||
156 | |||
157 | nv_wr32(dev, 0x2500, 0); | ||
158 | nv_wr32(dev, 0x3250, 0); | ||
159 | nv_wr32(dev, 0x3220, 0); | ||
160 | nv_wr32(dev, 0x3204, 0); | ||
161 | nv_wr32(dev, 0x3210, 0); | ||
162 | nv_wr32(dev, 0x3270, 0); | ||
163 | |||
164 | /* Enable dummy channels setup by nv50_instmem.c */ | ||
165 | nv50_fifo_channel_enable(dev, 0, true); | ||
166 | nv50_fifo_channel_enable(dev, 127, true); | ||
167 | } | ||
168 | |||
169 | int | ||
170 | nv50_fifo_init(struct drm_device *dev) | ||
171 | { | ||
172 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
173 | struct nv50_fifo_priv *priv; | ||
174 | int ret; | ||
175 | |||
176 | NV_DEBUG(dev, "\n"); | ||
177 | |||
178 | priv = dev_priv->engine.fifo.priv; | ||
179 | if (priv) { | ||
180 | priv->cur_thingo = !priv->cur_thingo; | ||
181 | goto just_reset; | ||
182 | } | ||
183 | |||
184 | priv = kzalloc(sizeof(*priv), GFP_KERNEL); | ||
185 | if (!priv) | ||
186 | return -ENOMEM; | ||
187 | dev_priv->engine.fifo.priv = priv; | ||
188 | |||
189 | ret = nouveau_gpuobj_new_ref(dev, NULL, NULL, 0, 128*4, 0x1000, | ||
190 | NVOBJ_FLAG_ZERO_ALLOC, &priv->thingo[0]); | ||
191 | if (ret) { | ||
192 | NV_ERROR(dev, "error creating thingo0: %d\n", ret); | ||
193 | return ret; | ||
194 | } | ||
195 | |||
196 | ret = nouveau_gpuobj_new_ref(dev, NULL, NULL, 0, 128*4, 0x1000, | ||
197 | NVOBJ_FLAG_ZERO_ALLOC, &priv->thingo[1]); | ||
198 | if (ret) { | ||
199 | NV_ERROR(dev, "error creating thingo1: %d\n", ret); | ||
200 | return ret; | ||
201 | } | ||
202 | |||
203 | just_reset: | ||
204 | nv50_fifo_init_reset(dev); | ||
205 | nv50_fifo_init_intr(dev); | ||
206 | nv50_fifo_init_context_table(dev); | ||
207 | nv50_fifo_init_regs__nv(dev); | ||
208 | nv50_fifo_init_regs(dev); | ||
209 | dev_priv->engine.fifo.enable(dev); | ||
210 | dev_priv->engine.fifo.reassign(dev, true); | ||
211 | |||
212 | return 0; | ||
213 | } | ||
214 | |||
215 | void | ||
216 | nv50_fifo_takedown(struct drm_device *dev) | ||
217 | { | ||
218 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
219 | struct nv50_fifo_priv *priv = dev_priv->engine.fifo.priv; | ||
220 | |||
221 | NV_DEBUG(dev, "\n"); | ||
222 | |||
223 | if (!priv) | ||
224 | return; | ||
225 | |||
226 | nouveau_gpuobj_ref_del(dev, &priv->thingo[0]); | ||
227 | nouveau_gpuobj_ref_del(dev, &priv->thingo[1]); | ||
228 | |||
229 | dev_priv->engine.fifo.priv = NULL; | ||
230 | kfree(priv); | ||
231 | } | ||
232 | |||
233 | int | ||
234 | nv50_fifo_channel_id(struct drm_device *dev) | ||
235 | { | ||
236 | return nv_rd32(dev, NV03_PFIFO_CACHE1_PUSH1) & | ||
237 | NV50_PFIFO_CACHE1_PUSH1_CHID_MASK; | ||
238 | } | ||
239 | |||
240 | int | ||
241 | nv50_fifo_create_context(struct nouveau_channel *chan) | ||
242 | { | ||
243 | struct drm_device *dev = chan->dev; | ||
244 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
245 | struct nouveau_gpuobj *ramfc = NULL; | ||
246 | unsigned long flags; | ||
247 | int ret; | ||
248 | |||
249 | NV_DEBUG(dev, "ch%d\n", chan->id); | ||
250 | |||
251 | if (IS_G80) { | ||
252 | uint32_t ramin_poffset = chan->ramin->gpuobj->im_pramin->start; | ||
253 | uint32_t ramin_voffset = chan->ramin->gpuobj->im_backing_start; | ||
254 | |||
255 | ret = nouveau_gpuobj_new_fake(dev, ramin_poffset, ramin_voffset, | ||
256 | 0x100, NVOBJ_FLAG_ZERO_ALLOC | | ||
257 | NVOBJ_FLAG_ZERO_FREE, &ramfc, | ||
258 | &chan->ramfc); | ||
259 | if (ret) | ||
260 | return ret; | ||
261 | |||
262 | ret = nouveau_gpuobj_new_fake(dev, ramin_poffset + 0x0400, | ||
263 | ramin_voffset + 0x0400, 4096, | ||
264 | 0, NULL, &chan->cache); | ||
265 | if (ret) | ||
266 | return ret; | ||
267 | } else { | ||
268 | ret = nouveau_gpuobj_new_ref(dev, chan, NULL, 0, 0x100, 256, | ||
269 | NVOBJ_FLAG_ZERO_ALLOC | | ||
270 | NVOBJ_FLAG_ZERO_FREE, | ||
271 | &chan->ramfc); | ||
272 | if (ret) | ||
273 | return ret; | ||
274 | ramfc = chan->ramfc->gpuobj; | ||
275 | |||
276 | ret = nouveau_gpuobj_new_ref(dev, chan, NULL, 0, 4096, 1024, | ||
277 | 0, &chan->cache); | ||
278 | if (ret) | ||
279 | return ret; | ||
280 | } | ||
281 | |||
282 | spin_lock_irqsave(&dev_priv->context_switch_lock, flags); | ||
283 | |||
284 | dev_priv->engine.instmem.prepare_access(dev, true); | ||
285 | |||
286 | nv_wo32(dev, ramfc, 0x48/4, chan->pushbuf->instance >> 4); | ||
287 | nv_wo32(dev, ramfc, 0x80/4, (0xc << 24) | (chan->ramht->instance >> 4)); | ||
288 | nv_wo32(dev, ramfc, 0x44/4, 0x2101ffff); | ||
289 | nv_wo32(dev, ramfc, 0x60/4, 0x7fffffff); | ||
290 | nv_wo32(dev, ramfc, 0x40/4, 0x00000000); | ||
291 | nv_wo32(dev, ramfc, 0x7c/4, 0x30000001); | ||
292 | nv_wo32(dev, ramfc, 0x78/4, 0x00000000); | ||
293 | nv_wo32(dev, ramfc, 0x3c/4, 0x403f6078); | ||
294 | nv_wo32(dev, ramfc, 0x50/4, chan->pushbuf_base + | ||
295 | chan->dma.ib_base * 4); | ||
296 | nv_wo32(dev, ramfc, 0x54/4, drm_order(chan->dma.ib_max + 1) << 16); | ||
297 | |||
298 | if (!IS_G80) { | ||
299 | nv_wo32(dev, chan->ramin->gpuobj, 0, chan->id); | ||
300 | nv_wo32(dev, chan->ramin->gpuobj, 1, | ||
301 | chan->ramfc->instance >> 8); | ||
302 | |||
303 | nv_wo32(dev, ramfc, 0x88/4, chan->cache->instance >> 10); | ||
304 | nv_wo32(dev, ramfc, 0x98/4, chan->ramin->instance >> 12); | ||
305 | } | ||
306 | |||
307 | dev_priv->engine.instmem.finish_access(dev); | ||
308 | |||
309 | ret = nv50_fifo_channel_enable(dev, chan->id, false); | ||
310 | if (ret) { | ||
311 | NV_ERROR(dev, "error enabling ch%d: %d\n", chan->id, ret); | ||
312 | spin_unlock_irqrestore(&dev_priv->context_switch_lock, flags); | ||
313 | nouveau_gpuobj_ref_del(dev, &chan->ramfc); | ||
314 | return ret; | ||
315 | } | ||
316 | |||
317 | spin_unlock_irqrestore(&dev_priv->context_switch_lock, flags); | ||
318 | return 0; | ||
319 | } | ||
320 | |||
321 | void | ||
322 | nv50_fifo_destroy_context(struct nouveau_channel *chan) | ||
323 | { | ||
324 | struct drm_device *dev = chan->dev; | ||
325 | struct nouveau_gpuobj_ref *ramfc = chan->ramfc; | ||
326 | |||
327 | NV_DEBUG(dev, "ch%d\n", chan->id); | ||
328 | |||
329 | /* This will ensure the channel is seen as disabled. */ | ||
330 | chan->ramfc = NULL; | ||
331 | nv50_fifo_channel_disable(dev, chan->id, false); | ||
332 | |||
333 | /* Dummy channel, also used on ch 127 */ | ||
334 | if (chan->id == 0) | ||
335 | nv50_fifo_channel_disable(dev, 127, false); | ||
336 | |||
337 | nouveau_gpuobj_ref_del(dev, &ramfc); | ||
338 | nouveau_gpuobj_ref_del(dev, &chan->cache); | ||
339 | } | ||
340 | |||
341 | int | ||
342 | nv50_fifo_load_context(struct nouveau_channel *chan) | ||
343 | { | ||
344 | struct drm_device *dev = chan->dev; | ||
345 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
346 | struct nouveau_gpuobj *ramfc = chan->ramfc->gpuobj; | ||
347 | struct nouveau_gpuobj *cache = chan->cache->gpuobj; | ||
348 | int ptr, cnt; | ||
349 | |||
350 | NV_DEBUG(dev, "ch%d\n", chan->id); | ||
351 | |||
352 | dev_priv->engine.instmem.prepare_access(dev, false); | ||
353 | |||
354 | nv_wr32(dev, 0x3330, nv_ro32(dev, ramfc, 0x00/4)); | ||
355 | nv_wr32(dev, 0x3334, nv_ro32(dev, ramfc, 0x04/4)); | ||
356 | nv_wr32(dev, 0x3240, nv_ro32(dev, ramfc, 0x08/4)); | ||
357 | nv_wr32(dev, 0x3320, nv_ro32(dev, ramfc, 0x0c/4)); | ||
358 | nv_wr32(dev, 0x3244, nv_ro32(dev, ramfc, 0x10/4)); | ||
359 | nv_wr32(dev, 0x3328, nv_ro32(dev, ramfc, 0x14/4)); | ||
360 | nv_wr32(dev, 0x3368, nv_ro32(dev, ramfc, 0x18/4)); | ||
361 | nv_wr32(dev, 0x336c, nv_ro32(dev, ramfc, 0x1c/4)); | ||
362 | nv_wr32(dev, 0x3370, nv_ro32(dev, ramfc, 0x20/4)); | ||
363 | nv_wr32(dev, 0x3374, nv_ro32(dev, ramfc, 0x24/4)); | ||
364 | nv_wr32(dev, 0x3378, nv_ro32(dev, ramfc, 0x28/4)); | ||
365 | nv_wr32(dev, 0x337c, nv_ro32(dev, ramfc, 0x2c/4)); | ||
366 | nv_wr32(dev, 0x3228, nv_ro32(dev, ramfc, 0x30/4)); | ||
367 | nv_wr32(dev, 0x3364, nv_ro32(dev, ramfc, 0x34/4)); | ||
368 | nv_wr32(dev, 0x32a0, nv_ro32(dev, ramfc, 0x38/4)); | ||
369 | nv_wr32(dev, 0x3224, nv_ro32(dev, ramfc, 0x3c/4)); | ||
370 | nv_wr32(dev, 0x324c, nv_ro32(dev, ramfc, 0x40/4)); | ||
371 | nv_wr32(dev, 0x2044, nv_ro32(dev, ramfc, 0x44/4)); | ||
372 | nv_wr32(dev, 0x322c, nv_ro32(dev, ramfc, 0x48/4)); | ||
373 | nv_wr32(dev, 0x3234, nv_ro32(dev, ramfc, 0x4c/4)); | ||
374 | nv_wr32(dev, 0x3340, nv_ro32(dev, ramfc, 0x50/4)); | ||
375 | nv_wr32(dev, 0x3344, nv_ro32(dev, ramfc, 0x54/4)); | ||
376 | nv_wr32(dev, 0x3280, nv_ro32(dev, ramfc, 0x58/4)); | ||
377 | nv_wr32(dev, 0x3254, nv_ro32(dev, ramfc, 0x5c/4)); | ||
378 | nv_wr32(dev, 0x3260, nv_ro32(dev, ramfc, 0x60/4)); | ||
379 | nv_wr32(dev, 0x3264, nv_ro32(dev, ramfc, 0x64/4)); | ||
380 | nv_wr32(dev, 0x3268, nv_ro32(dev, ramfc, 0x68/4)); | ||
381 | nv_wr32(dev, 0x326c, nv_ro32(dev, ramfc, 0x6c/4)); | ||
382 | nv_wr32(dev, 0x32e4, nv_ro32(dev, ramfc, 0x70/4)); | ||
383 | nv_wr32(dev, 0x3248, nv_ro32(dev, ramfc, 0x74/4)); | ||
384 | nv_wr32(dev, 0x2088, nv_ro32(dev, ramfc, 0x78/4)); | ||
385 | nv_wr32(dev, 0x2058, nv_ro32(dev, ramfc, 0x7c/4)); | ||
386 | nv_wr32(dev, 0x2210, nv_ro32(dev, ramfc, 0x80/4)); | ||
387 | |||
388 | cnt = nv_ro32(dev, ramfc, 0x84/4); | ||
389 | for (ptr = 0; ptr < cnt; ptr++) { | ||
390 | nv_wr32(dev, NV40_PFIFO_CACHE1_METHOD(ptr), | ||
391 | nv_ro32(dev, cache, (ptr * 2) + 0)); | ||
392 | nv_wr32(dev, NV40_PFIFO_CACHE1_DATA(ptr), | ||
393 | nv_ro32(dev, cache, (ptr * 2) + 1)); | ||
394 | } | ||
395 | nv_wr32(dev, NV03_PFIFO_CACHE1_PUT, cnt << 2); | ||
396 | nv_wr32(dev, NV03_PFIFO_CACHE1_GET, 0); | ||
397 | |||
398 | /* guessing that all the 0x34xx regs aren't on NV50 */ | ||
399 | if (!IS_G80) { | ||
400 | nv_wr32(dev, 0x340c, nv_ro32(dev, ramfc, 0x88/4)); | ||
401 | nv_wr32(dev, 0x3400, nv_ro32(dev, ramfc, 0x8c/4)); | ||
402 | nv_wr32(dev, 0x3404, nv_ro32(dev, ramfc, 0x90/4)); | ||
403 | nv_wr32(dev, 0x3408, nv_ro32(dev, ramfc, 0x94/4)); | ||
404 | nv_wr32(dev, 0x3410, nv_ro32(dev, ramfc, 0x98/4)); | ||
405 | } | ||
406 | |||
407 | dev_priv->engine.instmem.finish_access(dev); | ||
408 | |||
409 | nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, chan->id | (1<<16)); | ||
410 | return 0; | ||
411 | } | ||
412 | |||
413 | int | ||
414 | nv50_fifo_unload_context(struct drm_device *dev) | ||
415 | { | ||
416 | struct drm_nouveau_private *dev_priv = dev->dev_private; | ||
417 | struct nouveau_fifo_engine *pfifo = &dev_priv->engine.fifo; | ||
418 | struct nouveau_gpuobj *ramfc, *cache; | ||
419 | struct nouveau_channel *chan = NULL; | ||
420 | int chid, get, put, ptr; | ||
421 | |||
422 | NV_DEBUG(dev, "\n"); | ||
423 | |||
424 | chid = pfifo->channel_id(dev); | ||
425 | if (chid < 1 || chid >= dev_priv->engine.fifo.channels - 1) | ||
426 | return 0; | ||
427 | |||
428 | chan = dev_priv->fifos[chid]; | ||
429 | if (!chan) { | ||
430 | NV_ERROR(dev, "Inactive channel on PFIFO: %d\n", chid); | ||
431 | return -EINVAL; | ||
432 | } | ||
433 | NV_DEBUG(dev, "ch%d\n", chan->id); | ||
434 | ramfc = chan->ramfc->gpuobj; | ||
435 | cache = chan->cache->gpuobj; | ||
436 | |||
437 | dev_priv->engine.instmem.prepare_access(dev, true); | ||
438 | |||
439 | nv_wo32(dev, ramfc, 0x00/4, nv_rd32(dev, 0x3330)); | ||
440 | nv_wo32(dev, ramfc, 0x04/4, nv_rd32(dev, 0x3334)); | ||
441 | nv_wo32(dev, ramfc, 0x08/4, nv_rd32(dev, 0x3240)); | ||
442 | nv_wo32(dev, ramfc, 0x0c/4, nv_rd32(dev, 0x3320)); | ||
443 | nv_wo32(dev, ramfc, 0x10/4, nv_rd32(dev, 0x3244)); | ||
444 | nv_wo32(dev, ramfc, 0x14/4, nv_rd32(dev, 0x3328)); | ||
445 | nv_wo32(dev, ramfc, 0x18/4, nv_rd32(dev, 0x3368)); | ||
446 | nv_wo32(dev, ramfc, 0x1c/4, nv_rd32(dev, 0x336c)); | ||
447 | nv_wo32(dev, ramfc, 0x20/4, nv_rd32(dev, 0x3370)); | ||
448 | nv_wo32(dev, ramfc, 0x24/4, nv_rd32(dev, 0x3374)); | ||
449 | nv_wo32(dev, ramfc, 0x28/4, nv_rd32(dev, 0x3378)); | ||
450 | nv_wo32(dev, ramfc, 0x2c/4, nv_rd32(dev, 0x337c)); | ||
451 | nv_wo32(dev, ramfc, 0x30/4, nv_rd32(dev, 0x3228)); | ||
452 | nv_wo32(dev, ramfc, 0x34/4, nv_rd32(dev, 0x3364)); | ||
453 | nv_wo32(dev, ramfc, 0x38/4, nv_rd32(dev, 0x32a0)); | ||
454 | nv_wo32(dev, ramfc, 0x3c/4, nv_rd32(dev, 0x3224)); | ||
455 | nv_wo32(dev, ramfc, 0x40/4, nv_rd32(dev, 0x324c)); | ||
456 | nv_wo32(dev, ramfc, 0x44/4, nv_rd32(dev, 0x2044)); | ||
457 | nv_wo32(dev, ramfc, 0x48/4, nv_rd32(dev, 0x322c)); | ||
458 | nv_wo32(dev, ramfc, 0x4c/4, nv_rd32(dev, 0x3234)); | ||
459 | nv_wo32(dev, ramfc, 0x50/4, nv_rd32(dev, 0x3340)); | ||
460 | nv_wo32(dev, ramfc, 0x54/4, nv_rd32(dev, 0x3344)); | ||
461 | nv_wo32(dev, ramfc, 0x58/4, nv_rd32(dev, 0x3280)); | ||
462 | nv_wo32(dev, ramfc, 0x5c/4, nv_rd32(dev, 0x3254)); | ||
463 | nv_wo32(dev, ramfc, 0x60/4, nv_rd32(dev, 0x3260)); | ||
464 | nv_wo32(dev, ramfc, 0x64/4, nv_rd32(dev, 0x3264)); | ||
465 | nv_wo32(dev, ramfc, 0x68/4, nv_rd32(dev, 0x3268)); | ||
466 | nv_wo32(dev, ramfc, 0x6c/4, nv_rd32(dev, 0x326c)); | ||
467 | nv_wo32(dev, ramfc, 0x70/4, nv_rd32(dev, 0x32e4)); | ||
468 | nv_wo32(dev, ramfc, 0x74/4, nv_rd32(dev, 0x3248)); | ||
469 | nv_wo32(dev, ramfc, 0x78/4, nv_rd32(dev, 0x2088)); | ||
470 | nv_wo32(dev, ramfc, 0x7c/4, nv_rd32(dev, 0x2058)); | ||
471 | nv_wo32(dev, ramfc, 0x80/4, nv_rd32(dev, 0x2210)); | ||
472 | |||
473 | put = (nv_rd32(dev, NV03_PFIFO_CACHE1_PUT) & 0x7ff) >> 2; | ||
474 | get = (nv_rd32(dev, NV03_PFIFO_CACHE1_GET) & 0x7ff) >> 2; | ||
475 | ptr = 0; | ||
476 | while (put != get) { | ||
477 | nv_wo32(dev, cache, ptr++, | ||
478 | nv_rd32(dev, NV40_PFIFO_CACHE1_METHOD(get))); | ||
479 | nv_wo32(dev, cache, ptr++, | ||
480 | nv_rd32(dev, NV40_PFIFO_CACHE1_DATA(get))); | ||
481 | get = (get + 1) & 0x1ff; | ||
482 | } | ||
483 | |||
484 | /* guessing that all the 0x34xx regs aren't on NV50 */ | ||
485 | if (!IS_G80) { | ||
486 | nv_wo32(dev, ramfc, 0x84/4, ptr >> 1); | ||
487 | nv_wo32(dev, ramfc, 0x88/4, nv_rd32(dev, 0x340c)); | ||
488 | nv_wo32(dev, ramfc, 0x8c/4, nv_rd32(dev, 0x3400)); | ||
489 | nv_wo32(dev, ramfc, 0x90/4, nv_rd32(dev, 0x3404)); | ||
490 | nv_wo32(dev, ramfc, 0x94/4, nv_rd32(dev, 0x3408)); | ||
491 | nv_wo32(dev, ramfc, 0x98/4, nv_rd32(dev, 0x3410)); | ||
492 | } | ||
493 | |||
494 | dev_priv->engine.instmem.finish_access(dev); | ||
495 | |||
496 | /*XXX: probably reload ch127 (NULL) state back too */ | ||
497 | nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, 127); | ||
498 | return 0; | ||
499 | } | ||
500 | |||