diff options
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/vce_v2_0.c')
-rw-r--r-- | drivers/gpu/drm/amd/amdgpu/vce_v2_0.c | 451 |
1 files changed, 234 insertions, 217 deletions
diff --git a/drivers/gpu/drm/amd/amdgpu/vce_v2_0.c b/drivers/gpu/drm/amd/amdgpu/vce_v2_0.c index 38ed903dd6f8..9ea99348e493 100644 --- a/drivers/gpu/drm/amd/amdgpu/vce_v2_0.c +++ b/drivers/gpu/drm/amd/amdgpu/vce_v2_0.c | |||
@@ -42,10 +42,9 @@ | |||
42 | #define VCE_V2_0_DATA_SIZE (23552 * AMDGPU_MAX_VCE_HANDLES) | 42 | #define VCE_V2_0_DATA_SIZE (23552 * AMDGPU_MAX_VCE_HANDLES) |
43 | #define VCE_STATUS_VCPU_REPORT_FW_LOADED_MASK 0x02 | 43 | #define VCE_STATUS_VCPU_REPORT_FW_LOADED_MASK 0x02 |
44 | 44 | ||
45 | static void vce_v2_0_mc_resume(struct amdgpu_device *adev); | ||
46 | static void vce_v2_0_set_ring_funcs(struct amdgpu_device *adev); | 45 | static void vce_v2_0_set_ring_funcs(struct amdgpu_device *adev); |
47 | static void vce_v2_0_set_irq_funcs(struct amdgpu_device *adev); | 46 | static void vce_v2_0_set_irq_funcs(struct amdgpu_device *adev); |
48 | static int vce_v2_0_wait_for_idle(void *handle); | 47 | |
49 | /** | 48 | /** |
50 | * vce_v2_0_ring_get_rptr - get read pointer | 49 | * vce_v2_0_ring_get_rptr - get read pointer |
51 | * | 50 | * |
@@ -140,6 +139,86 @@ static int vce_v2_0_firmware_loaded(struct amdgpu_device *adev) | |||
140 | return -ETIMEDOUT; | 139 | return -ETIMEDOUT; |
141 | } | 140 | } |
142 | 141 | ||
142 | static void vce_v2_0_disable_cg(struct amdgpu_device *adev) | ||
143 | { | ||
144 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 7); | ||
145 | } | ||
146 | |||
147 | static void vce_v2_0_init_cg(struct amdgpu_device *adev) | ||
148 | { | ||
149 | u32 tmp; | ||
150 | |||
151 | tmp = RREG32(mmVCE_CLOCK_GATING_A); | ||
152 | tmp &= ~0xfff; | ||
153 | tmp |= ((0 << 0) | (4 << 4)); | ||
154 | tmp |= 0x40000; | ||
155 | WREG32(mmVCE_CLOCK_GATING_A, tmp); | ||
156 | |||
157 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
158 | tmp &= ~0xfff; | ||
159 | tmp |= ((0 << 0) | (4 << 4)); | ||
160 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
161 | |||
162 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
163 | tmp |= 0x10; | ||
164 | tmp &= ~0x100000; | ||
165 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
166 | } | ||
167 | |||
168 | static void vce_v2_0_mc_resume(struct amdgpu_device *adev) | ||
169 | { | ||
170 | uint64_t addr = adev->vce.gpu_addr; | ||
171 | uint32_t size; | ||
172 | |||
173 | WREG32_P(mmVCE_CLOCK_GATING_A, 0, ~(1 << 16)); | ||
174 | WREG32_P(mmVCE_UENC_CLOCK_GATING, 0x1FF000, ~0xFF9FF000); | ||
175 | WREG32_P(mmVCE_UENC_REG_CLOCK_GATING, 0x3F, ~0x3F); | ||
176 | WREG32(mmVCE_CLOCK_GATING_B, 0xf7); | ||
177 | |||
178 | WREG32(mmVCE_LMI_CTRL, 0x00398000); | ||
179 | WREG32_P(mmVCE_LMI_CACHE_CTRL, 0x0, ~0x1); | ||
180 | WREG32(mmVCE_LMI_SWAP_CNTL, 0); | ||
181 | WREG32(mmVCE_LMI_SWAP_CNTL1, 0); | ||
182 | WREG32(mmVCE_LMI_VM_CTRL, 0); | ||
183 | |||
184 | addr += AMDGPU_VCE_FIRMWARE_OFFSET; | ||
185 | size = VCE_V2_0_FW_SIZE; | ||
186 | WREG32(mmVCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff); | ||
187 | WREG32(mmVCE_VCPU_CACHE_SIZE0, size); | ||
188 | |||
189 | addr += size; | ||
190 | size = VCE_V2_0_STACK_SIZE; | ||
191 | WREG32(mmVCE_VCPU_CACHE_OFFSET1, addr & 0x7fffffff); | ||
192 | WREG32(mmVCE_VCPU_CACHE_SIZE1, size); | ||
193 | |||
194 | addr += size; | ||
195 | size = VCE_V2_0_DATA_SIZE; | ||
196 | WREG32(mmVCE_VCPU_CACHE_OFFSET2, addr & 0x7fffffff); | ||
197 | WREG32(mmVCE_VCPU_CACHE_SIZE2, size); | ||
198 | |||
199 | WREG32_P(mmVCE_LMI_CTRL2, 0x0, ~0x100); | ||
200 | WREG32_FIELD(VCE_SYS_INT_EN, VCE_SYS_INT_TRAP_INTERRUPT_EN, 1); | ||
201 | } | ||
202 | |||
203 | static bool vce_v2_0_is_idle(void *handle) | ||
204 | { | ||
205 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | ||
206 | |||
207 | return !(RREG32(mmSRBM_STATUS2) & SRBM_STATUS2__VCE_BUSY_MASK); | ||
208 | } | ||
209 | |||
210 | static int vce_v2_0_wait_for_idle(void *handle) | ||
211 | { | ||
212 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | ||
213 | unsigned i; | ||
214 | |||
215 | for (i = 0; i < adev->usec_timeout; i++) { | ||
216 | if (vce_v2_0_is_idle(handle)) | ||
217 | return 0; | ||
218 | } | ||
219 | return -ETIMEDOUT; | ||
220 | } | ||
221 | |||
143 | /** | 222 | /** |
144 | * vce_v2_0_start - start VCE block | 223 | * vce_v2_0_start - start VCE block |
145 | * | 224 | * |
@@ -152,11 +231,14 @@ static int vce_v2_0_start(struct amdgpu_device *adev) | |||
152 | struct amdgpu_ring *ring; | 231 | struct amdgpu_ring *ring; |
153 | int r; | 232 | int r; |
154 | 233 | ||
155 | vce_v2_0_mc_resume(adev); | ||
156 | |||
157 | /* set BUSY flag */ | 234 | /* set BUSY flag */ |
158 | WREG32_P(mmVCE_STATUS, 1, ~1); | 235 | WREG32_P(mmVCE_STATUS, 1, ~1); |
159 | 236 | ||
237 | vce_v2_0_init_cg(adev); | ||
238 | vce_v2_0_disable_cg(adev); | ||
239 | |||
240 | vce_v2_0_mc_resume(adev); | ||
241 | |||
160 | ring = &adev->vce.ring[0]; | 242 | ring = &adev->vce.ring[0]; |
161 | WREG32(mmVCE_RB_RPTR, ring->wptr); | 243 | WREG32(mmVCE_RB_RPTR, ring->wptr); |
162 | WREG32(mmVCE_RB_WPTR, ring->wptr); | 244 | WREG32(mmVCE_RB_WPTR, ring->wptr); |
@@ -189,6 +271,145 @@ static int vce_v2_0_start(struct amdgpu_device *adev) | |||
189 | return 0; | 271 | return 0; |
190 | } | 272 | } |
191 | 273 | ||
274 | static int vce_v2_0_stop(struct amdgpu_device *adev) | ||
275 | { | ||
276 | int i, j; | ||
277 | int status; | ||
278 | |||
279 | if (vce_v2_0_lmi_clean(adev)) { | ||
280 | DRM_INFO("vce is not idle \n"); | ||
281 | return 0; | ||
282 | } | ||
283 | /* | ||
284 | for (i = 0; i < 10; ++i) { | ||
285 | for (j = 0; j < 100; ++j) { | ||
286 | status = RREG32(mmVCE_FW_REG_STATUS); | ||
287 | if (!(status & 1)) | ||
288 | break; | ||
289 | mdelay(1); | ||
290 | } | ||
291 | break; | ||
292 | } | ||
293 | */ | ||
294 | if (vce_v2_0_wait_for_idle(adev)) { | ||
295 | DRM_INFO("VCE is busy, Can't set clock gateing"); | ||
296 | return 0; | ||
297 | } | ||
298 | |||
299 | /* Stall UMC and register bus before resetting VCPU */ | ||
300 | WREG32_P(mmVCE_LMI_CTRL2, 1 << 8, ~(1 << 8)); | ||
301 | |||
302 | for (i = 0; i < 10; ++i) { | ||
303 | for (j = 0; j < 100; ++j) { | ||
304 | status = RREG32(mmVCE_LMI_STATUS); | ||
305 | if (status & 0x240) | ||
306 | break; | ||
307 | mdelay(1); | ||
308 | } | ||
309 | break; | ||
310 | } | ||
311 | |||
312 | WREG32_P(mmVCE_VCPU_CNTL, 0, ~0x80001); | ||
313 | |||
314 | /* put LMI, VCPU, RBC etc... into reset */ | ||
315 | WREG32_P(mmVCE_SOFT_RESET, 1, ~0x1); | ||
316 | |||
317 | WREG32(mmVCE_STATUS, 0); | ||
318 | |||
319 | return 0; | ||
320 | } | ||
321 | |||
322 | static void vce_v2_0_set_sw_cg(struct amdgpu_device *adev, bool gated) | ||
323 | { | ||
324 | u32 tmp; | ||
325 | |||
326 | if (gated) { | ||
327 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
328 | tmp |= 0xe70000; | ||
329 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
330 | |||
331 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
332 | tmp |= 0xff000000; | ||
333 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
334 | |||
335 | tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING); | ||
336 | tmp &= ~0x3fc; | ||
337 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp); | ||
338 | |||
339 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0); | ||
340 | } else { | ||
341 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
342 | tmp |= 0xe7; | ||
343 | tmp &= ~0xe70000; | ||
344 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
345 | |||
346 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
347 | tmp |= 0x1fe000; | ||
348 | tmp &= ~0xff000000; | ||
349 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
350 | |||
351 | tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING); | ||
352 | tmp |= 0x3fc; | ||
353 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp); | ||
354 | } | ||
355 | } | ||
356 | |||
357 | static void vce_v2_0_set_dyn_cg(struct amdgpu_device *adev, bool gated) | ||
358 | { | ||
359 | u32 orig, tmp; | ||
360 | |||
361 | /* LMI_MC/LMI_UMC always set in dynamic, | ||
362 | * set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {0, 0} | ||
363 | */ | ||
364 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
365 | tmp &= ~0x00060006; | ||
366 | |||
367 | /* Exception for ECPU, IH, SEM, SYS blocks needs to be turned on/off by SW */ | ||
368 | if (gated) { | ||
369 | tmp |= 0xe10000; | ||
370 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
371 | } else { | ||
372 | tmp |= 0xe1; | ||
373 | tmp &= ~0xe10000; | ||
374 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
375 | } | ||
376 | |||
377 | orig = tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
378 | tmp &= ~0x1fe000; | ||
379 | tmp &= ~0xff000000; | ||
380 | if (tmp != orig) | ||
381 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
382 | |||
383 | orig = tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING); | ||
384 | tmp &= ~0x3fc; | ||
385 | if (tmp != orig) | ||
386 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp); | ||
387 | |||
388 | /* set VCE_UENC_REG_CLOCK_GATING always in dynamic mode */ | ||
389 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, 0x00); | ||
390 | |||
391 | if(gated) | ||
392 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0); | ||
393 | } | ||
394 | |||
395 | static void vce_v2_0_enable_mgcg(struct amdgpu_device *adev, bool enable, | ||
396 | bool sw_cg) | ||
397 | { | ||
398 | if (enable && (adev->cg_flags & AMD_CG_SUPPORT_VCE_MGCG)) { | ||
399 | if (sw_cg) | ||
400 | vce_v2_0_set_sw_cg(adev, true); | ||
401 | else | ||
402 | vce_v2_0_set_dyn_cg(adev, true); | ||
403 | } else { | ||
404 | vce_v2_0_disable_cg(adev); | ||
405 | |||
406 | if (sw_cg) | ||
407 | vce_v2_0_set_sw_cg(adev, false); | ||
408 | else | ||
409 | vce_v2_0_set_dyn_cg(adev, false); | ||
410 | } | ||
411 | } | ||
412 | |||
192 | static int vce_v2_0_early_init(void *handle) | 413 | static int vce_v2_0_early_init(void *handle) |
193 | { | 414 | { |
194 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | 415 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; |
@@ -254,11 +475,8 @@ static int vce_v2_0_hw_init(void *handle) | |||
254 | int r, i; | 475 | int r, i; |
255 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | 476 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; |
256 | 477 | ||
257 | r = vce_v2_0_start(adev); | 478 | amdgpu_asic_set_vce_clocks(adev, 10000, 10000); |
258 | /* this error mean vcpu not in running state, so just skip ring test, not stop driver initialize */ | 479 | vce_v2_0_enable_mgcg(adev, true, false); |
259 | if (r) | ||
260 | return 0; | ||
261 | |||
262 | for (i = 0; i < adev->vce.num_rings; i++) | 480 | for (i = 0; i < adev->vce.num_rings; i++) |
263 | adev->vce.ring[i].ready = false; | 481 | adev->vce.ring[i].ready = false; |
264 | 482 | ||
@@ -312,190 +530,6 @@ static int vce_v2_0_resume(void *handle) | |||
312 | return r; | 530 | return r; |
313 | } | 531 | } |
314 | 532 | ||
315 | static void vce_v2_0_set_sw_cg(struct amdgpu_device *adev, bool gated) | ||
316 | { | ||
317 | u32 tmp; | ||
318 | |||
319 | if (gated) { | ||
320 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
321 | tmp |= 0xe70000; | ||
322 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
323 | |||
324 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
325 | tmp |= 0xff000000; | ||
326 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
327 | |||
328 | tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING); | ||
329 | tmp &= ~0x3fc; | ||
330 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp); | ||
331 | |||
332 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0); | ||
333 | } else { | ||
334 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
335 | tmp |= 0xe7; | ||
336 | tmp &= ~0xe70000; | ||
337 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
338 | |||
339 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
340 | tmp |= 0x1fe000; | ||
341 | tmp &= ~0xff000000; | ||
342 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
343 | |||
344 | tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING); | ||
345 | tmp |= 0x3fc; | ||
346 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp); | ||
347 | } | ||
348 | } | ||
349 | |||
350 | static void vce_v2_0_set_dyn_cg(struct amdgpu_device *adev, bool gated) | ||
351 | { | ||
352 | if (vce_v2_0_wait_for_idle(adev)) { | ||
353 | DRM_INFO("VCE is busy, Can't set clock gateing"); | ||
354 | return; | ||
355 | } | ||
356 | |||
357 | WREG32_P(mmVCE_LMI_CTRL2, 0x100, ~0x100); | ||
358 | |||
359 | if (vce_v2_0_lmi_clean(adev)) { | ||
360 | DRM_INFO("LMI is busy, Can't set clock gateing"); | ||
361 | return; | ||
362 | } | ||
363 | |||
364 | WREG32_P(mmVCE_VCPU_CNTL, 0, ~VCE_VCPU_CNTL__CLK_EN_MASK); | ||
365 | WREG32_P(mmVCE_SOFT_RESET, | ||
366 | VCE_SOFT_RESET__ECPU_SOFT_RESET_MASK, | ||
367 | ~VCE_SOFT_RESET__ECPU_SOFT_RESET_MASK); | ||
368 | WREG32(mmVCE_STATUS, 0); | ||
369 | |||
370 | if (gated) | ||
371 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0); | ||
372 | /* LMI_MC/LMI_UMC always set in dynamic, set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {0, 0} */ | ||
373 | if (gated) { | ||
374 | /* Force CLOCK OFF , set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {*, 1} */ | ||
375 | WREG32(mmVCE_CLOCK_GATING_B, 0xe90010); | ||
376 | } else { | ||
377 | /* Force CLOCK ON, set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {1, 0} */ | ||
378 | WREG32(mmVCE_CLOCK_GATING_B, 0x800f1); | ||
379 | } | ||
380 | |||
381 | /* Set VCE_UENC_CLOCK_GATING always in dynamic mode {*_FORCE_ON, *_FORCE_OFF} = {0, 0}*/; | ||
382 | WREG32(mmVCE_UENC_CLOCK_GATING, 0x40); | ||
383 | |||
384 | /* set VCE_UENC_REG_CLOCK_GATING always in dynamic mode */ | ||
385 | WREG32(mmVCE_UENC_REG_CLOCK_GATING, 0x00); | ||
386 | |||
387 | WREG32_P(mmVCE_LMI_CTRL2, 0, ~0x100); | ||
388 | if(!gated) { | ||
389 | WREG32_P(mmVCE_VCPU_CNTL, VCE_VCPU_CNTL__CLK_EN_MASK, ~VCE_VCPU_CNTL__CLK_EN_MASK); | ||
390 | mdelay(100); | ||
391 | WREG32_P(mmVCE_SOFT_RESET, 0, ~VCE_SOFT_RESET__ECPU_SOFT_RESET_MASK); | ||
392 | |||
393 | vce_v2_0_firmware_loaded(adev); | ||
394 | WREG32_P(mmVCE_STATUS, 0, ~VCE_STATUS__JOB_BUSY_MASK); | ||
395 | } | ||
396 | } | ||
397 | |||
398 | static void vce_v2_0_disable_cg(struct amdgpu_device *adev) | ||
399 | { | ||
400 | WREG32(mmVCE_CGTT_CLK_OVERRIDE, 7); | ||
401 | } | ||
402 | |||
403 | static void vce_v2_0_enable_mgcg(struct amdgpu_device *adev, bool enable) | ||
404 | { | ||
405 | bool sw_cg = false; | ||
406 | |||
407 | if (enable && (adev->cg_flags & AMD_CG_SUPPORT_VCE_MGCG)) { | ||
408 | if (sw_cg) | ||
409 | vce_v2_0_set_sw_cg(adev, true); | ||
410 | else | ||
411 | vce_v2_0_set_dyn_cg(adev, true); | ||
412 | } else { | ||
413 | vce_v2_0_disable_cg(adev); | ||
414 | |||
415 | if (sw_cg) | ||
416 | vce_v2_0_set_sw_cg(adev, false); | ||
417 | else | ||
418 | vce_v2_0_set_dyn_cg(adev, false); | ||
419 | } | ||
420 | } | ||
421 | |||
422 | static void vce_v2_0_init_cg(struct amdgpu_device *adev) | ||
423 | { | ||
424 | u32 tmp; | ||
425 | |||
426 | tmp = RREG32(mmVCE_CLOCK_GATING_A); | ||
427 | tmp &= ~0xfff; | ||
428 | tmp |= ((0 << 0) | (4 << 4)); | ||
429 | tmp |= 0x40000; | ||
430 | WREG32(mmVCE_CLOCK_GATING_A, tmp); | ||
431 | |||
432 | tmp = RREG32(mmVCE_UENC_CLOCK_GATING); | ||
433 | tmp &= ~0xfff; | ||
434 | tmp |= ((0 << 0) | (4 << 4)); | ||
435 | WREG32(mmVCE_UENC_CLOCK_GATING, tmp); | ||
436 | |||
437 | tmp = RREG32(mmVCE_CLOCK_GATING_B); | ||
438 | tmp |= 0x10; | ||
439 | tmp &= ~0x100000; | ||
440 | WREG32(mmVCE_CLOCK_GATING_B, tmp); | ||
441 | } | ||
442 | |||
443 | static void vce_v2_0_mc_resume(struct amdgpu_device *adev) | ||
444 | { | ||
445 | uint64_t addr = adev->vce.gpu_addr; | ||
446 | uint32_t size; | ||
447 | |||
448 | WREG32_P(mmVCE_CLOCK_GATING_A, 0, ~(1 << 16)); | ||
449 | WREG32_P(mmVCE_UENC_CLOCK_GATING, 0x1FF000, ~0xFF9FF000); | ||
450 | WREG32_P(mmVCE_UENC_REG_CLOCK_GATING, 0x3F, ~0x3F); | ||
451 | WREG32(mmVCE_CLOCK_GATING_B, 0xf7); | ||
452 | |||
453 | WREG32(mmVCE_LMI_CTRL, 0x00398000); | ||
454 | WREG32_P(mmVCE_LMI_CACHE_CTRL, 0x0, ~0x1); | ||
455 | WREG32(mmVCE_LMI_SWAP_CNTL, 0); | ||
456 | WREG32(mmVCE_LMI_SWAP_CNTL1, 0); | ||
457 | WREG32(mmVCE_LMI_VM_CTRL, 0); | ||
458 | |||
459 | addr += AMDGPU_VCE_FIRMWARE_OFFSET; | ||
460 | size = VCE_V2_0_FW_SIZE; | ||
461 | WREG32(mmVCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff); | ||
462 | WREG32(mmVCE_VCPU_CACHE_SIZE0, size); | ||
463 | |||
464 | addr += size; | ||
465 | size = VCE_V2_0_STACK_SIZE; | ||
466 | WREG32(mmVCE_VCPU_CACHE_OFFSET1, addr & 0x7fffffff); | ||
467 | WREG32(mmVCE_VCPU_CACHE_SIZE1, size); | ||
468 | |||
469 | addr += size; | ||
470 | size = VCE_V2_0_DATA_SIZE; | ||
471 | WREG32(mmVCE_VCPU_CACHE_OFFSET2, addr & 0x7fffffff); | ||
472 | WREG32(mmVCE_VCPU_CACHE_SIZE2, size); | ||
473 | |||
474 | WREG32_P(mmVCE_LMI_CTRL2, 0x0, ~0x100); | ||
475 | WREG32_FIELD(VCE_SYS_INT_EN, VCE_SYS_INT_TRAP_INTERRUPT_EN, 1); | ||
476 | |||
477 | vce_v2_0_init_cg(adev); | ||
478 | } | ||
479 | |||
480 | static bool vce_v2_0_is_idle(void *handle) | ||
481 | { | ||
482 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | ||
483 | |||
484 | return !(RREG32(mmSRBM_STATUS2) & SRBM_STATUS2__VCE_BUSY_MASK); | ||
485 | } | ||
486 | |||
487 | static int vce_v2_0_wait_for_idle(void *handle) | ||
488 | { | ||
489 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | ||
490 | unsigned i; | ||
491 | |||
492 | for (i = 0; i < adev->usec_timeout; i++) { | ||
493 | if (vce_v2_0_is_idle(handle)) | ||
494 | return 0; | ||
495 | } | ||
496 | return -ETIMEDOUT; | ||
497 | } | ||
498 | |||
499 | static int vce_v2_0_soft_reset(void *handle) | 533 | static int vce_v2_0_soft_reset(void *handle) |
500 | { | 534 | { |
501 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | 535 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; |
@@ -539,33 +573,20 @@ static int vce_v2_0_process_interrupt(struct amdgpu_device *adev, | |||
539 | return 0; | 573 | return 0; |
540 | } | 574 | } |
541 | 575 | ||
542 | static void vce_v2_0_set_bypass_mode(struct amdgpu_device *adev, bool enable) | ||
543 | { | ||
544 | u32 tmp = RREG32_SMC(ixGCK_DFS_BYPASS_CNTL); | ||
545 | |||
546 | if (enable) | ||
547 | tmp |= GCK_DFS_BYPASS_CNTL__BYPASSECLK_MASK; | ||
548 | else | ||
549 | tmp &= ~GCK_DFS_BYPASS_CNTL__BYPASSECLK_MASK; | ||
550 | |||
551 | WREG32_SMC(ixGCK_DFS_BYPASS_CNTL, tmp); | ||
552 | } | ||
553 | |||
554 | |||
555 | static int vce_v2_0_set_clockgating_state(void *handle, | 576 | static int vce_v2_0_set_clockgating_state(void *handle, |
556 | enum amd_clockgating_state state) | 577 | enum amd_clockgating_state state) |
557 | { | 578 | { |
558 | bool gate = false; | 579 | bool gate = false; |
559 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | 580 | bool sw_cg = false; |
560 | bool enable = (state == AMD_CG_STATE_GATE) ? true : false; | ||
561 | |||
562 | 581 | ||
563 | vce_v2_0_set_bypass_mode(adev, enable); | 582 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; |
564 | 583 | ||
565 | if (state == AMD_CG_STATE_GATE) | 584 | if (state == AMD_CG_STATE_GATE) { |
566 | gate = true; | 585 | gate = true; |
586 | sw_cg = true; | ||
587 | } | ||
567 | 588 | ||
568 | vce_v2_0_enable_mgcg(adev, gate); | 589 | vce_v2_0_enable_mgcg(adev, gate, sw_cg); |
569 | 590 | ||
570 | return 0; | 591 | return 0; |
571 | } | 592 | } |
@@ -582,12 +603,8 @@ static int vce_v2_0_set_powergating_state(void *handle, | |||
582 | */ | 603 | */ |
583 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; | 604 | struct amdgpu_device *adev = (struct amdgpu_device *)handle; |
584 | 605 | ||
585 | if (!(adev->pg_flags & AMD_PG_SUPPORT_VCE)) | ||
586 | return 0; | ||
587 | |||
588 | if (state == AMD_PG_STATE_GATE) | 606 | if (state == AMD_PG_STATE_GATE) |
589 | /* XXX do we need a vce_v2_0_stop()? */ | 607 | return vce_v2_0_stop(adev); |
590 | return 0; | ||
591 | else | 608 | else |
592 | return vce_v2_0_start(adev); | 609 | return vce_v2_0_start(adev); |
593 | } | 610 | } |