summaryrefslogtreecommitdiffstats
path: root/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/gpu/nvgpu/gk20a/fifo_gk20a.c')
-rw-r--r--drivers/gpu/nvgpu/gk20a/fifo_gk20a.c21
1 files changed, 10 insertions, 11 deletions
diff --git a/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c b/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c
index 0c8bc6f4..05377c3d 100644
--- a/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c
+++ b/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c
@@ -1,6 +1,4 @@
1/* 1/*
2 * drivers/video/tegra/host/gk20a/fifo_gk20a.c
3 *
4 * GK20A Graphics FIFO (gr host) 2 * GK20A Graphics FIFO (gr host)
5 * 3 *
6 * Copyright (c) 2011-2014, NVIDIA CORPORATION. All rights reserved. 4 * Copyright (c) 2011-2014, NVIDIA CORPORATION. All rights reserved.
@@ -27,6 +25,7 @@
27 25
28#include "gk20a.h" 26#include "gk20a.h"
29#include "debug_gk20a.h" 27#include "debug_gk20a.h"
28#include "semaphore_gk20a.h"
30#include "hw_fifo_gk20a.h" 29#include "hw_fifo_gk20a.h"
31#include "hw_pbdma_gk20a.h" 30#include "hw_pbdma_gk20a.h"
32#include "hw_ccsr_gk20a.h" 31#include "hw_ccsr_gk20a.h"
@@ -917,11 +916,11 @@ static bool gk20a_fifo_set_ctx_mmu_error(struct gk20a *g,
917 * error condition. 916 * error condition.
918 * Don't overwrite error flag. */ 917 * Don't overwrite error flag. */
919 /* Fifo timeout debug spew is controlled by user */ 918 /* Fifo timeout debug spew is controlled by user */
920 if (err == NVHOST_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT) 919 if (err == NVGPU_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT)
921 verbose = ch->timeout_debug_dump; 920 verbose = ch->timeout_debug_dump;
922 } else { 921 } else {
923 gk20a_set_error_notifier(ch, 922 gk20a_set_error_notifier(ch,
924 NVHOST_CHANNEL_FIFO_ERROR_MMU_ERR_FLT); 923 NVGPU_CHANNEL_FIFO_ERROR_MMU_ERR_FLT);
925 } 924 }
926 } 925 }
927 /* mark channel as faulted */ 926 /* mark channel as faulted */
@@ -1294,13 +1293,13 @@ int gk20a_fifo_force_reset_ch(struct channel_gk20a *ch, bool verbose)
1294 mutex_lock(&tsg->ch_list_lock); 1293 mutex_lock(&tsg->ch_list_lock);
1295 list_for_each_entry(ch_tsg, &tsg->ch_list, ch_entry) { 1294 list_for_each_entry(ch_tsg, &tsg->ch_list, ch_entry) {
1296 gk20a_set_error_notifier(ch_tsg, 1295 gk20a_set_error_notifier(ch_tsg,
1297 NVHOST_CHANNEL_RESETCHANNEL_VERIF_ERROR); 1296 NVGPU_CHANNEL_RESETCHANNEL_VERIF_ERROR);
1298 } 1297 }
1299 mutex_unlock(&tsg->ch_list_lock); 1298 mutex_unlock(&tsg->ch_list_lock);
1300 gk20a_fifo_recover_tsg(ch->g, ch->tsgid, verbose); 1299 gk20a_fifo_recover_tsg(ch->g, ch->tsgid, verbose);
1301 } else { 1300 } else {
1302 gk20a_set_error_notifier(ch, 1301 gk20a_set_error_notifier(ch,
1303 NVHOST_CHANNEL_RESETCHANNEL_VERIF_ERROR); 1302 NVGPU_CHANNEL_RESETCHANNEL_VERIF_ERROR);
1304 gk20a_fifo_recover_ch(ch->g, ch->hw_chid, verbose); 1303 gk20a_fifo_recover_ch(ch->g, ch->hw_chid, verbose);
1305 } 1304 }
1306 1305
@@ -1364,7 +1363,7 @@ static bool gk20a_fifo_handle_sched_error(struct gk20a *g)
1364 if (gk20a_channel_update_and_check_timeout(ch, 1363 if (gk20a_channel_update_and_check_timeout(ch,
1365 GRFIFO_TIMEOUT_CHECK_PERIOD_US / 1000)) { 1364 GRFIFO_TIMEOUT_CHECK_PERIOD_US / 1000)) {
1366 gk20a_set_error_notifier(ch, 1365 gk20a_set_error_notifier(ch,
1367 NVHOST_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT); 1366 NVGPU_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT);
1368 gk20a_err(dev_from_gk20a(g), 1367 gk20a_err(dev_from_gk20a(g),
1369 "fifo sched ctxsw timeout error:" 1368 "fifo sched ctxsw timeout error:"
1370 "engine = %u, ch = %d", engine_id, id); 1369 "engine = %u, ch = %d", engine_id, id);
@@ -1504,7 +1503,7 @@ static u32 gk20a_fifo_handle_pbdma_intr(struct device *dev,
1504 struct channel_gk20a *ch = &f->channel[id]; 1503 struct channel_gk20a *ch = &f->channel[id];
1505 1504
1506 gk20a_set_error_notifier(ch, 1505 gk20a_set_error_notifier(ch,
1507 NVHOST_CHANNEL_PBDMA_ERROR); 1506 NVGPU_CHANNEL_PBDMA_ERROR);
1508 gk20a_fifo_recover_ch(g, id, true); 1507 gk20a_fifo_recover_ch(g, id, true);
1509 } else if (fifo_pbdma_status_id_type_v(status) 1508 } else if (fifo_pbdma_status_id_type_v(status)
1510 == fifo_pbdma_status_id_type_tsgid_v()) { 1509 == fifo_pbdma_status_id_type_tsgid_v()) {
@@ -1514,7 +1513,7 @@ static u32 gk20a_fifo_handle_pbdma_intr(struct device *dev,
1514 mutex_lock(&tsg->ch_list_lock); 1513 mutex_lock(&tsg->ch_list_lock);
1515 list_for_each_entry(ch, &tsg->ch_list, ch_entry) { 1514 list_for_each_entry(ch, &tsg->ch_list, ch_entry) {
1516 gk20a_set_error_notifier(ch, 1515 gk20a_set_error_notifier(ch,
1517 NVHOST_CHANNEL_PBDMA_ERROR); 1516 NVGPU_CHANNEL_PBDMA_ERROR);
1518 } 1517 }
1519 mutex_unlock(&tsg->ch_list_lock); 1518 mutex_unlock(&tsg->ch_list_lock);
1520 gk20a_fifo_recover_tsg(g, id, true); 1519 gk20a_fifo_recover_tsg(g, id, true);
@@ -1644,7 +1643,7 @@ static int __locked_fifo_preempt(struct gk20a *g, u32 id, bool is_tsg)
1644 mutex_lock(&tsg->ch_list_lock); 1643 mutex_lock(&tsg->ch_list_lock);
1645 list_for_each_entry(ch, &tsg->ch_list, ch_entry) { 1644 list_for_each_entry(ch, &tsg->ch_list, ch_entry) {
1646 gk20a_set_error_notifier(ch, 1645 gk20a_set_error_notifier(ch,
1647 NVHOST_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT); 1646 NVGPU_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT);
1648 } 1647 }
1649 mutex_unlock(&tsg->ch_list_lock); 1648 mutex_unlock(&tsg->ch_list_lock);
1650 gk20a_fifo_recover_tsg(g, id, true); 1649 gk20a_fifo_recover_tsg(g, id, true);
@@ -1655,7 +1654,7 @@ static int __locked_fifo_preempt(struct gk20a *g, u32 id, bool is_tsg)
1655 "preempt channel %d timeout\n", id); 1654 "preempt channel %d timeout\n", id);
1656 1655
1657 gk20a_set_error_notifier(ch, 1656 gk20a_set_error_notifier(ch,
1658 NVHOST_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT); 1657 NVGPU_CHANNEL_FIFO_ERROR_IDLE_TIMEOUT);
1659 gk20a_fifo_recover_ch(g, id, true); 1658 gk20a_fifo_recover_ch(g, id, true);
1660 } 1659 }
1661 } 1660 }