diff options
Diffstat (limited to 'drivers/gpu/nvgpu/gk20a/fifo_gk20a.c')
-rw-r--r-- | drivers/gpu/nvgpu/gk20a/fifo_gk20a.c | 32 |
1 files changed, 16 insertions, 16 deletions
diff --git a/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c b/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c index f2fc6234..05c13374 100644 --- a/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c +++ b/drivers/gpu/nvgpu/gk20a/fifo_gk20a.c | |||
@@ -152,7 +152,7 @@ u32 gk20a_fifo_get_gr_engine_id(struct gk20a *g) | |||
152 | 1, ENGINE_GR_GK20A); | 152 | 1, ENGINE_GR_GK20A); |
153 | 153 | ||
154 | if (!gr_engine_cnt) { | 154 | if (!gr_engine_cnt) { |
155 | nvgpu_err(g, "No GR engine available on this device!\n"); | 155 | nvgpu_err(g, "No GR engine available on this device!"); |
156 | } | 156 | } |
157 | 157 | ||
158 | return gr_engine_id; | 158 | return gr_engine_id; |
@@ -693,7 +693,7 @@ static int init_runlist(struct gk20a *g, struct fifo_gk20a *f) | |||
693 | int err = nvgpu_dma_alloc_sys(g, runlist_size, | 693 | int err = nvgpu_dma_alloc_sys(g, runlist_size, |
694 | &runlist->mem[i]); | 694 | &runlist->mem[i]); |
695 | if (err) { | 695 | if (err) { |
696 | nvgpu_err(g, "memory allocation failed\n"); | 696 | nvgpu_err(g, "memory allocation failed"); |
697 | goto clean_up_runlist; | 697 | goto clean_up_runlist; |
698 | } | 698 | } |
699 | } | 699 | } |
@@ -947,7 +947,7 @@ static int gk20a_init_fifo_setup_sw(struct gk20a *g) | |||
947 | err = nvgpu_dma_alloc_sys(g, f->userd_entry_size * | 947 | err = nvgpu_dma_alloc_sys(g, f->userd_entry_size * |
948 | f->num_channels, &f->userd); | 948 | f->num_channels, &f->userd); |
949 | if (err) { | 949 | if (err) { |
950 | nvgpu_err(g, "userd memory allocation failed\n"); | 950 | nvgpu_err(g, "userd memory allocation failed"); |
951 | goto clean_up; | 951 | goto clean_up; |
952 | } | 952 | } |
953 | gk20a_dbg(gpu_dbg_map, "userd gpu va = 0x%llx", f->userd.gpu_va); | 953 | gk20a_dbg(gpu_dbg_map, "userd gpu va = 0x%llx", f->userd.gpu_va); |
@@ -1001,7 +1001,7 @@ void gk20a_fifo_handle_runlist_event(struct gk20a *g) | |||
1001 | { | 1001 | { |
1002 | u32 runlist_event = gk20a_readl(g, fifo_intr_runlist_r()); | 1002 | u32 runlist_event = gk20a_readl(g, fifo_intr_runlist_r()); |
1003 | 1003 | ||
1004 | gk20a_dbg(gpu_dbg_intr, "runlist event %08x\n", | 1004 | gk20a_dbg(gpu_dbg_intr, "runlist event %08x", |
1005 | runlist_event); | 1005 | runlist_event); |
1006 | 1006 | ||
1007 | gk20a_writel(g, fifo_intr_runlist_r(), runlist_event); | 1007 | gk20a_writel(g, fifo_intr_runlist_r(), runlist_event); |
@@ -1259,7 +1259,7 @@ static void gk20a_fifo_handle_chsw_fault(struct gk20a *g) | |||
1259 | u32 intr; | 1259 | u32 intr; |
1260 | 1260 | ||
1261 | intr = gk20a_readl(g, fifo_intr_chsw_error_r()); | 1261 | intr = gk20a_readl(g, fifo_intr_chsw_error_r()); |
1262 | nvgpu_err(g, "chsw: %08x\n", intr); | 1262 | nvgpu_err(g, "chsw: %08x", intr); |
1263 | gk20a_fecs_dump_falcon_stats(g); | 1263 | gk20a_fecs_dump_falcon_stats(g); |
1264 | gk20a_writel(g, fifo_intr_chsw_error_r(), intr); | 1264 | gk20a_writel(g, fifo_intr_chsw_error_r(), intr); |
1265 | } | 1265 | } |
@@ -1545,7 +1545,7 @@ static bool gk20a_fifo_handle_mmu_fault( | |||
1545 | nvgpu_err(g, "%s mmu fault on engine %d, " | 1545 | nvgpu_err(g, "%s mmu fault on engine %d, " |
1546 | "engine subid %d (%s), client %d (%s), " | 1546 | "engine subid %d (%s), client %d (%s), " |
1547 | "addr 0x%08x:0x%08x, type %d (%s), info 0x%08x," | 1547 | "addr 0x%08x:0x%08x, type %d (%s), info 0x%08x," |
1548 | "inst_ptr 0x%llx\n", | 1548 | "inst_ptr 0x%llx", |
1549 | fake_fault ? "fake" : "", | 1549 | fake_fault ? "fake" : "", |
1550 | engine_id, | 1550 | engine_id, |
1551 | f.engine_subid_v, f.engine_subid_desc, | 1551 | f.engine_subid_v, f.engine_subid_desc, |
@@ -2136,7 +2136,7 @@ bool gk20a_fifo_handle_sched_error(struct gk20a *g) | |||
2136 | 2136 | ||
2137 | /* could not find the engine - should never happen */ | 2137 | /* could not find the engine - should never happen */ |
2138 | if (!gk20a_fifo_is_valid_engine_id(g, engine_id)) { | 2138 | if (!gk20a_fifo_is_valid_engine_id(g, engine_id)) { |
2139 | nvgpu_err(g, "fifo sched error : 0x%08x, failed to find engine\n", | 2139 | nvgpu_err(g, "fifo sched error : 0x%08x, failed to find engine", |
2140 | sched_error); | 2140 | sched_error); |
2141 | ret = false; | 2141 | ret = false; |
2142 | goto err; | 2142 | goto err; |
@@ -2193,7 +2193,7 @@ static u32 fifo_error_isr(struct gk20a *g, u32 fifo_intr) | |||
2193 | if (fifo_intr & fifo_intr_0_pio_error_pending_f()) { | 2193 | if (fifo_intr & fifo_intr_0_pio_error_pending_f()) { |
2194 | /* pio mode is unused. this shouldn't happen, ever. */ | 2194 | /* pio mode is unused. this shouldn't happen, ever. */ |
2195 | /* should we clear it or just leave it pending? */ | 2195 | /* should we clear it or just leave it pending? */ |
2196 | nvgpu_err(g, "fifo pio error!\n"); | 2196 | nvgpu_err(g, "fifo pio error!"); |
2197 | BUG_ON(1); | 2197 | BUG_ON(1); |
2198 | } | 2198 | } |
2199 | 2199 | ||
@@ -2547,7 +2547,7 @@ void __locked_fifo_preempt_timeout_rc(struct gk20a *g, u32 id, | |||
2547 | struct channel_gk20a *ch = NULL; | 2547 | struct channel_gk20a *ch = NULL; |
2548 | 2548 | ||
2549 | nvgpu_err(g, | 2549 | nvgpu_err(g, |
2550 | "preempt TSG %d timeout\n", id); | 2550 | "preempt TSG %d timeout", id); |
2551 | 2551 | ||
2552 | down_read(&tsg->ch_list_lock); | 2552 | down_read(&tsg->ch_list_lock); |
2553 | list_for_each_entry(ch, &tsg->ch_list, ch_entry) { | 2553 | list_for_each_entry(ch, &tsg->ch_list, ch_entry) { |
@@ -2563,7 +2563,7 @@ void __locked_fifo_preempt_timeout_rc(struct gk20a *g, u32 id, | |||
2563 | struct channel_gk20a *ch = &g->fifo.channel[id]; | 2563 | struct channel_gk20a *ch = &g->fifo.channel[id]; |
2564 | 2564 | ||
2565 | nvgpu_err(g, | 2565 | nvgpu_err(g, |
2566 | "preempt channel %d timeout\n", id); | 2566 | "preempt channel %d timeout", id); |
2567 | 2567 | ||
2568 | if (gk20a_channel_get(ch)) { | 2568 | if (gk20a_channel_get(ch)) { |
2569 | gk20a_set_error_notifier(ch, | 2569 | gk20a_set_error_notifier(ch, |
@@ -2746,7 +2746,7 @@ int gk20a_fifo_enable_all_engine_activity(struct gk20a *g) | |||
2746 | &g->fifo.engine_info[active_engine_id]); | 2746 | &g->fifo.engine_info[active_engine_id]); |
2747 | if (err) { | 2747 | if (err) { |
2748 | nvgpu_err(g, | 2748 | nvgpu_err(g, |
2749 | "failed to enable engine %d activity\n", active_engine_id); | 2749 | "failed to enable engine %d activity", active_engine_id); |
2750 | ret = err; | 2750 | ret = err; |
2751 | } | 2751 | } |
2752 | } | 2752 | } |
@@ -2819,7 +2819,7 @@ clean_up: | |||
2819 | gk20a_dbg_fn("failed"); | 2819 | gk20a_dbg_fn("failed"); |
2820 | if (gk20a_fifo_enable_engine_activity(g, eng_info)) | 2820 | if (gk20a_fifo_enable_engine_activity(g, eng_info)) |
2821 | nvgpu_err(g, | 2821 | nvgpu_err(g, |
2822 | "failed to enable gr engine activity\n"); | 2822 | "failed to enable gr engine activity"); |
2823 | } else { | 2823 | } else { |
2824 | gk20a_dbg_fn("done"); | 2824 | gk20a_dbg_fn("done"); |
2825 | } | 2825 | } |
@@ -2839,7 +2839,7 @@ int gk20a_fifo_disable_all_engine_activity(struct gk20a *g, | |||
2839 | &g->fifo.engine_info[active_engine_id], | 2839 | &g->fifo.engine_info[active_engine_id], |
2840 | wait_for_idle); | 2840 | wait_for_idle); |
2841 | if (err) { | 2841 | if (err) { |
2842 | nvgpu_err(g, "failed to disable engine %d activity\n", | 2842 | nvgpu_err(g, "failed to disable engine %d activity", |
2843 | active_engine_id); | 2843 | active_engine_id); |
2844 | ret = err; | 2844 | ret = err; |
2845 | break; | 2845 | break; |
@@ -2853,7 +2853,7 @@ int gk20a_fifo_disable_all_engine_activity(struct gk20a *g, | |||
2853 | &g->fifo.engine_info[active_engine_id]); | 2853 | &g->fifo.engine_info[active_engine_id]); |
2854 | if (err) | 2854 | if (err) |
2855 | nvgpu_err(g, | 2855 | nvgpu_err(g, |
2856 | "failed to re-enable engine %d activity\n", | 2856 | "failed to re-enable engine %d activity", |
2857 | active_engine_id); | 2857 | active_engine_id); |
2858 | } | 2858 | } |
2859 | } | 2859 | } |
@@ -4108,7 +4108,7 @@ int gk20a_fifo_set_timeslice(struct channel_gk20a *ch, u32 timeslice) | |||
4108 | struct gk20a *g = ch->g; | 4108 | struct gk20a *g = ch->g; |
4109 | 4109 | ||
4110 | if (gk20a_is_channel_marked_as_tsg(ch)) { | 4110 | if (gk20a_is_channel_marked_as_tsg(ch)) { |
4111 | nvgpu_err(g, "invalid operation for TSG!\n"); | 4111 | nvgpu_err(g, "invalid operation for TSG!"); |
4112 | return -EINVAL; | 4112 | return -EINVAL; |
4113 | } | 4113 | } |
4114 | 4114 | ||
@@ -4127,7 +4127,7 @@ int gk20a_fifo_set_timeslice(struct channel_gk20a *ch, u32 timeslice) | |||
4127 | int gk20a_fifo_set_priority(struct channel_gk20a *ch, u32 priority) | 4127 | int gk20a_fifo_set_priority(struct channel_gk20a *ch, u32 priority) |
4128 | { | 4128 | { |
4129 | if (gk20a_is_channel_marked_as_tsg(ch)) { | 4129 | if (gk20a_is_channel_marked_as_tsg(ch)) { |
4130 | nvgpu_err(ch->g, "invalid operation for TSG!\n"); | 4130 | nvgpu_err(ch->g, "invalid operation for TSG!"); |
4131 | return -EINVAL; | 4131 | return -EINVAL; |
4132 | } | 4132 | } |
4133 | 4133 | ||