aboutsummaryrefslogtreecommitdiffstats
path: root/include/os/linux/driver_common.c
diff options
context:
space:
mode:
Diffstat (limited to 'include/os/linux/driver_common.c')
-rw-r--r--include/os/linux/driver_common.c400
1 files changed, 0 insertions, 400 deletions
diff --git a/include/os/linux/driver_common.c b/include/os/linux/driver_common.c
deleted file mode 100644
index 8f5872d..0000000
--- a/include/os/linux/driver_common.c
+++ /dev/null
@@ -1,400 +0,0 @@
1/*
2 * Copyright (c) 2016-2022, NVIDIA CORPORATION. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
11 * more details.
12 *
13 * You should have received a copy of the GNU General Public License
14 * along with this program. If not, see <http://www.gnu.org/licenses/>.
15 */
16
17#include <linux/reboot.h>
18#include <linux/dma-mapping.h>
19#include <linux/mm.h>
20#include <linux/slab.h>
21#include <linux/of_platform.h>
22#include <uapi/linux/nvgpu.h>
23
24#include <nvgpu/defaults.h>
25#include <nvgpu/kmem.h>
26#include <nvgpu/nvgpu_common.h>
27#include <nvgpu/soc.h>
28#include <nvgpu/bug.h>
29#include <nvgpu/enabled.h>
30#include <nvgpu/debug.h>
31#include <nvgpu/sizes.h>
32#include <nvgpu/gk20a.h>
33
34#include "platform_gk20a.h"
35#include "module.h"
36#include "os_linux.h"
37#include "sysfs.h"
38#include "ioctl.h"
39#include "gk20a/regops_gk20a.h"
40
41#define EMC3D_DEFAULT_RATIO 750
42
43void nvgpu_kernel_restart(void *cmd)
44{
45 kernel_restart(cmd);
46}
47
48static void nvgpu_init_vars(struct gk20a *g)
49{
50 struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
51 struct device *dev = dev_from_gk20a(g);
52 struct gk20a_platform *platform = dev_get_drvdata(dev);
53
54 nvgpu_cond_init(&l->sw_irq_stall_last_handled_wq);
55 nvgpu_cond_init(&l->sw_irq_nonstall_last_handled_wq);
56
57 init_rwsem(&l->busy_lock);
58 nvgpu_rwsem_init(&g->deterministic_busy);
59
60 nvgpu_spinlock_init(&g->mc_enable_lock);
61
62 nvgpu_mutex_init(&platform->railgate_lock);
63 nvgpu_mutex_init(&g->dbg_sessions_lock);
64 nvgpu_mutex_init(&g->client_lock);
65 nvgpu_mutex_init(&g->power_lock);
66 nvgpu_mutex_init(&g->ctxsw_disable_lock);
67 nvgpu_mutex_init(&g->tpc_pg_lock);
68 nvgpu_mutex_init(&g->clk_arb_enable_lock);
69 nvgpu_mutex_init(&g->cg_pg_lock);
70
71 /* Init the clock req count to 0 */
72 nvgpu_atomic_set(&g->clk_arb_global_nr, 0);
73
74 nvgpu_mutex_init(&l->ctrl.privs_lock);
75 nvgpu_init_list_node(&l->ctrl.privs);
76
77 l->regs_saved = l->regs;
78 l->bar1_saved = l->bar1;
79
80 g->emc3d_ratio = EMC3D_DEFAULT_RATIO;
81
82 /* Set DMA parameters to allow larger sgt lists */
83 dev->dma_parms = &l->dma_parms;
84 dma_set_max_seg_size(dev, UINT_MAX);
85
86 /*
87 * A default of 16GB is the largest supported DMA size that is
88 * acceptable to all currently supported Tegra SoCs.
89 */
90 if (!platform->dma_mask)
91 platform->dma_mask = DMA_BIT_MASK(34);
92
93 dma_set_mask(dev, platform->dma_mask);
94 dma_set_coherent_mask(dev, platform->dma_mask);
95
96 nvgpu_init_list_node(&g->profiler_objects);
97
98 nvgpu_init_list_node(&g->boardobj_head);
99 nvgpu_init_list_node(&g->boardobjgrp_head);
100
101 __nvgpu_set_enabled(g, NVGPU_HAS_SYNCPOINTS, platform->has_syncpoints);
102}
103
104static void nvgpu_init_gr_vars(struct gk20a *g)
105{
106 gk20a_init_gr(g);
107
108 nvgpu_log_info(g, "total ram pages : %lu", totalram_pages);
109 g->gr.max_comptag_mem = totalram_size_in_mb;
110}
111
112static void nvgpu_init_timeout(struct gk20a *g)
113{
114 struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
115
116 g->timeouts_disabled_by_user = false;
117 nvgpu_atomic_set(&g->timeouts_disabled_refcount, 0);
118
119 if (nvgpu_platform_is_silicon(g)) {
120 g->gr_idle_timeout_default = NVGPU_DEFAULT_GR_IDLE_TIMEOUT;
121 } else if (nvgpu_platform_is_fpga(g)) {
122 g->gr_idle_timeout_default = GK20A_TIMEOUT_FPGA;
123 } else {
124 g->gr_idle_timeout_default = (u32)ULONG_MAX;
125 }
126 g->ch_wdt_timeout_ms = platform->ch_wdt_timeout_ms;
127 g->fifo_eng_timeout_us = GRFIFO_TIMEOUT_CHECK_PERIOD_US;
128}
129
130static void nvgpu_init_timeslice(struct gk20a *g)
131{
132 g->runlist_interleave = true;
133
134 g->timeslice_low_priority_us = 1300;
135 g->timeslice_medium_priority_us = 2600;
136 g->timeslice_high_priority_us = 5200;
137
138 g->min_timeslice_us = 1000;
139 g->max_timeslice_us = 50000;
140}
141
142static void nvgpu_init_pm_vars(struct gk20a *g)
143{
144 struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
145 u32 i = 0;
146
147 /*
148 * Set up initial power settings. For non-slicon platforms, disable
149 * power features and for silicon platforms, read from platform data
150 */
151 g->slcg_enabled =
152 nvgpu_platform_is_silicon(g) ? platform->enable_slcg : false;
153 g->blcg_enabled =
154 nvgpu_platform_is_silicon(g) ? platform->enable_blcg : false;
155 g->elcg_enabled =
156 nvgpu_platform_is_silicon(g) ? platform->enable_elcg : false;
157 g->elpg_enabled =
158 nvgpu_platform_is_silicon(g) ? platform->enable_elpg : false;
159 g->aelpg_enabled =
160 nvgpu_platform_is_silicon(g) ? platform->enable_aelpg : false;
161 g->mscg_enabled =
162 nvgpu_platform_is_silicon(g) ? platform->enable_mscg : false;
163 g->can_elpg =
164 nvgpu_platform_is_silicon(g) ? platform->can_elpg_init : false;
165
166 __nvgpu_set_enabled(g, NVGPU_GPU_CAN_ELCG,
167 nvgpu_platform_is_silicon(g) ? platform->can_elcg : false);
168 __nvgpu_set_enabled(g, NVGPU_GPU_CAN_SLCG,
169 nvgpu_platform_is_silicon(g) ? platform->can_slcg : false);
170 __nvgpu_set_enabled(g, NVGPU_GPU_CAN_BLCG,
171 nvgpu_platform_is_silicon(g) ? platform->can_blcg : false);
172
173 g->aggressive_sync_destroy = platform->aggressive_sync_destroy;
174 g->aggressive_sync_destroy_thresh = platform->aggressive_sync_destroy_thresh;
175#ifdef CONFIG_NVGPU_SUPPORT_CDE
176 g->has_cde = platform->has_cde;
177#endif
178 g->ptimer_src_freq = platform->ptimer_src_freq;
179 g->support_pmu = support_gk20a_pmu(dev_from_gk20a(g));
180 __nvgpu_set_enabled(g, NVGPU_CAN_RAILGATE, platform->can_railgate_init);
181 g->can_tpc_powergate = platform->can_tpc_powergate;
182
183 for (i = 0; i < MAX_TPC_PG_CONFIGS; i++)
184 g->valid_tpc_mask[i] = platform->valid_tpc_mask[i];
185
186 g->ldiv_slowdown_factor = platform->ldiv_slowdown_factor_init;
187 /* if default delay is not set, set default delay to 500msec */
188 if (platform->railgate_delay_init)
189 g->railgate_delay = platform->railgate_delay_init;
190 else
191 g->railgate_delay = NVGPU_DEFAULT_RAILGATE_IDLE_TIMEOUT;
192 __nvgpu_set_enabled(g, NVGPU_PMU_PERFMON, platform->enable_perfmon);
193
194 /* set default values to aelpg parameters */
195 g->pmu.aelpg_param[0] = APCTRL_SAMPLING_PERIOD_PG_DEFAULT_US;
196 g->pmu.aelpg_param[1] = APCTRL_MINIMUM_IDLE_FILTER_DEFAULT_US;
197 g->pmu.aelpg_param[2] = APCTRL_MINIMUM_TARGET_SAVING_DEFAULT_US;
198 g->pmu.aelpg_param[3] = APCTRL_POWER_BREAKEVEN_DEFAULT_US;
199 g->pmu.aelpg_param[4] = APCTRL_CYCLES_PER_SAMPLE_MAX_DEFAULT;
200
201 __nvgpu_set_enabled(g, NVGPU_SUPPORT_ASPM, !platform->disable_aspm);
202}
203
204static void nvgpu_init_vbios_vars(struct gk20a *g)
205{
206 struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
207
208 __nvgpu_set_enabled(g, NVGPU_PMU_RUN_PREOS, platform->run_preos);
209 g->vbios_min_version = platform->vbios_min_version;
210}
211
212static void nvgpu_init_ltc_vars(struct gk20a *g)
213{
214 struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
215
216 g->ltc_streamid = platform->ltc_streamid;
217}
218
219static void nvgpu_init_mm_vars(struct gk20a *g)
220{
221 struct gk20a_platform *platform = dev_get_drvdata(dev_from_gk20a(g));
222
223 g->mm.disable_bigpage = platform->disable_bigpage;
224 __nvgpu_set_enabled(g, NVGPU_MM_HONORS_APERTURE,
225 platform->honors_aperture);
226 __nvgpu_set_enabled(g, NVGPU_MM_UNIFIED_MEMORY,
227 platform->unified_memory);
228 __nvgpu_set_enabled(g, NVGPU_MM_UNIFY_ADDRESS_SPACES,
229 platform->unify_address_spaces);
230 __nvgpu_set_enabled(g, NVGPU_MM_FORCE_128K_PMU_VM,
231 platform->force_128K_pmu_vm);
232
233 nvgpu_mutex_init(&g->mm.tlb_lock);
234 nvgpu_mutex_init(&g->mm.priv_lock);
235}
236
237int nvgpu_probe(struct gk20a *g,
238 const char *debugfs_symlink,
239 const char *interface_name,
240 struct class *class)
241{
242 struct device *dev = dev_from_gk20a(g);
243 struct gk20a_platform *platform = dev_get_drvdata(dev);
244 int err = 0;
245 struct device_node *np = dev->of_node;
246 bool disable_l3_alloc = false;
247
248 nvgpu_init_vars(g);
249 nvgpu_init_gr_vars(g);
250 nvgpu_init_timeout(g);
251 nvgpu_init_timeslice(g);
252 nvgpu_init_pm_vars(g);
253 nvgpu_init_vbios_vars(g);
254 nvgpu_init_ltc_vars(g);
255 err = nvgpu_init_soc_vars(g);
256 if (err) {
257 nvgpu_err(g, "init soc vars failed");
258 return err;
259 }
260
261 /* Initialize the platform interface. */
262 err = platform->probe(dev);
263 if (err) {
264 if (err == -EPROBE_DEFER)
265 nvgpu_info(g, "platform probe failed");
266 else
267 nvgpu_err(g, "platform probe failed");
268 return err;
269 }
270
271 disable_l3_alloc = of_property_read_bool(np, "disable_l3_alloc");
272 if (disable_l3_alloc) {
273 nvgpu_log_info(g, "L3 alloc is disabled\n");
274 __nvgpu_set_enabled(g, NVGPU_DISABLE_L3_SUPPORT, true);
275 }
276
277 nvgpu_init_mm_vars(g);
278
279 /* platform probe can defer do user init only if probe succeeds */
280 err = gk20a_user_init(dev, interface_name, class);
281 if (err)
282 return err;
283
284 if (platform->late_probe) {
285 err = platform->late_probe(dev);
286 if (err) {
287 nvgpu_err(g, "late probe failed");
288 return err;
289 }
290 }
291
292 nvgpu_create_sysfs(dev);
293 gk20a_debug_init(g, debugfs_symlink);
294
295 g->dbg_regops_tmp_buf = nvgpu_kzalloc(g, SZ_4K);
296 if (!g->dbg_regops_tmp_buf) {
297 nvgpu_err(g, "couldn't allocate regops tmp buf");
298 return -ENOMEM;
299 }
300 g->dbg_regops_tmp_buf_ops =
301 SZ_4K / sizeof(g->dbg_regops_tmp_buf[0]);
302
303 g->remove_support = gk20a_remove_support;
304
305 nvgpu_ref_init(&g->refcount);
306
307 return 0;
308}
309
310/**
311 * cyclic_delta - Returns delta of cyclic integers a and b.
312 *
313 * @a - First integer
314 * @b - Second integer
315 *
316 * Note: if a is ahead of b, delta is positive.
317 */
318static int cyclic_delta(int a, int b)
319{
320 return a - b;
321}
322
323/**
324 * nvgpu_wait_for_stall_interrupts - Wait for the stalling interrupts to
325 * complete.
326 *
327 * @g - The GPU to wait on.
328 * @timeout - maximum time period to wait for.
329 *
330 * Waits until all stalling interrupt handlers that have been scheduled to run
331 * have completed.
332 */
333int nvgpu_wait_for_stall_interrupts(struct gk20a *g, u32 timeout)
334{
335 struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
336 int stall_irq_threshold = atomic_read(&l->hw_irq_stall_count);
337
338 /* wait until all stalling irqs are handled */
339 return NVGPU_COND_WAIT(&l->sw_irq_stall_last_handled_wq,
340 cyclic_delta(stall_irq_threshold,
341 atomic_read(&l->sw_irq_stall_last_handled))
342 <= 0, timeout);
343}
344
345/**
346 * nvgpu_wait_for_nonstall_interrupts - Wait for the nonstalling interrupts to
347 * complete.
348 *
349 * @g - The GPU to wait on.
350 * @timeout - maximum time period to wait for.
351 *
352 * Waits until all non-stalling interrupt handlers that have been scheduled to
353 * run have completed.
354 */
355int nvgpu_wait_for_nonstall_interrupts(struct gk20a *g, u32 timeout)
356{
357 struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
358 int nonstall_irq_threshold = atomic_read(&l->hw_irq_nonstall_count);
359
360 /* wait until all non-stalling irqs are handled */
361 return NVGPU_COND_WAIT(&l->sw_irq_nonstall_last_handled_wq,
362 cyclic_delta(nonstall_irq_threshold,
363 atomic_read(&l->sw_irq_nonstall_last_handled))
364 <= 0, timeout);
365}
366
367/**
368 * nvgpu_wait_for_deferred_interrupts - Wait for interrupts to complete
369 *
370 * @g - The GPU to wait on.
371 *
372 * Waits until all interrupt handlers that have been scheduled to run have
373 * completed.
374 */
375void nvgpu_wait_for_deferred_interrupts(struct gk20a *g)
376{
377 int ret;
378
379 ret = nvgpu_wait_for_stall_interrupts(g, 0U);
380 if (ret != 0) {
381 nvgpu_err(g, "wait for stall interrupts failed %d", ret);
382 }
383
384 ret = nvgpu_wait_for_nonstall_interrupts(g, 0U);
385 if (ret != 0) {
386 nvgpu_err(g, "wait for nonstall interrupts failed %d", ret);
387 }
388}
389
390static void nvgpu_free_gk20a(struct gk20a *g)
391{
392 struct nvgpu_os_linux *l = nvgpu_os_linux_from_gk20a(g);
393
394 kfree(l);
395}
396
397void nvgpu_init_gk20a(struct gk20a *g)
398{
399 g->free = nvgpu_free_gk20a;
400}