diff options
Diffstat (limited to 'drivers/gpu')
-rw-r--r-- | drivers/gpu/nvgpu/gp10b/mm_gp10b.c | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/drivers/gpu/nvgpu/gp10b/mm_gp10b.c b/drivers/gpu/nvgpu/gp10b/mm_gp10b.c index 0c76abb5..ae9c5c7e 100644 --- a/drivers/gpu/nvgpu/gp10b/mm_gp10b.c +++ b/drivers/gpu/nvgpu/gp10b/mm_gp10b.c | |||
@@ -152,7 +152,7 @@ static int update_gmmu_pde3_locked(struct vm_gk20a *vm, | |||
152 | struct scatterlist **sgl, | 152 | struct scatterlist **sgl, |
153 | u64 *offset, | 153 | u64 *offset, |
154 | u64 *iova, | 154 | u64 *iova, |
155 | u32 kind_v, u32 *ctag, | 155 | u32 kind_v, u64 *ctag, |
156 | bool cacheable, bool unmapped_pte, | 156 | bool cacheable, bool unmapped_pte, |
157 | int rw_flag, bool sparse, bool priv) | 157 | int rw_flag, bool sparse, bool priv) |
158 | { | 158 | { |
@@ -193,7 +193,7 @@ static int update_gmmu_pde0_locked(struct vm_gk20a *vm, | |||
193 | struct scatterlist **sgl, | 193 | struct scatterlist **sgl, |
194 | u64 *offset, | 194 | u64 *offset, |
195 | u64 *iova, | 195 | u64 *iova, |
196 | u32 kind_v, u32 *ctag, | 196 | u32 kind_v, u64 *ctag, |
197 | bool cacheable, bool unmapped_pte, | 197 | bool cacheable, bool unmapped_pte, |
198 | int rw_flag, bool sparse, bool priv) | 198 | int rw_flag, bool sparse, bool priv) |
199 | { | 199 | { |
@@ -249,13 +249,13 @@ static int update_gmmu_pte_locked(struct vm_gk20a *vm, | |||
249 | struct scatterlist **sgl, | 249 | struct scatterlist **sgl, |
250 | u64 *offset, | 250 | u64 *offset, |
251 | u64 *iova, | 251 | u64 *iova, |
252 | u32 kind_v, u32 *ctag, | 252 | u32 kind_v, u64 *ctag, |
253 | bool cacheable, bool unmapped_pte, | 253 | bool cacheable, bool unmapped_pte, |
254 | int rw_flag, bool sparse, bool priv) | 254 | int rw_flag, bool sparse, bool priv) |
255 | { | 255 | { |
256 | struct gk20a *g = vm->mm->g; | 256 | struct gk20a *g = vm->mm->g; |
257 | u32 page_size = vm->gmmu_page_sizes[gmmu_pgsz_idx]; | 257 | u32 page_size = vm->gmmu_page_sizes[gmmu_pgsz_idx]; |
258 | u32 ctag_granularity = g->ops.fb.compression_page_size(g); | 258 | u64 ctag_granularity = g->ops.fb.compression_page_size(g); |
259 | u32 pte_w[2] = {0, 0}; /* invalid pte */ | 259 | u32 pte_w[2] = {0, 0}; /* invalid pte */ |
260 | 260 | ||
261 | gk20a_dbg_fn(""); | 261 | gk20a_dbg_fn(""); |
@@ -274,7 +274,7 @@ static int update_gmmu_pte_locked(struct vm_gk20a *vm, | |||
274 | 274 | ||
275 | pte_w[1] = *iova >> (24 + gmmu_new_pte_address_shift_v()) | | 275 | pte_w[1] = *iova >> (24 + gmmu_new_pte_address_shift_v()) | |
276 | gmmu_new_pte_kind_f(kind_v) | | 276 | gmmu_new_pte_kind_f(kind_v) | |
277 | gmmu_new_pte_comptagline_f(*ctag / ctag_granularity); | 277 | gmmu_new_pte_comptagline_f((u32)(*ctag / ctag_granularity)); |
278 | 278 | ||
279 | if (rw_flag == gk20a_mem_flag_read_only) | 279 | if (rw_flag == gk20a_mem_flag_read_only) |
280 | pte_w[0] |= gmmu_new_pte_read_only_true_f(); | 280 | pte_w[0] |= gmmu_new_pte_read_only_true_f(); |
@@ -287,7 +287,7 @@ static int update_gmmu_pte_locked(struct vm_gk20a *vm, | |||
287 | " ctag=%d vol=%d" | 287 | " ctag=%d vol=%d" |
288 | " [0x%08x, 0x%08x]", | 288 | " [0x%08x, 0x%08x]", |
289 | i, *iova, | 289 | i, *iova, |
290 | kind_v, *ctag / ctag_granularity, !cacheable, | 290 | kind_v, (u32)(*ctag / ctag_granularity), !cacheable, |
291 | pte_w[1], pte_w[0]); | 291 | pte_w[1], pte_w[0]); |
292 | 292 | ||
293 | if (*ctag) | 293 | if (*ctag) |