diff options
author | Alex Deucher <alexander.deucher@amd.com> | 2015-04-20 16:55:21 -0400 |
---|---|---|
committer | Alex Deucher <alexander.deucher@amd.com> | 2015-06-03 21:03:15 -0400 |
commit | d38ceaf99ed015f2a0b9af3499791bd3a3daae21 (patch) | |
tree | c8e237ea218e8ed8a5f64c1654fc01fe5d2239cb /drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | |
parent | 97b2e202fba05b87d720318a6500a337100dab4d (diff) |
drm/amdgpu: add core driver (v4)
This adds the non-asic specific core driver code.
v2: remove extra kconfig option
v3: implement minor fixes from Fengguang Wu
v4: fix cast in amdgpu_ucode.c
Acked-by: Christian König <christian.koenig@amd.com>
Acked-by: Jammy Zhou <Jammy.Zhou@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/amdgpu_object.c')
-rw-r--r-- | drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | 646 |
1 files changed, 646 insertions, 0 deletions
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c new file mode 100644 index 000000000000..b51582714c21 --- /dev/null +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.c | |||
@@ -0,0 +1,646 @@ | |||
1 | /* | ||
2 | * Copyright 2009 Jerome Glisse. | ||
3 | * All Rights Reserved. | ||
4 | * | ||
5 | * Permission is hereby granted, free of charge, to any person obtaining a | ||
6 | * copy of this software and associated documentation files (the | ||
7 | * "Software"), to deal in the Software without restriction, including | ||
8 | * without limitation the rights to use, copy, modify, merge, publish, | ||
9 | * distribute, sub license, and/or sell copies of the Software, and to | ||
10 | * permit persons to whom the Software is furnished to do so, subject to | ||
11 | * the following conditions: | ||
12 | * | ||
13 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
14 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
15 | * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL | ||
16 | * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, | ||
17 | * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR | ||
18 | * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE | ||
19 | * USE OR OTHER DEALINGS IN THE SOFTWARE. | ||
20 | * | ||
21 | * The above copyright notice and this permission notice (including the | ||
22 | * next paragraph) shall be included in all copies or substantial portions | ||
23 | * of the Software. | ||
24 | * | ||
25 | */ | ||
26 | /* | ||
27 | * Authors: | ||
28 | * Jerome Glisse <glisse@freedesktop.org> | ||
29 | * Thomas Hellstrom <thomas-at-tungstengraphics-dot-com> | ||
30 | * Dave Airlie | ||
31 | */ | ||
32 | #include <linux/list.h> | ||
33 | #include <linux/slab.h> | ||
34 | #include <drm/drmP.h> | ||
35 | #include <drm/amdgpu_drm.h> | ||
36 | #include "amdgpu.h" | ||
37 | #include "amdgpu_trace.h" | ||
38 | |||
39 | |||
40 | int amdgpu_ttm_init(struct amdgpu_device *adev); | ||
41 | void amdgpu_ttm_fini(struct amdgpu_device *adev); | ||
42 | |||
43 | static u64 amdgpu_get_vis_part_size(struct amdgpu_device *adev, | ||
44 | struct ttm_mem_reg * mem) | ||
45 | { | ||
46 | u64 ret = 0; | ||
47 | if (mem->start << PAGE_SHIFT < adev->mc.visible_vram_size) { | ||
48 | ret = (u64)((mem->start << PAGE_SHIFT) + mem->size) > | ||
49 | adev->mc.visible_vram_size ? | ||
50 | adev->mc.visible_vram_size - (mem->start << PAGE_SHIFT): | ||
51 | mem->size; | ||
52 | } | ||
53 | return ret; | ||
54 | } | ||
55 | |||
56 | static void amdgpu_update_memory_usage(struct amdgpu_device *adev, | ||
57 | struct ttm_mem_reg *old_mem, | ||
58 | struct ttm_mem_reg *new_mem) | ||
59 | { | ||
60 | u64 vis_size; | ||
61 | if (!adev) | ||
62 | return; | ||
63 | |||
64 | if (new_mem) { | ||
65 | switch (new_mem->mem_type) { | ||
66 | case TTM_PL_TT: | ||
67 | atomic64_add(new_mem->size, &adev->gtt_usage); | ||
68 | break; | ||
69 | case TTM_PL_VRAM: | ||
70 | atomic64_add(new_mem->size, &adev->vram_usage); | ||
71 | vis_size = amdgpu_get_vis_part_size(adev, new_mem); | ||
72 | atomic64_add(vis_size, &adev->vram_vis_usage); | ||
73 | break; | ||
74 | } | ||
75 | } | ||
76 | |||
77 | if (old_mem) { | ||
78 | switch (old_mem->mem_type) { | ||
79 | case TTM_PL_TT: | ||
80 | atomic64_sub(old_mem->size, &adev->gtt_usage); | ||
81 | break; | ||
82 | case TTM_PL_VRAM: | ||
83 | atomic64_sub(old_mem->size, &adev->vram_usage); | ||
84 | vis_size = amdgpu_get_vis_part_size(adev, old_mem); | ||
85 | atomic64_sub(vis_size, &adev->vram_vis_usage); | ||
86 | break; | ||
87 | } | ||
88 | } | ||
89 | } | ||
90 | |||
91 | static void amdgpu_ttm_bo_destroy(struct ttm_buffer_object *tbo) | ||
92 | { | ||
93 | struct amdgpu_bo *bo; | ||
94 | |||
95 | bo = container_of(tbo, struct amdgpu_bo, tbo); | ||
96 | |||
97 | amdgpu_update_memory_usage(bo->adev, &bo->tbo.mem, NULL); | ||
98 | amdgpu_mn_unregister(bo); | ||
99 | |||
100 | mutex_lock(&bo->adev->gem.mutex); | ||
101 | list_del_init(&bo->list); | ||
102 | mutex_unlock(&bo->adev->gem.mutex); | ||
103 | drm_gem_object_release(&bo->gem_base); | ||
104 | kfree(bo->metadata); | ||
105 | kfree(bo); | ||
106 | } | ||
107 | |||
108 | bool amdgpu_ttm_bo_is_amdgpu_bo(struct ttm_buffer_object *bo) | ||
109 | { | ||
110 | if (bo->destroy == &amdgpu_ttm_bo_destroy) | ||
111 | return true; | ||
112 | return false; | ||
113 | } | ||
114 | |||
115 | void amdgpu_ttm_placement_from_domain(struct amdgpu_bo *rbo, u32 domain) | ||
116 | { | ||
117 | u32 c = 0, i; | ||
118 | rbo->placement.placement = rbo->placements; | ||
119 | rbo->placement.busy_placement = rbo->placements; | ||
120 | |||
121 | if (domain & AMDGPU_GEM_DOMAIN_VRAM) { | ||
122 | if (rbo->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS && | ||
123 | rbo->adev->mc.visible_vram_size < rbo->adev->mc.real_vram_size) { | ||
124 | rbo->placements[c].fpfn = | ||
125 | rbo->adev->mc.visible_vram_size >> PAGE_SHIFT; | ||
126 | rbo->placements[c++].flags = TTM_PL_FLAG_WC | TTM_PL_FLAG_UNCACHED | | ||
127 | TTM_PL_FLAG_VRAM; | ||
128 | } | ||
129 | rbo->placements[c].fpfn = 0; | ||
130 | rbo->placements[c++].flags = TTM_PL_FLAG_WC | TTM_PL_FLAG_UNCACHED | | ||
131 | TTM_PL_FLAG_VRAM; | ||
132 | } | ||
133 | |||
134 | if (domain & AMDGPU_GEM_DOMAIN_GTT) { | ||
135 | if (rbo->flags & AMDGPU_GEM_CREATE_CPU_GTT_UC) { | ||
136 | rbo->placements[c].fpfn = 0; | ||
137 | rbo->placements[c++].flags = TTM_PL_FLAG_UNCACHED | TTM_PL_FLAG_TT; | ||
138 | } else if (rbo->flags & AMDGPU_GEM_CREATE_CPU_GTT_WC) { | ||
139 | rbo->placements[c].fpfn = 0; | ||
140 | rbo->placements[c++].flags = TTM_PL_FLAG_WC | TTM_PL_FLAG_TT | | ||
141 | TTM_PL_FLAG_UNCACHED; | ||
142 | } else { | ||
143 | rbo->placements[c].fpfn = 0; | ||
144 | rbo->placements[c++].flags = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_TT; | ||
145 | } | ||
146 | } | ||
147 | |||
148 | if (domain & AMDGPU_GEM_DOMAIN_CPU) { | ||
149 | if (rbo->flags & AMDGPU_GEM_CREATE_CPU_GTT_UC) { | ||
150 | rbo->placements[c].fpfn = 0; | ||
151 | rbo->placements[c++].flags = TTM_PL_FLAG_UNCACHED | TTM_PL_FLAG_SYSTEM; | ||
152 | } else if (rbo->flags & AMDGPU_GEM_CREATE_CPU_GTT_WC) { | ||
153 | rbo->placements[c].fpfn = 0; | ||
154 | rbo->placements[c++].flags = TTM_PL_FLAG_WC | TTM_PL_FLAG_SYSTEM | | ||
155 | TTM_PL_FLAG_UNCACHED; | ||
156 | } else { | ||
157 | rbo->placements[c].fpfn = 0; | ||
158 | rbo->placements[c++].flags = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_SYSTEM; | ||
159 | } | ||
160 | } | ||
161 | |||
162 | if (domain & AMDGPU_GEM_DOMAIN_GDS) { | ||
163 | rbo->placements[c++].flags = TTM_PL_FLAG_UNCACHED | | ||
164 | AMDGPU_PL_FLAG_GDS; | ||
165 | } | ||
166 | if (domain & AMDGPU_GEM_DOMAIN_GWS) { | ||
167 | rbo->placements[c++].flags = TTM_PL_FLAG_UNCACHED | | ||
168 | AMDGPU_PL_FLAG_GWS; | ||
169 | } | ||
170 | if (domain & AMDGPU_GEM_DOMAIN_OA) { | ||
171 | rbo->placements[c++].flags = TTM_PL_FLAG_UNCACHED | | ||
172 | AMDGPU_PL_FLAG_OA; | ||
173 | } | ||
174 | |||
175 | if (!c) { | ||
176 | rbo->placements[c].fpfn = 0; | ||
177 | rbo->placements[c++].flags = TTM_PL_MASK_CACHING | | ||
178 | TTM_PL_FLAG_SYSTEM; | ||
179 | } | ||
180 | rbo->placement.num_placement = c; | ||
181 | rbo->placement.num_busy_placement = c; | ||
182 | |||
183 | for (i = 0; i < c; i++) { | ||
184 | if ((rbo->flags & AMDGPU_GEM_CREATE_CPU_ACCESS_REQUIRED) && | ||
185 | (rbo->placements[i].flags & TTM_PL_FLAG_VRAM) && | ||
186 | !rbo->placements[i].fpfn) | ||
187 | rbo->placements[i].lpfn = | ||
188 | rbo->adev->mc.visible_vram_size >> PAGE_SHIFT; | ||
189 | else | ||
190 | rbo->placements[i].lpfn = 0; | ||
191 | } | ||
192 | |||
193 | if (rbo->tbo.mem.size > 512 * 1024) { | ||
194 | for (i = 0; i < c; i++) { | ||
195 | rbo->placements[i].flags |= TTM_PL_FLAG_TOPDOWN; | ||
196 | } | ||
197 | } | ||
198 | } | ||
199 | |||
200 | int amdgpu_bo_create(struct amdgpu_device *adev, | ||
201 | unsigned long size, int byte_align, bool kernel, u32 domain, u64 flags, | ||
202 | struct sg_table *sg, struct amdgpu_bo **bo_ptr) | ||
203 | { | ||
204 | struct amdgpu_bo *bo; | ||
205 | enum ttm_bo_type type; | ||
206 | unsigned long page_align; | ||
207 | size_t acc_size; | ||
208 | int r; | ||
209 | |||
210 | /* VI has a hw bug where VM PTEs have to be allocated in groups of 8. | ||
211 | * do this as a temporary workaround | ||
212 | */ | ||
213 | if (!(domain & (AMDGPU_GEM_DOMAIN_GDS | AMDGPU_GEM_DOMAIN_GWS | AMDGPU_GEM_DOMAIN_OA))) { | ||
214 | if (adev->asic_type >= CHIP_TOPAZ) { | ||
215 | if (byte_align & 0x7fff) | ||
216 | byte_align = ALIGN(byte_align, 0x8000); | ||
217 | if (size & 0x7fff) | ||
218 | size = ALIGN(size, 0x8000); | ||
219 | } | ||
220 | } | ||
221 | |||
222 | page_align = roundup(byte_align, PAGE_SIZE) >> PAGE_SHIFT; | ||
223 | size = ALIGN(size, PAGE_SIZE); | ||
224 | |||
225 | if (kernel) { | ||
226 | type = ttm_bo_type_kernel; | ||
227 | } else if (sg) { | ||
228 | type = ttm_bo_type_sg; | ||
229 | } else { | ||
230 | type = ttm_bo_type_device; | ||
231 | } | ||
232 | *bo_ptr = NULL; | ||
233 | |||
234 | acc_size = ttm_bo_dma_acc_size(&adev->mman.bdev, size, | ||
235 | sizeof(struct amdgpu_bo)); | ||
236 | |||
237 | bo = kzalloc(sizeof(struct amdgpu_bo), GFP_KERNEL); | ||
238 | if (bo == NULL) | ||
239 | return -ENOMEM; | ||
240 | r = drm_gem_object_init(adev->ddev, &bo->gem_base, size); | ||
241 | if (unlikely(r)) { | ||
242 | kfree(bo); | ||
243 | return r; | ||
244 | } | ||
245 | bo->adev = adev; | ||
246 | INIT_LIST_HEAD(&bo->list); | ||
247 | INIT_LIST_HEAD(&bo->va); | ||
248 | bo->initial_domain = domain & (AMDGPU_GEM_DOMAIN_VRAM | | ||
249 | AMDGPU_GEM_DOMAIN_GTT | | ||
250 | AMDGPU_GEM_DOMAIN_CPU | | ||
251 | AMDGPU_GEM_DOMAIN_GDS | | ||
252 | AMDGPU_GEM_DOMAIN_GWS | | ||
253 | AMDGPU_GEM_DOMAIN_OA); | ||
254 | |||
255 | bo->flags = flags; | ||
256 | amdgpu_ttm_placement_from_domain(bo, domain); | ||
257 | /* Kernel allocation are uninterruptible */ | ||
258 | down_read(&adev->pm.mclk_lock); | ||
259 | r = ttm_bo_init(&adev->mman.bdev, &bo->tbo, size, type, | ||
260 | &bo->placement, page_align, !kernel, NULL, | ||
261 | acc_size, sg, NULL, &amdgpu_ttm_bo_destroy); | ||
262 | up_read(&adev->pm.mclk_lock); | ||
263 | if (unlikely(r != 0)) { | ||
264 | return r; | ||
265 | } | ||
266 | *bo_ptr = bo; | ||
267 | |||
268 | trace_amdgpu_bo_create(bo); | ||
269 | |||
270 | return 0; | ||
271 | } | ||
272 | |||
273 | int amdgpu_bo_kmap(struct amdgpu_bo *bo, void **ptr) | ||
274 | { | ||
275 | bool is_iomem; | ||
276 | int r; | ||
277 | |||
278 | if (bo->kptr) { | ||
279 | if (ptr) { | ||
280 | *ptr = bo->kptr; | ||
281 | } | ||
282 | return 0; | ||
283 | } | ||
284 | r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); | ||
285 | if (r) { | ||
286 | return r; | ||
287 | } | ||
288 | bo->kptr = ttm_kmap_obj_virtual(&bo->kmap, &is_iomem); | ||
289 | if (ptr) { | ||
290 | *ptr = bo->kptr; | ||
291 | } | ||
292 | return 0; | ||
293 | } | ||
294 | |||
295 | void amdgpu_bo_kunmap(struct amdgpu_bo *bo) | ||
296 | { | ||
297 | if (bo->kptr == NULL) | ||
298 | return; | ||
299 | bo->kptr = NULL; | ||
300 | ttm_bo_kunmap(&bo->kmap); | ||
301 | } | ||
302 | |||
303 | struct amdgpu_bo *amdgpu_bo_ref(struct amdgpu_bo *bo) | ||
304 | { | ||
305 | if (bo == NULL) | ||
306 | return NULL; | ||
307 | |||
308 | ttm_bo_reference(&bo->tbo); | ||
309 | return bo; | ||
310 | } | ||
311 | |||
312 | void amdgpu_bo_unref(struct amdgpu_bo **bo) | ||
313 | { | ||
314 | struct ttm_buffer_object *tbo; | ||
315 | |||
316 | if ((*bo) == NULL) | ||
317 | return; | ||
318 | |||
319 | tbo = &((*bo)->tbo); | ||
320 | ttm_bo_unref(&tbo); | ||
321 | if (tbo == NULL) | ||
322 | *bo = NULL; | ||
323 | } | ||
324 | |||
325 | int amdgpu_bo_pin_restricted(struct amdgpu_bo *bo, u32 domain, u64 max_offset, | ||
326 | u64 *gpu_addr) | ||
327 | { | ||
328 | int r, i; | ||
329 | |||
330 | if (amdgpu_ttm_tt_has_userptr(bo->tbo.ttm)) | ||
331 | return -EPERM; | ||
332 | |||
333 | if (bo->pin_count) { | ||
334 | bo->pin_count++; | ||
335 | if (gpu_addr) | ||
336 | *gpu_addr = amdgpu_bo_gpu_offset(bo); | ||
337 | |||
338 | if (max_offset != 0) { | ||
339 | u64 domain_start; | ||
340 | |||
341 | if (domain == AMDGPU_GEM_DOMAIN_VRAM) | ||
342 | domain_start = bo->adev->mc.vram_start; | ||
343 | else | ||
344 | domain_start = bo->adev->mc.gtt_start; | ||
345 | WARN_ON_ONCE(max_offset < | ||
346 | (amdgpu_bo_gpu_offset(bo) - domain_start)); | ||
347 | } | ||
348 | |||
349 | return 0; | ||
350 | } | ||
351 | amdgpu_ttm_placement_from_domain(bo, domain); | ||
352 | for (i = 0; i < bo->placement.num_placement; i++) { | ||
353 | /* force to pin into visible video ram */ | ||
354 | if ((bo->placements[i].flags & TTM_PL_FLAG_VRAM) && | ||
355 | !(bo->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS) && | ||
356 | (!max_offset || max_offset > bo->adev->mc.visible_vram_size)) | ||
357 | bo->placements[i].lpfn = | ||
358 | bo->adev->mc.visible_vram_size >> PAGE_SHIFT; | ||
359 | else | ||
360 | bo->placements[i].lpfn = max_offset >> PAGE_SHIFT; | ||
361 | |||
362 | bo->placements[i].flags |= TTM_PL_FLAG_NO_EVICT; | ||
363 | } | ||
364 | |||
365 | r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); | ||
366 | if (likely(r == 0)) { | ||
367 | bo->pin_count = 1; | ||
368 | if (gpu_addr != NULL) | ||
369 | *gpu_addr = amdgpu_bo_gpu_offset(bo); | ||
370 | if (domain == AMDGPU_GEM_DOMAIN_VRAM) | ||
371 | bo->adev->vram_pin_size += amdgpu_bo_size(bo); | ||
372 | else | ||
373 | bo->adev->gart_pin_size += amdgpu_bo_size(bo); | ||
374 | } else { | ||
375 | dev_err(bo->adev->dev, "%p pin failed\n", bo); | ||
376 | } | ||
377 | return r; | ||
378 | } | ||
379 | |||
380 | int amdgpu_bo_pin(struct amdgpu_bo *bo, u32 domain, u64 *gpu_addr) | ||
381 | { | ||
382 | return amdgpu_bo_pin_restricted(bo, domain, 0, gpu_addr); | ||
383 | } | ||
384 | |||
385 | int amdgpu_bo_unpin(struct amdgpu_bo *bo) | ||
386 | { | ||
387 | int r, i; | ||
388 | |||
389 | if (!bo->pin_count) { | ||
390 | dev_warn(bo->adev->dev, "%p unpin not necessary\n", bo); | ||
391 | return 0; | ||
392 | } | ||
393 | bo->pin_count--; | ||
394 | if (bo->pin_count) | ||
395 | return 0; | ||
396 | for (i = 0; i < bo->placement.num_placement; i++) { | ||
397 | bo->placements[i].lpfn = 0; | ||
398 | bo->placements[i].flags &= ~TTM_PL_FLAG_NO_EVICT; | ||
399 | } | ||
400 | r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); | ||
401 | if (likely(r == 0)) { | ||
402 | if (bo->tbo.mem.mem_type == TTM_PL_VRAM) | ||
403 | bo->adev->vram_pin_size -= amdgpu_bo_size(bo); | ||
404 | else | ||
405 | bo->adev->gart_pin_size -= amdgpu_bo_size(bo); | ||
406 | } else { | ||
407 | dev_err(bo->adev->dev, "%p validate failed for unpin\n", bo); | ||
408 | } | ||
409 | return r; | ||
410 | } | ||
411 | |||
412 | int amdgpu_bo_evict_vram(struct amdgpu_device *adev) | ||
413 | { | ||
414 | /* late 2.6.33 fix IGP hibernate - we need pm ops to do this correct */ | ||
415 | if (0 && (adev->flags & AMDGPU_IS_APU)) { | ||
416 | /* Useless to evict on IGP chips */ | ||
417 | return 0; | ||
418 | } | ||
419 | return ttm_bo_evict_mm(&adev->mman.bdev, TTM_PL_VRAM); | ||
420 | } | ||
421 | |||
422 | void amdgpu_bo_force_delete(struct amdgpu_device *adev) | ||
423 | { | ||
424 | struct amdgpu_bo *bo, *n; | ||
425 | |||
426 | if (list_empty(&adev->gem.objects)) { | ||
427 | return; | ||
428 | } | ||
429 | dev_err(adev->dev, "Userspace still has active objects !\n"); | ||
430 | list_for_each_entry_safe(bo, n, &adev->gem.objects, list) { | ||
431 | mutex_lock(&adev->ddev->struct_mutex); | ||
432 | dev_err(adev->dev, "%p %p %lu %lu force free\n", | ||
433 | &bo->gem_base, bo, (unsigned long)bo->gem_base.size, | ||
434 | *((unsigned long *)&bo->gem_base.refcount)); | ||
435 | mutex_lock(&bo->adev->gem.mutex); | ||
436 | list_del_init(&bo->list); | ||
437 | mutex_unlock(&bo->adev->gem.mutex); | ||
438 | /* this should unref the ttm bo */ | ||
439 | drm_gem_object_unreference(&bo->gem_base); | ||
440 | mutex_unlock(&adev->ddev->struct_mutex); | ||
441 | } | ||
442 | } | ||
443 | |||
444 | int amdgpu_bo_init(struct amdgpu_device *adev) | ||
445 | { | ||
446 | /* Add an MTRR for the VRAM */ | ||
447 | adev->mc.vram_mtrr = arch_phys_wc_add(adev->mc.aper_base, | ||
448 | adev->mc.aper_size); | ||
449 | DRM_INFO("Detected VRAM RAM=%lluM, BAR=%lluM\n", | ||
450 | adev->mc.mc_vram_size >> 20, | ||
451 | (unsigned long long)adev->mc.aper_size >> 20); | ||
452 | DRM_INFO("RAM width %dbits DDR\n", | ||
453 | adev->mc.vram_width); | ||
454 | return amdgpu_ttm_init(adev); | ||
455 | } | ||
456 | |||
457 | void amdgpu_bo_fini(struct amdgpu_device *adev) | ||
458 | { | ||
459 | amdgpu_ttm_fini(adev); | ||
460 | arch_phys_wc_del(adev->mc.vram_mtrr); | ||
461 | } | ||
462 | |||
463 | int amdgpu_bo_fbdev_mmap(struct amdgpu_bo *bo, | ||
464 | struct vm_area_struct *vma) | ||
465 | { | ||
466 | return ttm_fbdev_mmap(vma, &bo->tbo); | ||
467 | } | ||
468 | |||
469 | int amdgpu_bo_set_tiling_flags(struct amdgpu_bo *bo, u64 tiling_flags) | ||
470 | { | ||
471 | unsigned bankw, bankh, mtaspect, tilesplit, stilesplit; | ||
472 | |||
473 | bankw = (tiling_flags >> AMDGPU_TILING_EG_BANKW_SHIFT) & AMDGPU_TILING_EG_BANKW_MASK; | ||
474 | bankh = (tiling_flags >> AMDGPU_TILING_EG_BANKH_SHIFT) & AMDGPU_TILING_EG_BANKH_MASK; | ||
475 | mtaspect = (tiling_flags >> AMDGPU_TILING_EG_MACRO_TILE_ASPECT_SHIFT) & AMDGPU_TILING_EG_MACRO_TILE_ASPECT_MASK; | ||
476 | tilesplit = (tiling_flags >> AMDGPU_TILING_EG_TILE_SPLIT_SHIFT) & AMDGPU_TILING_EG_TILE_SPLIT_MASK; | ||
477 | stilesplit = (tiling_flags >> AMDGPU_TILING_EG_STENCIL_TILE_SPLIT_SHIFT) & AMDGPU_TILING_EG_STENCIL_TILE_SPLIT_MASK; | ||
478 | switch (bankw) { | ||
479 | case 0: | ||
480 | case 1: | ||
481 | case 2: | ||
482 | case 4: | ||
483 | case 8: | ||
484 | break; | ||
485 | default: | ||
486 | return -EINVAL; | ||
487 | } | ||
488 | switch (bankh) { | ||
489 | case 0: | ||
490 | case 1: | ||
491 | case 2: | ||
492 | case 4: | ||
493 | case 8: | ||
494 | break; | ||
495 | default: | ||
496 | return -EINVAL; | ||
497 | } | ||
498 | switch (mtaspect) { | ||
499 | case 0: | ||
500 | case 1: | ||
501 | case 2: | ||
502 | case 4: | ||
503 | case 8: | ||
504 | break; | ||
505 | default: | ||
506 | return -EINVAL; | ||
507 | } | ||
508 | if (tilesplit > 6) { | ||
509 | return -EINVAL; | ||
510 | } | ||
511 | if (stilesplit > 6) { | ||
512 | return -EINVAL; | ||
513 | } | ||
514 | |||
515 | bo->tiling_flags = tiling_flags; | ||
516 | return 0; | ||
517 | } | ||
518 | |||
519 | void amdgpu_bo_get_tiling_flags(struct amdgpu_bo *bo, u64 *tiling_flags) | ||
520 | { | ||
521 | lockdep_assert_held(&bo->tbo.resv->lock.base); | ||
522 | |||
523 | if (tiling_flags) | ||
524 | *tiling_flags = bo->tiling_flags; | ||
525 | } | ||
526 | |||
527 | int amdgpu_bo_set_metadata (struct amdgpu_bo *bo, void *metadata, | ||
528 | uint32_t metadata_size, uint64_t flags) | ||
529 | { | ||
530 | void *buffer; | ||
531 | |||
532 | if (!metadata_size) { | ||
533 | if (bo->metadata_size) { | ||
534 | kfree(bo->metadata); | ||
535 | bo->metadata_size = 0; | ||
536 | } | ||
537 | return 0; | ||
538 | } | ||
539 | |||
540 | if (metadata == NULL) | ||
541 | return -EINVAL; | ||
542 | |||
543 | buffer = kzalloc(metadata_size, GFP_KERNEL); | ||
544 | if (buffer == NULL) | ||
545 | return -ENOMEM; | ||
546 | |||
547 | memcpy(buffer, metadata, metadata_size); | ||
548 | |||
549 | kfree(bo->metadata); | ||
550 | bo->metadata_flags = flags; | ||
551 | bo->metadata = buffer; | ||
552 | bo->metadata_size = metadata_size; | ||
553 | |||
554 | return 0; | ||
555 | } | ||
556 | |||
557 | int amdgpu_bo_get_metadata(struct amdgpu_bo *bo, void *buffer, | ||
558 | size_t buffer_size, uint32_t *metadata_size, | ||
559 | uint64_t *flags) | ||
560 | { | ||
561 | if (!buffer && !metadata_size) | ||
562 | return -EINVAL; | ||
563 | |||
564 | if (buffer) { | ||
565 | if (buffer_size < bo->metadata_size) | ||
566 | return -EINVAL; | ||
567 | |||
568 | if (bo->metadata_size) | ||
569 | memcpy(buffer, bo->metadata, bo->metadata_size); | ||
570 | } | ||
571 | |||
572 | if (metadata_size) | ||
573 | *metadata_size = bo->metadata_size; | ||
574 | if (flags) | ||
575 | *flags = bo->metadata_flags; | ||
576 | |||
577 | return 0; | ||
578 | } | ||
579 | |||
580 | void amdgpu_bo_move_notify(struct ttm_buffer_object *bo, | ||
581 | struct ttm_mem_reg *new_mem) | ||
582 | { | ||
583 | struct amdgpu_bo *rbo; | ||
584 | |||
585 | if (!amdgpu_ttm_bo_is_amdgpu_bo(bo)) | ||
586 | return; | ||
587 | |||
588 | rbo = container_of(bo, struct amdgpu_bo, tbo); | ||
589 | amdgpu_vm_bo_invalidate(rbo->adev, rbo); | ||
590 | |||
591 | /* update statistics */ | ||
592 | if (!new_mem) | ||
593 | return; | ||
594 | |||
595 | /* move_notify is called before move happens */ | ||
596 | amdgpu_update_memory_usage(rbo->adev, &bo->mem, new_mem); | ||
597 | } | ||
598 | |||
599 | int amdgpu_bo_fault_reserve_notify(struct ttm_buffer_object *bo) | ||
600 | { | ||
601 | struct amdgpu_device *adev; | ||
602 | struct amdgpu_bo *rbo; | ||
603 | unsigned long offset, size; | ||
604 | int r; | ||
605 | |||
606 | if (!amdgpu_ttm_bo_is_amdgpu_bo(bo)) | ||
607 | return 0; | ||
608 | rbo = container_of(bo, struct amdgpu_bo, tbo); | ||
609 | adev = rbo->adev; | ||
610 | if (bo->mem.mem_type == TTM_PL_VRAM) { | ||
611 | size = bo->mem.num_pages << PAGE_SHIFT; | ||
612 | offset = bo->mem.start << PAGE_SHIFT; | ||
613 | if ((offset + size) > adev->mc.visible_vram_size) { | ||
614 | /* hurrah the memory is not visible ! */ | ||
615 | amdgpu_ttm_placement_from_domain(rbo, AMDGPU_GEM_DOMAIN_VRAM); | ||
616 | rbo->placements[0].lpfn = adev->mc.visible_vram_size >> PAGE_SHIFT; | ||
617 | r = ttm_bo_validate(bo, &rbo->placement, false, false); | ||
618 | if (unlikely(r != 0)) | ||
619 | return r; | ||
620 | offset = bo->mem.start << PAGE_SHIFT; | ||
621 | /* this should not happen */ | ||
622 | if ((offset + size) > adev->mc.visible_vram_size) | ||
623 | return -EINVAL; | ||
624 | } | ||
625 | } | ||
626 | return 0; | ||
627 | } | ||
628 | |||
629 | /** | ||
630 | * amdgpu_bo_fence - add fence to buffer object | ||
631 | * | ||
632 | * @bo: buffer object in question | ||
633 | * @fence: fence to add | ||
634 | * @shared: true if fence should be added shared | ||
635 | * | ||
636 | */ | ||
637 | void amdgpu_bo_fence(struct amdgpu_bo *bo, struct amdgpu_fence *fence, | ||
638 | bool shared) | ||
639 | { | ||
640 | struct reservation_object *resv = bo->tbo.resv; | ||
641 | |||
642 | if (shared) | ||
643 | reservation_object_add_shared_fence(resv, &fence->base); | ||
644 | else | ||
645 | reservation_object_add_excl_fence(resv, &fence->base); | ||
646 | } | ||