diff options
author | Dave Airlie <airlied@redhat.com> | 2018-11-18 20:07:52 -0500 |
---|---|---|
committer | Dave Airlie <airlied@redhat.com> | 2018-11-18 20:07:52 -0500 |
commit | 9235dd441af43599b9cdcce599a3da4083fcad3c (patch) | |
tree | 5f8a79cc2d378f05e807c6c5e388394b8e86319d /drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c | |
parent | d7563c55ef9fc1fd2301b8708b3c1f53509d6745 (diff) | |
parent | 36b486bc682114a2f1001cbf1a87f21ae381bfc1 (diff) |
Merge branch 'drm-next-4.21' of git://people.freedesktop.org/~agd5f/linux into drm-next
New features for 4.21:
amdgpu:
- Support for SDMA paging queue on vega
- Put compute EOP buffers into vram for better performance
- Share more code with amdkfd
- Support for scanout with DCC on gfx9
- Initial kerneldoc for DC
- Updated SMU firmware support for gfx8 chips
- Rework CSA handling for eventual support for preemption
- XGMI PSP support
- Clean up RLC handling
- Enable GPU reset by default on VI, SOC15 dGPUs
- Ring and IB test cleanups
amdkfd:
- Share more code with amdgpu
ttm:
- Move global init out of the drivers
scheduler:
- Track if schedulers are ready for work
- Timeout/fault handling changes to facilitate GPU recovery
Signed-off-by: Dave Airlie <airlied@redhat.com>
From: Alex Deucher <alexdeucher@gmail.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181114165113.3751-1-alexander.deucher@amd.com
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c')
-rw-r--r-- | drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c | 140 |
1 files changed, 42 insertions, 98 deletions
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c index f2f358aa0597..cfee74732edb 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c | |||
@@ -23,16 +23,6 @@ | |||
23 | 23 | ||
24 | #include "amdgpu.h" | 24 | #include "amdgpu.h" |
25 | 25 | ||
26 | uint64_t amdgpu_csa_vaddr(struct amdgpu_device *adev) | ||
27 | { | ||
28 | uint64_t addr = adev->vm_manager.max_pfn << AMDGPU_GPU_PAGE_SHIFT; | ||
29 | |||
30 | addr -= AMDGPU_VA_RESERVED_SIZE; | ||
31 | addr = amdgpu_gmc_sign_extend(addr); | ||
32 | |||
33 | return addr; | ||
34 | } | ||
35 | |||
36 | bool amdgpu_virt_mmio_blocked(struct amdgpu_device *adev) | 26 | bool amdgpu_virt_mmio_blocked(struct amdgpu_device *adev) |
37 | { | 27 | { |
38 | /* By now all MMIO pages except mailbox are blocked */ | 28 | /* By now all MMIO pages except mailbox are blocked */ |
@@ -41,88 +31,6 @@ bool amdgpu_virt_mmio_blocked(struct amdgpu_device *adev) | |||
41 | return RREG32_NO_KIQ(0xc040) == 0xffffffff; | 31 | return RREG32_NO_KIQ(0xc040) == 0xffffffff; |
42 | } | 32 | } |
43 | 33 | ||
44 | int amdgpu_allocate_static_csa(struct amdgpu_device *adev) | ||
45 | { | ||
46 | int r; | ||
47 | void *ptr; | ||
48 | |||
49 | r = amdgpu_bo_create_kernel(adev, AMDGPU_CSA_SIZE, PAGE_SIZE, | ||
50 | AMDGPU_GEM_DOMAIN_VRAM, &adev->virt.csa_obj, | ||
51 | &adev->virt.csa_vmid0_addr, &ptr); | ||
52 | if (r) | ||
53 | return r; | ||
54 | |||
55 | memset(ptr, 0, AMDGPU_CSA_SIZE); | ||
56 | return 0; | ||
57 | } | ||
58 | |||
59 | void amdgpu_free_static_csa(struct amdgpu_device *adev) { | ||
60 | amdgpu_bo_free_kernel(&adev->virt.csa_obj, | ||
61 | &adev->virt.csa_vmid0_addr, | ||
62 | NULL); | ||
63 | } | ||
64 | |||
65 | /* | ||
66 | * amdgpu_map_static_csa should be called during amdgpu_vm_init | ||
67 | * it maps virtual address amdgpu_csa_vaddr() to this VM, and each command | ||
68 | * submission of GFX should use this virtual address within META_DATA init | ||
69 | * package to support SRIOV gfx preemption. | ||
70 | */ | ||
71 | int amdgpu_map_static_csa(struct amdgpu_device *adev, struct amdgpu_vm *vm, | ||
72 | struct amdgpu_bo_va **bo_va) | ||
73 | { | ||
74 | uint64_t csa_addr = amdgpu_csa_vaddr(adev) & AMDGPU_GMC_HOLE_MASK; | ||
75 | struct ww_acquire_ctx ticket; | ||
76 | struct list_head list; | ||
77 | struct amdgpu_bo_list_entry pd; | ||
78 | struct ttm_validate_buffer csa_tv; | ||
79 | int r; | ||
80 | |||
81 | INIT_LIST_HEAD(&list); | ||
82 | INIT_LIST_HEAD(&csa_tv.head); | ||
83 | csa_tv.bo = &adev->virt.csa_obj->tbo; | ||
84 | csa_tv.shared = true; | ||
85 | |||
86 | list_add(&csa_tv.head, &list); | ||
87 | amdgpu_vm_get_pd_bo(vm, &list, &pd); | ||
88 | |||
89 | r = ttm_eu_reserve_buffers(&ticket, &list, true, NULL); | ||
90 | if (r) { | ||
91 | DRM_ERROR("failed to reserve CSA,PD BOs: err=%d\n", r); | ||
92 | return r; | ||
93 | } | ||
94 | |||
95 | *bo_va = amdgpu_vm_bo_add(adev, vm, adev->virt.csa_obj); | ||
96 | if (!*bo_va) { | ||
97 | ttm_eu_backoff_reservation(&ticket, &list); | ||
98 | DRM_ERROR("failed to create bo_va for static CSA\n"); | ||
99 | return -ENOMEM; | ||
100 | } | ||
101 | |||
102 | r = amdgpu_vm_alloc_pts(adev, (*bo_va)->base.vm, csa_addr, | ||
103 | AMDGPU_CSA_SIZE); | ||
104 | if (r) { | ||
105 | DRM_ERROR("failed to allocate pts for static CSA, err=%d\n", r); | ||
106 | amdgpu_vm_bo_rmv(adev, *bo_va); | ||
107 | ttm_eu_backoff_reservation(&ticket, &list); | ||
108 | return r; | ||
109 | } | ||
110 | |||
111 | r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, AMDGPU_CSA_SIZE, | ||
112 | AMDGPU_PTE_READABLE | AMDGPU_PTE_WRITEABLE | | ||
113 | AMDGPU_PTE_EXECUTABLE); | ||
114 | |||
115 | if (r) { | ||
116 | DRM_ERROR("failed to do bo_map on static CSA, err=%d\n", r); | ||
117 | amdgpu_vm_bo_rmv(adev, *bo_va); | ||
118 | ttm_eu_backoff_reservation(&ticket, &list); | ||
119 | return r; | ||
120 | } | ||
121 | |||
122 | ttm_eu_backoff_reservation(&ticket, &list); | ||
123 | return 0; | ||
124 | } | ||
125 | |||
126 | void amdgpu_virt_init_setting(struct amdgpu_device *adev) | 34 | void amdgpu_virt_init_setting(struct amdgpu_device *adev) |
127 | { | 35 | { |
128 | /* enable virtual display */ | 36 | /* enable virtual display */ |
@@ -162,9 +70,7 @@ uint32_t amdgpu_virt_kiq_rreg(struct amdgpu_device *adev, uint32_t reg) | |||
162 | if (r < 1 && (adev->in_gpu_reset || in_interrupt())) | 70 | if (r < 1 && (adev->in_gpu_reset || in_interrupt())) |
163 | goto failed_kiq_read; | 71 | goto failed_kiq_read; |
164 | 72 | ||
165 | if (in_interrupt()) | 73 | might_sleep(); |
166 | might_sleep(); | ||
167 | |||
168 | while (r < 1 && cnt++ < MAX_KIQ_REG_TRY) { | 74 | while (r < 1 && cnt++ < MAX_KIQ_REG_TRY) { |
169 | msleep(MAX_KIQ_REG_BAILOUT_INTERVAL); | 75 | msleep(MAX_KIQ_REG_BAILOUT_INTERVAL); |
170 | r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); | 76 | r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); |
@@ -210,9 +116,7 @@ void amdgpu_virt_kiq_wreg(struct amdgpu_device *adev, uint32_t reg, uint32_t v) | |||
210 | if (r < 1 && (adev->in_gpu_reset || in_interrupt())) | 116 | if (r < 1 && (adev->in_gpu_reset || in_interrupt())) |
211 | goto failed_kiq_write; | 117 | goto failed_kiq_write; |
212 | 118 | ||
213 | if (in_interrupt()) | 119 | might_sleep(); |
214 | might_sleep(); | ||
215 | |||
216 | while (r < 1 && cnt++ < MAX_KIQ_REG_TRY) { | 120 | while (r < 1 && cnt++ < MAX_KIQ_REG_TRY) { |
217 | 121 | ||
218 | msleep(MAX_KIQ_REG_BAILOUT_INTERVAL); | 122 | msleep(MAX_KIQ_REG_BAILOUT_INTERVAL); |
@@ -228,6 +132,46 @@ failed_kiq_write: | |||
228 | pr_err("failed to write reg:%x\n", reg); | 132 | pr_err("failed to write reg:%x\n", reg); |
229 | } | 133 | } |
230 | 134 | ||
135 | void amdgpu_virt_kiq_reg_write_reg_wait(struct amdgpu_device *adev, | ||
136 | uint32_t reg0, uint32_t reg1, | ||
137 | uint32_t ref, uint32_t mask) | ||
138 | { | ||
139 | struct amdgpu_kiq *kiq = &adev->gfx.kiq; | ||
140 | struct amdgpu_ring *ring = &kiq->ring; | ||
141 | signed long r, cnt = 0; | ||
142 | unsigned long flags; | ||
143 | uint32_t seq; | ||
144 | |||
145 | spin_lock_irqsave(&kiq->ring_lock, flags); | ||
146 | amdgpu_ring_alloc(ring, 32); | ||
147 | amdgpu_ring_emit_reg_write_reg_wait(ring, reg0, reg1, | ||
148 | ref, mask); | ||
149 | amdgpu_fence_emit_polling(ring, &seq); | ||
150 | amdgpu_ring_commit(ring); | ||
151 | spin_unlock_irqrestore(&kiq->ring_lock, flags); | ||
152 | |||
153 | r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); | ||
154 | |||
155 | /* don't wait anymore for IRQ context */ | ||
156 | if (r < 1 && in_interrupt()) | ||
157 | goto failed_kiq; | ||
158 | |||
159 | might_sleep(); | ||
160 | while (r < 1 && cnt++ < MAX_KIQ_REG_TRY) { | ||
161 | |||
162 | msleep(MAX_KIQ_REG_BAILOUT_INTERVAL); | ||
163 | r = amdgpu_fence_wait_polling(ring, seq, MAX_KIQ_REG_WAIT); | ||
164 | } | ||
165 | |||
166 | if (cnt > MAX_KIQ_REG_TRY) | ||
167 | goto failed_kiq; | ||
168 | |||
169 | return; | ||
170 | |||
171 | failed_kiq: | ||
172 | pr_err("failed to write reg %x wait reg %x\n", reg0, reg1); | ||
173 | } | ||
174 | |||
231 | /** | 175 | /** |
232 | * amdgpu_virt_request_full_gpu() - request full gpu access | 176 | * amdgpu_virt_request_full_gpu() - request full gpu access |
233 | * @amdgpu: amdgpu device. | 177 | * @amdgpu: amdgpu device. |