diff options
Diffstat (limited to 'drivers/gpu/drm/ttm/ttm_execbuf_util.c')
-rw-r--r-- | drivers/gpu/drm/ttm/ttm_execbuf_util.c | 169 |
1 files changed, 145 insertions, 24 deletions
diff --git a/drivers/gpu/drm/ttm/ttm_execbuf_util.c b/drivers/gpu/drm/ttm/ttm_execbuf_util.c index c285c2902d15..3832fe10b4df 100644 --- a/drivers/gpu/drm/ttm/ttm_execbuf_util.c +++ b/drivers/gpu/drm/ttm/ttm_execbuf_util.c | |||
@@ -32,7 +32,7 @@ | |||
32 | #include <linux/sched.h> | 32 | #include <linux/sched.h> |
33 | #include <linux/module.h> | 33 | #include <linux/module.h> |
34 | 34 | ||
35 | void ttm_eu_backoff_reservation(struct list_head *list) | 35 | static void ttm_eu_backoff_reservation_locked(struct list_head *list) |
36 | { | 36 | { |
37 | struct ttm_validate_buffer *entry; | 37 | struct ttm_validate_buffer *entry; |
38 | 38 | ||
@@ -41,10 +41,77 @@ void ttm_eu_backoff_reservation(struct list_head *list) | |||
41 | if (!entry->reserved) | 41 | if (!entry->reserved) |
42 | continue; | 42 | continue; |
43 | 43 | ||
44 | if (entry->removed) { | ||
45 | ttm_bo_add_to_lru(bo); | ||
46 | entry->removed = false; | ||
47 | |||
48 | } | ||
44 | entry->reserved = false; | 49 | entry->reserved = false; |
45 | ttm_bo_unreserve(bo); | 50 | atomic_set(&bo->reserved, 0); |
51 | wake_up_all(&bo->event_queue); | ||
52 | } | ||
53 | } | ||
54 | |||
55 | static void ttm_eu_del_from_lru_locked(struct list_head *list) | ||
56 | { | ||
57 | struct ttm_validate_buffer *entry; | ||
58 | |||
59 | list_for_each_entry(entry, list, head) { | ||
60 | struct ttm_buffer_object *bo = entry->bo; | ||
61 | if (!entry->reserved) | ||
62 | continue; | ||
63 | |||
64 | if (!entry->removed) { | ||
65 | entry->put_count = ttm_bo_del_from_lru(bo); | ||
66 | entry->removed = true; | ||
67 | } | ||
46 | } | 68 | } |
47 | } | 69 | } |
70 | |||
71 | static void ttm_eu_list_ref_sub(struct list_head *list) | ||
72 | { | ||
73 | struct ttm_validate_buffer *entry; | ||
74 | |||
75 | list_for_each_entry(entry, list, head) { | ||
76 | struct ttm_buffer_object *bo = entry->bo; | ||
77 | |||
78 | if (entry->put_count) { | ||
79 | ttm_bo_list_ref_sub(bo, entry->put_count, true); | ||
80 | entry->put_count = 0; | ||
81 | } | ||
82 | } | ||
83 | } | ||
84 | |||
85 | static int ttm_eu_wait_unreserved_locked(struct list_head *list, | ||
86 | struct ttm_buffer_object *bo) | ||
87 | { | ||
88 | struct ttm_bo_global *glob = bo->glob; | ||
89 | int ret; | ||
90 | |||
91 | ttm_eu_del_from_lru_locked(list); | ||
92 | spin_unlock(&glob->lru_lock); | ||
93 | ret = ttm_bo_wait_unreserved(bo, true); | ||
94 | spin_lock(&glob->lru_lock); | ||
95 | if (unlikely(ret != 0)) | ||
96 | ttm_eu_backoff_reservation_locked(list); | ||
97 | return ret; | ||
98 | } | ||
99 | |||
100 | |||
101 | void ttm_eu_backoff_reservation(struct list_head *list) | ||
102 | { | ||
103 | struct ttm_validate_buffer *entry; | ||
104 | struct ttm_bo_global *glob; | ||
105 | |||
106 | if (list_empty(list)) | ||
107 | return; | ||
108 | |||
109 | entry = list_first_entry(list, struct ttm_validate_buffer, head); | ||
110 | glob = entry->bo->glob; | ||
111 | spin_lock(&glob->lru_lock); | ||
112 | ttm_eu_backoff_reservation_locked(list); | ||
113 | spin_unlock(&glob->lru_lock); | ||
114 | } | ||
48 | EXPORT_SYMBOL(ttm_eu_backoff_reservation); | 115 | EXPORT_SYMBOL(ttm_eu_backoff_reservation); |
49 | 116 | ||
50 | /* | 117 | /* |
@@ -59,37 +126,76 @@ EXPORT_SYMBOL(ttm_eu_backoff_reservation); | |||
59 | * buffers in different orders. | 126 | * buffers in different orders. |
60 | */ | 127 | */ |
61 | 128 | ||
62 | int ttm_eu_reserve_buffers(struct list_head *list, uint32_t val_seq) | 129 | int ttm_eu_reserve_buffers(struct list_head *list) |
63 | { | 130 | { |
131 | struct ttm_bo_global *glob; | ||
64 | struct ttm_validate_buffer *entry; | 132 | struct ttm_validate_buffer *entry; |
65 | int ret; | 133 | int ret; |
134 | uint32_t val_seq; | ||
135 | |||
136 | if (list_empty(list)) | ||
137 | return 0; | ||
138 | |||
139 | list_for_each_entry(entry, list, head) { | ||
140 | entry->reserved = false; | ||
141 | entry->put_count = 0; | ||
142 | entry->removed = false; | ||
143 | } | ||
144 | |||
145 | entry = list_first_entry(list, struct ttm_validate_buffer, head); | ||
146 | glob = entry->bo->glob; | ||
66 | 147 | ||
67 | retry: | 148 | retry: |
149 | spin_lock(&glob->lru_lock); | ||
150 | val_seq = entry->bo->bdev->val_seq++; | ||
151 | |||
68 | list_for_each_entry(entry, list, head) { | 152 | list_for_each_entry(entry, list, head) { |
69 | struct ttm_buffer_object *bo = entry->bo; | 153 | struct ttm_buffer_object *bo = entry->bo; |
70 | 154 | ||
71 | entry->reserved = false; | 155 | retry_this_bo: |
72 | ret = ttm_bo_reserve(bo, true, false, true, val_seq); | 156 | ret = ttm_bo_reserve_locked(bo, true, true, true, val_seq); |
73 | if (ret != 0) { | 157 | switch (ret) { |
74 | ttm_eu_backoff_reservation(list); | 158 | case 0: |
75 | if (ret == -EAGAIN) { | 159 | break; |
76 | ret = ttm_bo_wait_unreserved(bo, true); | 160 | case -EBUSY: |
77 | if (unlikely(ret != 0)) | 161 | ret = ttm_eu_wait_unreserved_locked(list, bo); |
78 | return ret; | 162 | if (unlikely(ret != 0)) { |
79 | goto retry; | 163 | spin_unlock(&glob->lru_lock); |
80 | } else | 164 | ttm_eu_list_ref_sub(list); |
81 | return ret; | 165 | return ret; |
166 | } | ||
167 | goto retry_this_bo; | ||
168 | case -EAGAIN: | ||
169 | ttm_eu_backoff_reservation_locked(list); | ||
170 | spin_unlock(&glob->lru_lock); | ||
171 | ttm_eu_list_ref_sub(list); | ||
172 | ret = ttm_bo_wait_unreserved(bo, true); | ||
173 | if (unlikely(ret != 0)) | ||
174 | return ret; | ||
175 | goto retry; | ||
176 | default: | ||
177 | ttm_eu_backoff_reservation_locked(list); | ||
178 | spin_unlock(&glob->lru_lock); | ||
179 | ttm_eu_list_ref_sub(list); | ||
180 | return ret; | ||
82 | } | 181 | } |
83 | 182 | ||
84 | entry->reserved = true; | 183 | entry->reserved = true; |
85 | if (unlikely(atomic_read(&bo->cpu_writers) > 0)) { | 184 | if (unlikely(atomic_read(&bo->cpu_writers) > 0)) { |
86 | ttm_eu_backoff_reservation(list); | 185 | ttm_eu_backoff_reservation_locked(list); |
186 | spin_unlock(&glob->lru_lock); | ||
187 | ttm_eu_list_ref_sub(list); | ||
87 | ret = ttm_bo_wait_cpu(bo, false); | 188 | ret = ttm_bo_wait_cpu(bo, false); |
88 | if (ret) | 189 | if (ret) |
89 | return ret; | 190 | return ret; |
90 | goto retry; | 191 | goto retry; |
91 | } | 192 | } |
92 | } | 193 | } |
194 | |||
195 | ttm_eu_del_from_lru_locked(list); | ||
196 | spin_unlock(&glob->lru_lock); | ||
197 | ttm_eu_list_ref_sub(list); | ||
198 | |||
93 | return 0; | 199 | return 0; |
94 | } | 200 | } |
95 | EXPORT_SYMBOL(ttm_eu_reserve_buffers); | 201 | EXPORT_SYMBOL(ttm_eu_reserve_buffers); |
@@ -97,21 +203,36 @@ EXPORT_SYMBOL(ttm_eu_reserve_buffers); | |||
97 | void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj) | 203 | void ttm_eu_fence_buffer_objects(struct list_head *list, void *sync_obj) |
98 | { | 204 | { |
99 | struct ttm_validate_buffer *entry; | 205 | struct ttm_validate_buffer *entry; |
206 | struct ttm_buffer_object *bo; | ||
207 | struct ttm_bo_global *glob; | ||
208 | struct ttm_bo_device *bdev; | ||
209 | struct ttm_bo_driver *driver; | ||
100 | 210 | ||
101 | list_for_each_entry(entry, list, head) { | 211 | if (list_empty(list)) |
102 | struct ttm_buffer_object *bo = entry->bo; | 212 | return; |
103 | struct ttm_bo_driver *driver = bo->bdev->driver; | 213 | |
104 | void *old_sync_obj; | 214 | bo = list_first_entry(list, struct ttm_validate_buffer, head)->bo; |
215 | bdev = bo->bdev; | ||
216 | driver = bdev->driver; | ||
217 | glob = bo->glob; | ||
105 | 218 | ||
106 | spin_lock(&bo->lock); | 219 | spin_lock(&bdev->fence_lock); |
107 | old_sync_obj = bo->sync_obj; | 220 | spin_lock(&glob->lru_lock); |
221 | |||
222 | list_for_each_entry(entry, list, head) { | ||
223 | bo = entry->bo; | ||
224 | entry->old_sync_obj = bo->sync_obj; | ||
108 | bo->sync_obj = driver->sync_obj_ref(sync_obj); | 225 | bo->sync_obj = driver->sync_obj_ref(sync_obj); |
109 | bo->sync_obj_arg = entry->new_sync_obj_arg; | 226 | bo->sync_obj_arg = entry->new_sync_obj_arg; |
110 | spin_unlock(&bo->lock); | 227 | ttm_bo_unreserve_locked(bo); |
111 | ttm_bo_unreserve(bo); | ||
112 | entry->reserved = false; | 228 | entry->reserved = false; |
113 | if (old_sync_obj) | 229 | } |
114 | driver->sync_obj_unref(&old_sync_obj); | 230 | spin_unlock(&glob->lru_lock); |
231 | spin_unlock(&bdev->fence_lock); | ||
232 | |||
233 | list_for_each_entry(entry, list, head) { | ||
234 | if (entry->old_sync_obj) | ||
235 | driver->sync_obj_unref(&entry->old_sync_obj); | ||
115 | } | 236 | } |
116 | } | 237 | } |
117 | EXPORT_SYMBOL(ttm_eu_fence_buffer_objects); | 238 | EXPORT_SYMBOL(ttm_eu_fence_buffer_objects); |