diff options
author | Keith Packard <keithp@keithp.com> | 2008-10-30 22:38:48 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-10-31 05:12:40 -0400 |
commit | 0839ccb8ac6a9e2d5e175a4ae9c82b5c574d510d (patch) | |
tree | c231d0492e4c03ae8d33f2101f21ff2928e56710 /drivers | |
parent | 9663f2e6a6cf3f82b06d8fb699b11b80f92553ba (diff) |
i915: use io-mapping interfaces instead of a variety of mapping kludges
Impact: optimize/clean-up the IO mapping implementation of the i915 DRM driver
Switch the i915 device aperture mapping to the io-mapping interface, taking
advantage of the cleaner API to extend it across all of the mapping uses,
including both pwrite and relocation updates.
This dramatically improves performance on 64-bit kernels which were using
the same slow path as 32-bit non-HIGHMEM kernels prior to this patch.
Signed-off-by: Keith Packard <keithp@keithp.com>
Signed-off-by: Eric Anholt <eric@anholt.net>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/gpu/drm/i915/i915_drv.h | 3 | ||||
-rw-r--r-- | drivers/gpu/drm/i915/i915_gem.c | 174 |
2 files changed, 83 insertions, 94 deletions
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index f20ffe17df71..126b2f13258c 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
@@ -31,6 +31,7 @@ | |||
31 | #define _I915_DRV_H_ | 31 | #define _I915_DRV_H_ |
32 | 32 | ||
33 | #include "i915_reg.h" | 33 | #include "i915_reg.h" |
34 | #include <linux/io-mapping.h> | ||
34 | 35 | ||
35 | /* General customization: | 36 | /* General customization: |
36 | */ | 37 | */ |
@@ -246,6 +247,8 @@ typedef struct drm_i915_private { | |||
246 | struct { | 247 | struct { |
247 | struct drm_mm gtt_space; | 248 | struct drm_mm gtt_space; |
248 | 249 | ||
250 | struct io_mapping *gtt_mapping; | ||
251 | |||
249 | /** | 252 | /** |
250 | * List of objects currently involved in rendering from the | 253 | * List of objects currently involved in rendering from the |
251 | * ringbuffer. | 254 | * ringbuffer. |
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 17ae330ff269..61183b95b108 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
@@ -171,35 +171,50 @@ i915_gem_pread_ioctl(struct drm_device *dev, void *data, | |||
171 | return 0; | 171 | return 0; |
172 | } | 172 | } |
173 | 173 | ||
174 | /* | 174 | /* This is the fast write path which cannot handle |
175 | * Try to write quickly with an atomic kmap. Return true on success. | 175 | * page faults in the source data |
176 | * | ||
177 | * If this fails (which includes a partial write), we'll redo the whole | ||
178 | * thing with the slow version. | ||
179 | * | ||
180 | * This is a workaround for the low performance of iounmap (approximate | ||
181 | * 10% cpu cost on normal 3D workloads). kmap_atomic on HIGHMEM kernels | ||
182 | * happens to let us map card memory without taking IPIs. When the vmap | ||
183 | * rework lands we should be able to dump this hack. | ||
184 | */ | 176 | */ |
185 | static inline int fast_user_write(unsigned long pfn, char __user *user_data, | 177 | |
186 | int l, int o) | 178 | static inline int |
179 | fast_user_write(struct io_mapping *mapping, | ||
180 | loff_t page_base, int page_offset, | ||
181 | char __user *user_data, | ||
182 | int length) | ||
187 | { | 183 | { |
188 | #ifdef CONFIG_HIGHMEM | ||
189 | unsigned long unwritten; | ||
190 | char *vaddr_atomic; | 184 | char *vaddr_atomic; |
185 | unsigned long unwritten; | ||
191 | 186 | ||
192 | vaddr_atomic = kmap_atomic_pfn(pfn, KM_USER0); | 187 | vaddr_atomic = io_mapping_map_atomic_wc(mapping, page_base); |
193 | #if WATCH_PWRITE | 188 | unwritten = __copy_from_user_inatomic_nocache(vaddr_atomic + page_offset, |
194 | DRM_INFO("pwrite i %d o %d l %d pfn %ld vaddr %p\n", | 189 | user_data, length); |
195 | i, o, l, pfn, vaddr_atomic); | 190 | io_mapping_unmap_atomic(vaddr_atomic); |
196 | #endif | 191 | if (unwritten) |
197 | unwritten = __copy_from_user_inatomic_nocache(vaddr_atomic + o, user_data, l); | 192 | return -EFAULT; |
198 | kunmap_atomic(vaddr_atomic, KM_USER0); | 193 | return 0; |
199 | return !unwritten; | 194 | } |
200 | #else | 195 | |
196 | /* Here's the write path which can sleep for | ||
197 | * page faults | ||
198 | */ | ||
199 | |||
200 | static inline int | ||
201 | slow_user_write(struct io_mapping *mapping, | ||
202 | loff_t page_base, int page_offset, | ||
203 | char __user *user_data, | ||
204 | int length) | ||
205 | { | ||
206 | char __iomem *vaddr; | ||
207 | unsigned long unwritten; | ||
208 | |||
209 | vaddr = io_mapping_map_wc(mapping, page_base); | ||
210 | if (vaddr == NULL) | ||
211 | return -EFAULT; | ||
212 | unwritten = __copy_from_user(vaddr + page_offset, | ||
213 | user_data, length); | ||
214 | io_mapping_unmap(vaddr); | ||
215 | if (unwritten) | ||
216 | return -EFAULT; | ||
201 | return 0; | 217 | return 0; |
202 | #endif | ||
203 | } | 218 | } |
204 | 219 | ||
205 | static int | 220 | static int |
@@ -208,10 +223,12 @@ i915_gem_gtt_pwrite(struct drm_device *dev, struct drm_gem_object *obj, | |||
208 | struct drm_file *file_priv) | 223 | struct drm_file *file_priv) |
209 | { | 224 | { |
210 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 225 | struct drm_i915_gem_object *obj_priv = obj->driver_private; |
226 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
211 | ssize_t remain; | 227 | ssize_t remain; |
212 | loff_t offset; | 228 | loff_t offset, page_base; |
213 | char __user *user_data; | 229 | char __user *user_data; |
214 | int ret = 0; | 230 | int page_offset, page_length; |
231 | int ret; | ||
215 | 232 | ||
216 | user_data = (char __user *) (uintptr_t) args->data_ptr; | 233 | user_data = (char __user *) (uintptr_t) args->data_ptr; |
217 | remain = args->size; | 234 | remain = args->size; |
@@ -235,57 +252,37 @@ i915_gem_gtt_pwrite(struct drm_device *dev, struct drm_gem_object *obj, | |||
235 | obj_priv->dirty = 1; | 252 | obj_priv->dirty = 1; |
236 | 253 | ||
237 | while (remain > 0) { | 254 | while (remain > 0) { |
238 | unsigned long pfn; | ||
239 | int i, o, l; | ||
240 | |||
241 | /* Operation in this page | 255 | /* Operation in this page |
242 | * | 256 | * |
243 | * i = page number | 257 | * page_base = page offset within aperture |
244 | * o = offset within page | 258 | * page_offset = offset within page |
245 | * l = bytes to copy | 259 | * page_length = bytes to copy for this page |
246 | */ | 260 | */ |
247 | i = offset >> PAGE_SHIFT; | 261 | page_base = (offset & ~(PAGE_SIZE-1)); |
248 | o = offset & (PAGE_SIZE-1); | 262 | page_offset = offset & (PAGE_SIZE-1); |
249 | l = remain; | 263 | page_length = remain; |
250 | if ((o + l) > PAGE_SIZE) | 264 | if ((page_offset + remain) > PAGE_SIZE) |
251 | l = PAGE_SIZE - o; | 265 | page_length = PAGE_SIZE - page_offset; |
252 | 266 | ||
253 | pfn = (dev->agp->base >> PAGE_SHIFT) + i; | 267 | ret = fast_user_write (dev_priv->mm.gtt_mapping, page_base, |
254 | 268 | page_offset, user_data, page_length); | |
255 | if (!fast_user_write(pfn, user_data, l, o)) { | 269 | |
256 | unsigned long unwritten; | 270 | /* If we get a fault while copying data, then (presumably) our |
257 | char __iomem *vaddr; | 271 | * source page isn't available. In this case, use the |
258 | 272 | * non-atomic function | |
259 | vaddr = ioremap_wc(pfn << PAGE_SHIFT, PAGE_SIZE); | 273 | */ |
260 | #if WATCH_PWRITE | 274 | if (ret) { |
261 | DRM_INFO("pwrite slow i %d o %d l %d " | 275 | ret = slow_user_write (dev_priv->mm.gtt_mapping, |
262 | "pfn %ld vaddr %p\n", | 276 | page_base, page_offset, |
263 | i, o, l, pfn, vaddr); | 277 | user_data, page_length); |
264 | #endif | 278 | if (ret) |
265 | if (vaddr == NULL) { | ||
266 | ret = -EFAULT; | ||
267 | goto fail; | ||
268 | } | ||
269 | unwritten = __copy_from_user(vaddr + o, user_data, l); | ||
270 | #if WATCH_PWRITE | ||
271 | DRM_INFO("unwritten %ld\n", unwritten); | ||
272 | #endif | ||
273 | iounmap(vaddr); | ||
274 | if (unwritten) { | ||
275 | ret = -EFAULT; | ||
276 | goto fail; | 279 | goto fail; |
277 | } | ||
278 | } | 280 | } |
279 | 281 | ||
280 | remain -= l; | 282 | remain -= page_length; |
281 | user_data += l; | 283 | user_data += page_length; |
282 | offset += l; | 284 | offset += page_length; |
283 | } | 285 | } |
284 | #if WATCH_PWRITE && 1 | ||
285 | i915_gem_clflush_object(obj); | ||
286 | i915_gem_dump_object(obj, args->offset + args->size, __func__, ~0); | ||
287 | i915_gem_clflush_object(obj); | ||
288 | #endif | ||
289 | 286 | ||
290 | fail: | 287 | fail: |
291 | i915_gem_object_unpin(obj); | 288 | i915_gem_object_unpin(obj); |
@@ -1503,12 +1500,12 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
1503 | struct drm_i915_gem_exec_object *entry) | 1500 | struct drm_i915_gem_exec_object *entry) |
1504 | { | 1501 | { |
1505 | struct drm_device *dev = obj->dev; | 1502 | struct drm_device *dev = obj->dev; |
1503 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
1506 | struct drm_i915_gem_relocation_entry reloc; | 1504 | struct drm_i915_gem_relocation_entry reloc; |
1507 | struct drm_i915_gem_relocation_entry __user *relocs; | 1505 | struct drm_i915_gem_relocation_entry __user *relocs; |
1508 | struct drm_i915_gem_object *obj_priv = obj->driver_private; | 1506 | struct drm_i915_gem_object *obj_priv = obj->driver_private; |
1509 | int i, ret; | 1507 | int i, ret; |
1510 | uint32_t last_reloc_offset = -1; | 1508 | void __iomem *reloc_page; |
1511 | void __iomem *reloc_page = NULL; | ||
1512 | 1509 | ||
1513 | /* Choose the GTT offset for our buffer and put it there. */ | 1510 | /* Choose the GTT offset for our buffer and put it there. */ |
1514 | ret = i915_gem_object_pin(obj, (uint32_t) entry->alignment); | 1511 | ret = i915_gem_object_pin(obj, (uint32_t) entry->alignment); |
@@ -1631,26 +1628,11 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
1631 | * perform. | 1628 | * perform. |
1632 | */ | 1629 | */ |
1633 | reloc_offset = obj_priv->gtt_offset + reloc.offset; | 1630 | reloc_offset = obj_priv->gtt_offset + reloc.offset; |
1634 | if (reloc_page == NULL || | 1631 | reloc_page = io_mapping_map_atomic_wc(dev_priv->mm.gtt_mapping, |
1635 | (last_reloc_offset & ~(PAGE_SIZE - 1)) != | 1632 | (reloc_offset & |
1636 | (reloc_offset & ~(PAGE_SIZE - 1))) { | 1633 | ~(PAGE_SIZE - 1))); |
1637 | if (reloc_page != NULL) | ||
1638 | iounmap(reloc_page); | ||
1639 | |||
1640 | reloc_page = ioremap_wc(dev->agp->base + | ||
1641 | (reloc_offset & | ||
1642 | ~(PAGE_SIZE - 1)), | ||
1643 | PAGE_SIZE); | ||
1644 | last_reloc_offset = reloc_offset; | ||
1645 | if (reloc_page == NULL) { | ||
1646 | drm_gem_object_unreference(target_obj); | ||
1647 | i915_gem_object_unpin(obj); | ||
1648 | return -ENOMEM; | ||
1649 | } | ||
1650 | } | ||
1651 | |||
1652 | reloc_entry = (uint32_t __iomem *)(reloc_page + | 1634 | reloc_entry = (uint32_t __iomem *)(reloc_page + |
1653 | (reloc_offset & (PAGE_SIZE - 1))); | 1635 | (reloc_offset & (PAGE_SIZE - 1))); |
1654 | reloc_val = target_obj_priv->gtt_offset + reloc.delta; | 1636 | reloc_val = target_obj_priv->gtt_offset + reloc.delta; |
1655 | 1637 | ||
1656 | #if WATCH_BUF | 1638 | #if WATCH_BUF |
@@ -1659,6 +1641,7 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
1659 | readl(reloc_entry), reloc_val); | 1641 | readl(reloc_entry), reloc_val); |
1660 | #endif | 1642 | #endif |
1661 | writel(reloc_val, reloc_entry); | 1643 | writel(reloc_val, reloc_entry); |
1644 | io_mapping_unmap_atomic(reloc_page); | ||
1662 | 1645 | ||
1663 | /* Write the updated presumed offset for this entry back out | 1646 | /* Write the updated presumed offset for this entry back out |
1664 | * to the user. | 1647 | * to the user. |
@@ -1674,9 +1657,6 @@ i915_gem_object_pin_and_relocate(struct drm_gem_object *obj, | |||
1674 | drm_gem_object_unreference(target_obj); | 1657 | drm_gem_object_unreference(target_obj); |
1675 | } | 1658 | } |
1676 | 1659 | ||
1677 | if (reloc_page != NULL) | ||
1678 | iounmap(reloc_page); | ||
1679 | |||
1680 | #if WATCH_BUF | 1660 | #if WATCH_BUF |
1681 | if (0) | 1661 | if (0) |
1682 | i915_gem_dump_object(obj, 128, __func__, ~0); | 1662 | i915_gem_dump_object(obj, 128, __func__, ~0); |
@@ -2518,6 +2498,10 @@ i915_gem_entervt_ioctl(struct drm_device *dev, void *data, | |||
2518 | if (ret != 0) | 2498 | if (ret != 0) |
2519 | return ret; | 2499 | return ret; |
2520 | 2500 | ||
2501 | dev_priv->mm.gtt_mapping = io_mapping_create_wc(dev->agp->base, | ||
2502 | dev->agp->agp_info.aper_size | ||
2503 | * 1024 * 1024); | ||
2504 | |||
2521 | mutex_lock(&dev->struct_mutex); | 2505 | mutex_lock(&dev->struct_mutex); |
2522 | BUG_ON(!list_empty(&dev_priv->mm.active_list)); | 2506 | BUG_ON(!list_empty(&dev_priv->mm.active_list)); |
2523 | BUG_ON(!list_empty(&dev_priv->mm.flushing_list)); | 2507 | BUG_ON(!list_empty(&dev_priv->mm.flushing_list)); |
@@ -2535,11 +2519,13 @@ int | |||
2535 | i915_gem_leavevt_ioctl(struct drm_device *dev, void *data, | 2519 | i915_gem_leavevt_ioctl(struct drm_device *dev, void *data, |
2536 | struct drm_file *file_priv) | 2520 | struct drm_file *file_priv) |
2537 | { | 2521 | { |
2522 | drm_i915_private_t *dev_priv = dev->dev_private; | ||
2538 | int ret; | 2523 | int ret; |
2539 | 2524 | ||
2540 | ret = i915_gem_idle(dev); | 2525 | ret = i915_gem_idle(dev); |
2541 | drm_irq_uninstall(dev); | 2526 | drm_irq_uninstall(dev); |
2542 | 2527 | ||
2528 | io_mapping_free(dev_priv->mm.gtt_mapping); | ||
2543 | return ret; | 2529 | return ret; |
2544 | } | 2530 | } |
2545 | 2531 | ||