diff options
| -rw-r--r-- | mm/dmapool.c | 119 |
1 files changed, 58 insertions, 61 deletions
diff --git a/mm/dmapool.c b/mm/dmapool.c index e2ea4543abb4..72e7ece7ee9d 100644 --- a/mm/dmapool.c +++ b/mm/dmapool.c | |||
| @@ -17,7 +17,9 @@ | |||
| 17 | * The current design of this allocator is fairly simple. The pool is | 17 | * The current design of this allocator is fairly simple. The pool is |
| 18 | * represented by the 'struct dma_pool' which keeps a doubly-linked list of | 18 | * represented by the 'struct dma_pool' which keeps a doubly-linked list of |
| 19 | * allocated pages. Each page in the page_list is split into blocks of at | 19 | * allocated pages. Each page in the page_list is split into blocks of at |
| 20 | * least 'size' bytes. | 20 | * least 'size' bytes. Free blocks are tracked in an unsorted singly-linked |
| 21 | * list of free blocks within the page. Used blocks aren't tracked, but we | ||
| 22 | * keep a count of how many are currently allocated from each page. | ||
| 21 | */ | 23 | */ |
| 22 | 24 | ||
| 23 | #include <linux/device.h> | 25 | #include <linux/device.h> |
| @@ -38,7 +40,6 @@ | |||
| 38 | struct dma_pool { /* the pool */ | 40 | struct dma_pool { /* the pool */ |
| 39 | struct list_head page_list; | 41 | struct list_head page_list; |
| 40 | spinlock_t lock; | 42 | spinlock_t lock; |
| 41 | size_t blocks_per_page; | ||
| 42 | size_t size; | 43 | size_t size; |
| 43 | struct device *dev; | 44 | struct device *dev; |
| 44 | size_t allocation; | 45 | size_t allocation; |
| @@ -51,8 +52,8 @@ struct dma_page { /* cacheable header for 'allocation' bytes */ | |||
| 51 | struct list_head page_list; | 52 | struct list_head page_list; |
| 52 | void *vaddr; | 53 | void *vaddr; |
| 53 | dma_addr_t dma; | 54 | dma_addr_t dma; |
| 54 | unsigned in_use; | 55 | unsigned int in_use; |
| 55 | unsigned long bitmap[0]; | 56 | unsigned int offset; |
| 56 | }; | 57 | }; |
| 57 | 58 | ||
| 58 | #define POOL_TIMEOUT_JIFFIES ((100 /* msec */ * HZ) / 1000) | 59 | #define POOL_TIMEOUT_JIFFIES ((100 /* msec */ * HZ) / 1000) |
| @@ -87,8 +88,8 @@ show_pools(struct device *dev, struct device_attribute *attr, char *buf) | |||
| 87 | 88 | ||
| 88 | /* per-pool info, no real statistics yet */ | 89 | /* per-pool info, no real statistics yet */ |
| 89 | temp = scnprintf(next, size, "%-16s %4u %4Zu %4Zu %2u\n", | 90 | temp = scnprintf(next, size, "%-16s %4u %4Zu %4Zu %2u\n", |
| 90 | pool->name, | 91 | pool->name, blocks, |
| 91 | blocks, pages * pool->blocks_per_page, | 92 | pages * (pool->allocation / pool->size), |
| 92 | pool->size, pages); | 93 | pool->size, pages); |
| 93 | size -= temp; | 94 | size -= temp; |
| 94 | next += temp; | 95 | next += temp; |
| @@ -132,8 +133,11 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev, | |||
| 132 | return NULL; | 133 | return NULL; |
| 133 | } | 134 | } |
| 134 | 135 | ||
| 135 | if (size == 0) | 136 | if (size == 0) { |
| 136 | return NULL; | 137 | return NULL; |
| 138 | } else if (size < 4) { | ||
| 139 | size = 4; | ||
| 140 | } | ||
| 137 | 141 | ||
| 138 | if ((size % align) != 0) | 142 | if ((size % align) != 0) |
| 139 | size = ALIGN(size, align); | 143 | size = ALIGN(size, align); |
| @@ -160,7 +164,6 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev, | |||
| 160 | spin_lock_init(&retval->lock); | 164 | spin_lock_init(&retval->lock); |
| 161 | retval->size = size; | 165 | retval->size = size; |
| 162 | retval->allocation = allocation; | 166 | retval->allocation = allocation; |
| 163 | retval->blocks_per_page = allocation / size; | ||
| 164 | init_waitqueue_head(&retval->waitq); | 167 | init_waitqueue_head(&retval->waitq); |
| 165 | 168 | ||
| 166 | if (dev) { | 169 | if (dev) { |
| @@ -186,28 +189,36 @@ struct dma_pool *dma_pool_create(const char *name, struct device *dev, | |||
| 186 | } | 189 | } |
| 187 | EXPORT_SYMBOL(dma_pool_create); | 190 | EXPORT_SYMBOL(dma_pool_create); |
| 188 | 191 | ||
| 192 | static void pool_initialise_page(struct dma_pool *pool, struct dma_page *page) | ||
| 193 | { | ||
| 194 | unsigned int offset = 0; | ||
| 195 | |||
| 196 | do { | ||
| 197 | unsigned int next = offset + pool->size; | ||
| 198 | if (unlikely((next + pool->size) >= pool->allocation)) | ||
| 199 | next = pool->allocation; | ||
| 200 | *(int *)(page->vaddr + offset) = next; | ||
| 201 | offset = next; | ||
| 202 | } while (offset < pool->allocation); | ||
| 203 | } | ||
| 204 | |||
| 189 | static struct dma_page *pool_alloc_page(struct dma_pool *pool, gfp_t mem_flags) | 205 | static struct dma_page *pool_alloc_page(struct dma_pool *pool, gfp_t mem_flags) |
| 190 | { | 206 | { |
| 191 | struct dma_page *page; | 207 | struct dma_page *page; |
| 192 | int mapsize; | ||
| 193 | |||
| 194 | mapsize = pool->blocks_per_page; | ||
| 195 | mapsize = (mapsize + BITS_PER_LONG - 1) / BITS_PER_LONG; | ||
| 196 | mapsize *= sizeof(long); | ||
| 197 | 208 | ||
| 198 | page = kmalloc(mapsize + sizeof *page, mem_flags); | 209 | page = kmalloc(sizeof(*page), mem_flags); |
| 199 | if (!page) | 210 | if (!page) |
| 200 | return NULL; | 211 | return NULL; |
| 201 | page->vaddr = dma_alloc_coherent(pool->dev, | 212 | page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, |
| 202 | pool->allocation, | ||
| 203 | &page->dma, mem_flags); | 213 | &page->dma, mem_flags); |
| 204 | if (page->vaddr) { | 214 | if (page->vaddr) { |
| 205 | memset(page->bitmap, 0xff, mapsize); /* bit set == free */ | ||
| 206 | #ifdef CONFIG_DEBUG_SLAB | 215 | #ifdef CONFIG_DEBUG_SLAB |
| 207 | memset(page->vaddr, POOL_POISON_FREED, pool->allocation); | 216 | memset(page->vaddr, POOL_POISON_FREED, pool->allocation); |
| 208 | #endif | 217 | #endif |
| 218 | pool_initialise_page(pool, page); | ||
| 209 | list_add(&page->page_list, &pool->page_list); | 219 | list_add(&page->page_list, &pool->page_list); |
| 210 | page->in_use = 0; | 220 | page->in_use = 0; |
| 221 | page->offset = 0; | ||
| 211 | } else { | 222 | } else { |
| 212 | kfree(page); | 223 | kfree(page); |
| 213 | page = NULL; | 224 | page = NULL; |
| @@ -215,14 +226,9 @@ static struct dma_page *pool_alloc_page(struct dma_pool *pool, gfp_t mem_flags) | |||
| 215 | return page; | 226 | return page; |
| 216 | } | 227 | } |
| 217 | 228 | ||
| 218 | static inline int is_page_busy(int blocks, unsigned long *bitmap) | 229 | static inline int is_page_busy(struct dma_page *page) |
| 219 | { | 230 | { |
| 220 | while (blocks > 0) { | 231 | return page->in_use != 0; |
| 221 | if (*bitmap++ != ~0UL) | ||
| 222 | return 1; | ||
| 223 | blocks -= BITS_PER_LONG; | ||
| 224 | } | ||
| 225 | return 0; | ||
| 226 | } | 232 | } |
| 227 | 233 | ||
| 228 | static void pool_free_page(struct dma_pool *pool, struct dma_page *page) | 234 | static void pool_free_page(struct dma_pool *pool, struct dma_page *page) |
| @@ -257,7 +263,7 @@ void dma_pool_destroy(struct dma_pool *pool) | |||
| 257 | struct dma_page *page; | 263 | struct dma_page *page; |
| 258 | page = list_entry(pool->page_list.next, | 264 | page = list_entry(pool->page_list.next, |
| 259 | struct dma_page, page_list); | 265 | struct dma_page, page_list); |
| 260 | if (is_page_busy(pool->blocks_per_page, page->bitmap)) { | 266 | if (is_page_busy(page)) { |
| 261 | if (pool->dev) | 267 | if (pool->dev) |
| 262 | dev_err(pool->dev, | 268 | dev_err(pool->dev, |
| 263 | "dma_pool_destroy %s, %p busy\n", | 269 | "dma_pool_destroy %s, %p busy\n", |
| @@ -292,27 +298,14 @@ void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, | |||
| 292 | { | 298 | { |
| 293 | unsigned long flags; | 299 | unsigned long flags; |
| 294 | struct dma_page *page; | 300 | struct dma_page *page; |
| 295 | int map, block; | ||
| 296 | size_t offset; | 301 | size_t offset; |
| 297 | void *retval; | 302 | void *retval; |
| 298 | 303 | ||
| 299 | spin_lock_irqsave(&pool->lock, flags); | 304 | spin_lock_irqsave(&pool->lock, flags); |
| 300 | restart: | 305 | restart: |
| 301 | list_for_each_entry(page, &pool->page_list, page_list) { | 306 | list_for_each_entry(page, &pool->page_list, page_list) { |
| 302 | int i; | 307 | if (page->offset < pool->allocation) |
| 303 | /* only cachable accesses here ... */ | 308 | goto ready; |
| 304 | for (map = 0, i = 0; | ||
| 305 | i < pool->blocks_per_page; i += BITS_PER_LONG, map++) { | ||
| 306 | if (page->bitmap[map] == 0) | ||
| 307 | continue; | ||
| 308 | block = ffz(~page->bitmap[map]); | ||
| 309 | if ((i + block) < pool->blocks_per_page) { | ||
| 310 | clear_bit(block, &page->bitmap[map]); | ||
| 311 | offset = (BITS_PER_LONG * map) + block; | ||
| 312 | offset *= pool->size; | ||
| 313 | goto ready; | ||
| 314 | } | ||
| 315 | } | ||
| 316 | } | 309 | } |
| 317 | page = pool_alloc_page(pool, GFP_ATOMIC); | 310 | page = pool_alloc_page(pool, GFP_ATOMIC); |
| 318 | if (!page) { | 311 | if (!page) { |
| @@ -333,10 +326,10 @@ void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags, | |||
| 333 | goto done; | 326 | goto done; |
| 334 | } | 327 | } |
| 335 | 328 | ||
| 336 | clear_bit(0, &page->bitmap[0]); | ||
| 337 | offset = 0; | ||
| 338 | ready: | 329 | ready: |
| 339 | page->in_use++; | 330 | page->in_use++; |
| 331 | offset = page->offset; | ||
| 332 | page->offset = *(int *)(page->vaddr + offset); | ||
| 340 | retval = offset + page->vaddr; | 333 | retval = offset + page->vaddr; |
| 341 | *handle = offset + page->dma; | 334 | *handle = offset + page->dma; |
| 342 | #ifdef CONFIG_DEBUG_SLAB | 335 | #ifdef CONFIG_DEBUG_SLAB |
| @@ -379,7 +372,7 @@ void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma) | |||
| 379 | { | 372 | { |
| 380 | struct dma_page *page; | 373 | struct dma_page *page; |
| 381 | unsigned long flags; | 374 | unsigned long flags; |
| 382 | int map, block; | 375 | unsigned int offset; |
| 383 | 376 | ||
| 384 | page = pool_find_page(pool, dma); | 377 | page = pool_find_page(pool, dma); |
| 385 | if (!page) { | 378 | if (!page) { |
| @@ -393,13 +386,9 @@ void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma) | |||
| 393 | return; | 386 | return; |
| 394 | } | 387 | } |
| 395 | 388 | ||
| 396 | block = dma - page->dma; | 389 | offset = vaddr - page->vaddr; |
| 397 | block /= pool->size; | ||
| 398 | map = block / BITS_PER_LONG; | ||
| 399 | block %= BITS_PER_LONG; | ||
| 400 | |||
| 401 | #ifdef CONFIG_DEBUG_SLAB | 390 | #ifdef CONFIG_DEBUG_SLAB |
| 402 | if (((dma - page->dma) + (void *)page->vaddr) != vaddr) { | 391 | if ((dma - page->dma) != offset) { |
| 403 | if (pool->dev) | 392 | if (pool->dev) |
| 404 | dev_err(pool->dev, | 393 | dev_err(pool->dev, |
| 405 | "dma_pool_free %s, %p (bad vaddr)/%Lx\n", | 394 | "dma_pool_free %s, %p (bad vaddr)/%Lx\n", |
| @@ -410,28 +399,36 @@ void dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma) | |||
| 410 | pool->name, vaddr, (unsigned long long)dma); | 399 | pool->name, vaddr, (unsigned long long)dma); |
| 411 | return; | 400 | return; |
| 412 | } | 401 | } |
| 413 | if (page->bitmap[map] & (1UL << block)) { | 402 | { |
| 414 | if (pool->dev) | 403 | unsigned int chain = page->offset; |
| 415 | dev_err(pool->dev, | 404 | while (chain < pool->allocation) { |
| 416 | "dma_pool_free %s, dma %Lx already free\n", | 405 | if (chain != offset) { |
| 417 | pool->name, (unsigned long long)dma); | 406 | chain = *(int *)(page->vaddr + chain); |
| 418 | else | 407 | continue; |
| 419 | printk(KERN_ERR | 408 | } |
| 420 | "dma_pool_free %s, dma %Lx already free\n", | 409 | if (pool->dev) |
| 421 | pool->name, (unsigned long long)dma); | 410 | dev_err(pool->dev, "dma_pool_free %s, dma %Lx " |
| 422 | return; | 411 | "already free\n", pool->name, |
| 412 | (unsigned long long)dma); | ||
| 413 | else | ||
| 414 | printk(KERN_ERR "dma_pool_free %s, dma %Lx " | ||
| 415 | "already free\n", pool->name, | ||
| 416 | (unsigned long long)dma); | ||
| 417 | return; | ||
| 418 | } | ||
| 423 | } | 419 | } |
| 424 | memset(vaddr, POOL_POISON_FREED, pool->size); | 420 | memset(vaddr, POOL_POISON_FREED, pool->size); |
| 425 | #endif | 421 | #endif |
| 426 | 422 | ||
| 427 | spin_lock_irqsave(&pool->lock, flags); | 423 | spin_lock_irqsave(&pool->lock, flags); |
| 428 | page->in_use--; | 424 | page->in_use--; |
| 429 | set_bit(block, &page->bitmap[map]); | 425 | *(int *)vaddr = page->offset; |
| 426 | page->offset = offset; | ||
| 430 | if (waitqueue_active(&pool->waitq)) | 427 | if (waitqueue_active(&pool->waitq)) |
| 431 | wake_up_locked(&pool->waitq); | 428 | wake_up_locked(&pool->waitq); |
| 432 | /* | 429 | /* |
| 433 | * Resist a temptation to do | 430 | * Resist a temptation to do |
| 434 | * if (!is_page_busy(bpp, page->bitmap)) pool_free_page(pool, page); | 431 | * if (!is_page_busy(page)) pool_free_page(pool, page); |
| 435 | * Better have a few empty pages hang around. | 432 | * Better have a few empty pages hang around. |
| 436 | */ | 433 | */ |
| 437 | spin_unlock_irqrestore(&pool->lock, flags); | 434 | spin_unlock_irqrestore(&pool->lock, flags); |
