diff options
Diffstat (limited to 'mm/vmalloc.c')
-rw-r--r-- | mm/vmalloc.c | 158 |
1 files changed, 104 insertions, 54 deletions
diff --git a/mm/vmalloc.c b/mm/vmalloc.c index f9b166732e70..5d6030235d7a 100644 --- a/mm/vmalloc.c +++ b/mm/vmalloc.c | |||
@@ -261,8 +261,15 @@ struct vmap_area { | |||
261 | }; | 261 | }; |
262 | 262 | ||
263 | static DEFINE_SPINLOCK(vmap_area_lock); | 263 | static DEFINE_SPINLOCK(vmap_area_lock); |
264 | static struct rb_root vmap_area_root = RB_ROOT; | ||
265 | static LIST_HEAD(vmap_area_list); | 264 | static LIST_HEAD(vmap_area_list); |
265 | static struct rb_root vmap_area_root = RB_ROOT; | ||
266 | |||
267 | /* The vmap cache globals are protected by vmap_area_lock */ | ||
268 | static struct rb_node *free_vmap_cache; | ||
269 | static unsigned long cached_hole_size; | ||
270 | static unsigned long cached_vstart; | ||
271 | static unsigned long cached_align; | ||
272 | |||
266 | static unsigned long vmap_area_pcpu_hole; | 273 | static unsigned long vmap_area_pcpu_hole; |
267 | 274 | ||
268 | static struct vmap_area *__find_vmap_area(unsigned long addr) | 275 | static struct vmap_area *__find_vmap_area(unsigned long addr) |
@@ -331,9 +338,11 @@ static struct vmap_area *alloc_vmap_area(unsigned long size, | |||
331 | struct rb_node *n; | 338 | struct rb_node *n; |
332 | unsigned long addr; | 339 | unsigned long addr; |
333 | int purged = 0; | 340 | int purged = 0; |
341 | struct vmap_area *first; | ||
334 | 342 | ||
335 | BUG_ON(!size); | 343 | BUG_ON(!size); |
336 | BUG_ON(size & ~PAGE_MASK); | 344 | BUG_ON(size & ~PAGE_MASK); |
345 | BUG_ON(!is_power_of_2(align)); | ||
337 | 346 | ||
338 | va = kmalloc_node(sizeof(struct vmap_area), | 347 | va = kmalloc_node(sizeof(struct vmap_area), |
339 | gfp_mask & GFP_RECLAIM_MASK, node); | 348 | gfp_mask & GFP_RECLAIM_MASK, node); |
@@ -341,79 +350,106 @@ static struct vmap_area *alloc_vmap_area(unsigned long size, | |||
341 | return ERR_PTR(-ENOMEM); | 350 | return ERR_PTR(-ENOMEM); |
342 | 351 | ||
343 | retry: | 352 | retry: |
344 | addr = ALIGN(vstart, align); | ||
345 | |||
346 | spin_lock(&vmap_area_lock); | 353 | spin_lock(&vmap_area_lock); |
347 | if (addr + size - 1 < addr) | 354 | /* |
348 | goto overflow; | 355 | * Invalidate cache if we have more permissive parameters. |
356 | * cached_hole_size notes the largest hole noticed _below_ | ||
357 | * the vmap_area cached in free_vmap_cache: if size fits | ||
358 | * into that hole, we want to scan from vstart to reuse | ||
359 | * the hole instead of allocating above free_vmap_cache. | ||
360 | * Note that __free_vmap_area may update free_vmap_cache | ||
361 | * without updating cached_hole_size or cached_align. | ||
362 | */ | ||
363 | if (!free_vmap_cache || | ||
364 | size < cached_hole_size || | ||
365 | vstart < cached_vstart || | ||
366 | align < cached_align) { | ||
367 | nocache: | ||
368 | cached_hole_size = 0; | ||
369 | free_vmap_cache = NULL; | ||
370 | } | ||
371 | /* record if we encounter less permissive parameters */ | ||
372 | cached_vstart = vstart; | ||
373 | cached_align = align; | ||
374 | |||
375 | /* find starting point for our search */ | ||
376 | if (free_vmap_cache) { | ||
377 | first = rb_entry(free_vmap_cache, struct vmap_area, rb_node); | ||
378 | addr = ALIGN(first->va_end + PAGE_SIZE, align); | ||
379 | if (addr < vstart) | ||
380 | goto nocache; | ||
381 | if (addr + size - 1 < addr) | ||
382 | goto overflow; | ||
383 | |||
384 | } else { | ||
385 | addr = ALIGN(vstart, align); | ||
386 | if (addr + size - 1 < addr) | ||
387 | goto overflow; | ||
349 | 388 | ||
350 | /* XXX: could have a last_hole cache */ | 389 | n = vmap_area_root.rb_node; |
351 | n = vmap_area_root.rb_node; | 390 | first = NULL; |
352 | if (n) { | ||
353 | struct vmap_area *first = NULL; | ||
354 | 391 | ||
355 | do { | 392 | while (n) { |
356 | struct vmap_area *tmp; | 393 | struct vmap_area *tmp; |
357 | tmp = rb_entry(n, struct vmap_area, rb_node); | 394 | tmp = rb_entry(n, struct vmap_area, rb_node); |
358 | if (tmp->va_end >= addr) { | 395 | if (tmp->va_end >= addr) { |
359 | if (!first && tmp->va_start < addr + size) | ||
360 | first = tmp; | ||
361 | n = n->rb_left; | ||
362 | } else { | ||
363 | first = tmp; | 396 | first = tmp; |
397 | if (tmp->va_start <= addr) | ||
398 | break; | ||
399 | n = n->rb_left; | ||
400 | } else | ||
364 | n = n->rb_right; | 401 | n = n->rb_right; |
365 | } | 402 | } |
366 | } while (n); | ||
367 | 403 | ||
368 | if (!first) | 404 | if (!first) |
369 | goto found; | 405 | goto found; |
370 | |||
371 | if (first->va_end < addr) { | ||
372 | n = rb_next(&first->rb_node); | ||
373 | if (n) | ||
374 | first = rb_entry(n, struct vmap_area, rb_node); | ||
375 | else | ||
376 | goto found; | ||
377 | } | ||
378 | |||
379 | while (addr + size > first->va_start && addr + size <= vend) { | ||
380 | addr = ALIGN(first->va_end + PAGE_SIZE, align); | ||
381 | if (addr + size - 1 < addr) | ||
382 | goto overflow; | ||
383 | |||
384 | n = rb_next(&first->rb_node); | ||
385 | if (n) | ||
386 | first = rb_entry(n, struct vmap_area, rb_node); | ||
387 | else | ||
388 | goto found; | ||
389 | } | ||
390 | } | 406 | } |
391 | found: | 407 | |
392 | if (addr + size > vend) { | 408 | /* from the starting point, walk areas until a suitable hole is found */ |
393 | overflow: | 409 | while (addr + size >= first->va_start && addr + size <= vend) { |
394 | spin_unlock(&vmap_area_lock); | 410 | if (addr + cached_hole_size < first->va_start) |
395 | if (!purged) { | 411 | cached_hole_size = first->va_start - addr; |
396 | purge_vmap_area_lazy(); | 412 | addr = ALIGN(first->va_end + PAGE_SIZE, align); |
397 | purged = 1; | 413 | if (addr + size - 1 < addr) |
398 | goto retry; | 414 | goto overflow; |
399 | } | 415 | |
400 | if (printk_ratelimit()) | 416 | n = rb_next(&first->rb_node); |
401 | printk(KERN_WARNING | 417 | if (n) |
402 | "vmap allocation for size %lu failed: " | 418 | first = rb_entry(n, struct vmap_area, rb_node); |
403 | "use vmalloc=<size> to increase size.\n", size); | 419 | else |
404 | kfree(va); | 420 | goto found; |
405 | return ERR_PTR(-EBUSY); | ||
406 | } | 421 | } |
407 | 422 | ||
408 | BUG_ON(addr & (align-1)); | 423 | found: |
424 | if (addr + size > vend) | ||
425 | goto overflow; | ||
409 | 426 | ||
410 | va->va_start = addr; | 427 | va->va_start = addr; |
411 | va->va_end = addr + size; | 428 | va->va_end = addr + size; |
412 | va->flags = 0; | 429 | va->flags = 0; |
413 | __insert_vmap_area(va); | 430 | __insert_vmap_area(va); |
431 | free_vmap_cache = &va->rb_node; | ||
414 | spin_unlock(&vmap_area_lock); | 432 | spin_unlock(&vmap_area_lock); |
415 | 433 | ||
434 | BUG_ON(va->va_start & (align-1)); | ||
435 | BUG_ON(va->va_start < vstart); | ||
436 | BUG_ON(va->va_end > vend); | ||
437 | |||
416 | return va; | 438 | return va; |
439 | |||
440 | overflow: | ||
441 | spin_unlock(&vmap_area_lock); | ||
442 | if (!purged) { | ||
443 | purge_vmap_area_lazy(); | ||
444 | purged = 1; | ||
445 | goto retry; | ||
446 | } | ||
447 | if (printk_ratelimit()) | ||
448 | printk(KERN_WARNING | ||
449 | "vmap allocation for size %lu failed: " | ||
450 | "use vmalloc=<size> to increase size.\n", size); | ||
451 | kfree(va); | ||
452 | return ERR_PTR(-EBUSY); | ||
417 | } | 453 | } |
418 | 454 | ||
419 | static void rcu_free_va(struct rcu_head *head) | 455 | static void rcu_free_va(struct rcu_head *head) |
@@ -426,6 +462,22 @@ static void rcu_free_va(struct rcu_head *head) | |||
426 | static void __free_vmap_area(struct vmap_area *va) | 462 | static void __free_vmap_area(struct vmap_area *va) |
427 | { | 463 | { |
428 | BUG_ON(RB_EMPTY_NODE(&va->rb_node)); | 464 | BUG_ON(RB_EMPTY_NODE(&va->rb_node)); |
465 | |||
466 | if (free_vmap_cache) { | ||
467 | if (va->va_end < cached_vstart) { | ||
468 | free_vmap_cache = NULL; | ||
469 | } else { | ||
470 | struct vmap_area *cache; | ||
471 | cache = rb_entry(free_vmap_cache, struct vmap_area, rb_node); | ||
472 | if (va->va_start <= cache->va_start) { | ||
473 | free_vmap_cache = rb_prev(&va->rb_node); | ||
474 | /* | ||
475 | * We don't try to update cached_hole_size or | ||
476 | * cached_align, but it won't go very wrong. | ||
477 | */ | ||
478 | } | ||
479 | } | ||
480 | } | ||
429 | rb_erase(&va->rb_node, &vmap_area_root); | 481 | rb_erase(&va->rb_node, &vmap_area_root); |
430 | RB_CLEAR_NODE(&va->rb_node); | 482 | RB_CLEAR_NODE(&va->rb_node); |
431 | list_del_rcu(&va->list); | 483 | list_del_rcu(&va->list); |
@@ -1951,8 +2003,6 @@ finished: | |||
1951 | * should know vmalloc() area is valid and can use memcpy(). | 2003 | * should know vmalloc() area is valid and can use memcpy(). |
1952 | * This is for routines which have to access vmalloc area without | 2004 | * This is for routines which have to access vmalloc area without |
1953 | * any informaion, as /dev/kmem. | 2005 | * any informaion, as /dev/kmem. |
1954 | * | ||
1955 | * The caller should guarantee KM_USER1 is not used. | ||
1956 | */ | 2006 | */ |
1957 | 2007 | ||
1958 | long vwrite(char *buf, char *addr, unsigned long count) | 2008 | long vwrite(char *buf, char *addr, unsigned long count) |