aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/lib/rheap.c
diff options
context:
space:
mode:
authorLi Yang <leoli@freescale.com>2007-06-18 07:29:21 -0400
committerKumar Gala <galak@kernel.crashing.org>2007-06-19 23:35:53 -0400
commit7c8545e98468c53809fc06788a3b9a34dff05240 (patch)
tree11f7cff9f7f0f67b04db8234c41de5d6bc871b4c /arch/powerpc/lib/rheap.c
parent7b7a57c77dccddd84b6aa02a38deee7ad97c977a (diff)
[POWERPC] rheap - eliminates internal fragments caused by alignment
The patch adds fragments caused by rh_alloc_align() back to free list, instead of allocating the whole chunk of memory. This will greatly improve memory utilization managed by rheap. It solves MURAM not enough problem with 3 UCCs enabled on MPC8323. Signed-off-by: Li Yang <leoli@freescale.com> Acked-by: Joakim Tjernlund <joakim.tjernlund@transmode.se> Signed-off-by: Kumar Gala <galak@kernel.crashing.org>
Diffstat (limited to 'arch/powerpc/lib/rheap.c')
-rw-r--r--arch/powerpc/lib/rheap.c48
1 files changed, 29 insertions, 19 deletions
diff --git a/arch/powerpc/lib/rheap.c b/arch/powerpc/lib/rheap.c
index 180ee2933ab9..2f24ea0d723a 100644
--- a/arch/powerpc/lib/rheap.c
+++ b/arch/powerpc/lib/rheap.c
@@ -437,27 +437,26 @@ unsigned long rh_alloc_align(rh_info_t * info, int size, int alignment, const ch
437 struct list_head *l; 437 struct list_head *l;
438 rh_block_t *blk; 438 rh_block_t *blk;
439 rh_block_t *newblk; 439 rh_block_t *newblk;
440 unsigned long start; 440 unsigned long start, sp_size;
441 441
442 /* Validate size, and alignment must be power of two */ 442 /* Validate size, and alignment must be power of two */
443 if (size <= 0 || (alignment & (alignment - 1)) != 0) 443 if (size <= 0 || (alignment & (alignment - 1)) != 0)
444 return (unsigned long) -EINVAL; 444 return (unsigned long) -EINVAL;
445 445
446 /* given alignment larger that default rheap alignment */
447 if (alignment > info->alignment)
448 size += alignment - 1;
449
450 /* Align to configured alignment */ 446 /* Align to configured alignment */
451 size = (size + (info->alignment - 1)) & ~(info->alignment - 1); 447 size = (size + (info->alignment - 1)) & ~(info->alignment - 1);
452 448
453 if (assure_empty(info, 1) < 0) 449 if (assure_empty(info, 2) < 0)
454 return (unsigned long) -ENOMEM; 450 return (unsigned long) -ENOMEM;
455 451
456 blk = NULL; 452 blk = NULL;
457 list_for_each(l, &info->free_list) { 453 list_for_each(l, &info->free_list) {
458 blk = list_entry(l, rh_block_t, list); 454 blk = list_entry(l, rh_block_t, list);
459 if (size <= blk->size) 455 if (size <= blk->size) {
460 break; 456 start = (blk->start + alignment - 1) & ~(alignment - 1);
457 if (start + size <= blk->start + blk->size)
458 break;
459 }
461 blk = NULL; 460 blk = NULL;
462 } 461 }
463 462
@@ -470,25 +469,36 @@ unsigned long rh_alloc_align(rh_info_t * info, int size, int alignment, const ch
470 list_del(&blk->list); 469 list_del(&blk->list);
471 newblk = blk; 470 newblk = blk;
472 } else { 471 } else {
472 /* Fragment caused, split if needed */
473 /* Create block for fragment in the beginning */
474 sp_size = start - blk->start;
475 if (sp_size) {
476 rh_block_t *spblk;
477
478 spblk = get_slot(info);
479 spblk->start = blk->start;
480 spblk->size = sp_size;
481 /* add before the blk */
482 list_add(&spblk->list, blk->list.prev);
483 }
473 newblk = get_slot(info); 484 newblk = get_slot(info);
474 newblk->start = blk->start; 485 newblk->start = start;
475 newblk->size = size; 486 newblk->size = size;
476 487
477 /* blk still in free list, with updated start, size */ 488 /* blk still in free list, with updated start and size
478 blk->start += size; 489 * for fragment in the end */
479 blk->size -= size; 490 blk->start = start + size;
491 blk->size -= sp_size + size;
492 /* No fragment in the end, remove blk */
493 if (blk->size == 0) {
494 list_del(&blk->list);
495 release_slot(info, blk);
496 }
480 } 497 }
481 498
482 newblk->owner = owner; 499 newblk->owner = owner;
483 start = newblk->start;
484 attach_taken_block(info, newblk); 500 attach_taken_block(info, newblk);
485 501
486 /* for larger alignment return fixed up pointer */
487 /* this is no problem with the deallocator since */
488 /* we scan for pointers that lie in the blocks */
489 if (alignment > info->alignment)
490 start = (start + alignment - 1) & ~(alignment - 1);
491
492 return start; 502 return start;
493} 503}
494 504