aboutsummaryrefslogtreecommitdiffstats
path: root/mm/slub.c
Commit message (Expand)AuthorAge
* SLUB's ksize() fails for size > 2048Vegard Nossum2007-12-05
* SLUB: killed the unused "end" variableDenis Cheng2007-11-12
* SLUB: Fix memory leak by not reusing cpu_slabChristoph Lameter2007-11-05
* missing atomic_read_long() in slub.cAl Viro2007-10-29
* memory hotplug: make kmem_cache_node for SLUB on memory online avoid panicYasunori Goto2007-10-22
* Slab API: remove useless ctor parameter and reorder parametersChristoph Lameter2007-10-17
* SLUB: simplify IRQ off handlingChristoph Lameter2007-10-17
* slub: list_locations() can use GFP_TEMPORARYAndrew Morton2007-10-16
* SLUB: Optimize cacheline use for zeroingChristoph Lameter2007-10-16
* SLUB: Place kmem_cache_cpu structures in a NUMA aware wayChristoph Lameter2007-10-16
* SLUB: Avoid touching page struct when freeing to per cpu slabChristoph Lameter2007-10-16
* SLUB: Move page->offset to kmem_cache_cpu->offsetChristoph Lameter2007-10-16
* SLUB: Do not use page->mappingChristoph Lameter2007-10-16
* SLUB: Avoid page struct cacheline bouncing due to remote frees to cpu slabChristoph Lameter2007-10-16
* Group short-lived and reclaimable kernel allocationsMel Gorman2007-10-16
* Categorize GFP flagsChristoph Lameter2007-10-16
* Memoryless nodes: SLUB supportChristoph Lameter2007-10-16
* Slab allocators: fail if ksize is called with a NULL parameterChristoph Lameter2007-10-16
* {slub, slob}: use unlikely() for kfree(ZERO_OR_NULL_PTR) checkSatyam Sharma2007-10-16
* SLUB: direct pass through of page size or higher kmalloc requestsChristoph Lameter2007-10-16
* slub.c:early_kmem_cache_node_alloc() shouldn't be __initAdrian Bunk2007-10-16
* SLUB: accurately compare debug flags during slab cache mergeChristoph Lameter2007-09-11
* slub: do not fail if we cannot register a slab with sysfsChristoph Lameter2007-08-31
* SLUB: do not fail on broken memory configurationsChristoph Lameter2007-08-22
* SLUB: use atomic_long_read for atomic_long variablesChristoph Lameter2007-08-22
* SLUB: Fix dynamic dma kmalloc cache creationChristoph Lameter2007-08-10
* SLUB: Remove checks for MAX_PARTIAL from kmem_cache_shrinkChristoph Lameter2007-08-10
* slub: fix bug in slub debug supportPeter Zijlstra2007-07-30
* slub: add lock debugging checkPeter Zijlstra2007-07-30
* mm: Remove slab destructors from kmem_cache_create().Paul Mundt2007-07-19
* slub: fix ksize() for zero-sized pointersLinus Torvalds2007-07-19
* SLUB: Fix CONFIG_SLUB_DEBUG use for CONFIG_NUMAChristoph Lameter2007-07-17
* SLUB: Move sysfs operations outside of slub_lockChristoph Lameter2007-07-17
* SLUB: Do not allocate object bit array on stackChristoph Lameter2007-07-17
* Slab allocators: Cleanup zeroing allocationsChristoph Lameter2007-07-17
* SLUB: Do not use length parameter in slab_alloc()Christoph Lameter2007-07-17
* SLUB: Style fix up the loop to disable small slabsChristoph Lameter2007-07-17
* mm/slub.c: make code staticAdrian Bunk2007-07-17
* SLUB: Simplify dma index -> size calculationChristoph Lameter2007-07-17
* SLUB: faster more efficient slab determination for __kmallocChristoph Lameter2007-07-17
* SLUB: do proper locking during dma slab creationChristoph Lameter2007-07-17
* SLUB: extract dma_kmalloc_cache from get_cache.Christoph Lameter2007-07-17
* SLUB: add some more inlines and #ifdef CONFIG_SLUB_DEBUGChristoph Lameter2007-07-17
* Slab allocators: support __GFP_ZERO in all allocatorsChristoph Lameter2007-07-17
* Slab allocators: consistent ZERO_SIZE_PTR support and NULL result semanticsChristoph Lameter2007-07-17
* Slab allocators: consolidate code for krealloc in mm/util.cChristoph Lameter2007-07-17
* SLUB Debug: fix initial object debug state of NUMA bootstrap objectsChristoph Lameter2007-07-17
* SLUB: ensure that the number of objects per slab stays low for high ordersChristoph Lameter2007-07-17
* SLUB slab validation: Move tracking information alloc outside of lockChristoph Lameter2007-07-17
* SLUB: use list_for_each_entry for loops over all slabsChristoph Lameter2007-07-17