On Wed, 16 May 2007 01:33:47 +0200 Nick Piggin wrote: > On Tue, May 15, 2007 at 10:17:31AM -0500, Matt Mackall wrote: > > On Tue, May 15, 2007 at 10:43:05AM +0200, Nick Piggin wrote: > > > This patch goes on top of my previous RCU patch, and has various > > > improvements for slob I noticed while implementing said patch ;) > > > > > > Comments? > > > > I'm warming to this. Please check that the comment block at the top is > > still accurate. > > It wasn't, fixed. > > ... > > Here is an updated version. > --- > > Improve slob by turning the freelist into a list of pages using struct page > fields, then each page has a singly linked freelist of slob blocks via a > pointer in the struct page. > > - The first benefit is that the slob freelists can be indexed by a smaller > type (2 bytes, if the PAGE_SIZE is reasonable). > > - Next is that freeing is much quicker because it does not have to traverse > the entire freelist. Allocation can be slightly faster too, because we can > skip almost-full freelist pages completely. > > - Slob pages are then freed immediately when they become empty, rather than > having a periodic timer try to free them. This gives efficiency and memory > consumption improvement. > > > Then, we don't encode seperate size and next fields into each slob block, > rather we use the sign bit to distinguish between "size" or "next". Then > size 1 blocks contain a "next" offset, and others contain the "size" in > the first unit and "next" in the second unit. > > - This allows minimum slob allocation alignment to go from 8 bytes to 2 > bytes on 32-bit and 12 bytes to 2 bytes on 64-bit. In practice, it is > best to align them to word size, however some architectures (eg. cris) > could gain space savings from turning off this extra alignment. > > > Then, make kmalloc use its own slob_block at the front of the allocation > in order to encode allocation size, rather than rely on not overwriting > slob's existing header block. > > - This reduces kmalloc allocation overhead similarly to alignment reductions. > > - Decouples kmalloc layer from the slob allocator. > > > Then, add a page flag specific to slob pages. > > - This means kfree of a page aligned slob block doesn't have to traverse > the bigblock list. > > > I would get benchmarks, but my test box's network doesn't come up with > slob before this patch. I think something is timing out. Anyway, things > are faster after the patch. > > Code size goes up about 1K, however dynamic memory usage _should_ be > lower even on relatively small memory systems. > > Signed-off-by: Nick Piggin <[EMAIL PROTECTED]> > > --- > Index: linux-2.6/init/Kconfig > =================================================================== > --- linux-2.6.orig/init/Kconfig > +++ linux-2.6/init/Kconfig > @@ -529,7 +529,7 @@ config SLUB > way and has enhanced diagnostics. > > config SLOB > - depends on EMBEDDED && !SPARSEMEM > + depends on EMBEDDED && !SPARSEMEM && !ARCH_USES_SLAB_PAGE_STRUCT > bool "SLOB (Simple Allocator)" > help > SLOB replaces the SLAB allocator with a drastically simpler
Is this patch going in? I have a randconfig (2.6.22) with CONFIG_SLOB=y CONFIG_NUMA=y CONFIG_SMP=y The kernel build fails with these config symbols set like that. It looks like SLOB needs that additional && !ARCH_USES_SLAB_PAGE_STRUCT (or && !SMP) --- ~Randy *** Remember to use Documentation/SubmitChecklist when testing your code *** - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/