Re: [00/17] [RFC] Virtual Compound Page Support

2007-09-19 Thread Eric Dumazet
On Wed, 19 Sep 2007 08:34:47 +0100
Anton Altaparmakov [EMAIL PROTECTED] wrote:

 Hi Christoph,
 
 On 19 Sep 2007, at 04:36, Christoph Lameter wrote:
  Currently there is a strong tendency to avoid larger page  
  allocations in
  the kernel because of past fragmentation issues and the current
  defragmentation methods are still evolving. It is not clear to what  
  extend
  they can provide reliable allocations for higher order pages (plus the
  definition of reliable seems to be in the eye of the beholder).
 
  Currently we use vmalloc allocations in many locations to provide a  
  safe
  way to allocate larger arrays. That is due to the danger of higher  
  order
  allocations failing. Virtual Compound pages allow the use of regular
  page allocator allocations that will fall back only if there is an  
  actual
  problem with acquiring a higher order page.
 
  This patch set provides a way for a higher page allocation to fall  
  back.
  Instead of a physically contiguous page a virtually contiguous page
  is provided. The functionality of the vmalloc layer is used to provide
  the necessary page tables and control structures to establish a  
  virtually
  contiguous area.
 
 I like this a lot.  It will get rid of all the silly games we have to  
 play when needing both large allocations and efficient allocations  
 where possible.  In NTFS I can then just allocated higher order pages  
 instead of having to mess about with the allocation size and  
 allocating a single page if the requested size is = PAGE_SIZE or  
 using vmalloc() if the size is bigger.  And it will make it faster  
 because a lot of the time a higher order page allocation will succeed  
 with your patchset without resorting to vmalloc() so that will be a  
 lot faster.
 
 So where I currently have fs/ntfs/malloc.h the below mess I could get  
 rid of it completely and just use the normal page allocator/ 
 deallocator instead...
 
 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask)
 {
  if (likely(size = PAGE_SIZE)) {
  BUG_ON(!size);
  /* kmalloc() has per-CPU caches so is faster for  
 now. */
  return kmalloc(PAGE_SIZE, gfp_mask  ~__GFP_HIGHMEM);
  /* return (void *)__get_free_page(gfp_mask); */
  }
  if (likely(size  PAGE_SHIFT  num_physpages))
  return __vmalloc(size, gfp_mask, PAGE_KERNEL);
  return NULL;
 }
 
 And other places in the kernel can make use of the same.  I think XFS  
 does very similar things to NTFS in terms of larger allocations at  
 least and there are probably more places I don't know about off the  
 top of my head...
 
 I am looking forward to your patchset going into mainline.  (-:

Sure, it sounds *really* good. But...

1) Only power of two allocations are good candidates, or we waste RAM

2) On i386 machines, we have a small vmalloc window. (128 MB default value)
  Many servers with 4GB memory (PAE) like to boot with vmalloc=32M option to 
get 992MB of LOWMEM.
  If we allow some slub caches to fallback to vmalloc land, we'll have problems 
to tune this.

3) A fallback to vmalloc means an allocation of one vm_struct per compound page.

4) vmalloc() currently uses a linked list of vm_struct. Might need something 
more scalable.

-
To unsubscribe from this list: send the line unsubscribe linux-fsdevel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [00/17] [RFC] Virtual Compound Page Support

2007-09-19 Thread Christoph Lameter
On Wed, 19 Sep 2007, Eric Dumazet wrote:

 1) Only power of two allocations are good candidates, or we waste RAM

Correct.

 2) On i386 machines, we have a small vmalloc window. (128 MB default value)
   Many servers with 4GB memory (PAE) like to boot with vmalloc=32M option to 
 get 992MB of LOWMEM.
   If we allow some slub caches to fallback to vmalloc land, we'll have 
 problems to tune this.

We would first do the vmalloc conversion to GFP_VFALLBACK which would 
reduce the vmalloc requirements of drivers and core significantly. The 
patchset should actually reduce the vmalloc space requirements 
significantly. They are only needed in situations where the page allocator 
cannot provide a contiguous mapping and that gets rarer the better Mel's 
antifrag code works.
 
 4) vmalloc() currently uses a linked list of vm_struct. Might need something 
 more scalable.

If its rarely used then its not that big of a deal. The better the anti 
fragmentation measures the less vmalloc use.
-
To unsubscribe from this list: send the line unsubscribe linux-fsdevel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [00/17] [RFC] Virtual Compound Page Support

2007-09-19 Thread Christoph Lameter
On Wed, 19 Sep 2007, Andi Kleen wrote:

  alloc_page(GFP_VFALLBACK, order, )
 
 Is there a reason this needs to be a GFP flag versus a wrapper
 around alloc_page/free_page ?  page_alloc.c is already too complicated
 and it's better to keep new features separated. The only drawback
 would be that free_pages would need a different call, but that
 doesn't seem like a big problem.

I tried to make this a wrapper but there is a lot of logic in 
__alloc_pages() that would have to be replicated. Also there are specific 
places in __alloc_pages() were we can establish that we have enough memory
but its the memory fragmentation that prevents us from satisfying the 
request for a larger page.
 
 I'm also a little dubious about your attempts to do vmalloc in
 interrupt context. Is that really needed? GFP_ATOMIC allocations of
 large areas seem to be extremly unreliable to me and not design. Even
 if it works sometimes free probably wouldn't work there due to the
 flushes, which is very nasty. It would be better to drop that.

The flushes are only done on virtuall mapped architectures (xtensa) and 
are simple ASM code that can run in an interrupt context
-
To unsubscribe from this list: send the line unsubscribe linux-fsdevel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[00/17] [RFC] Virtual Compound Page Support

2007-09-18 Thread Christoph Lameter
Currently there is a strong tendency to avoid larger page allocations in
the kernel because of past fragmentation issues and the current
defragmentation methods are still evolving. It is not clear to what extend
they can provide reliable allocations for higher order pages (plus the
definition of reliable seems to be in the eye of the beholder).

Currently we use vmalloc allocations in many locations to provide a safe
way to allocate larger arrays. That is due to the danger of higher order
allocations failing. Virtual Compound pages allow the use of regular
page allocator allocations that will fall back only if there is an actual
problem with acquiring a higher order page.

This patch set provides a way for a higher page allocation to fall back.
Instead of a physically contiguous page a virtually contiguous page
is provided. The functionality of the vmalloc layer is used to provide
the necessary page tables and control structures to establish a virtually
contiguous area.

Advantages:

- If higher order allocations are failing then virtual compound pages
  consisting of a series of order-0 pages can stand in for those
  allocations.

- Reliability as long as the vmalloc layer can provide virtual mappings.

- Ability to reduce the use of vmalloc layer significantly by using
  physically contiguous memory instead of virtual contiguous memory.
  Most uses of vmalloc() can be converted to page allocator calls.

- The use of physically contiguous memory instead of vmalloc may allow the
  use larger TLB entries thus reducing TLB pressure. Also reduces the need
  for page table walks.

Disadvantages:

- In order to use fall back the logic accessing the memory must be
  aware that the memory could be backed by a virtual mapping and take
  precautions. virt_to_page() and page_address() may not work and
  vmalloc_to_page() and vmalloc_address() (introduced through this
  patch set) may have to be called.

- Virtual mappings are less efficient than physical mappings.
  Performance will drop once virtual fall back occurs.

- Virtual mappings have more memory overhead. vm_area control structures
  page tables, page arrays etc need to be allocated and managed to provide
  virtual mappings.

The patchset provides this functionality in stages. Stage 1 introduces
the basic fall back mechanism necessary to replace vmalloc allocations
with

alloc_page(GFP_VFALLBACK, order, )

which signifies to the page allocator that a higher order is to be found
but a virtual mapping may stand in if there is an issue with fragmentation.

Stage 1 functionality does not allow allocation and freeing of virtual
mappings from interrupt contexts.

The stage 1 series ends with the conversion of a few key uses of vmalloc
in the VM to alloc_pages() for the allocation of sparsemems memmap table
and the wait table in each zone. Other uses of vmalloc could be converted
in the same way.


Stage 2 functionality enhances the fallback even more allowing allocation
and frees in interrupt context.

SLUB is then modified to use the virtual mappings for slab caches
that are marked with SLAB_VFALLBACK. If a slab cache is marked this way
then we drop all the restraints regarding page order and allocate
good large memory areas that fit lots of objects so that we rarely
have to use the slow paths.

Two slab caches--the dentry cache and the buffer_heads--are then flagged
that way. Others could be converted in the same way.

The patch set also provides a debugging aid through setting

CONFIG_VFALLBACK_ALWAYS

If set then all GFP_VFALLBACK allocations fall back to the virtual
mappings. This is useful for verification tests. The test of this
patch set was done by enabling that options and compiling a kernel.


Stage 3 functionality could be the adding of support for the large
buffer size patchset. Not done yet and not sure if it would be useful
to do.

Much of this patchset may only be needed for special cases in which the
existing defragmentation methods fail for some reason. It may be better to
have the system operate without such a safety net and make sure that the
page allocator can return large orders in a reliable way.

The initial idea for this patchset came from Nick Piggin's fsblock
and from his arguments about reliability and guarantees. Since his
fsblock uses the virtual mappings I think it is legitimate to
generalize the use of virtual mappings to support higher order
allocations in this way. The application of these ideas to the large
block size patchset etc are straightforward. If wanted I can base
the next rev of the largebuffer patchset on this one and implement
fallback.

Contrary to Nick, I still doubt that any of this provides a guarantee.
Have said that I have to deal with various failure scenarios in the VM
daily and I'd certainly like to see it work in a more reliable manner.

IMHO getting rid of the various workarounds to deal with the small 4k
pages and avoiding additional layers that group these pages in subsystem
specific