On Tue, 14 Aug 2007, Peter Zijlstra wrote:

> > Ok but that could be addressed by making sure that a certain portion of 
> > memory is reserved for clean file backed pages.
> 
> Which gets us back to the initial problem of sizing this portion and
> ensuring it is big enough to service the need.

Clean file backed pages dominate memory on most boxes. They can be 
calculated by NR_FILE_PAGES - NR_FILE_DIRTY

On my 2G system that is 

Cached:        1731480 kB
Dirty:             424 kB

So for most load the patch as is will fix your issues. The problem arises 
if you have extreme loads that are making the majority of pages anonymous.

We could change min_free_kbytes to specify the number of free + clean 
pages required (if we can do atomic reclaim then we do not need it 
anymore). Then we can specify a large portion of memory for 
min_free_kbytes. 20%? That would give you 400M on my box which would 
certainly suffice.

If the amount of clean file backed pages falls below that limit then do 
the usual reclaim. If we write anonymous pages out to swap then they 
can also become clean and reclaimable.


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to