On Mon, 1 Aug 2005, Joerg Bruehe wrote:
> As a result, the allocation succeeds, but some process gets killed when 
> the paging space cannot take such an additional page. To the affected 
> process, this looks like a crash.

Linux 2.4 and 2.6 kernels have a setting for their overcommitment 
behaviour under /proc/sys/vm/overcommit_memory. The different settings 
are:

0       -       Heuristic overcommit handling. Obvious overcommits of
                address space are refused. Used for a typical system. It
                ensures a seriously wild allocation fails while allowing
                overcommit to reduce swap usage.  root is allowed to
                allocate slighly more memory in this mode. This is the
                default.

1       -       Always overcommit. Appropriate for some scientific
                applications.

2       -       Don't overcommit. The total address space commit
                for the system is not permitted to exceed swap + a
                configurable percentage (default is 50) of physical RAM.
                Depending on the percentage you use, in most situations
                this means a process will not be killed while accessing
                pages but will receive errors on memory allocation as
                appropriate.

Heuristic overcommit handling seems to be the default, and my problem is 
in the 'Obvious overcommits of address space are refused'. For some (to me 
unknown) reason the kernel looks at a single 7GB malloc as if it be an 
'obvious overcommit' while 100 2GB mallocs (200GB total) is no problem. :P

For now I've set this to '2' which means the kernel won't overcommit 
anymore, just like any other proper OS... ;-) This makes things far more 
simple as I can only allocate as much memory as there is physically 
available now. However it does force me to be a bit more conservative. I 
have configured InnoDB with a 4GB buffer pool now, which leaves about 3GB 
for connections (about 300 with my current MySQL settings). Now this seems 
reasonable.

One final question though: my experience with InnoDB is that it really,
really likes to be able to fit all of it's data and keys into the buffer
pool. This would limit the maximum size of my database to roughly 4GB in
this case, correct?  This is in a website hosting environment where the
database is hit with about 1000 queries/s (mixed read/write).


> I am a bit surprised that the Linux kernel management will only allocate 
> memory if a single chunk of sufficient size is available. My 
> understanding was that in a paging system this is not necessary.
> 
> If this is (becoming) standard Linux policy, it might be necessary to 
> demand memory piecewise. One drawback of this approach is increased 
> bookeeping, if it ever needs to be released.
> 
> I have no idea how the developers view this issue - you might open a 
> change request if you consider this Linux kernel policy definite.
> 
> You wrote that if a mysql server start fails, you can run "fillmem", and 
> after its exit the memory will be available. I am not sure whether 
> Rick's explanation addresses this issue as well - it might be the 
> "memory defragger" he refers to. If not, the once used chunks might 
> still be considered "active".

I think it all refers to the IMHO buggy (hey, even the manpages state it!)  
VM memory allocation scheme. As stated I have disabled the overcommitment
behaviour for now, which seems to fit better to a dedicated database
server.


Best regards,

-- 
Matthijs van der Klip
Systeembeheerder

Spill E-Projects BV
Arendstraat 1-3
1223 RE  Hilversum
Tel. 035-6478248



-- 
MySQL General Mailing List
For list archives: http://lists.mysql.com/mysql
To unsubscribe:    http://lists.mysql.com/[EMAIL PROTECTED]

Reply via email to