Hi all,

I've been scratching my head a bit with a performance issue on a live system
and was wondering if anyone could point me in the right direction for
resolving?


I have a load balancer using 2.4.32 stock kernel running LVS with HTTP &
HTTPS in DR mode with 10 real servers each.
The load balancer is a 2.4GHz Celeron with 512Mb RAM, broadcom gigabit nics.

free -m shows 300MB used + 200MB cached (no swap file)

When the active connections climb to 60,000 and throughput is about 100MB+
we get a sudden and massive drop of in throughput...
We get a corresponding drop in system CPU and increase in Kernel CPU...
I have a selection of graphs here including throughput and cpu usage etc.

http://www.loadbalancer.org/test/

No errors in syslog relating to NICs or netfilter etc.

The actual number of connections is actually about 9K or 11K/sec (Direct
routing exagerating them), I was looking at using ipvsadm --set to reduce
the TCP timeout and therefore connection table size but I don't think that
the memory is an issue.

Any other ideas?

*NB. I've tested similar hardware to 800MB+ and 50K cons/sec (obviously with
small/large packet sizes though not live data)*

Many thanks in advance for any insight :-).

-- 
Regards,

Malcolm Turnbull.

Loadbalancer.org Ltd.
Phone: +44 (0)870 443 8779
http://www.loadbalancer.org/
_______________________________________________
Please read the documentation before posting - it's available at:
http://www.linuxvirtualserver.org/

LinuxVirtualServer.org mailing list - [email protected]
Send requests to [email protected]
or go to http://lists.graemef.net/mailman/listinfo/lvs-users

Reply via email to