Greetings,
[EMAIL PROTECTED] wrote:
At Wed, 30 Jan 2008 19:13:07 +0200,
Stefan Lambrev wrote:
Greetings,
After playing with many settings and testing various configuration, now
I'm able to to receive on bridge more then 800,000 packets/s
without errors, which is amazing!
Unfortunately the server behind bridge can't handle more then 250,000
packets/s
Please advise how I can increase those limits?
Is is possible?
The servers are with 82573E Gigabit Ethernet Controller (quad port)
So far I tried with lagg and ng_fec, but with them I see more problems
then benefits :)
Tried polling with kern.polling.user_frac from 5 to 95,
different HZ, but nothing helped.
Increase the size of your socket buffers.
Increase the amount of mbufs in the system.
Best,
George
Here is what I put in my sysctl.conf:
kern.random.sys.harvest.ethernet=0
kern.ipc.nmbclusters=262144
kern.ipc.maxsockbuf=2097152
kern.ipc.maxsockets=98624
kern.ipc.somaxconn=1024
and in /boot/loader.conf:
vm.kmem_size="1024M"
kern.hz="500"
this is from netstat -m
516/774/1290 mbufs in use (current/cache/total)
513/411/924/262144 mbuf clusters in use (current/cache/total/max)
513/383 mbuf+clusters out of packet secondary zone in use (current/cache)
0/2/2/12800 4k (page size) jumbo clusters in use (current/cache/total/max)
0/0/0/6400 9k jumbo clusters in use (current/cache/total/max)
0/0/0/3200 16k jumbo clusters in use (current/cache/total/max)
1155K/1023K/2178K bytes allocated to network (current/cache/total)
0/0/0 requests for mbufs denied (mbufs/clusters/mbuf+clusters)
0/0/0 requests for jumbo clusters denied (4k/9k/16k)
0/0/0 sfbufs in use (current/peak/max)
0 requests for sfbufs denied
0 requests for sfbufs delayed
0 requests for I/O initiated by sendfile
0 calls to protocol drain routines
But still netstat -w1 -I em0 shows:
input (em0) output
packets errs bytes packets errs bytes colls
273877 113313 16432620 254270 0 14746500 0
273397 109905 16403820 253946 0 14728810 0
273945 113337 16436700 254285 0 14750560 0
What bothers me is the output of top -S:
PID USERNAME THR PRI NICE SIZE RES STATE C TIME WCPU COMMAND
22 root 1 -68 - 0K 16K CPU1 1 12:11 100.00% em0 taskq
11 root 1 171 ki31 0K 16K RUN 0 21:56 99.17% idle: cpu0
10 root 1 171 ki31 0K 16K RUN 1 9:16 0.00% idle: cpu1
14 root 1 -44 - 0K 16K WAIT 0 0:07 0.00% swi1: net
and vmstat:
procs memory page disk faults cpu
r b w avm fre flt re pi po fr sr ad4 in sy cs us
sy id
1 0 0 67088 1939700 0 0 0 0 0 0 0 2759 119 1325 0
50 50
0 0 0 67088 1939700 0 0 0 0 0 0 0 2760 127 1178 0
50 50
0 0 0 67088 1939700 0 0 0 0 0 0 0 2761 120 1269 0
50 50
What I'm missing?
_______________________________________________
freebsd-performance@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-performance
To unsubscribe, send any mail to "[EMAIL PROTECTED]"