Martin,
I'm running pfdnacluster_master on an interface that averages between 3-10
Gbps. The traffic is copied to 28 queues (0-27). The 28th queue contains a
copy of all of the traffic. I don't have any issues with packets being dropped.
How are you initializing the ixbe and pf_ring drivers in your
/etc/modprobe.d/*.conf file? Mine looks something like:
options igb RSS=8,8
options ixgbe MQ=0,0 num_rx_slots=32768
options pf_ring min_num_slots=65536 transparent_mode=1
install ixgbe /sbin/modprobe pf_ring $CMDLINE_OPTS; /sbin/modprobe
--ignore-install ixgbe $CMDLINE_OPTS
How are you bringing up the interface? I'm using DNA/Libzero for Snort, so I
bring up the interface in the Snort init script with something like:
function adapter_settings() {
ifconfig dna0 up promisc
ethtool -K dna0 tso off &>/dev/null
ethtool -K dna0 gro off &>/dev/null
ethtool -K dna0 lro off &>/dev/null
ethtool -K dna0 gso off &>/dev/null
ethtool -G dna0 tx 32768 &>/dev/null
ethtool -G dna0 rx 32768 &>/dev/null
}
Thanks.
Craig
-----Original Message-----
From: [email protected]
[mailto:[email protected]] On Behalf Of Martin Kummer
Sent: Thursday, March 14, 2013 10:24 AM
To: [email protected]
Subject: [Ntop-misc] libzero performance
Hi everyone.
For my bachelor thesis I'm modifying Vermont
(https://github.com/constcast/vermont/wiki) to use PF_RING/libzero instead of
pcap.
To test libzero I used one of the examples on the website: On one host I used
pfdnacluster_master and pfcount. From another host I sent 100M packets at
10Gbps. (ca. 12Mpps)
There are two issues:
• When I use just one slave (pfdancluster_master [...] -n 1) the performance is
very good. I get about 99% of the packets. When I split the data between two
slaves that number drops to about 90%. The last line of "pfdnacluster_master -n
2" output:
Absolute Stats: RX 90'276'085 pkts [2'314'438.07 pkt/sec] Processed
90'276'085 pkts [2'314'438.07 pkt/sec] The more slaves, the lower the RX value
gets.
• Even with just one slave pfdnacluster_master uses almost 100% of a CPU core.
While not yet a problem for me, this is likely to be the next bottleneck.
Is there a way to increase the performance of a dnacluster, when there are
several slaves?
The software used:
- the current svn chockout of ntop
- the DNA ixgbe driver by ntop
The hardware used:
- Core i7-3930K (6C, HT, 3.2 GHz)
- Ethernet controller: Intel Corporation 82599EB 10-Gigabit SFI/SFP+ Network
Connection (rev 01)
Subsystem: Intel Corporation Ethernet Server Adapter X520-2
best regards,
Martin
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc