Martin,

I'm running pfdnacluster_master on an interface that averages between 3-10 
Gbps.  The traffic is copied to 28 queues (0-27).  The 28th queue contains a 
copy of all of the traffic.  I don't have any issues with packets being dropped.

How are you initializing the ixbe and pf_ring drivers in your 
/etc/modprobe.d/*.conf file?  Mine looks something like:

options igb RSS=8,8
options ixgbe MQ=0,0 num_rx_slots=32768
options pf_ring min_num_slots=65536 transparent_mode=1
install ixgbe /sbin/modprobe pf_ring $CMDLINE_OPTS; /sbin/modprobe 
--ignore-install ixgbe $CMDLINE_OPTS

How are you bringing up the interface?  I'm using DNA/Libzero for Snort, so I 
bring up the interface in the Snort init script with something like:


function adapter_settings() {
    ifconfig dna0 up promisc
    ethtool -K dna0 tso off &>/dev/null
    ethtool -K dna0 gro off &>/dev/null
    ethtool -K dna0 lro off &>/dev/null
    ethtool -K dna0 gso off &>/dev/null
    ethtool -G dna0 tx 32768 &>/dev/null
    ethtool -G dna0 rx 32768 &>/dev/null
}

Thanks.

Craig

-----Original Message-----
From: [email protected] 
[mailto:[email protected]] On Behalf Of Martin Kummer
Sent: Thursday, March 14, 2013 10:24 AM
To: [email protected]
Subject: [Ntop-misc] libzero performance

Hi everyone.

For my bachelor thesis I'm modifying Vermont 
(https://github.com/constcast/vermont/wiki) to use PF_RING/libzero instead of 
pcap.

To test libzero I used one of the examples on the website: On one host I used 
pfdnacluster_master and pfcount. From another host I sent 100M packets at 
10Gbps. (ca. 12Mpps)

There are two issues:
• When I use just one slave (pfdancluster_master [...] -n 1) the performance is 
very good. I get about 99% of the packets. When I split the data between two 
slaves that number drops to about 90%. The last line of "pfdnacluster_master -n 
2" output:
    Absolute Stats: RX 90'276'085 pkts [2'314'438.07 pkt/sec] Processed 
90'276'085 pkts [2'314'438.07 pkt/sec] The more slaves, the lower the RX value 
gets.

• Even with just one slave pfdnacluster_master uses almost 100% of a CPU core. 
While not yet a problem for me, this is likely to be the next bottleneck.

Is there a way to increase the performance of a dnacluster, when there are 
several slaves?

The software used:
 - the current svn chockout of ntop
 - the DNA ixgbe driver by ntop

The hardware used:
 - Core i7-3930K (6C, HT, 3.2 GHz)
 - Ethernet controller: Intel Corporation 82599EB 10-Gigabit SFI/SFP+ Network 
Connection (rev 01)
    Subsystem: Intel Corporation Ethernet Server Adapter X520-2

best regards,
Martin
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Reply via email to