That load script already loads pf_ring first and deactivates RSS, according to the README.
I tried these settings. Basically there was no change. I also tried: - use "pfdnacluster_multithread -n N" instead of "pfdnacluster_master -n N" and N times "pfcount". - set transparent mode to 2 - use active waiting None of it had any noticable effect. Martin On Thu, Mar 14, 2013 at 06:58:49PM +0000, Craig Merchant wrote: > From my understanding and experience, you don't use RSS with DNA/Libzero. > The RSS queues are limited to 16 queues. > > The major value of using DNA/Libzero is that it lets you use more queues than > RSS. > > Try the settings below... It's been a while since I set this up, but I > remember having some issues that required me to force pf_ring to load before > the ixgbe driver. > > options ixgbe MQ=0,0 num_rx_slots=32768 > options pf_ring min_num_slots=65536 transparent_mode=1 > install ixgbe /sbin/modprobe pf_ring $CMDLINE_OPTS; /sbin/modprobe > --ignore-install ixgbe $CMDLINE_OPTS > > -----Original Message----- > From: [email protected] > [mailto:[email protected]] On Behalf Of Martin Kummer > Sent: Thursday, March 14, 2013 11:18 AM > To: [email protected] > Subject: Re: [Ntop-misc] libzero performance > > Wow, thx for the quick answer. > > I just use the provided script > (drivers/DNA/ixgbe-3.10.16-DNA/src/load_dna_driver.sh) to load the drivers. > In short it does: > insmod ../../../../kernel/pf_ring.ko > insmod ./ixgbe.ko RSS=1,1,1,1 > ifconfig dna1 up > bash ../scripts/set_irq_affinity.sh ${IF[index]} > > The sysadmin has forbidden me to install these drivers permanently so there's > nothing in /etc/modprobe/*.conf > > Martin > > > On Thu, Mar 14, 2013 at 05:44:48PM +0000, Craig Merchant wrote: > > Martin, > > > > I'm running pfdnacluster_master on an interface that averages between 3-10 > > Gbps. The traffic is copied to 28 queues (0-27). The 28th queue contains > > a copy of all of the traffic. I don't have any issues with packets being > > dropped. > > > > How are you initializing the ixbe and pf_ring drivers in your > > /etc/modprobe.d/*.conf file? Mine looks something like: > > > > options igb RSS=8,8 > > options ixgbe MQ=0,0 num_rx_slots=32768 options pf_ring > > min_num_slots=65536 transparent_mode=1 install ixgbe /sbin/modprobe > > pf_ring $CMDLINE_OPTS; /sbin/modprobe --ignore-install ixgbe > > $CMDLINE_OPTS > > > > How are you bringing up the interface? I'm using DNA/Libzero for Snort, so > > I bring up the interface in the Snort init script with something like: > > > > > > function adapter_settings() { > > ifconfig dna0 up promisc > > ethtool -K dna0 tso off &>/dev/null > > ethtool -K dna0 gro off &>/dev/null > > ethtool -K dna0 lro off &>/dev/null > > ethtool -K dna0 gso off &>/dev/null > > ethtool -G dna0 tx 32768 &>/dev/null > > ethtool -G dna0 rx 32768 &>/dev/null } > > > > Thanks. > > > > Craig > > > > -----Original Message----- > > From: [email protected] > > [mailto:[email protected]] On Behalf Of Martin > > Kummer > > Sent: Thursday, March 14, 2013 10:24 AM > > To: [email protected] > > Subject: [Ntop-misc] libzero performance > > > > Hi everyone. > > > > For my bachelor thesis I'm modifying Vermont > > (https://github.com/constcast/vermont/wiki) to use PF_RING/libzero instead > > of pcap. > > > > To test libzero I used one of the examples on the website: On one host > > I used pfdnacluster_master and pfcount. From another host I sent 100M > > packets at 10Gbps. (ca. 12Mpps) > > > > There are two issues: > > • When I use just one slave (pfdancluster_master [...] -n 1) the > > performance is very good. I get about 99% of the packets. When I split the > > data between two slaves that number drops to about 90%. The last line of > > "pfdnacluster_master -n 2" output: > > Absolute Stats: RX 90'276'085 pkts [2'314'438.07 pkt/sec] Processed > > 90'276'085 pkts [2'314'438.07 pkt/sec] The more slaves, the lower the RX > > value gets. > > > > • Even with just one slave pfdnacluster_master uses almost 100% of a CPU > > core. While not yet a problem for me, this is likely to be the next > > bottleneck. > > > > Is there a way to increase the performance of a dnacluster, when there are > > several slaves? > > > > The software used: > > - the current svn chockout of ntop > > - the DNA ixgbe driver by ntop > > > > The hardware used: > > - Core i7-3930K (6C, HT, 3.2 GHz) > > - Ethernet controller: Intel Corporation 82599EB 10-Gigabit SFI/SFP+ > > Network Connection (rev 01) > > Subsystem: Intel Corporation Ethernet Server Adapter X520-2 > > > > best regards, > > Martin > > _______________________________________________ > > Ntop-misc mailing list > > [email protected] > > http://listgateway.unipi.it/mailman/listinfo/ntop-misc > > _______________________________________________ > > Ntop-misc mailing list > > [email protected] > > http://listgateway.unipi.it/mailman/listinfo/ntop-misc > _______________________________________________ > Ntop-misc mailing list > [email protected] > http://listgateway.unipi.it/mailman/listinfo/ntop-misc > _______________________________________________ > Ntop-misc mailing list > [email protected] > http://listgateway.unipi.it/mailman/listinfo/ntop-misc _______________________________________________ Ntop-misc mailing list [email protected] http://listgateway.unipi.it/mailman/listinfo/ntop-misc
