That load script already loads pf_ring first and deactivates RSS, according to 
the README.

I tried these settings. Basically there was no change.

I also tried:
 - use "pfdnacluster_multithread -n N" instead of "pfdnacluster_master -n N" 
and N times "pfcount".
 - set transparent mode to 2
 - use active waiting
None of it had any noticable effect.

Martin

On Thu, Mar 14, 2013 at 06:58:49PM +0000, Craig Merchant wrote:
> From my understanding and experience, you don't use RSS with DNA/Libzero.  
> The RSS queues are limited to 16 queues.  
> 
> The major value of using DNA/Libzero is that it lets you use more queues than 
> RSS.
> 
> Try the settings below...  It's been a while since I set this up, but I 
> remember having some issues that required me to force pf_ring to load before 
> the ixgbe driver.  
> 
> options ixgbe MQ=0,0 num_rx_slots=32768 
> options pf_ring min_num_slots=65536 transparent_mode=1 
> install ixgbe /sbin/modprobe pf_ring $CMDLINE_OPTS; /sbin/modprobe 
> --ignore-install ixgbe $CMDLINE_OPTS
> 
> -----Original Message-----
> From: [email protected] 
> [mailto:[email protected]] On Behalf Of Martin Kummer
> Sent: Thursday, March 14, 2013 11:18 AM
> To: [email protected]
> Subject: Re: [Ntop-misc] libzero performance
> 
> Wow, thx for the quick answer.
> 
> I just use the provided script 
> (drivers/DNA/ixgbe-3.10.16-DNA/src/load_dna_driver.sh) to load the drivers. 
> In short it does:
> insmod ../../../../kernel/pf_ring.ko
> insmod ./ixgbe.ko RSS=1,1,1,1
> ifconfig dna1 up
> bash ../scripts/set_irq_affinity.sh ${IF[index]}
> 
> The sysadmin has forbidden me to install these drivers permanently so there's 
> nothing in /etc/modprobe/*.conf
> 
> Martin
> 
> 
> On Thu, Mar 14, 2013 at 05:44:48PM +0000, Craig Merchant wrote:
> > Martin,
> > 
> > I'm running pfdnacluster_master on an interface that averages between 3-10 
> > Gbps.  The traffic is copied to 28 queues (0-27).  The 28th queue contains 
> > a copy of all of the traffic.  I don't have any issues with packets being 
> > dropped.
> > 
> > How are you initializing the ixbe and pf_ring drivers in your 
> > /etc/modprobe.d/*.conf file?  Mine looks something like:
> > 
> > options igb RSS=8,8
> > options ixgbe MQ=0,0 num_rx_slots=32768 options pf_ring 
> > min_num_slots=65536 transparent_mode=1 install ixgbe /sbin/modprobe 
> > pf_ring $CMDLINE_OPTS; /sbin/modprobe --ignore-install ixgbe 
> > $CMDLINE_OPTS
> > 
> > How are you bringing up the interface?  I'm using DNA/Libzero for Snort, so 
> > I bring up the interface in the Snort init script with something like:
> > 
> > 
> > function adapter_settings() {
> >     ifconfig dna0 up promisc
> >     ethtool -K dna0 tso off &>/dev/null
> >     ethtool -K dna0 gro off &>/dev/null
> >     ethtool -K dna0 lro off &>/dev/null
> >     ethtool -K dna0 gso off &>/dev/null
> >     ethtool -G dna0 tx 32768 &>/dev/null
> >     ethtool -G dna0 rx 32768 &>/dev/null }
> > 
> > Thanks.
> > 
> > Craig
> > 
> > -----Original Message-----
> > From: [email protected] 
> > [mailto:[email protected]] On Behalf Of Martin 
> > Kummer
> > Sent: Thursday, March 14, 2013 10:24 AM
> > To: [email protected]
> > Subject: [Ntop-misc] libzero performance
> > 
> > Hi everyone.
> > 
> > For my bachelor thesis I'm modifying Vermont 
> > (https://github.com/constcast/vermont/wiki) to use PF_RING/libzero instead 
> > of pcap.
> > 
> > To test libzero I used one of the examples on the website: On one host 
> > I used pfdnacluster_master and pfcount. From another host I sent 100M 
> > packets at 10Gbps. (ca. 12Mpps)
> > 
> > There are two issues:
> > • When I use just one slave (pfdancluster_master [...] -n 1) the 
> > performance is very good. I get about 99% of the packets. When I split the 
> > data between two slaves that number drops to about 90%. The last line of 
> > "pfdnacluster_master -n 2" output:
> >     Absolute Stats: RX 90'276'085 pkts [2'314'438.07 pkt/sec] Processed 
> > 90'276'085 pkts [2'314'438.07 pkt/sec] The more slaves, the lower the RX 
> > value gets.
> > 
> > • Even with just one slave pfdnacluster_master uses almost 100% of a CPU 
> > core. While not yet a problem for me, this is likely to be the next 
> > bottleneck.
> > 
> > Is there a way to increase the performance of a dnacluster, when there are 
> > several slaves?
> > 
> > The software used:
> >  - the current svn chockout of ntop
> >  - the DNA ixgbe driver by ntop
> > 
> > The hardware used:
> >  - Core i7-3930K (6C, HT, 3.2 GHz)
> >  - Ethernet controller: Intel Corporation 82599EB 10-Gigabit SFI/SFP+ 
> > Network Connection (rev 01)
> >     Subsystem: Intel Corporation Ethernet Server Adapter X520-2
> > 
> > best regards,
> > Martin
> > _______________________________________________
> > Ntop-misc mailing list
> > [email protected]
> > http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> > _______________________________________________
> > Ntop-misc mailing list
> > [email protected]
> > http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> _______________________________________________
> Ntop-misc mailing list
> [email protected]
> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> _______________________________________________
> Ntop-misc mailing list
> [email protected]
> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Reply via email to