On 17/10/12 17:39, Alfredo Cardigliano wrote:
> Chris
> please see inline
>
> On Oct 17, 2012, at 6:00 PM, Chris Wakelin <[email protected]> wrote:
>
>> I still can't get more than 12 cores used with Suricata on my Ubuntu
>> 12.04 machine with ixgbe. Even with DNA + RSS and Suricata using dna0@0
>> ... dna0@15, it fails for pfring_open on dna0@12 to dna0@15 (though
>> pfcount_aggregator manages the 16 queues in that case).
>
> You mean standard DNA (no DNA cluster, etc), right?
> This is definitely strange as DNA memory is allocated when loading the driver.
Yes I meant standard DNA.
Hmm. Strangely it's working now! Last night it didn't but I can't see
why. I tried again this morning but thought it failed, when now it seems
it didn't (silly me). I was probably mistaken about the discrepancy with
pfcount_aggregator.
>
>> How is memory allocated in DNA? Are there kernel options I'm missing?
>
> No, there is no configuration for that.
>
>> With DNA clusters, I can't get pfdnacluster_master to manage more than
>> 16 queues either. I would have expected my custom one with duplication
>> should actually only use as much memory as it does without duplication
>> as the duplicates are of course the same packets and therefore the same
>> memory.
>
> Even if you are using duplication, memory with DNA clusters is allocated when
> opening the socket.
> Actually on my test system with 4Gb of RAM I can run up to two cluster with
> 32 queues each.
> Anyway memory management in libzero is something we are working on (there is
> space for improvements).
Is that with all the sockets open? I can certainly start
pfdnacluster_master with that many queues, but the applications fail.
How much memory is used per socket? Strangely I didn't have problems
with testing e1000e DNA +libzero on another 64-bit system with less
memory (16GB instead of 32) but running Ubuntu 10.04 instead of 12.04.
Is the memory used what is shown in ifconfig? :-
dna0 Link encap:Ethernet HWaddr 00:1b:21:cd:a2:74
inet6 addr: fe80::21b:21ff:fecd:a274/64 Scope:Link
UP BROADCAST RUNNING PROMISC MULTICAST MTU:1522 Metric:1
RX packets:195271292 errors:0 dropped:0 overruns:0 frame:0
TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:157968626881 (157.9 GB) TX bytes:0 (0.0 B)
Memory:d9280000-d9300000
Does the setting of "ethtool -g" make a difference (presumably the same
as num_rx_slots= in the module parameters)?
Sorry for the inquisition :-) but it would be nice to understand what's
possible.
Best Wishes,
Chris
--
--+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+-
Christopher Wakelin, [email protected]
IT Services Centre, The University of Reading, Tel: +44 (0)118 378 2908
Whiteknights, Reading, RG6 6AF, UK Fax: +44 (0)118 975 3094
_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc