Hi Jeremy
please read inline:

> On 26 Jul 2016, at 14:24, Jeremy Ashton <[email protected]> wrote:
> 
> AC> (rough) number on ingress interfaces, expected ingress total rate
> JA> Ingress interfaces will be 8x 10Gbit.  With these we form 80Gbps
> aggregate link to our edge switches.  The switches are then configured
> to send any of the monitor sessions down the aggregate link.
> Sustained traffic volume ~<7Gbps.  Ideal capture capacity for ~20Gbps
> before starting to drop packets.

Ok got it.

> AC> Cento already does shunting on aggregated queue (for feeding n2disk).
> JA> I understand it already offers shunting on aggregated queue.  The
> question would be, how do I go about using cento to aggregate the
> interfaces into multiple aggregate egress queues.

At the moment only one aggregated queue is supported, we will discuss
this internally and probably add it to the roadmap.

> Of which, we could
> then use cento to read from the aggregated egress queue then shunt the
> traffic for use by n2disk.

Why do you want to use 2 cento processes in a 2-stage pipeline instead of
doing everything in a single process?

> AC> Cento already does load balancing, but not after aggregation (it
> distribute to multiple egress queues traffic coming from each -i).
> Please note you can specify an interface pair with -i ethic thY if you
> have two directions from two ingress interfaces (i.e. from a TAP).  Is
> it ok for your use case?
> JA> I think what I am asking is if one can use cento as a consumer of
> packets that were output from a separate cento process providing an
> aggregated queue.

This is not an expected use case and it’s not optimal (e.g. it requires
1 packet copy passing from one cento process to the other), even though
it should “work”

> I am not sure if this is making sense.  Let me know if you require
> clarification.

It could make sense, but I am trying to figure out what is the reason for
pipelining 2 cento processes, probably I am missing something.

Regards
Alfredo

> On Tue, Jul 26, 2016 at 5:01 PM, Alfredo Cardigliano
> <[email protected]> wrote:
>> Hi Jeremy
>> please read below.
>> 
>>> On 26 Jul 2016, at 11:50, Jeremy Ashton <[email protected]> wrote:
>>> 
>>> I am looking to implement a packet processing pipeline that first
>>> aggregates a number of interfaces, generates ipfix flows and outputs
>>> to multiple aggregated queues (cento-ids).
>> 
>> Please provide:
>> - (rough) number on ingress interfaces
>> - expected ingress total rate
>> 
>>> Afterwards, I am looking
>>> to have (cento?) read from one of the aggregated queues and apply
>>> packet shunting before passing it off to n2disk.
>> 
>> Cento already does shunting on aggregated queue (for feeding n2disk).
>> 
>>> The other aggregated
>>> queues would be used to feed other analysis tools via load balanced
>>> output queues.
>> 
>> Cento already does load balancing, but not after aggregation (it distribute 
>> to multiple
>> egress queues traffic coming from each -i). Please note you can specify an 
>> interface
>> pair with -i ethX,ethY if you have two directions from two ingress 
>> interfaces (i.e. from a TAP).
>> Is it ok for your use case?
>> 
>> Alfredo
>> 
>>> I am currently using zbalance_ipc for the initial “capture” and
>>> distribution.  But it seems cento is the way it should be done.  From
>>> what I can tell, I would require two new features to be added:
>>> 
>>> 1) Adding the ability to create multiple aggregated queues as output
>>> from cento.  (https://github.com/ntop/nProbe/issues/86).
>>> 
>>> 2) Add the ability to bypass flow generation in cento and use it just
>>> for aggregation and distribution of packets.
>>> 
>>> I am not sure if this is the correct approach.  I wonder if you might
>>> offer some advice as to the best approach for doing this.
>>> 
>>> Thanks!
>>> _______________________________________________
>>> Ntop-misc mailing list
>>> [email protected]
>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>> 
>> 
>> _______________________________________________
>> Ntop-misc mailing list
>> [email protected]
>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> _______________________________________________
> Ntop-misc mailing list
> [email protected]
> http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Reply via email to