On Sep 5, 2013, at 3:00 PM, Dennis Gamayunov <[email protected]> wrote:
> Hi, Alfredo! > > We were thinking about DNA, but it arises one more question: if we have > a filtering module implemented as PF_RING module (kernelspace), we'll > have to rewrite it as a userspace app to work with DNA, won't we? Yes > Or > it's possible to get PF_RING modules somewhere on the path between > capturing interface and userspace? No Regards Alfredo > > Regards, > > Dennis > > 05.09.2013 13:12, Alfredo Cardigliano пишет: >> Hi Dennis >> you should not expect a significant performance boost using pfring aware >> drivers for this application, the linux bridge is at the very bottom of the >> network stack thus overhead is negligible. You should use DNA drivers for >> better performance (take a look at pfdnabounce in userland/examples). >> >> Best Regards >> Alfredo >> >> On Sep 5, 2013, at 10:55 AM, Dennis Gamayunov <[email protected]> wrote: >> >>> Hi all, >>> >>> We're trying to get L2 tansparent bridge using PF_RING 5.6, PF_RING aware >>> ixgbe, Linux 3.2 (debian wheezy) and PF_RING transparent mode 2. >>> >>> Testing with pktgen and PF_RING show that bridging small (64 byte) packets >>> hits the same threshold that we see when using standart linux bridge or >>> simply test with netperf (UDP), i.e. about 480Mbit/s (about 1.5Mpps). >>> >>> The detailed description of hardware used and testing conditions are given >>> below. >>> >>> Is this an expected behavior or we missed something in PF_RING >>> configuration and/or ixgbe and kernel tuning? PF_RING in this config >>> behaves almost the same way as original Linux stack, and that concerns me >>> and makes think that we missed something in configuration. >>> >>> Hardware configuration: >>> >>> System: Supermicro X9DRW >>> Processors: 2 x Intel Xeon CPU E5-2665 @ 2.40GHz. 2 x 8 cores with HT. >>> Chipset: Intel C602 >>> Memory: 8x8Gb DIMM DDR3-1600 - 64 Gb total >>> Network adapters: 2 x 82599EB 10 Gigabit TN Network Connection (one adapter >>> equipped with twisted pair, other with optical SFP+ modules). Each adapter >>> has two ports. >>> >>> Testing configuration: >>> Vanilla pf_ring (version 5.6.0) with pf_ring-aware ixgbe (3.11.33) drivers >>> and transparent_mode=2 (non-transparent) is used. >>> LRO, gro features disabled. >>> irq-affinities spread across all processors >>> >>> No DNA present. >>> Traffic generated with pktgen reporting approximately 4 Mpps. >>> Adapter with twisted pair media used and pf_ring is used to bridge traffic >>> within two ports of one network adapter. >>> >>> We are using rings in userspace for bridging in following manner: >>> >>> .... >>> ring1 = pfring_open (interface1, snapshot, 0 | (promiscuous ? >>> PF_RING_PROMISC : 0)) >>> ring2 = pfring_open (interface2, snapshot, 0 | (promiscuous ? >>> PF_RING_PROMISC : 0)) >>> memset(&rule1, 0, sizeof(rule)); >>> memset(&rule2, 0, sizeof(rule)); >>> rule1.rule_id = 1; >>> rule2.rule_id = 2; >>> strcpy(rule1.reflector_device_ >>> name, interface2); >>> strcpy(rule2.reflector_device_name, interface1); >>> rule1.rule_action = reflect_packet_and_stop_rule_evaluation; >>> rule2.rule_action = reflect_packet_and_stop_rule_evaluation; >>> pfring_enable_ring(ring1); >>> pfring_enable_ring(ring2); >>> while(true) sleep(1); >>> ..... >>> >>> On bridge output interface we have only 37.5% of generated traffic, so >>> 62.5% drop. >>> If userspace application opens 8 rings for each reflect direction we have >>> 43.5% of bridged drivers, and 57.5 % drop. This gives approximately the >>> same 1.5-2Mpps. >>> >>> Kind regards, >>> >>> Dennis Gamayunov >>> _______________________________________________ >>> Ntop-misc mailing list >>> [email protected] >>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc >> _______________________________________________ >> Ntop-misc mailing list >> [email protected] >> http://listgateway.unipi.it/mailman/listinfo/ntop-misc > > _______________________________________________ > Ntop-misc mailing list > [email protected] > http://listgateway.unipi.it/mailman/listinfo/ntop-misc _______________________________________________ Ntop-misc mailing list [email protected] http://listgateway.unipi.it/mailman/listinfo/ntop-misc
