Hi Doug
we released 5.5.3 a few days ago, it is likely we refresh that tarball.

Regards
Alfredo

On Jun 5, 2013, at 5:45 PM, Doug Burks <[email protected]> wrote:

> Hi Alfredo,
> 
> Thanks for the fix!  I did a few quick tests and it appears to be
> working properly.
> 
> Will you be releasing a 5.5.4 tarball soon?
> 
> Thanks,
> Doug
> 
> On Wed, Jun 5, 2013 at 11:33 AM, Alfredo Cardigliano
> <[email protected]> wrote:
>> Hi Doug
>> thank you for the info you reported, as you guessed it was a bug in the 
>> fragments handling.
>> Please update from svn, it should be fixed now.
>> 
>> Regards
>> Alfredo
>> 
>> On Jun 5, 2013, at 4:20 PM, Doug Burks <[email protected]> wrote:
>> 
>>> Here's an additional test on 5.5.3 with http_gzip.pcap (with 10 packets):
>>> 
>>> - running a single instance of pfcount everything works properly and
>>> /proc/net/pf_ring/info shows:
>>> PF_RING Version          : 5.5.3 ($Revision: exported$)
>>> Total rings              : 1
>>> 
>>> Standard (non DNA) Options
>>> Ring slots               : 4096
>>> Slot version             : 15
>>> Capture TX               : Yes [RX+TX]
>>> IP Defragment            : No
>>> Socket Mode              : Standard
>>> Transparent mode         : Yes [mode 0]
>>> Total plugins            : 0
>>> Cluster Fragment Queue   : 0
>>> Cluster Fragment Discard : 0
>>> 
>>> - running two instances of pfcount with the same clusterId shows no
>>> packets in either pfcount instance and /proc/net/pf_ring/info shows:
>>> PF_RING Version          : 5.5.3 ($Revision: exported$)
>>> Total rings              : 2
>>> 
>>> Standard (non DNA) Options
>>> Ring slots               : 4096
>>> Slot version             : 15
>>> Capture TX               : Yes [RX+TX]
>>> IP Defragment            : No
>>> Socket Mode              : Standard
>>> Transparent mode         : Yes [mode 0]
>>> Total plugins            : 0
>>> Cluster Fragment Queue   : 0
>>> Cluster Fragment Discard : 10
>>> 
>>> Any ideas why it's discarding the 10 packets?
>>> 
>>> Thanks!
>>> 
>>> Doug
>>> 
>>> On Wed, Jun 5, 2013 at 8:44 AM, Doug Burks <[email protected]> wrote:
>>>> I repeated these tests on the same Ubuntu VM using PF_RING 5.5.2
>>>> kernel/userland and clustering works as expected without the packet
>>>> loss seen in 5.5.3.
>>>> 
>>>> Perhaps I'm missing something, but it appears that there is a bug in
>>>> the 5.5.3 cluster code.
>>>> 
>>>> Can somebody confirm, please?
>>>> 
>>>> Thanks!
>>>> 
>>>> Doug
>>>> 
>>>> 
>>>> On Wed, Jun 5, 2013 at 8:06 AM, Doug Burks <[email protected]> wrote:
>>>>> I just did a new test as follows:
>>>>> 
>>>>> - started with a fresh installation of Ubuntu 12.04
>>>>> - downloaded the PF_RING 5.5.3 tarball
>>>>> - compiled and inserted the kernel module
>>>>> - changed pfcount.c to use cluster_per_flow_2_tuple:
>>>>>   rc = pfring_set_cluster(pd, clusterId, cluster_per_flow_2_tuple);
>>>>> - compiled pfcount
>>>>> 
>>>>> TEST #1
>>>>> - downloaded http.cap from wireshark.org:
>>>>> http://wiki.wireshark.org/SampleCaptures?action=AttachFile&do=get&target=http.cap
>>>>> - capinfos reports 43 packets in the file:
>>>>> capinfos -c http.cap
>>>>> File name:           http.cap
>>>>> Number of packets:   43
>>>>> - replayed pcap using:
>>>>> sudo tcpreplay -ieth0 -t http.cap
>>>>> - running a single instance of pfcount results in all 43 packets received
>>>>> - adding a second instance of pfcount with the same clusterId results
>>>>> in all 43 packets received by the first instance
>>>>> - adding a third instance of pfcount results in only 2 packets being
>>>>> seen by the first instance, 7 packets being seen by the second
>>>>> instance, and 0 packets being seen by the third instance
>>>>> 
>>>>> TEST #2
>>>>> - downloaded http_gzip.cap from wireshark.org:
>>>>> http://wiki.wireshark.org/SampleCaptures?action=AttachFile&do=get&target=http_gzip.cap
>>>>> - capinfos reports 10 packets in the file:
>>>>> capinfos -c http_gzip.cap
>>>>> File name:           http_gzip.cap
>>>>> Number of packets:   10
>>>>> - replayed pcap using:
>>>>> sudo tcpreplay -ieth0 -t http_gzip.cap
>>>>> - running a single instance of pfcount results in all 10 packets received
>>>>> - adding a second instance of pfcount with the same clusterId results
>>>>> in 0 packets received by both instances
>>>>> - adding a third instance of pfcount with the same clusterId results
>>>>> in 0 packets received by all three instances
>>>>> 
>>>>> What am I missing?
>>>>> 
>>>>> Can somebody please try the tests above and let me know what results you 
>>>>> get?
>>>>> 
>>>>> Thanks!
>>>>> 
>>>>> Doug
>>>>> 
>>>>> On Tue, Jun 4, 2013 at 7:27 PM, Doug Burks <[email protected]> wrote:
>>>>>> I pulled new code from svn, compiled and inserted the new kernel
>>>>>> module, and verified that I get the same results.
>>>>>> 
>>>>>> I see this in the 5.5.3 Changelog:
>>>>>> 
>>>>>> - Added ability to balance tunneled/fragmented packets with the cluster
>>>>>> 
>>>>>> Is it possible that this change is affecting the hashing mechanism?
>>>>>> 
>>>>>> Anything else I can try?
>>>>>> 
>>>>>> Thanks,
>>>>>> Doug
>>>>>> 
>>>>>> 
>>>>>> On Tue, Jun 4, 2013 at 6:40 AM, Alfredo Cardigliano
>>>>>> <[email protected]> wrote:
>>>>>>> Good morning Doug
>>>>>>> I received the pcap but I was traveling, I will check them asap
>>>>>>> 
>>>>>>> Thanks
>>>>>>> Alfredo
>>>>>>> 
>>>>>>> On Jun 4, 2013, at 12:30 PM, Doug Burks <[email protected]> wrote:
>>>>>>> 
>>>>>>>> Good morning Alfredo,
>>>>>>>> 
>>>>>>>> Just wanted to follow up and confirm that you received the 5 pcaps I
>>>>>>>> sent off-list yesterday.
>>>>>>>> 
>>>>>>>> Is there anything else I can provide to help troubleshoot this issue?
>>>>>>>> 
>>>>>>>> Thanks!
>>>>>>>> 
>>>>>>>> Doug
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Mon, Jun 3, 2013 at 7:24 AM, Doug Burks <[email protected]> 
>>>>>>>> wrote:
>>>>>>>>> On Mon, Jun 3, 2013 at 3:39 AM, Alfredo Cardigliano
>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>> Doug
>>>>>>>>>> I don't think the support for packet injection is going to interfere 
>>>>>>>>>> your test.
>>>>>>>>>> Could you try sending packets from another interface?
>>>>>>>>> 
>>>>>>>>> I've confirmed this behavior using tcpreplay in a VM and also on a
>>>>>>>>> physical sensor connected to a tap.
>>>>>>>>> 
>>>>>>>>>> Could you provide me the original pcap you are using and the 
>>>>>>>>>> produced pcaps?
>>>>>>>>> 
>>>>>>>>> Sent off-list.
>>>>>>>>> 
>>>>>>>>> Please let me know if there is anything else I can provide to help
>>>>>>>>> troubleshoot this issue.
>>>>>>>>> 
>>>>>>>>> Thanks!
>>>>>>>>> 
>>>>>>>>> Doug
>>>>>>>>>> 
>>>>>>>>>> Thanks
>>>>>>>>>> Alfredo
>>>>>>>>>> 
>>>>>>>>>> On Jun 2, 2013, at 11:40 PM, Doug Burks <[email protected]> wrote:
>>>>>>>>>> 
>>>>>>>>>>> I see this in the Changelog:
>>>>>>>>>>> 
>>>>>>>>>>> - Support for injecting packets to the stack
>>>>>>>>>>> 
>>>>>>>>>>> Is it possible that this change could have an impact on my test 
>>>>>>>>>>> since
>>>>>>>>>>> I'm using tcpreplay?
>>>>>>>>>>> 
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Doug
>>>>>>>>>>> 
>>>>>>>>>>> On Sun, Jun 2, 2013 at 2:59 PM, Doug Burks <[email protected]> 
>>>>>>>>>>> wrote:
>>>>>>>>>>>> cat /proc/net/pf_ring/info
>>>>>>>>>>>> 
>>>>>>>>>>>> PF_RING Version          : 5.5.3 ($Revision: $)
>>>>>>>>>>>> Total rings              : 2
>>>>>>>>>>>> 
>>>>>>>>>>>> Standard (non DNA) Options
>>>>>>>>>>>> Ring slots               : 4096
>>>>>>>>>>>> Slot version             : 15
>>>>>>>>>>>> Capture TX               : Yes [RX+TX]
>>>>>>>>>>>> IP Defragment            : No
>>>>>>>>>>>> Socket Mode              : Standard
>>>>>>>>>>>> Transparent mode         : Yes [mode 0]
>>>>>>>>>>>> Total plugins            : 0
>>>>>>>>>>>> Cluster Fragment Queue   : 0
>>>>>>>>>>>> Cluster Fragment Discard : 16830
>>>>>>>>>>>> 
>>>>>>>>>>>> I've tried a few different pcaps, some of them are like my 
>>>>>>>>>>>> testmyids
>>>>>>>>>>>> sample in that no packets make it to pfdump, others work perfectly,
>>>>>>>>>>>> while for others it looks like only some of the packets are making 
>>>>>>>>>>>> it
>>>>>>>>>>>> into pfdump:
>>>>>>>>>>>> 
>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 
>>>>>>>>>>>> /opt/samples/markofu/honeynet_suspicious-time.pcap
>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>> processing file: /opt/samples/markofu/honeynet_suspicious-time.pcap
>>>>>>>>>>>> Actual: 745 packets (293958 bytes) sent in 0.32 seconds
>>>>>>>>>>>> Rated: 918618.8 bps, 7.01 Mbps, 2328.12 pps
>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>> Attempted packets:         745
>>>>>>>>>>>> Successful packets:        745
>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>> 
>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 11 sec pkts 257 drop 0 bytes 81262 | pkts 257 bytes 81262 drop 0
>>>>>>>>>>>> 12 sec pkts 136 drop 0 bytes 72265 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 15 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 16 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>> 18 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>> 10 sec pkts 21 drop 0 bytes 6352 | pkts 21 bytes 6352 drop 0
>>>>>>>>>>>> 11 sec pkts 15 drop 0 bytes 3640 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 15 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 16 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>> 
>>>>>>>>>>>> What else can I test?
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks!
>>>>>>>>>>>> 
>>>>>>>>>>>> Doug
>>>>>>>>>>>> 
>>>>>>>>>>>> On Sun, Jun 2, 2013 at 2:07 PM, Alfredo Cardigliano
>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>> Doug
>>>>>>>>>>>>> I ran a test using curl + pfcount and it is working for me.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> $ curl testmyids.com
>>>>>>>>>>>>> 
>>>>>>>>>>>>> (first instance)
>>>>>>>>>>>>> $ ./pfcount -i eth0 -c 99 -v 1 -m
>>>>>>>>>>>>> ...
>>>>>>>>>>>>> Absolute Stats: [0 pkts rcvd][0 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>> 
>>>>>>>>>>>>> (second instance)
>>>>>>>>>>>>> $ ./pfcount -i eth0 -c 99 -v 1 -m
>>>>>>>>>>>>> ...
>>>>>>>>>>>>> Absolute Stats: [11 pkts rcvd][11 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Please make sure tx capture is enabled in your test (cat 
>>>>>>>>>>>>> /proc/net/pf_ring/info)
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Alfredo
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On Jun 2, 2013, at 7:43 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi Alfredo,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks for your suggestion!
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I've changed pfdump.c to use cluster_per_flow_2_tuple:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> if(clusterId > 0) {
>>>>>>>>>>>>>> rc = pfring_set_cluster(pd, clusterId, cluster_per_flow_2_tuple);
>>>>>>>>>>>>>> printf("pfring_set_cluster returned %d\n", rc);
>>>>>>>>>>>>>> }
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I then re-ran the test as follows:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Replayed a TCP stream with 11 packets onto eth1:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 testmyids.pcap
>>>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>>>> processing file: testmyids.pcap
>>>>>>>>>>>>>> Actual: 11 packets (1062 bytes) sent in 0.00 seconds
>>>>>>>>>>>>>> Rated: inf bps, inf Mbps, inf pps
>>>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>>>> Attempted packets:         11
>>>>>>>>>>>>>> Successful packets:        11
>>>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Ran two instances of pfdump on eth1 with same clusterId but 
>>>>>>>>>>>>>> neither of
>>>>>>>>>>>>>> them saw traffic this time:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 11 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 11 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> tcpdump -nnvvr instance1.pcap
>>>>>>>>>>>>>> reading from file instance1.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> tcpdump -nnvvr instance2.pcap
>>>>>>>>>>>>>> reading from file instance2.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I've repeated this a few times and get the same result each time.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Any ideas why cluster_per_flow_2_tuple wouldn't be passing the 
>>>>>>>>>>>>>> traffic?
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 12:41 PM, Alfredo Cardigliano
>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>> Hi Doug
>>>>>>>>>>>>>>> the code in pfcount sets  the cluster mode to round-robin,
>>>>>>>>>>>>>>> for flow coherency you should change it to (for instance)
>>>>>>>>>>>>>>> cluster_per_flow_2_tuple.
>>>>>>>>>>>>>>> The daq-pfring code sets the cluster mode to 
>>>>>>>>>>>>>>> cluster_per_flow_2_tuple by
>>>>>>>>>>>>>>> default.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>> Alfredo
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Index: pfcount.c
>>>>>>>>>>>>>>> ===================================================================
>>>>>>>>>>>>>>> --- pfcount.c (revisione 6336)
>>>>>>>>>>>>>>> +++ pfcount.c (copia locale)
>>>>>>>>>>>>>>> @@ -924,7 +924,7 @@
>>>>>>>>>>>>>>> #endif
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> if(clusterId > 0) {
>>>>>>>>>>>>>>> -    rc = pfring_set_cluster(pd, clusterId, 
>>>>>>>>>>>>>>> cluster_round_robin);
>>>>>>>>>>>>>>> +    rc = pfring_set_cluster(pd, clusterId, 
>>>>>>>>>>>>>>> cluster_per_flow_2_tuple);
>>>>>>>>>>>>>>> printf("pfring_set_cluster returned %d\n", rc);
>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Jun 2, 2013, at 2:54 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I copied the clusterId code from pfcount and pasted into pfdump 
>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>> compiled it.  Then tested with a fresh pcap of "curl 
>>>>>>>>>>>>>>> testmyids.com":
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> tcpdump -nnr testmyids.pcap
>>>>>>>>>>>>>>> reading from file testmyids.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>> 12:37:21.846561 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [S],
>>>>>>>>>>>>>>> seq 2183306783, win 42340, options [mss 1460,sackOK,TS val 
>>>>>>>>>>>>>>> 13599714
>>>>>>>>>>>>>>> ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>> 12:37:21.963023 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [S.], seq 3354284181, ack 2183306784, win 64240, options [mss 
>>>>>>>>>>>>>>> 1460],
>>>>>>>>>>>>>>> length 0
>>>>>>>>>>>>>>> 12:37:21.963070 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 1, win 42340, length 0
>>>>>>>>>>>>>>> 12:37:21.963268 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 42340, length 165
>>>>>>>>>>>>>>> 12:37:21.963423 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 166, win 64240, length 0
>>>>>>>>>>>>>>> 12:37:22.083864 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 64240, length 259
>>>>>>>>>>>>>>> 12:37:22.083906 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 260, win 42081, length 0
>>>>>>>>>>>>>>> 12:37:22.084118 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 42081, length 0
>>>>>>>>>>>>>>> 12:37:22.085362 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 167, win 64239, length 0
>>>>>>>>>>>>>>> 12:37:22.202741 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [FP.], seq 260, ack 167, win 64239, length 0
>>>>>>>>>>>>>>> 12:37:22.202786 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 261, win 42081, length 0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I then started the two instances of pfdump using the same 
>>>>>>>>>>>>>>> clusterId
>>>>>>>>>>>>>>> and then replayed the 11 packets with tcpreplay:
>>>>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 testmyids.pcap
>>>>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>>>>> processing file: testmyids.pcap
>>>>>>>>>>>>>>> Actual: 11 packets (1062 bytes) sent in 0.01 seconds
>>>>>>>>>>>>>>> Rated: 106200.0 bps, 0.81 Mbps, 1100.00 pps
>>>>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>>>>> Attempted packets:         11
>>>>>>>>>>>>>>> Successful packets:        11
>>>>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> FIRST INSTANCE OF PFDUMP
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>> 241 sec pkts 6 drop 0 bytes 500 | pkts 6 bytes 500 drop 0
>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> tcpdump -nnr instance1.pcap
>>>>>>>>>>>>>>> reading from file instance1.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>> 12:38:55.886037 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [S],
>>>>>>>>>>>>>>> seq 2183306783, win 42340, options [mss 1460,sackOK,TS val 
>>>>>>>>>>>>>>> 13599714
>>>>>>>>>>>>>>> ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>> 12:38:55.886889 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 3354284182, win 42340, length 0
>>>>>>>>>>>>>>> 12:38:55.887325 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 165, win 64240, length 0
>>>>>>>>>>>>>>> 12:38:55.887986 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 260, win 42081, length 0
>>>>>>>>>>>>>>> 12:38:55.888306 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 166, win 64239, length 0
>>>>>>>>>>>>>>> 12:38:55.888741 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>> ack 261, win 42081, length 0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> SECOND INSTANCE OF PFDUMP
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>> 16 sec pkts 5 drop 0 bytes 826 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>> 18 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>> 19 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>>> 20 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> tcpdump -nnr instance2.pcap
>>>>>>>>>>>>>>> reading from file instance2.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>> 12:38:55.886499 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [S.], seq 3354284181, ack 2183306784, win 64240, options [mss 
>>>>>>>>>>>>>>> 1460],
>>>>>>>>>>>>>>> length 0
>>>>>>>>>>>>>>> 12:38:55.887129 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 42340, length 165
>>>>>>>>>>>>>>> 12:38:55.887666 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 64240, length 259
>>>>>>>>>>>>>>> 12:38:55.888117 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 42081, length 0
>>>>>>>>>>>>>>> 12:38:55.888530 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [FP.], seq 260, ack 167, win 64239, length 0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> As you can see, the first instance sees 6 packets and the second
>>>>>>>>>>>>>>> instance sees 5 packets.  Shouldn't all 11 packets in that TCP 
>>>>>>>>>>>>>>> stream
>>>>>>>>>>>>>>> be sent to the same instance?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 7:11 AM, Doug Burks 
>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi Luca,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I can repeat the test with pfdump when I'm back at my computer, 
>>>>>>>>>>>>>>> but is there
>>>>>>>>>>>>>>> something in particular you're looking for that wasn't in the 
>>>>>>>>>>>>>>> pfcount output
>>>>>>>>>>>>>>> I provided?  Shouldn't all the traffic from that one TCP stream 
>>>>>>>>>>>>>>> be sent to
>>>>>>>>>>>>>>> one instance of pfcount?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Sunday, June 2, 2013, Luca Deri wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>>> You're right. We need to add it: you can c&p the code from 
>>>>>>>>>>>>>>> pfcount in the
>>>>>>>>>>>>>>> meantime
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Luca
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Jun 2, 2013, at 1:54 AM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I have pfdump now but I don't see a cluster-id option.  Did you 
>>>>>>>>>>>>>>> mean
>>>>>>>>>>>>>>> pfcount?  If I run 2 instances of pfcount with the same 
>>>>>>>>>>>>>>> cluster-id and
>>>>>>>>>>>>>>> then replay a pcap with 10 packets all belonging to the same TCP
>>>>>>>>>>>>>>> stream, I get 5 packets being sent to each pfcount instance.
>>>>>>>>>>>>>>> Shouldn't all 10 packets be sent to 1 instance?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> First instance:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> sudo ./pfcount -c77 -i eth1
>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> Absolute Stats: [5 pkts rcvd][5 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>> Total Pkts=5/Dropped=0.0 %
>>>>>>>>>>>>>>> 5 pkts - 434 bytes [0.38 pkt/sec - 0.00 Mbit/sec]
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> Actual Stats: 5 pkts [1'000.75 ms][5.00 pps/0.00 Gbps]
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Second instance:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> sudo ./pfcount -c77 -i eth1
>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> Absolute Stats: [5 pkts rcvd][5 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>> Total Pkts=5/Dropped=0.0 %
>>>>>>>>>>>>>>> 5 pkts - 834 bytes [0.62 pkt/sec - 0.00 Mbit/sec]
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> Actual Stats: 5 pkts [1'001.39 ms][4.99 pps/0.00 Gbps]
>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> The replayed pcap is just ten packets that result from "curl
>>>>>>>>>>>>>>> testmyids.com":
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> tcpdump -nnr testmyids.pcap
>>>>>>>>>>>>>>> reading from file testmyids.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>> 11:46:11.691648 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [S], seq 3840903154, win 42340, options [mss 1460,sackOK,TS val
>>>>>>>>>>>>>>> 20137183 ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>> 11:46:11.808833 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [S.], seq 2859277445, ack 3840903155, win 5840, options [mss
>>>>>>>>>>>>>>> 1460,nop,wscale 7], length 0
>>>>>>>>>>>>>>> 11:46:11.808854 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [.], ack 1, win 21, length 0
>>>>>>>>>>>>>>> 11:46:11.809083 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 21, length 165
>>>>>>>>>>>>>>> 11:46:11.927518 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [.], ack 166, win 54, length 0
>>>>>>>>>>>>>>> 11:46:12.036708 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 54, length 259
>>>>>>>>>>>>>>> 11:46:12.036956 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [.], ack 260, win 21, length 0
>>>>>>>>>>>>>>> 11:46:12.037206 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 21, length 0
>>>>>>>>>>>>>>> 11:46:12.154641 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [F.], seq 260, ack 167, win 54, length 0
>>>>>>>>>>>>>>> 11:46:12.154888 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>> [.], ack 261, win 21, length 0
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Any ideas?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Sat, Jun 1, 2013 at 5:48 PM, Doug Burks 
>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Sat, Jun 1, 2013 at 10:24 AM, Luca Deri <[email protected]> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi Doug
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Jun 1, 2013, at 6:59 AM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hello all,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I recently packaged PF_RING 5.5.3 for my Security Onion distro:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> http://securityonion.blogspot.com/2013/05/pfring-553-packages-now-available.html
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Perhaps I'm missing something, but I'm seeing some behavior I 
>>>>>>>>>>>>>>> don't
>>>>>>>>>>>>>>> remember seeing in 5.5.2 or previous versions of PF_RING.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Here are my testing parameters:
>>>>>>>>>>>>>>> - starting off with a good test, if I run just one instance of 
>>>>>>>>>>>>>>> snort,
>>>>>>>>>>>>>>> I get an alert from rule 2100498 for EACH time I run "curl
>>>>>>>>>>>>>>> testmyids.com"
>>>>>>>>>>>>>>> - if I increase to two instances of snort with the same 
>>>>>>>>>>>>>>> cluster-id, I
>>>>>>>>>>>>>>> get NO alerts when running "curl testmyids.com"
>>>>>>>>>>>>>>> - if I set the daq clustermode to 2, I get NO alerts when 
>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>> "curl > _______________________________________________
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>> 
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> --
>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> --
>>>>>>>>>>> Doug Burks
>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>> [email protected]
>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>> 
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>> [email protected]
>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> --
>>>>>>>>> Doug Burks
>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Doug Burks
>>>>>>>> http://securityonion.blogspot.com
>>>>>>>> _______________________________________________
>>>>>>>> Ntop-misc mailing list
>>>>>>>> [email protected]
>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> Ntop-misc mailing list
>>>>>>> [email protected]
>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> --
>>>>>> Doug Burks
>>>>>> http://securityonion.blogspot.com
>>>>> 
>>>>> 
>>>>> 
>>>>> --
>>>>> Doug Burks
>>>>> http://securityonion.blogspot.com
>>>> 
>>>> 
>>>> 
>>>> --
>>>> Doug Burks
>>>> http://securityonion.blogspot.com
>>> 
>>> 
>>> 
>>> --
>>> Doug Burks
>>> http://securityonion.blogspot.com
>>> _______________________________________________
>>> Ntop-misc mailing list
>>> [email protected]
>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>> 
>> _______________________________________________
>> Ntop-misc mailing list
>> [email protected]
>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> 
> 
> 
> -- 
> Doug Burks
> http://securityonion.blogspot.com
> _______________________________________________
> Ntop-misc mailing list
> [email protected]
> http://listgateway.unipi.it/mailman/listinfo/ntop-misc

_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Reply via email to