On Jun 5, 2013, at 6:29 PM, Doug Burks <[email protected]> wrote:

> OK, so you're just going to use the pf_ring.c from svn and keep
> userland as it was?
> 
> In other words, for my Security Onion distro, I can safely keep our
> existing 5.5.3 userland package and just update our kernel module
> package with the pf_ring.c from svn?

Yes

Alfredo

> 
> Thanks!
> 
> Doug
> 
> On Wed, Jun 5, 2013 at 11:51 AM, Alfredo Cardigliano
> <[email protected]> wrote:
>> Hi Doug
>> we released 5.5.3 a few days ago, it is likely we refresh that tarball.
>> 
>> Regards
>> Alfredo
>> 
>> On Jun 5, 2013, at 5:45 PM, Doug Burks <[email protected]> wrote:
>> 
>>> Hi Alfredo,
>>> 
>>> Thanks for the fix!  I did a few quick tests and it appears to be
>>> working properly.
>>> 
>>> Will you be releasing a 5.5.4 tarball soon?
>>> 
>>> Thanks,
>>> Doug
>>> 
>>> On Wed, Jun 5, 2013 at 11:33 AM, Alfredo Cardigliano
>>> <[email protected]> wrote:
>>>> Hi Doug
>>>> thank you for the info you reported, as you guessed it was a bug in the 
>>>> fragments handling.
>>>> Please update from svn, it should be fixed now.
>>>> 
>>>> Regards
>>>> Alfredo
>>>> 
>>>> On Jun 5, 2013, at 4:20 PM, Doug Burks <[email protected]> wrote:
>>>> 
>>>>> Here's an additional test on 5.5.3 with http_gzip.pcap (with 10 packets):
>>>>> 
>>>>> - running a single instance of pfcount everything works properly and
>>>>> /proc/net/pf_ring/info shows:
>>>>> PF_RING Version          : 5.5.3 ($Revision: exported$)
>>>>> Total rings              : 1
>>>>> 
>>>>> Standard (non DNA) Options
>>>>> Ring slots               : 4096
>>>>> Slot version             : 15
>>>>> Capture TX               : Yes [RX+TX]
>>>>> IP Defragment            : No
>>>>> Socket Mode              : Standard
>>>>> Transparent mode         : Yes [mode 0]
>>>>> Total plugins            : 0
>>>>> Cluster Fragment Queue   : 0
>>>>> Cluster Fragment Discard : 0
>>>>> 
>>>>> - running two instances of pfcount with the same clusterId shows no
>>>>> packets in either pfcount instance and /proc/net/pf_ring/info shows:
>>>>> PF_RING Version          : 5.5.3 ($Revision: exported$)
>>>>> Total rings              : 2
>>>>> 
>>>>> Standard (non DNA) Options
>>>>> Ring slots               : 4096
>>>>> Slot version             : 15
>>>>> Capture TX               : Yes [RX+TX]
>>>>> IP Defragment            : No
>>>>> Socket Mode              : Standard
>>>>> Transparent mode         : Yes [mode 0]
>>>>> Total plugins            : 0
>>>>> Cluster Fragment Queue   : 0
>>>>> Cluster Fragment Discard : 10
>>>>> 
>>>>> Any ideas why it's discarding the 10 packets?
>>>>> 
>>>>> Thanks!
>>>>> 
>>>>> Doug
>>>>> 
>>>>> On Wed, Jun 5, 2013 at 8:44 AM, Doug Burks <[email protected]> wrote:
>>>>>> I repeated these tests on the same Ubuntu VM using PF_RING 5.5.2
>>>>>> kernel/userland and clustering works as expected without the packet
>>>>>> loss seen in 5.5.3.
>>>>>> 
>>>>>> Perhaps I'm missing something, but it appears that there is a bug in
>>>>>> the 5.5.3 cluster code.
>>>>>> 
>>>>>> Can somebody confirm, please?
>>>>>> 
>>>>>> Thanks!
>>>>>> 
>>>>>> Doug
>>>>>> 
>>>>>> 
>>>>>> On Wed, Jun 5, 2013 at 8:06 AM, Doug Burks <[email protected]> wrote:
>>>>>>> I just did a new test as follows:
>>>>>>> 
>>>>>>> - started with a fresh installation of Ubuntu 12.04
>>>>>>> - downloaded the PF_RING 5.5.3 tarball
>>>>>>> - compiled and inserted the kernel module
>>>>>>> - changed pfcount.c to use cluster_per_flow_2_tuple:
>>>>>>>  rc = pfring_set_cluster(pd, clusterId, cluster_per_flow_2_tuple);
>>>>>>> - compiled pfcount
>>>>>>> 
>>>>>>> TEST #1
>>>>>>> - downloaded http.cap from wireshark.org:
>>>>>>> http://wiki.wireshark.org/SampleCaptures?action=AttachFile&do=get&target=http.cap
>>>>>>> - capinfos reports 43 packets in the file:
>>>>>>> capinfos -c http.cap
>>>>>>> File name:           http.cap
>>>>>>> Number of packets:   43
>>>>>>> - replayed pcap using:
>>>>>>> sudo tcpreplay -ieth0 -t http.cap
>>>>>>> - running a single instance of pfcount results in all 43 packets 
>>>>>>> received
>>>>>>> - adding a second instance of pfcount with the same clusterId results
>>>>>>> in all 43 packets received by the first instance
>>>>>>> - adding a third instance of pfcount results in only 2 packets being
>>>>>>> seen by the first instance, 7 packets being seen by the second
>>>>>>> instance, and 0 packets being seen by the third instance
>>>>>>> 
>>>>>>> TEST #2
>>>>>>> - downloaded http_gzip.cap from wireshark.org:
>>>>>>> http://wiki.wireshark.org/SampleCaptures?action=AttachFile&do=get&target=http_gzip.cap
>>>>>>> - capinfos reports 10 packets in the file:
>>>>>>> capinfos -c http_gzip.cap
>>>>>>> File name:           http_gzip.cap
>>>>>>> Number of packets:   10
>>>>>>> - replayed pcap using:
>>>>>>> sudo tcpreplay -ieth0 -t http_gzip.cap
>>>>>>> - running a single instance of pfcount results in all 10 packets 
>>>>>>> received
>>>>>>> - adding a second instance of pfcount with the same clusterId results
>>>>>>> in 0 packets received by both instances
>>>>>>> - adding a third instance of pfcount with the same clusterId results
>>>>>>> in 0 packets received by all three instances
>>>>>>> 
>>>>>>> What am I missing?
>>>>>>> 
>>>>>>> Can somebody please try the tests above and let me know what results 
>>>>>>> you get?
>>>>>>> 
>>>>>>> Thanks!
>>>>>>> 
>>>>>>> Doug
>>>>>>> 
>>>>>>> On Tue, Jun 4, 2013 at 7:27 PM, Doug Burks <[email protected]> wrote:
>>>>>>>> I pulled new code from svn, compiled and inserted the new kernel
>>>>>>>> module, and verified that I get the same results.
>>>>>>>> 
>>>>>>>> I see this in the 5.5.3 Changelog:
>>>>>>>> 
>>>>>>>> - Added ability to balance tunneled/fragmented packets with the cluster
>>>>>>>> 
>>>>>>>> Is it possible that this change is affecting the hashing mechanism?
>>>>>>>> 
>>>>>>>> Anything else I can try?
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> Doug
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Tue, Jun 4, 2013 at 6:40 AM, Alfredo Cardigliano
>>>>>>>> <[email protected]> wrote:
>>>>>>>>> Good morning Doug
>>>>>>>>> I received the pcap but I was traveling, I will check them asap
>>>>>>>>> 
>>>>>>>>> Thanks
>>>>>>>>> Alfredo
>>>>>>>>> 
>>>>>>>>> On Jun 4, 2013, at 12:30 PM, Doug Burks <[email protected]> wrote:
>>>>>>>>> 
>>>>>>>>>> Good morning Alfredo,
>>>>>>>>>> 
>>>>>>>>>> Just wanted to follow up and confirm that you received the 5 pcaps I
>>>>>>>>>> sent off-list yesterday.
>>>>>>>>>> 
>>>>>>>>>> Is there anything else I can provide to help troubleshoot this issue?
>>>>>>>>>> 
>>>>>>>>>> Thanks!
>>>>>>>>>> 
>>>>>>>>>> Doug
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> On Mon, Jun 3, 2013 at 7:24 AM, Doug Burks <[email protected]> 
>>>>>>>>>> wrote:
>>>>>>>>>>> On Mon, Jun 3, 2013 at 3:39 AM, Alfredo Cardigliano
>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>> Doug
>>>>>>>>>>>> I don't think the support for packet injection is going to 
>>>>>>>>>>>> interfere your test.
>>>>>>>>>>>> Could you try sending packets from another interface?
>>>>>>>>>>> 
>>>>>>>>>>> I've confirmed this behavior using tcpreplay in a VM and also on a
>>>>>>>>>>> physical sensor connected to a tap.
>>>>>>>>>>> 
>>>>>>>>>>>> Could you provide me the original pcap you are using and the 
>>>>>>>>>>>> produced pcaps?
>>>>>>>>>>> 
>>>>>>>>>>> Sent off-list.
>>>>>>>>>>> 
>>>>>>>>>>> Please let me know if there is anything else I can provide to help
>>>>>>>>>>> troubleshoot this issue.
>>>>>>>>>>> 
>>>>>>>>>>> Thanks!
>>>>>>>>>>> 
>>>>>>>>>>> Doug
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks
>>>>>>>>>>>> Alfredo
>>>>>>>>>>>> 
>>>>>>>>>>>> On Jun 2, 2013, at 11:40 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>> I see this in the Changelog:
>>>>>>>>>>>>> 
>>>>>>>>>>>>> - Support for injecting packets to the stack
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Is it possible that this change could have an impact on my test 
>>>>>>>>>>>>> since
>>>>>>>>>>>>> I'm using tcpreplay?
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Doug
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 2:59 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>> cat /proc/net/pf_ring/info
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> PF_RING Version          : 5.5.3 ($Revision: $)
>>>>>>>>>>>>>> Total rings              : 2
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Standard (non DNA) Options
>>>>>>>>>>>>>> Ring slots               : 4096
>>>>>>>>>>>>>> Slot version             : 15
>>>>>>>>>>>>>> Capture TX               : Yes [RX+TX]
>>>>>>>>>>>>>> IP Defragment            : No
>>>>>>>>>>>>>> Socket Mode              : Standard
>>>>>>>>>>>>>> Transparent mode         : Yes [mode 0]
>>>>>>>>>>>>>> Total plugins            : 0
>>>>>>>>>>>>>> Cluster Fragment Queue   : 0
>>>>>>>>>>>>>> Cluster Fragment Discard : 16830
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I've tried a few different pcaps, some of them are like my 
>>>>>>>>>>>>>> testmyids
>>>>>>>>>>>>>> sample in that no packets make it to pfdump, others work 
>>>>>>>>>>>>>> perfectly,
>>>>>>>>>>>>>> while for others it looks like only some of the packets are 
>>>>>>>>>>>>>> making it
>>>>>>>>>>>>>> into pfdump:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 
>>>>>>>>>>>>>> /opt/samples/markofu/honeynet_suspicious-time.pcap
>>>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>>>> processing file: 
>>>>>>>>>>>>>> /opt/samples/markofu/honeynet_suspicious-time.pcap
>>>>>>>>>>>>>> Actual: 745 packets (293958 bytes) sent in 0.32 seconds
>>>>>>>>>>>>>> Rated: 918618.8 bps, 7.01 Mbps, 2328.12 pps
>>>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>>>> Attempted packets:         745
>>>>>>>>>>>>>> Successful packets:        745
>>>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 11 sec pkts 257 drop 0 bytes 81262 | pkts 257 bytes 81262 drop 0
>>>>>>>>>>>>>> 12 sec pkts 136 drop 0 bytes 72265 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 15 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 16 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>> 18 sec pkts 0 drop 0 bytes 0 | pkts 393 bytes 153527 drop 0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>> 10 sec pkts 21 drop 0 bytes 6352 | pkts 21 bytes 6352 drop 0
>>>>>>>>>>>>>> 11 sec pkts 15 drop 0 bytes 3640 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 15 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 16 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 36 bytes 9992 drop 0
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> What else can I test?
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 2:07 PM, Alfredo Cardigliano
>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>> I ran a test using curl + pfcount and it is working for me.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> $ curl testmyids.com
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> (first instance)
>>>>>>>>>>>>>>> $ ./pfcount -i eth0 -c 99 -v 1 -m
>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>> Absolute Stats: [0 pkts rcvd][0 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> (second instance)
>>>>>>>>>>>>>>> $ ./pfcount -i eth0 -c 99 -v 1 -m
>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>> Absolute Stats: [11 pkts rcvd][11 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Please make sure tx capture is enabled in your test (cat 
>>>>>>>>>>>>>>> /proc/net/pf_ring/info)
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Alfredo
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Jun 2, 2013, at 7:43 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Hi Alfredo,
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Thanks for your suggestion!
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> I've changed pfdump.c to use cluster_per_flow_2_tuple:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> if(clusterId > 0) {
>>>>>>>>>>>>>>>> rc = pfring_set_cluster(pd, clusterId, 
>>>>>>>>>>>>>>>> cluster_per_flow_2_tuple);
>>>>>>>>>>>>>>>> printf("pfring_set_cluster returned %d\n", rc);
>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> I then re-ran the test as follows:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Replayed a TCP stream with 11 packets onto eth1:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 testmyids.pcap
>>>>>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>>>>>> processing file: testmyids.pcap
>>>>>>>>>>>>>>>> Actual: 11 packets (1062 bytes) sent in 0.00 seconds
>>>>>>>>>>>>>>>> Rated: inf bps, inf Mbps, inf pps
>>>>>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>>>>>> Attempted packets:         11
>>>>>>>>>>>>>>>> Successful packets:        11
>>>>>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Ran two instances of pfdump on eth1 with same clusterId but 
>>>>>>>>>>>>>>>> neither of
>>>>>>>>>>>>>>>> them saw traffic this time:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 11 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>>>> 14 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>>> 1 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 2 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 3 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 4 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 5 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 6 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 7 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 8 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 9 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 10 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 11 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 12 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>>>> 13 sec pkts 0 drop 0 bytes 0 | pkts 0 bytes 0 drop 0
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> tcpdump -nnvvr instance1.pcap
>>>>>>>>>>>>>>>> reading from file instance1.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> tcpdump -nnvvr instance2.pcap
>>>>>>>>>>>>>>>> reading from file instance2.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> I've repeated this a few times and get the same result each 
>>>>>>>>>>>>>>>> time.
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Any ideas why cluster_per_flow_2_tuple wouldn't be passing the 
>>>>>>>>>>>>>>>> traffic?
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 12:41 PM, Alfredo Cardigliano
>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>> Hi Doug
>>>>>>>>>>>>>>>>> the code in pfcount sets  the cluster mode to round-robin,
>>>>>>>>>>>>>>>>> for flow coherency you should change it to (for instance)
>>>>>>>>>>>>>>>>> cluster_per_flow_2_tuple.
>>>>>>>>>>>>>>>>> The daq-pfring code sets the cluster mode to 
>>>>>>>>>>>>>>>>> cluster_per_flow_2_tuple by
>>>>>>>>>>>>>>>>> default.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>> Alfredo
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Index: pfcount.c
>>>>>>>>>>>>>>>>> ===================================================================
>>>>>>>>>>>>>>>>> --- pfcount.c (revisione 6336)
>>>>>>>>>>>>>>>>> +++ pfcount.c (copia locale)
>>>>>>>>>>>>>>>>> @@ -924,7 +924,7 @@
>>>>>>>>>>>>>>>>> #endif
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> if(clusterId > 0) {
>>>>>>>>>>>>>>>>> -    rc = pfring_set_cluster(pd, clusterId, 
>>>>>>>>>>>>>>>>> cluster_round_robin);
>>>>>>>>>>>>>>>>> +    rc = pfring_set_cluster(pd, clusterId, 
>>>>>>>>>>>>>>>>> cluster_per_flow_2_tuple);
>>>>>>>>>>>>>>>>> printf("pfring_set_cluster returned %d\n", rc);
>>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Jun 2, 2013, at 2:54 PM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> I copied the clusterId code from pfcount and pasted into 
>>>>>>>>>>>>>>>>> pfdump and
>>>>>>>>>>>>>>>>> compiled it.  Then tested with a fresh pcap of "curl 
>>>>>>>>>>>>>>>>> testmyids.com":
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> tcpdump -nnr testmyids.pcap
>>>>>>>>>>>>>>>>> reading from file testmyids.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>>> 12:37:21.846561 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [S],
>>>>>>>>>>>>>>>>> seq 2183306783, win 42340, options [mss 1460,sackOK,TS val 
>>>>>>>>>>>>>>>>> 13599714
>>>>>>>>>>>>>>>>> ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>>>> 12:37:21.963023 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [S.], seq 3354284181, ack 2183306784, win 64240, options [mss 
>>>>>>>>>>>>>>>>> 1460],
>>>>>>>>>>>>>>>>> length 0
>>>>>>>>>>>>>>>>> 12:37:21.963070 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 1, win 42340, length 0
>>>>>>>>>>>>>>>>> 12:37:21.963268 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 42340, length 165
>>>>>>>>>>>>>>>>> 12:37:21.963423 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 166, win 64240, length 0
>>>>>>>>>>>>>>>>> 12:37:22.083864 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 64240, length 259
>>>>>>>>>>>>>>>>> 12:37:22.083906 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 260, win 42081, length 0
>>>>>>>>>>>>>>>>> 12:37:22.084118 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 42081, length 0
>>>>>>>>>>>>>>>>> 12:37:22.085362 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 167, win 64239, length 0
>>>>>>>>>>>>>>>>> 12:37:22.202741 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [FP.], seq 260, ack 167, win 64239, length 0
>>>>>>>>>>>>>>>>> 12:37:22.202786 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 261, win 42081, length 0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> I then started the two instances of pfdump using the same 
>>>>>>>>>>>>>>>>> clusterId
>>>>>>>>>>>>>>>>> and then replayed the 11 packets with tcpreplay:
>>>>>>>>>>>>>>>>> sudo tcpreplay -i eth1 -M10 testmyids.pcap
>>>>>>>>>>>>>>>>> sending out eth1
>>>>>>>>>>>>>>>>> processing file: testmyids.pcap
>>>>>>>>>>>>>>>>> Actual: 11 packets (1062 bytes) sent in 0.01 seconds
>>>>>>>>>>>>>>>>> Rated: 106200.0 bps, 0.81 Mbps, 1100.00 pps
>>>>>>>>>>>>>>>>> Statistics for network device: eth1
>>>>>>>>>>>>>>>>> Attempted packets:         11
>>>>>>>>>>>>>>>>> Successful packets:        11
>>>>>>>>>>>>>>>>> Failed packets:            0
>>>>>>>>>>>>>>>>> Retried packets (ENOBUFS): 0
>>>>>>>>>>>>>>>>> Retried packets (EAGAIN):  0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> FIRST INSTANCE OF PFDUMP
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance1.pcap
>>>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>>>> 241 sec pkts 6 drop 0 bytes 500 | pkts 6 bytes 500 drop 0
>>>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> tcpdump -nnr instance1.pcap
>>>>>>>>>>>>>>>>> reading from file instance1.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>>> 12:38:55.886037 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [S],
>>>>>>>>>>>>>>>>> seq 2183306783, win 42340, options [mss 1460,sackOK,TS val 
>>>>>>>>>>>>>>>>> 13599714
>>>>>>>>>>>>>>>>> ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>>>> 12:38:55.886889 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 3354284182, win 42340, length 0
>>>>>>>>>>>>>>>>> 12:38:55.887325 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 165, win 64240, length 0
>>>>>>>>>>>>>>>>> 12:38:55.887986 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 260, win 42081, length 0
>>>>>>>>>>>>>>>>> 12:38:55.888306 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 166, win 64239, length 0
>>>>>>>>>>>>>>>>> 12:38:55.888741 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags [.],
>>>>>>>>>>>>>>>>> ack 261, win 42081, length 0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> SECOND INSTANCE OF PFDUMP
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> sudo ./pfdump -l77 -i eth1 -w instance2.pcap
>>>>>>>>>>>>>>>>> Using PF_RING v.5.5.3
>>>>>>>>>>>>>>>>> Capturing from eth1 [00:0C:29:5F:58:D8][ifIndex: 3]
>>>>>>>>>>>>>>>>> # Device RX channels: 1
>>>>>>>>>>>>>>>>> pfring_set_cluster returned 0
>>>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>>>> 16 sec pkts 5 drop 0 bytes 826 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>>>> 17 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>>>> 18 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>>>> 19 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>>>> ^CLeaving...
>>>>>>>>>>>>>>>>> 20 sec pkts 0 drop 0 bytes 0 | pkts 5 bytes 826 drop 0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> tcpdump -nnr instance2.pcap
>>>>>>>>>>>>>>>>> reading from file instance2.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>>> 12:38:55.886499 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [S.], seq 3354284181, ack 2183306784, win 64240, options [mss 
>>>>>>>>>>>>>>>>> 1460],
>>>>>>>>>>>>>>>>> length 0
>>>>>>>>>>>>>>>>> 12:38:55.887129 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 42340, length 165
>>>>>>>>>>>>>>>>> 12:38:55.887666 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 64240, length 259
>>>>>>>>>>>>>>>>> 12:38:55.888117 IP 172.16.116.128.44229 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 42081, length 0
>>>>>>>>>>>>>>>>> 12:38:55.888530 IP 217.160.51.31.80 > 172.16.116.128.44229: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [FP.], seq 260, ack 167, win 64239, length 0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> As you can see, the first instance sees 6 packets and the 
>>>>>>>>>>>>>>>>> second
>>>>>>>>>>>>>>>>> instance sees 5 packets.  Shouldn't all 11 packets in that 
>>>>>>>>>>>>>>>>> TCP stream
>>>>>>>>>>>>>>>>> be sent to the same instance?
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Sun, Jun 2, 2013 at 7:11 AM, Doug Burks 
>>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Hi Luca,
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> I can repeat the test with pfdump when I'm back at my 
>>>>>>>>>>>>>>>>> computer, but is there
>>>>>>>>>>>>>>>>> something in particular you're looking for that wasn't in the 
>>>>>>>>>>>>>>>>> pfcount output
>>>>>>>>>>>>>>>>> I provided?  Shouldn't all the traffic from that one TCP 
>>>>>>>>>>>>>>>>> stream be sent to
>>>>>>>>>>>>>>>>> one instance of pfcount?
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Sunday, June 2, 2013, Luca Deri wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>>>>> You're right. We need to add it: you can c&p the code from 
>>>>>>>>>>>>>>>>> pfcount in the
>>>>>>>>>>>>>>>>> meantime
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Luca
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Jun 2, 2013, at 1:54 AM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> I have pfdump now but I don't see a cluster-id option.  Did 
>>>>>>>>>>>>>>>>> you mean
>>>>>>>>>>>>>>>>> pfcount?  If I run 2 instances of pfcount with the same 
>>>>>>>>>>>>>>>>> cluster-id and
>>>>>>>>>>>>>>>>> then replay a pcap with 10 packets all belonging to the same 
>>>>>>>>>>>>>>>>> TCP
>>>>>>>>>>>>>>>>> stream, I get 5 packets being sent to each pfcount instance.
>>>>>>>>>>>>>>>>> Shouldn't all 10 packets be sent to 1 instance?
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> First instance:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> sudo ./pfcount -c77 -i eth1
>>>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> Absolute Stats: [5 pkts rcvd][5 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>>>> Total Pkts=5/Dropped=0.0 %
>>>>>>>>>>>>>>>>> 5 pkts - 434 bytes [0.38 pkt/sec - 0.00 Mbit/sec]
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> Actual Stats: 5 pkts [1'000.75 ms][5.00 pps/0.00 Gbps]
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Second instance:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> sudo ./pfcount -c77 -i eth1
>>>>>>>>>>>>>>>>> <snip>
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> Absolute Stats: [5 pkts rcvd][5 pkts filtered][0 pkts dropped]
>>>>>>>>>>>>>>>>> Total Pkts=5/Dropped=0.0 %
>>>>>>>>>>>>>>>>> 5 pkts - 834 bytes [0.62 pkt/sec - 0.00 Mbit/sec]
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> Actual Stats: 5 pkts [1'001.39 ms][4.99 pps/0.00 Gbps]
>>>>>>>>>>>>>>>>> =========================
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> The replayed pcap is just ten packets that result from "curl
>>>>>>>>>>>>>>>>> testmyids.com":
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> tcpdump -nnr testmyids.pcap
>>>>>>>>>>>>>>>>> reading from file testmyids.pcap, link-type EN10MB (Ethernet)
>>>>>>>>>>>>>>>>> 11:46:11.691648 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [S], seq 3840903154, win 42340, options [mss 1460,sackOK,TS 
>>>>>>>>>>>>>>>>> val
>>>>>>>>>>>>>>>>> 20137183 ecr 0,nop,wscale 11], length 0
>>>>>>>>>>>>>>>>> 11:46:11.808833 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [S.], seq 2859277445, ack 3840903155, win 5840, options [mss
>>>>>>>>>>>>>>>>> 1460,nop,wscale 7], length 0
>>>>>>>>>>>>>>>>> 11:46:11.808854 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [.], ack 1, win 21, length 0
>>>>>>>>>>>>>>>>> 11:46:11.809083 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:166, ack 1, win 21, length 165
>>>>>>>>>>>>>>>>> 11:46:11.927518 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [.], ack 166, win 54, length 0
>>>>>>>>>>>>>>>>> 11:46:12.036708 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [P.], seq 1:260, ack 166, win 54, length 259
>>>>>>>>>>>>>>>>> 11:46:12.036956 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [.], ack 260, win 21, length 0
>>>>>>>>>>>>>>>>> 11:46:12.037206 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [F.], seq 166, ack 260, win 21, length 0
>>>>>>>>>>>>>>>>> 11:46:12.154641 IP 217.160.51.31.80 > 192.168.111.111.50154: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [F.], seq 260, ack 167, win 54, length 0
>>>>>>>>>>>>>>>>> 11:46:12.154888 IP 192.168.111.111.50154 > 217.160.51.31.80: 
>>>>>>>>>>>>>>>>> Flags
>>>>>>>>>>>>>>>>> [.], ack 261, win 21, length 0
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Any ideas?
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> Doug
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Sat, Jun 1, 2013 at 5:48 PM, Doug Burks 
>>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Sat, Jun 1, 2013 at 10:24 AM, Luca Deri <[email protected]> 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Hi Doug
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Jun 1, 2013, at 6:59 AM, Doug Burks <[email protected]> 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Hello all,
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> I recently packaged PF_RING 5.5.3 for my Security Onion 
>>>>>>>>>>>>>>>>> distro:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> http://securityonion.blogspot.com/2013/05/pfring-553-packages-now-available.html
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Perhaps I'm missing something, but I'm seeing some behavior I 
>>>>>>>>>>>>>>>>> don't
>>>>>>>>>>>>>>>>> remember seeing in 5.5.2 or previous versions of PF_RING.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Here are my testing parameters:
>>>>>>>>>>>>>>>>> - starting off with a good test, if I run just one instance 
>>>>>>>>>>>>>>>>> of snort,
>>>>>>>>>>>>>>>>> I get an alert from rule 2100498 for EACH time I run "curl
>>>>>>>>>>>>>>>>> testmyids.com"
>>>>>>>>>>>>>>>>> - if I increase to two instances of snort with the same 
>>>>>>>>>>>>>>>>> cluster-id, I
>>>>>>>>>>>>>>>>> get NO alerts when running "curl testmyids.com"
>>>>>>>>>>>>>>>>> - if I set the daq clustermode to 2, I get NO alerts when 
>>>>>>>>>>>>>>>>> running
>>>>>>>>>>>>>>>>> "curl > _______________________________________________
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Doug Burks
>>>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>>> [email protected]
>>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>>> 
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>>>> [email protected]
>>>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> --
>>>>>>>>>>> Doug Burks
>>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> --
>>>>>>>>>> Doug Burks
>>>>>>>>>> http://securityonion.blogspot.com
>>>>>>>>>> _______________________________________________
>>>>>>>>>> Ntop-misc mailing list
>>>>>>>>>> [email protected]
>>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> Ntop-misc mailing list
>>>>>>>>> [email protected]
>>>>>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Doug Burks
>>>>>>>> http://securityonion.blogspot.com
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> --
>>>>>>> Doug Burks
>>>>>>> http://securityonion.blogspot.com
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> --
>>>>>> Doug Burks
>>>>>> http://securityonion.blogspot.com
>>>>> 
>>>>> 
>>>>> 
>>>>> --
>>>>> Doug Burks
>>>>> http://securityonion.blogspot.com
>>>>> _______________________________________________
>>>>> Ntop-misc mailing list
>>>>> [email protected]
>>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>>> 
>>>> _______________________________________________
>>>> Ntop-misc mailing list
>>>> [email protected]
>>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>>> 
>>> 
>>> 
>>> --
>>> Doug Burks
>>> http://securityonion.blogspot.com
>>> _______________________________________________
>>> Ntop-misc mailing list
>>> [email protected]
>>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
>> 
>> _______________________________________________
>> Ntop-misc mailing list
>> [email protected]
>> http://listgateway.unipi.it/mailman/listinfo/ntop-misc
> 
> 
> 
> -- 
> Doug Burks
> http://securityonion.blogspot.com
> _______________________________________________
> Ntop-misc mailing list
> [email protected]
> http://listgateway.unipi.it/mailman/listinfo/ntop-misc

_______________________________________________
Ntop-misc mailing list
[email protected]
http://listgateway.unipi.it/mailman/listinfo/ntop-misc

Reply via email to