Hi all,

Peter Haag napsal(a):
> Hi Marek,
>
>
>
> The speed of the filter engine depends on various things, and a 
> general answer is rather complex. The filter tries to
> optimise as good as possible and evaluates only required expressions 
> and skips non relevant terms. you need many hundred
> or thousand of terms until a noticeable speed penalty occurs, as other 
> system resources are more relevant for speed,
> such as i.O.
>
> The worst filter term is a single chain of .. OR .. OR .. OR .. . If 
> this reaches many hundred terms, the list syntax is
> way more efficient:
> .. in [ list ]
>
> So, there is no general answer, however, the filter engine is seldom 
> the bottleneck. I have very large filters and
> nfdump performs well.
>
>
I would like to add my experience with filtering of huge data amounts. A 
year ago, we used 32b nfsen/nfdump on dual core cpu and 4 GB RAM with 
RAID5 to collect data with 75k flows per second on 10G network. Then we 
applied a filter with approx. 2000 net masks in global OR (for Marek: 
NIX traffic filter;). Nfprofile took about 8 minutes to finish its work 
making the filter not appliable (as nfsen process data every 5 minutes), 
we had to use sampling 1:10. But what was very strange: when I applied 
the NOT on the whole filter, it took several times more time to finish 
this filter. Why?

During this test it was not a problem with RAM which was not completely 
used. Bottle neck was a CPU and especially the HDD.

Next problem I faced was the old nfsen behaviour, when it ran nfprofile 
sequently for every profile so only 1 cpu core was used. As far as I 
know, this should be fixed in actual version, so extra memory and extra 
cpu cores makes the better performance. Also use the fastest HDDs as you 
can.

But what is important: It is also not very important how much traffic 
you meassure. The main factor slowing the netflow data processing is the 
number of flows. Try to decrease their number by increasing the flow 
cache on your netflow generators and increasing the active and inactive 
timeouts.

As far as I know, NfSen is the only tool able to process netflow from 
saturated 10GB traffic with more than 40k flows per second. Good work, 
Peter!

Regards,
    HOP
>
>
> Please note, if you really process high volume flow files, you should 
> consider to migrate to a 64bit system. nfdump
> takes as much memory as it requires, which is the most influencing 
> factor for speed. Memory can not be replaced except
> by memory! having enough memory an I.O capacity is the credo.
>
> Hope this helps
>
>     - Peter
>
>
>
>
>

------------------------------------------------------------------------------
Enter the BlackBerry Developer Challenge 
This is your chance to win up to $100,000 in prizes! For a limited time,
vendors submitting new applications to BlackBerry App World(TM) will have
the opportunity to enter the BlackBerry Developer Challenge. See full 
prize 
details at: http://p.sf.net/sfu/Challenge
_______________________________________________
Nfdump-discuss mailing list
Nfdump-discuss@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/nfdump-discuss

-- 
Jan Pazdera <pazdera at invea.cz>
INVEA-TECH a.s.
U Vodarny 2965/2, 61600 Brno, Czech Republic
Tel: +420 541 142 051
www.invea-tech.com
Key 0x89F62F78: 41A7 28C2 C624 FBD1 E236  6827 42EB 3694 89F6 2F78



------------------------------------------------------------------------------
Enter the BlackBerry Developer Challenge  
This is your chance to win up to $100,000 in prizes! For a limited time, 
vendors submitting new applications to BlackBerry App World(TM) will have
the opportunity to enter the BlackBerry Developer Challenge. See full prize  
details at: http://p.sf.net/sfu/Challenge
_______________________________________________
Nfdump-discuss mailing list
Nfdump-discuss@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/nfdump-discuss

Reply via email to