Hi Adam,

You are right, there is a bug lying around 1.5.0rc1 when not setting
an explicit value for plugin_pipe_size and/or plugin_buffer_size. The
issue was already fixed in the CVS code:

http://www.mail-archive.com/pmacct-commits@pmacct.net/msg00896.html

Cheers,
Paolo

On Thu, Nov 21, 2013 at 12:58:20AM -0500, Adam Jacob Muller wrote:
> To continue,
> 
> The instance of nfacctd that I said had no issues and was getting
> netflow data just had the same behavior.
> This instance was running without the buffer/pipe sizes, it handles
> a proportionally smaller amount of traffic (probably an order of
> magnitude) so perhaps just took longer to enter the same looping
> state.
> 
> The instance configured with buffer/pipe sizes was running for about
> three hours on 1.5 code without issues, it would previously only run
> for about 10 minutes.
> 
> I think the culprit here lies with pipe/buffer sizes and tee.
> 
> -Adam
> 
> On 11/20/13, 11:17 PM, Adam Jacob Muller wrote:
> >Gentoo and 64-bit.
> >
> >Compiled from sources that I just downloaded.
> >
> >I just downloaded 0.14.3 to test, and so far 0.14.3 appears to
> >work correctly (so far, running for about 10m).
> >
> >With a single odd exception, 0.14.3 refused to start because:
> >
> >mmap(NULL, 18446744072050714384, PROT_READ|PROT_WRITE,
> >MAP_SHARED|MAP_ANONYMOUS, -1, 0) = -1 ENOMEM (Cannot allocate
> >memory)
> >
> >It's trying to mmap() 18 exabytes of memory? I am not the NSA, I
> >don't have that much RAM :)
> >
> >Setting:
> >plugin_buffer_size: 102400
> >plugin_pipe_size[all]: 102400
> >
> >Seems to have resolved that, its humming along for about 15m now.
> >
> >-Adam
> >
> >On 11/20/13, 11:02 PM, Brent Van Dussen wrote:
> >>Hi Adam,
> >>
> >>Just for a point on the curve we have Juniper MX IPFIX ->
> >>nfacctd 1.5.0rc1 tee replicating out to a few different
> >>destinations (one local, two remote) and see exactly double the
> >>output traffic as input traffic as expected. This is on a Linux
> >>Debian 7 box.
> >>
> >>What OS/Arch are you running?
> >>
> >>-Brent
> >>
> >>-----Original Message-----
> >>From: pmacct-discussion
> >>[mailto:pmacct-discussion-boun...@pmacct.net] On Behalf Of Adam
> >>Jacob Muller
> >>Sent: Wednesday, November 20, 2013 7:36 PM
> >>To: pmacct-discussion@pmacct.net
> >>Subject: Re: [pmacct-discussion] nfacctd, ipfix and tee transparent mode
> >>
> >>Hi,
> >>That was actually one of my first thoughts. But my topology is
> >>extremely simple here and I only see the excess traffic on the
> >>replicator server from server->network, not network->server,
> >>which I would expect in the case of a loop.
> >>
> >>The single collector in this case is another nfacctd process
> >>that has no tee configured, it shouldn't even be possible to
> >>loop back (I think the version of nfacctd there is to old to
> >>even have 'tee' actually).
> >>
> >>Also tcpdump definitely sees a MUCH higher rate of outbound packets.
> >>
> >>-Adam
> >>
> >>On 11/20/13, 10:23 PM, Nathan Kennedy wrote:
> >>>Hi Adam,
> >>>
> >>>I'm guessing you've already thought of this, but is it at all
> >>>possible that one of the destinations you're tee-ing the
> >>>packets to (e.f.g.h) is then feeding it back (to a.b.c.d)?
> >>>
> >>>Thanks,
> >>>Nathan.
> >>>
> >>>-----Original Message-----
> >>>From: pmacct-discussion [mailto:pmacct-discussion-boun...@pmacct.net]
> >>>On Behalf Of Adam Jacob Muller
> >>>Sent: Thursday, 21 November 2013 4:05 p.m.
> >>>To: pmacct-discussion@pmacct.net
> >>>Subject: [pmacct-discussion] nfacctd, ipfix and tee transparent mode
> >>>
> >>>Hi,
> >>>I have an interesting issue that I think perhaps results from
> >>>a perhaps unique configuration.
> >>>
> >>>I have a very simple nfacctd setup on one box, its goal would
> >>>be to receive ipfix data from two sources (Juniper MX) and
> >>>replicate it out to a few places.
> >>>
> >>>The configuration is -very- simple:
> >>>nfacctd_ip:a.b.c.d
> >>>nfacctd_port: 2101
> >>>plugins: tee[all]
> >>>tee_receiver[all]: e.f.g.h:2100
> >>>tee_transparent: true
> >>>
> >>>When I turn up the data feed to this source, everything works fine for
> >>>a few minutes and then nfacctd will suddenly get into what looks like
> >>>a loop internally and start rebroadcasting out the same packets [I
> >>>think
> >>>-- I did not specifically confirm this] (not a single packet,
> >>>but perhaps the same group) as quickly as possible. Like, line
> >>>rate pegging the servers gigabit uplink.
> >>>
> >>>Some (hopefully) useful data points:
> >>>
> >>>I have another nfacctd instance teeing with an almost
> >>>identical configuration, except that the source is NetFlow v5
> >>>(Cisco).
> >>>
> >>>tee_transparent has no effect, I prefer it but it still breaks
> >>>with it off.
> >>>
> >>>Disabling the netflow source does not stop the packets.
> >>>nfacctd continues to rebroadcast the same (again, presumably
> >>>the same) set of packets over and over again until I kill the
> >>>process (or ctrl-c, that still works fine).
> >>>
> >>>This seems very unusual, because I assume this would be
> >>>obvious in testing / development if things were this badly
> >>>broken but my configuration is also exceedingly simple and I
> >>>don't particularly see where I did (or even could) go wrong.
> >>>
> >>>Thanks in advance for any advice you can offer,
> >>>
> >>>-Adam
> >>>
> >>>
> >>>_______________________________________________
> >>>pmacct-discussion mailing list
> >>>http://www.pmacct.net/#mailinglists
> >>>
> >>>_______________________________________________
> >>>pmacct-discussion mailing list
> >>>http://www.pmacct.net/#mailinglists
> >>
> >>_______________________________________________
> >>pmacct-discussion mailing list
> >>http://www.pmacct.net/#mailinglists
> >>
> >>_______________________________________________
> >>pmacct-discussion mailing list
> >>http://www.pmacct.net/#mailinglists
> >
> >
> >_______________________________________________
> >pmacct-discussion mailing list
> >http://www.pmacct.net/#mailinglists
> 
> 
> _______________________________________________
> pmacct-discussion mailing list
> http://www.pmacct.net/#mailinglists

_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists

Reply via email to