To continue,
The instance of nfacctd that I said had no issues and was getting
netflow data just had the same behavior.
This instance was running without the buffer/pipe sizes, it handles a
proportionally smaller amount of traffic (probably an order of
magnitude) so perhaps just took longer to enter the same looping state.
The instance configured with buffer/pipe sizes was running for about
three hours on 1.5 code without issues, it would previously only run for
about 10 minutes.
I think the culprit here lies with pipe/buffer sizes and tee.
-Adam
On 11/20/13, 11:17 PM, Adam Jacob Muller wrote:
Gentoo and 64-bit.
Compiled from sources that I just downloaded.
I just downloaded 0.14.3 to test, and so far 0.14.3 appears to work
correctly (so far, running for about 10m).
With a single odd exception, 0.14.3 refused to start because:
mmap(NULL, 18446744072050714384, PROT_READ|PROT_WRITE,
MAP_SHARED|MAP_ANONYMOUS, -1, 0) = -1 ENOMEM (Cannot allocate memory)
It's trying to mmap() 18 exabytes of memory? I am not the NSA, I don't
have that much RAM :)
Setting:
plugin_buffer_size: 102400
plugin_pipe_size[all]: 102400
Seems to have resolved that, its humming along for about 15m now.
-Adam
On 11/20/13, 11:02 PM, Brent Van Dussen wrote:
Hi Adam,
Just for a point on the curve we have Juniper MX IPFIX -> nfacctd
1.5.0rc1 tee replicating out to a few different destinations (one
local, two remote) and see exactly double the output traffic as input
traffic as expected. This is on a Linux Debian 7 box.
What OS/Arch are you running?
-Brent
-----Original Message-----
From: pmacct-discussion [mailto:[email protected]]
On Behalf Of Adam Jacob Muller
Sent: Wednesday, November 20, 2013 7:36 PM
To: [email protected]
Subject: Re: [pmacct-discussion] nfacctd, ipfix and tee transparent mode
Hi,
That was actually one of my first thoughts. But my topology is
extremely simple here and I only see the excess traffic on the
replicator server from server->network, not network->server, which I
would expect in the case of a loop.
The single collector in this case is another nfacctd process that has
no tee configured, it shouldn't even be possible to loop back (I
think the version of nfacctd there is to old to even have 'tee'
actually).
Also tcpdump definitely sees a MUCH higher rate of outbound packets.
-Adam
On 11/20/13, 10:23 PM, Nathan Kennedy wrote:
Hi Adam,
I'm guessing you've already thought of this, but is it at all
possible that one of the destinations you're tee-ing the packets to
(e.f.g.h) is then feeding it back (to a.b.c.d)?
Thanks,
Nathan.
-----Original Message-----
From: pmacct-discussion [mailto:[email protected]]
On Behalf Of Adam Jacob Muller
Sent: Thursday, 21 November 2013 4:05 p.m.
To: [email protected]
Subject: [pmacct-discussion] nfacctd, ipfix and tee transparent mode
Hi,
I have an interesting issue that I think perhaps results from a
perhaps unique configuration.
I have a very simple nfacctd setup on one box, its goal would be to
receive ipfix data from two sources (Juniper MX) and replicate it
out to a few places.
The configuration is -very- simple:
nfacctd_ip:a.b.c.d
nfacctd_port: 2101
plugins: tee[all]
tee_receiver[all]: e.f.g.h:2100
tee_transparent: true
When I turn up the data feed to this source, everything works fine for
a few minutes and then nfacctd will suddenly get into what looks like
a loop internally and start rebroadcasting out the same packets [I
think
-- I did not specifically confirm this] (not a single packet, but
perhaps the same group) as quickly as possible. Like, line rate
pegging the servers gigabit uplink.
Some (hopefully) useful data points:
I have another nfacctd instance teeing with an almost identical
configuration, except that the source is NetFlow v5 (Cisco).
tee_transparent has no effect, I prefer it but it still breaks with
it off.
Disabling the netflow source does not stop the packets. nfacctd
continues to rebroadcast the same (again, presumably the same) set
of packets over and over again until I kill the process (or ctrl-c,
that still works fine).
This seems very unusual, because I assume this would be obvious in
testing / development if things were this badly broken but my
configuration is also exceedingly simple and I don't particularly
see where I did (or even could) go wrong.
Thanks in advance for any advice you can offer,
-Adam
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists
_______________________________________________
pmacct-discussion mailing list
http://www.pmacct.net/#mailinglists