Hi Brian,
You are most probably looking for this: https://github.com/pmacct/pmacct/blob/master/QUICKSTART#L2644-#L2659 Should that not work, ie. too many input flows for the available resources, you have a couple load-balancing strategies possible: one is to configure a replicator (tee plugin, see in QUICKSTART). Paolo On Sun, Feb 24, 2019 at 05:31:55PM +0000, Brian Solar wrote: > Is there a way to adjust the UDP buffer receive size ? > > Are there any other indications of nfacctd not keeping up? > > cat /proc/net/udp |egrep drops\|0835 > > sl local_address rem_address st tx_queue rx_queue tr tm->when retrnsmt > uid timeout inode ref pointer drops > > 52366: 00000000:0835 00000000:0000 07 00000000:00034B80 00:00000000 00000000 > 0 0 20175528 2 ffff89993febd940 7495601 > > 7495601 drops w/ a buffer of 0x0034B80 or 214528 > > sysctl -a |fgrep mem > > net.core.optmem_max = 20480 > > net.core.rmem_default = 212992 > > net.core.rmem_max = 2147483647 > > net.core.wmem_default = 212992 > > net.core.wmem_max = 212992 > > net.ipv4.igmp_max_memberships = 20 > > net.ipv4.tcp_mem = 9249771 12333028 18499542 > > net.ipv4.tcp_rmem = 4096 87380 6291456 > > net.ipv4.tcp_wmem = 4096 16384 4194304 > > net.ipv4.udp_mem = 9252429 12336573 18504858 > > net.ipv4.udp_rmem_min = 4096 > > net.ipv4.udp_wmem_min = 4096 > > vm.lowmem_reserve_ratio = 256 256 32 > > vm.memory_failure_early_kill = 0 > > vm.memory_failure_recovery = 1 > > vm.nr_hugepages_mempolicy = 0 > > vm.overcommit_memory = 0 > _______________________________________________ > pmacct-discussion mailing list > http://www.pmacct.net/#mailinglists _______________________________________________ pmacct-discussion mailing list http://www.pmacct.net/#mailinglists