When conducting a more comprehensive assessment, one of our NMS tools managed to record CPU usage of both FPCs averaging 95% over a 5-minute period, with a certainty that it reached 100% at some point.
Em seg., 26 de ago. de 2024 às 14:45, Gustavo Santos <[email protected]> escreveu: > Thanks Jeff /Saku for the tips. > > As it is a "joint" solution by Corero / Juniper, they were supposed to > know how to optimize or know platform limitations. > Opened a Case with Corero to verify if they know something about that. > > > Saku, > > There is a match filter / term for counting all allowed network traffic, > after that there is another one to match attack traffic and count > and another one to discard and count matched traffic. > > As soon as I get a reply from the supplier I will update you on this > thread. > > > > > Em seg., 26 de ago. de 2024 às 12:31, Jeff Haas <[email protected]> > escreveu: > >> Implementation of firewall varies somewhat by platform and configuration >> mode. While I have a fair amount of experience in some of this via >> flowspec, such performance issues are usually "consult appropriate >> platform/pfe expert". >> >> That said, a general heuristic is that if firewall terms cannot be >> optimized together, some platforms will start hitting high CPU in a >> relatively small number of rules. The reason is relatively simple compiler >> optimization theory: If you can't optimize it, you've created a filter that >> must pass through the full filter set to do its work. You're now making >> every packet hit that. >> >> What will work nicely: Common layer4+ patterns with a prefix list. >> >> -- Jeff >> >> >> >> On 8/26/24, 09:43, "juniper-nsp on behalf of Gustavo Santos via >> juniper-nsp" <[email protected] <mailto: >> [email protected]> on behalf of >> [email protected] <mailto:[email protected]>> wrote: >> >> >> [External Email. Be cautious of content] >> >> >> >> >> Awesome, thanks for the info! >> Rules are like the one below. >> after adjusting the detection engine to handle as /24 network instead of >> /32 hosts the issue is gone.. >> As you said the issue was not caused by pps as the attack traffic was just >> about 30Mpps and with adjusted rules to /24 networks >> there were not more dropped packets from PFE. >> >> >> Did you know or have how to check the PPE information that should show >> what >> may have happened? >> >> >> >> >> Below a sample rule that was generated ( about 300 of them via netconf >> that >> caused the slowdown). >> >> >> set term e558d83516833f77dea28e0bd5e65871-match from >> destination-address 131.0.245.143/32 >> set term e558d83516833f77dea28e0bd5e65871-match from >> protocol 6 >> set term >> e558d83516833f77dea28e0bd5e65871-match from source-port 443 >> set term e558d83516833f77dea28e0bd5e65871-match from >> packet-length 32-63 >> set term e558d83516833f77dea28e0bd5e65871-match from >> tcp-flags "syn & ack & !fin & !rst & !psh" >> set term e558d83516833f77dea28e0bd5e65871-match then count >> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-match port-mirror next >> term >> set term e558d83516833f77dea28e0bd5e65871-action from >> destination-address 131.0.245.143/32 >> set term e558d83516833f77dea28e0bd5e65871-action from >> protocol 6 >> set term >> e558d83516833f77dea28e0bd5e65871-action from source-port 443 >> set term e558d83516833f77dea28e0bd5e65871-action from >> packet-length 32-63 >> set term e558d83516833f77dea28e0bd5e65871-action from >> tcp-flags "syn & ack & !fin & !rst & !psh" >> set term e558d83516833f77dea28e0bd5e65871-action then count >> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-discard discard >> >> >> >> >> >> Juniper Business Use Only >> Em dom., 25 de ago. de 2024 às 02:36, Saku Ytti <[email protected] <mailto: >> [email protected]>> escreveu: >> >> >> > The RE and LC CPU have nothing to do with this, you'd need to check >> > the Trio PPE congestion levels to figure out if you're running out of >> > cycles for ucode execution. >> > >> > This might improve your performance: >> > >> > >> https://www.juniper.net/documentation/us/en/software/junos/cli-reference/topics/ref/statement/firewall-fast-lookup-filter.html >> < >> https://www.juniper.net/documentation/us/en/software/junos/cli-reference/topics/ref/statement/firewall-fast-lookup-filter.html >> > >> > >> > There is also old and new trio ucode, new being 'hyper mode', but this >> > may already be default on, depending on your release. Hyper mode >> > should give a bit more PPS. >> > >> > There is precious little information available, like what exactly are >> > your filters doing, what kind of PPS are you pushing in the Trio >> > experiencing this, where are you seeing the drops, if you are >> > dropping, they are absolutely accounted for somewhere. >> > >> > Unless you are really pushing very heavy PPS, I have difficulties >> > seeing 100 sensible FW rules impacting performance, not saying it is >> > impossible, but suspecting there is a lot more here. We'd need to deep >> > dive into the rules, PPE configuration and load. >> > >> > On Sat, 24 Aug 2024 at 23:35, Gustavo Santos via juniper-nsp >> > <[email protected] <mailto:[email protected]>> >> wrote: >> > > >> > > Hi, >> > > >> > > We have noticed that when a not so large number of firewall filters >> > terms >> > > are generated and pushed to edge routers via via NETCONF into a >> triplet >> > of >> > > MX10003 , >> > > we start receiving customer complaints. These issues seem to be >> related >> > to >> > > the router's FPC limiting overall network traffic. To resolve the >> > problem, >> > > we simply deactivate the ephemeral configuration database that >> contains >> > the >> > > rules, which removes all the rules, >> > > and the traffic flow returns to normal. Is there any known limitation >> or >> > > bug that could cause this type of issue? >> > > We typically observe this problem with more than 100 rules; with a >> > smaller >> > > number of rules, we don't experience the same issue, even with much >> > larger >> > > attacks. Is there any known bug or limitation? >> > > >> > > As it is a customer traffic issue I didn't have the time to check fpc >> > > memory or fpc shell. I just checked the routing engine and fpc cpu and >> > > they are all fine ( under 50% fpc and under 10% RE). >> > > >> > > Any thoughts? >> > > >> > > Regards. >> > > _______________________________________________ >> > > juniper-nsp mailing list [email protected] <mailto: >> [email protected]> >> > > >> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$ >> < >> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$ >> > >> > >> > >> > >> > -- >> > ++ytti >> > >> _______________________________________________ >> juniper-nsp mailing list [email protected] <mailto: >> [email protected]> >> >> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$ >> < >> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$ >> > >> >> >> >> _______________________________________________ juniper-nsp mailing list [email protected] https://puck.nether.net/mailman/listinfo/juniper-nsp

