You don't have a counter term for drops? So we can't be entirely sure,
your terms aren't responsible for it? Please add counters for the
discard terms.


You can view the global NPU load % from the PFE CLI, as well as to a
degree individual PPE.

I'd say

- let's ensure it's not the filter dropping the packets as requested
- find out where the drops are reported (interface/extensive counters,
PFE stream counters, QoS counters, NPU exception counters, MQ FI/FO
counters ...)

On Mon, 26 Aug 2024 at 16:43, Gustavo Santos <[email protected]> wrote:
>
> Awesome, thanks for the info!
> Rules are like the one below.
> after adjusting the detection engine to handle as /24 network instead of /32 
> hosts the issue is gone..
> As you said the issue was not caused by pps as the attack traffic was just 
> about 30Mpps and with adjusted rules to /24 networks
> there were not more dropped packets from PFE.
>
> Did you know or have how to check the PPE information that should show what 
> may have happened?
>
>
> Below a sample rule that was generated ( about 300 of them via netconf that 
> caused the slowdown).
>
>       set term e558d83516833f77dea28e0bd5e65871-match from 
> destination-address 131.0.245.143/32
>                 set term e558d83516833f77dea28e0bd5e65871-match from protocol 
> 6
>                                 set term 
> e558d83516833f77dea28e0bd5e65871-match from source-port 443
>                 set term e558d83516833f77dea28e0bd5e65871-match from 
> packet-length 32-63
>                 set term e558d83516833f77dea28e0bd5e65871-match from 
> tcp-flags "syn &amp; ack &amp; !fin &amp; !rst &amp; !psh"
>                 set term e558d83516833f77dea28e0bd5e65871-match then count 
> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-match port-mirror next term
>         set term e558d83516833f77dea28e0bd5e65871-action from 
> destination-address 131.0.245.143/32
>                 set term e558d83516833f77dea28e0bd5e65871-action from 
> protocol 6
>                                 set term 
> e558d83516833f77dea28e0bd5e65871-action from source-port 443
>                 set term e558d83516833f77dea28e0bd5e65871-action from 
> packet-length 32-63
>                 set term e558d83516833f77dea28e0bd5e65871-action from 
> tcp-flags "syn &amp; ack &amp; !fin &amp; !rst &amp; !psh"
>                 set term e558d83516833f77dea28e0bd5e65871-action then count 
> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-discard  discard
>
>
> Em dom., 25 de ago. de 2024 às 02:36, Saku Ytti <[email protected]> escreveu:
>>
>> The RE and LC CPU have nothing to do with this, you'd need to check
>> the Trio PPE congestion levels to figure out if you're running out of
>> cycles for ucode execution.
>>
>> This might improve your performance:
>> https://www.juniper.net/documentation/us/en/software/junos/cli-reference/topics/ref/statement/firewall-fast-lookup-filter.html
>>
>> There is also old and new trio ucode, new being 'hyper mode', but this
>> may already be default on, depending on your release. Hyper mode
>> should give a bit more PPS.
>>
>> There is precious little information available, like what exactly are
>> your filters doing, what kind of PPS are you pushing in the Trio
>> experiencing this, where are you seeing the drops, if you are
>> dropping, they are absolutely accounted for somewhere.
>>
>> Unless you are really pushing very heavy PPS, I have difficulties
>> seeing 100 sensible FW rules impacting performance, not saying it is
>> impossible, but suspecting there is a lot more here. We'd need to deep
>> dive into the rules, PPE configuration and load.
>>
>> On Sat, 24 Aug 2024 at 23:35, Gustavo Santos via juniper-nsp
>> <[email protected]> wrote:
>> >
>> > Hi,
>> >
>> > We have noticed that when a not so large  number of firewall filters terms
>> > are generated and pushed to edge routers via  via NETCONF into a triplet of
>> > MX10003 ,
>> > we start receiving customer complaints. These issues seem to be related to
>> > the router's FPC limiting overall network traffic. To resolve the problem,
>> > we simply deactivate the ephemeral configuration database that contains the
>> > rules, which removes all the rules,
>> > and the traffic flow returns to normal. Is there any known limitation or
>> > bug that could cause this type of issue?
>> > We typically observe this problem with more than 100 rules; with a smaller
>> > number of rules, we don't experience the same issue, even with much larger
>> > attacks. Is there any known bug or limitation?
>> >
>> > As it is a customer traffic issue I didn't have the time to check fpc
>> > memory or fpc shell.  I just checked the routing engine and fpc cpu and
>> > they are all fine ( under 50% fpc and under 10% RE).
>> >
>> > Any thoughts?
>> >
>> > Regards.
>> > _______________________________________________
>> > juniper-nsp mailing list [email protected]
>> > https://puck.nether.net/mailman/listinfo/juniper-nsp
>>
>>
>>
>> --
>>   ++ytti



-- 
  ++ytti
_______________________________________________
juniper-nsp mailing list [email protected]
https://puck.nether.net/mailman/listinfo/juniper-nsp

Reply via email to