[j-nsp] A low number of Firewall filters reducing the bandwidth capacity.
Gustavo Santos
gustkiller at gmail.com
Mon Aug 26 14:03:22 EDT 2024
When conducting a more comprehensive assessment, one of our NMS tools
managed to record CPU usage of both FPCs averaging 95% over a 5-minute
period, with a certainty that it reached 100% at some point.
Em seg., 26 de ago. de 2024 às 14:45, Gustavo Santos <gustkiller at gmail.com>
escreveu:
> Thanks Jeff /Saku for the tips.
>
> As it is a "joint" solution by Corero / Juniper, they were supposed to
> know how to optimize or know platform limitations.
> Opened a Case with Corero to verify if they know something about that.
>
>
> Saku,
>
> There is a match filter / term for counting all allowed network traffic,
> after that there is another one to match attack traffic and count
> and another one to discard and count matched traffic.
>
> As soon as I get a reply from the supplier I will update you on this
> thread.
>
>
>
>
> Em seg., 26 de ago. de 2024 às 12:31, Jeff Haas <jhaas at juniper.net>
> escreveu:
>
>> Implementation of firewall varies somewhat by platform and configuration
>> mode. While I have a fair amount of experience in some of this via
>> flowspec, such performance issues are usually "consult appropriate
>> platform/pfe expert".
>>
>> That said, a general heuristic is that if firewall terms cannot be
>> optimized together, some platforms will start hitting high CPU in a
>> relatively small number of rules. The reason is relatively simple compiler
>> optimization theory: If you can't optimize it, you've created a filter that
>> must pass through the full filter set to do its work. You're now making
>> every packet hit that.
>>
>> What will work nicely: Common layer4+ patterns with a prefix list.
>>
>> -- Jeff
>>
>>
>>
>> On 8/26/24, 09:43, "juniper-nsp on behalf of Gustavo Santos via
>> juniper-nsp" <juniper-nsp-bounces at puck.nether.net <mailto:
>> juniper-nsp-bounces at puck.nether.net> on behalf of
>> juniper-nsp at puck.nether.net <mailto:juniper-nsp at puck.nether.net>> wrote:
>>
>>
>> [External Email. Be cautious of content]
>>
>>
>>
>>
>> Awesome, thanks for the info!
>> Rules are like the one below.
>> after adjusting the detection engine to handle as /24 network instead of
>> /32 hosts the issue is gone..
>> As you said the issue was not caused by pps as the attack traffic was just
>> about 30Mpps and with adjusted rules to /24 networks
>> there were not more dropped packets from PFE.
>>
>>
>> Did you know or have how to check the PPE information that should show
>> what
>> may have happened?
>>
>>
>>
>>
>> Below a sample rule that was generated ( about 300 of them via netconf
>> that
>> caused the slowdown).
>>
>>
>> set term e558d83516833f77dea28e0bd5e65871-match from
>> destination-address 131.0.245.143/32
>> set term e558d83516833f77dea28e0bd5e65871-match from
>> protocol 6
>> set term
>> e558d83516833f77dea28e0bd5e65871-match from source-port 443
>> set term e558d83516833f77dea28e0bd5e65871-match from
>> packet-length 32-63
>> set term e558d83516833f77dea28e0bd5e65871-match from
>> tcp-flags "syn & ack & !fin & !rst & !psh"
>> set term e558d83516833f77dea28e0bd5e65871-match then count
>> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-match port-mirror next
>> term
>> set term e558d83516833f77dea28e0bd5e65871-action from
>> destination-address 131.0.245.143/32
>> set term e558d83516833f77dea28e0bd5e65871-action from
>> protocol 6
>> set term
>> e558d83516833f77dea28e0bd5e65871-action from source-port 443
>> set term e558d83516833f77dea28e0bd5e65871-action from
>> packet-length 32-63
>> set term e558d83516833f77dea28e0bd5e65871-action from
>> tcp-flags "syn & ack & !fin & !rst & !psh"
>> set term e558d83516833f77dea28e0bd5e65871-action then count
>> Corero-auto-block-e558d83516833f77dea28e0bd5e65871-discard discard
>>
>>
>>
>>
>>
>> Juniper Business Use Only
>> Em dom., 25 de ago. de 2024 às 02:36, Saku Ytti <saku at ytti.fi <mailto:
>> saku at ytti.fi>> escreveu:
>>
>>
>> > The RE and LC CPU have nothing to do with this, you'd need to check
>> > the Trio PPE congestion levels to figure out if you're running out of
>> > cycles for ucode execution.
>> >
>> > This might improve your performance:
>> >
>> >
>> https://www.juniper.net/documentation/us/en/software/junos/cli-reference/topics/ref/statement/firewall-fast-lookup-filter.html
>> <
>> https://www.juniper.net/documentation/us/en/software/junos/cli-reference/topics/ref/statement/firewall-fast-lookup-filter.html
>> >
>> >
>> > There is also old and new trio ucode, new being 'hyper mode', but this
>> > may already be default on, depending on your release. Hyper mode
>> > should give a bit more PPS.
>> >
>> > There is precious little information available, like what exactly are
>> > your filters doing, what kind of PPS are you pushing in the Trio
>> > experiencing this, where are you seeing the drops, if you are
>> > dropping, they are absolutely accounted for somewhere.
>> >
>> > Unless you are really pushing very heavy PPS, I have difficulties
>> > seeing 100 sensible FW rules impacting performance, not saying it is
>> > impossible, but suspecting there is a lot more here. We'd need to deep
>> > dive into the rules, PPE configuration and load.
>> >
>> > On Sat, 24 Aug 2024 at 23:35, Gustavo Santos via juniper-nsp
>> > <juniper-nsp at puck.nether.net <mailto:juniper-nsp at puck.nether.net>>
>> wrote:
>> > >
>> > > Hi,
>> > >
>> > > We have noticed that when a not so large number of firewall filters
>> > terms
>> > > are generated and pushed to edge routers via via NETCONF into a
>> triplet
>> > of
>> > > MX10003 ,
>> > > we start receiving customer complaints. These issues seem to be
>> related
>> > to
>> > > the router's FPC limiting overall network traffic. To resolve the
>> > problem,
>> > > we simply deactivate the ephemeral configuration database that
>> contains
>> > the
>> > > rules, which removes all the rules,
>> > > and the traffic flow returns to normal. Is there any known limitation
>> or
>> > > bug that could cause this type of issue?
>> > > We typically observe this problem with more than 100 rules; with a
>> > smaller
>> > > number of rules, we don't experience the same issue, even with much
>> > larger
>> > > attacks. Is there any known bug or limitation?
>> > >
>> > > As it is a customer traffic issue I didn't have the time to check fpc
>> > > memory or fpc shell. I just checked the routing engine and fpc cpu and
>> > > they are all fine ( under 50% fpc and under 10% RE).
>> > >
>> > > Any thoughts?
>> > >
>> > > Regards.
>> > > _______________________________________________
>> > > juniper-nsp mailing list juniper-nsp at puck.nether.net <mailto:
>> juniper-nsp at puck.nether.net>
>> > >
>> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$
>> <
>> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$
>> >
>> >
>> >
>> >
>> > --
>> > ++ytti
>> >
>> _______________________________________________
>> juniper-nsp mailing list juniper-nsp at puck.nether.net <mailto:
>> juniper-nsp at puck.nether.net>
>>
>> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$
>> <
>> https://urldefense.com/v3/__https://puck.nether.net/mailman/listinfo/juniper-nsp__;!!NEt6yMaO-gk!Ezcjh-l_yFs02hNP37D-_DcpmIMhpEWtZAZPiGvrZZ5ZG3VkSndHrYcfiTW-I_Chpvpge2OR-yZaUXOZB9JrtyI$
>> >
>>
>>
>>
>>
More information about the juniper-nsp
mailing list