[j-nsp] QFX CRB
Cristian Cardoso
cristian.cardoso11 at gmail.com
Thu Nov 12 08:20:00 EST 2020
When there was an exhaustion of IPv4 LPM routes in spine2, I noticed a
greater increase in IRQs. I analyzed tcpdump and found nothing strange
on the server itself.
It reduces the L3 routes sent to the spines and released more slots in
the L2 profile below.
Profile active: l2-profile-three
Type Max Used Free % free
----------------------------------------------------
IPv4 Host 147456 1389 144923 98.28
IPv4 LPM 24576 18108 6286 25.58
IPv4 Mcast 73728 0 72462 98.28
IPv6 Host 73728 572 72462 98.28
IPv6 LPM(< 64) 12288 91 3143 25.58
IPv6 LPM(> 64) 2048 7 2041 99.66
IPv6 Mcast 36864 0 36231 98.28
Em qua., 11 de nov. de 2020 às 23:36, Laurent Dumont
<laurentfdumont at gmail.com> escreveu:
>
> How are you measuring IRQ on the servers? If it's network related IRQs, it should be seen during a packet capture.
>
> On Tue, Nov 10, 2020 at 4:40 PM Cristian Cardoso <cristian.cardoso11 at gmail.com> wrote:
>>
>> I running 19.1R2.8 version on Junos.
>> Today I was in contact with Juniper support about a route depletion
>> problem and it seems to be related to the IRQs problem.
>> When the IPv4 / IPv6 routes are exhausted in the LTM table, the IRQ
>> increment begins.
>> I did an analysis of the packages trafficked on the servers, but I
>> found nothing out of the ordinary.
>>
>> Em ter., 10 de nov. de 2020 às 17:47, Nitzan Tzelniker
>> <nitzan.tzelniker at gmail.com> escreveu:
>> >
>> > Looks ok to me
>> > Which junos version you are running ? and which devices ?
>> > Did you capture on the servers to see what is the traffic that causes the high CPU utilization ?
>> >
>> >
>> > On Tue, Nov 10, 2020 at 9:07 PM Cristian Cardoso <cristian.cardoso11 at gmail.com> wrote:
>> >>
>> >> > show configuration protocols evpn
>> >> vni-options {
>> >> vni 810 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 815 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 821 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 822 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 827 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 830 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 832 {
>> >> vrf-target target:888:888;
>> >> }
>> >> vni 910 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 915 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 921 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 922 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 927 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 930 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 932 {
>> >> vrf-target target:666:666;
>> >> }
>> >> vni 4018 {
>> >> vrf-target target:4018:4018;
>> >> }
>> >> }
>> >> encapsulation vxlan;
>> >> default-gateway no-gateway-community;
>> >> extended-vni-list all;
>> >>
>> >>
>> >> An example of configuring the interfaces follows, all follow this
>> >> pattern with more or less IP's.
>> >> > show configuration interfaces irb.810
>> >> proxy-macip-advertisement;
>> >> virtual-gateway-accept-data;
>> >> family inet {
>> >> mtu 9000;
>> >> address 10.19.11.253/22 {
>> >> preferred;
>> >> virtual-gateway-address 10.19.8.1;
>> >> }
>> >> }
>> >>
>> >> Em ter., 10 de nov. de 2020 às 15:16, Nitzan Tzelniker
>> >> <nitzan.tzelniker at gmail.com> escreveu:
>> >> >
>> >> > Can you show your irb and protocols evpn configuration please
>> >> >
>> >> > Nitzan
>> >> >
>> >> > On Tue, Nov 10, 2020 at 3:26 PM Cristian Cardoso <cristian.cardoso11 at gmail.com> wrote:
>> >> >>
>> >> >> Does anyone use EVPN-VXLAN in the Centrally-Routed and Bridging topology?
>> >> >> I have two spine switches and two leaf switches, when I use the
>> >> >> virtual-gateway in active / active mode in the spines, the servers
>> >> >> connected only in leaf1 have a large increase in IRQ's, generating
>> >> >> higher CPU consumption in the servers.
>> >> >> I did a test by deactivating spine2 and leaving only the gateway
>> >> >> spine1, and the IRQ was zeroed out.
>> >> >> Did anyone happen to go through this?
>> >> >> _______________________________________________
>> >> >> juniper-nsp mailing list juniper-nsp at puck.nether.net
>> >> >> https://puck.nether.net/mailman/listinfo/juniper-nsp
>> _______________________________________________
>> juniper-nsp mailing list juniper-nsp at puck.nether.net
>> https://puck.nether.net/mailman/listinfo/juniper-nsp
More information about the juniper-nsp
mailing list