[j-nsp] iBGP convergence time

andy andy at shady.org
Mon Feb 19 10:27:40 EST 2007


We are seeing the same on M10i's RE-850 after upgrading to 7.6R2.6 and also M7i's (same software and RE).

A fix would be nice indeed.

cheers


On Mon, Feb 19, 2007 at 02:25:55PM +0000, Ian MacKinnon wrote:
> Well we are running M40 and seeing very slow BGP convergence times, that
> this thread seems to describe well.
> 
> We first noticed this after we upgraded to 8.0
> 
> A fix would be nice :-)
> 
> 
> 
> Richard A Steenbergen wrote:
> > On Fri, Feb 02, 2007 at 04:13:27AM -0500, Richard A Steenbergen wrote:
> >> You know I haven't had any free time to run real tests or anything, but I 
> >> noticed a significant increase in BGP convergence time when upgrading from 
> >> JUNOS 7.2/7.3 (and I think some 7.4) to 7.6. When you make a policy 
> >> change, the routes take several minutes (from 2 to 7) to install. If you 
> >> do a show route you can see the new routes sitting in + and the old routes 
> >> sitting in - for minutes, RPD is spinning its heels at 100% cpu, and the 
> >> packets continue to forward over the old path while it is processing.
> > 
> > Ok so, after about a dozen people contacted me privately to confirm that 
> > they were seeing similar issues that hadn't been fully acknowledged, I ran 
> > off and did a little testing to replicate the issue. The one thing I can 
> > definitely confirm right now is that it only appears to affect M160 (or at 
> > least, not M5-M40).
> > 
> > On an RE-2.0 on a single switch board platform, performance is about what 
> > you would expect from a 7+ year old routing engine running modern code on 
> > a modern routing table. It syncs a full table inbound (from empty) on an 
> > otherwise unused RE-2.0 in just under 7 minutes, and it processes a switch 
> > of best path and installed routes on a full table is just barely under 2 
> > minutes (policy-statement with then local-preference only, no other 
> > processing). However, on an M160 the switch of best path which leads to 
> > installing new routes in the HW takes between 8-15 minutes in my tests. I 
> > haven't yet had time to go through every version one at a time to find 
> > exactly where there starts, but the behavior is definitely evident.
> > 
> > The following is based on absolutely nothing except my random guess as to 
> > what is happening, so someone please let me know if I'm warm or cold. It 
> > seems that the easiest way to replicate the state of new route showing up 
> > as "+" and the old route showing up as "-" is to intentionally break 
> > connectivity between the RE and PFE (easy with an m40 :P). My guess is 
> > that this is a kind of transactional FIB installation system, where the RE 
> > doesn't update its RIB to reflect that the new route has been installed 
> > until the switch board processes it and confirms it (and allowing it to 
> > retry the install if necessary), to prevent Customer Enragement Feature 
> > with Juniper's move towards distributed forwarding tables on the Gibson 
> > architecture. Whatever is going on with the M160 on RE-2.0 however, it is 
> > significantly slower. Maybe this just wasn't sufficiently regression 
> > tested on the M160 platform, or maybe it is just a natural effect of 
> > having 4 switch boards which all need to be updated, but it is very 
> > noticable. The offical Juniper line seems to be "just upgrade your REs", 
> > but it would be nice if we had an alternate option.
> > 
> > So, two things. The most obvious question is, is there a way to turn this 
> > behavior off or revert it to the previous behavior (if infact my guess as 
> > to the cause is correct :P)? The next question is, I noticed in the 
> > release notes for 8.2 that there is a new option to support indirect 
> > next-hops which may significantly reduce the number of FIB updates. My 
> > take on this feature is that you are changing from installing BGP route -> 
> > physical next-hop to BGP route -> BGP next-hop and BGP next-hop -> 
> > physical next-hop, so that when you make a routing change to the BGP 
> > nexthop you only have to update the 1 entry instead of the potentially 
> > thousands of entries for the BGP route itself (which I kinda thought 
> > Juniper had done since forever :P). Am I correct in that interpretation, 
> > or is there something else going on there?
> > 
> 
> -- 
> 
> This email and any files transmitted with it are confidential and intended 
> solely for the use of the individual or entity to whom they are addressed.  
> If you have received this email in error please notify the sender. Any 
> offers or quotation of service are subject to formal specification.  
> Errors and omissions excepted.  Please note that any views or opinions 
> presented in this email are solely those of the author and do not 
> necessarily represent those of Lumison, nplusone or lightershade ltd.  
> Finally, the recipient should check this email and any attachments for the 
> presence of viruses.  Lumison, nplusone and lightershade ltd accepts no 
> liability for any damage caused by any virus transmitted by this email.
> 
> -- 
> -- 
> Virus scanned by Lumison.
> _______________________________________________
> juniper-nsp mailing list juniper-nsp at puck.nether.net
> https://puck.nether.net/mailman/listinfo/juniper-nsp
> 

-- 
andy    andy at shady.org
-----------------------------------------------
Never argue with an idiot. They drag you down 
to their level, then beat you with experience.
----------------------------------------------- 


More information about the juniper-nsp mailing list