[c-nsp] Multilink PPP (MLPPP) Asymmetrical Throughput Problem NxT1

Rodney Dunn rodunn at cisco.com
Wed Jun 6 15:47:41 EDT 2007


Then that means you are overunning the BW and the drops
are having a more severe impact on the application than
the delay you introduce with the longer queue depth.

On Wed, Jun 06, 2007 at 03:41:00PM -0400, Sean Shepard wrote:
> Resolution was indeed increasing the output queue size.  Looks like around
> 160 to 240 (for bonded 2xT1 and 3xT1) seemed to do the trick.  Testing today
> has been tremendously positive.  CEF does appear to be okay on MLPPP that
> far back (woo-hoo!).
> 
> Thanks for your assistance and feedback!
> Sean
> 
> 
> 
> -----Original Message-----
> From: Rodney Dunn [mailto:rodunn at cisco.com] 
> Sent: Wednesday, June 06, 2007 10:51 AM
> To: Sean Shepard
> Cc: 'Rodney Dunn'; cisco-nsp at puck.nether.net
> Subject: Re: [c-nsp] Multilink PPP (MLPPP) Asymmetrical Throughput Problem
> NxT1
> 
> On Tue, Jun 05, 2007 at 11:02:07PM -0400, Sean Shepard wrote:
> > Thank you for the reply on this.  We did exactly what you mention here
> > (trying to isolate channels) and found the performance metrics didn't
> change
> > very much except that there seemed to be little impairment with just a
> > single T-1.
> 
> Good test.
> 
>   We do not believe that variance in latency exists to the point
> > that we should be having a severe issue and it has since reoccurred on a
> > couple of other bundled connections (on this same particular router - see
> > below).
> 
> Fair enough. There were a lot of MLPPP bugs in older releases too.
> MLPPP can be pretty complicated too becuase there are a lot of dependencies
> on the driver code to report backpressure correctly to the bundle.
> There is no queueing on the interface level so if the driver code doesn't
> put the backpressure to the MLPPP virtual interface correctly you will
> have probelems.
> 
> > 
> > None of the T-1s seem to take errors in any of the bundles.  We do see a
> lot
> > of output queue drops on the Multilink interfaces but not sure how
> > concerning that really is.
> 
> That's a problem. If they are valid drops you are overrunning the bundle
> member links. 
> 
> > 
> > The only difference between this device and similar ones on our network is
> > that we have exceeded the number of fast interfaces (4 vs. recommended 3 -
> > but the card in question is in the middle and should be getting its SRAM
> > allotment okay) and we do terminate some ATM/PPPoE/L2TP sessions on this
> > device.  The system is:
> 
> I'd be amazed if that had anything to do with it.
> 
> Did you disable MLPPP fragmentation "no ppp multilink fragmentation"
> or it's one with the disable CLI. We changed it at some point along the
> way.
> 
> > 
> > 7206 (non-VXR)
> > NPE-200 with IO-FE
> > IOS 12.2(31) [bootldr 12.0(13)S]
> >   (is there perhaps an issue in 12.2(31) with MLPPP?
> >    I'd like to go to a 12.3 release but need to verify 
> >    Support for the CT3/4T1 for two of our boxes).
> > 
> > We're using the older CT3/4T1 cards on this edge device and haven't had
> > problems with MLPPP in the past on a similar system (running 12.2(23)c).
> 
> See above. There are driver dependencies for each card for MLPPP to work.
> Can you get 'sh controller' just to see if it shows anything interesting
> that's different between the two?
> 
> > 
> > Download speed continues to perform okay in most tests but uploads get
> > woefully bad and we start losing packets above 1.6 to 2.0 mbps (2%
> observed
> > today as things crept over 2mbps) regardless of the number of bundled
> trunks
> > [2 or 3].  It "seems" that performance improves in the evenings when there
> > is less traffic going through the device, it's lightly loaded even during
> > the day (maybe a total of 10 mbps being handled on this one system).
> 
> To really isolate that you first need to determine direction of loss/latency
> and then narrow down the debugging. That's easier said than done.
> 
> > 
> > I considered tweaking the buffers, but if it's an issue of emptying the
> > queues fast enough (perhaps because it's servicing one too many high speed
> > interfaces?) than putting more in the buffers that it can't get to might
> > just make things worse.
> 
> My experience would say that's pretty much surely not the case. But I've
> been wrong before. I don't know if we even have CEF support for MLPPP back
> that far. In 'sh int stat' what does it look like for the bundle interface?
> 
> > 
> > We have several customers utilizing VoIP and have some policy-maps on
> those
> > interfaces, none of them using MLPPP [yet] but a few on the same box and
> > even the same card in question here.  No complaints about lost packets or
> > voice quality there so the overall system seems sound and CPU utilization
> is
> > generally in the low double digits.  Various debug outputs don't seem to
> > barking either.
> 
> It gets complicated but you would have to get the multilink debugs and
> compare to see if you are seeing loss/delay for the fragments.
> 
> does sh ppp multilink show anything when you are doing a transfer
> that is slow?
> 
> > 
> > Any suggestions are appreciated.  I think I'm close to just dropping
> another
> > chassis in with this DS3 on it and seeing if the problem cleans up.
> 
> Get some upgraded code (late 12.3 or 12.4) would be a good recommendation.
> 
> --history snip--


More information about the cisco-nsp mailing list