[j-nsp] Auto-bandwidth Accuracy
Danny Vernals
danny.vernals at gmail.com
Tue May 25 05:17:46 EDT 2010
On Sun, May 23, 2010 at 7:52 AM, Richard A Steenbergen <ras at e-gerbil.net> wrote:
> Recently I've been noticing some really odd auto-bandwidth behavior on
> several different routers, and I'm wondering if anybody knows if this is
> a known bug or if I'm doing something really wrong in my autobw config.
>
> Specifically, I'm seeing many cases where the rsvp reservations on an
> interface are vastly higher than the actual traffic going over it. I
> started comparing autobw measures bandwidth value vs rsvp resv bandwidth
> across my LSPs (with an op script :P), and noticed that a large number
> of LSPs that were ingress on Juniper routers were consistently reserving
> more bandwidth than they were actually passing.
>
> To troubleshoot this further, I picked one LSP at random and followed it
> through the course of an entire adjust-interval. I also watched it in
> "monitor label-switched-path", and followed the bandwidth recorded for
> it in the mpls stats file. The mpls stats file pretty consistently
> recorded a bandwidth of around 900Mbps. Some samples were up to 1G, some
> were down in the 800Mb's, but nothing was significantly outside this
> range:
>
> xxx.xxxx-xxx.xxxx-BRONZE-1 20442770 pkt 21800398308 Byte 91864 pps 97826023 Bps Util 43.47%
> xxx.xxxx-xxx.xxxx-BRONZE-1 25748678 pkt 27500224526 Byte 89930 pps 96607224 Bps Util 42.93%
> xxx.xxxx-xxx.xxxx-BRONZE-1 31309754 pkt 33516047564 Byte 95880 pps 103721086 Bps Util 46.09%
> xxx.xxxx-xxx.xxxx-BRONZE-1 36934965 pkt 39389728013 Byte 90729 pps 94736781 Bps Util 42.10%
> xxx.xxxx-xxx.xxxx-BRONZE-1 41323164 pkt 44001156442 Byte 86043 pps 90420165 Bps Util 40.18%
> xxx.xxxx-xxx.xxxx-BRONZE-1 46229207 pkt 49166295068 Byte 84586 pps 89054114 Bps Util 39.58%
> xxx.xxxx-xxx.xxxx-BRONZE-1 51764861 pkt 55023074603 Byte 92260 pps 97612992 Bps Util 43.38%
> xxx.xxxx-xxx.xxxx-BRONZE-1 57091315 pkt 60691783494 Byte 90278 pps 96079811 Bps Util 42.70%
> xxx.xxxx-xxx.xxxx-BRONZE-1 62138489 pkt 66009079194 Byte 90128 pps 94951708 Bps Util 42.20%
> xxx.xxxx-xxx.xxxx-BRONZE-1 67697838 pkt 72030553645 Byte 92655 pps 100357907 Bps Util 44.60%
> xxx.xxxx-xxx.xxxx-BRONZE-1 73083250 pkt 77870203449 Byte 89756 pps 97327496 Bps Util 43.25%
> xxx.xxxx-xxx.xxxx-BRONZE-1 78530642 pkt 83799427998 Byte 90789 pps 98820409 Bps Util 43.91%
> xxx.xxxx-xxx.xxxx-BRONZE-1 84166327 pkt 89767404007 Byte 85389 pps 90423878 Bps Util 40.18%
> xxx.xxxx-xxx.xxxx-BRONZE-1 89990750 pkt 96052103366 Byte 85653 pps 92422049 Bps Util 41.07%
> xxx.xxxx-xxx.xxxx-BRONZE-1 94808838 pkt 101299936674 Byte 87601 pps 95415151 Bps Util 42.40%
> xxx.xxxx-xxx.xxxx-BRONZE-1 100044983 pkt 106918990604 Byte 83113 pps 89191332 Bps Util 39.64%
> xxx.xxxx-xxx.xxxx-BRONZE-1 104706036 pkt 111928263183 Byte 86315 pps 92764307 Bps Util 41.22%
> xxx.xxxx-xxx.xxxx-BRONZE-1 109664547 pkt 117256403183 Byte 81287 pps 87346557 Bps Util 38.82%
> xxx.xxxx-xxx.xxxx-BRONZE-1 115001230 pkt 123065374817 Byte 84709 pps 92205898 Bps Util 40.98%
> xxx.xxxx-xxx.xxxx-BRONZE-1 120197917 pkt 128761293505 Byte 85191 pps 93375716 Bps Util 41.50%
> xxx.xxxx-xxx.xxxx-BRONZE-1 124790487 pkt 133783111501 Byte 79182 pps 86583068 Bps Util 38.48%
> xxx.xxxx-xxx.xxxx-BRONZE-1 129450091 pkt 138908431043 Byte 84720 pps 93187628 Bps Util 41.41%
> xxx.xxxx-xxx.xxxx-BRONZE-1 134048794 pkt 143940227806 Byte 82119 pps 89853513 Bps Util 39.93%
> xxx.xxxx-xxx.xxxx-BRONZE-1 138900130 pkt 149257983679 Byte 80855 pps 88629264 Bps Util 39.39%
> xxx.xxxx-xxx.xxxx-BRONZE-1 143665805 pkt 154447812210 Byte 79427 pps 86497142 Bps Util 38.44%
> xxx.xxxx-xxx.xxxx-BRONZE-1 148501587 pkt 159667032930 Byte 80596 pps 86987012 Bps Util 38.66%
> xxx.xxxx-xxx.xxxx-BRONZE-1 153971586 pkt 165650360517 Byte 78142 pps 85476108 Bps Util 37.99%
>
> Next, I watched the output of "show mpls lsp name BLAH detail", looking
> at the autobw measured amount (Max AvgBW) and the reserved bandwidth.
> I'm using a stats interval of 60 seconds, an adjust-interval of 900
> seconds, and in this instance no overflow samples occured. After the
> previous adjust-interval completes the measured bw is reset to 0, and
> then starts updating again after the first 60 sec stats interval is up.
> For around the first 700 seconds the Max AvgBW was pretty close to what
> one would expect (around 900Mbps), then it jumped to ~1.6Gbps for no
> reason that I can determine. The stats file for this LSP (above) never
> showed anything above 1.0G, and a monitor of the lsp never showed any
> sample thatever got anywhere near that high (let alone enough to make an
> entire 60 sec sample interval report that high). At the end of the 900
> seconds, te 1.6G value is what was signaled to RSVP, and the cycle
> repeated itself. I watched it for several more cycles, and saw the same
> behavior happening over and over again, with measured values of 1.8G
> plus, while the stats file continued to show an average of around
> 800-900Mbps and no sample that ever went above 1G.
>
I've seen something similar on 9.5R2 although I didn't pay it much
heed at the time as I was investigating other issues. My guess (and
it is definitely a guess) is that there is an internal data structure
which stores the LSP usage which is then divided by the sampling
interval and written to the statistics file after the sampling
interval. If something (rpd scheduling issue, CPU at 100%?) prevents
this value from being written to the statistics file after the
sampling interval it gets a default value of 0. The data structure
keeps the stats from the previous sampling interval and is added to.
When the next sampling interval expires this value is then divided by
1 x sampling interval leading to an average bps value roughly double
what it should be.
I'll keep an eye out and report back if I see this behaviour again.
> This particular router is running 9.4R3, but I've seen similar behavior
> on some other 9.5R4 routers as well. This really seems like some kind of
> bug, but honestly I'd sooner slit my wrists with a rusty PIC than try to
> explain the above to JTAC (besides, they would probably just ask me for
> 50 irrelevent log files then do nothing for the next 6 months like all
> of my other cases :P). I'm wondering if this is some kind of known
> issue, or if there is some reason why this config wouldn't work well.
>
> The stats interval of 60 seconds is because I snmp poll and graph the
> mplsLspOctets every 60 seconds, and snmp is updated based on the stats
> interval. Any value other than 60 secs makes the graphs wildly jitter.
> But in the JUNOS documentation for auto-bandwidth, there is the
> following warning:
>
> http://www.juniper.net/techpubs/en_US/junos9.5/information-products/topic-collections/config-guide-mpls-applications/mpls-configuring-automatic-bandwidth-allocation-for-lsps.html
>
> Note: To prevent unnecessary resignaling of LSPs, it is best to
> configure an MPLS automatic bandwidth statistics interval of no more
> than one third the corresponding LSP adjustment interval. For example,
> if you configure a value of 30 seconds for the interval statement at the
> [edit protocols mpls statistics] hierarchy level, you should configure a
> value of no more than 90 seconds for the adjust-interval statement at
> the [edit protocols mpls label-switched-path label-switched-path-name
> auto-bandwidth] hierarchy level.
>
> I could never figure this one out, and personally I always thought it
> was some kind of documentation error. What possible reason could there
> be for not having an adjust-interval of more than 3x the statistics
> value? I'm running 900 sec adjust-intervals with 300 sec overflow
> detection (the lowest value you can configure) to try and reduce RSVP
> resignaling load on the network. Every time an LDP resignals, it tears
> down the bypass LSPs as well, and at one point (prior to 9.4 I think) it
> took over 50 seconds before JUNOS would even try to start resignaling
> the bypass LSPs. There were some optimizations made to make it kick off
> the bypass LSP resignal within ~15 secs instead of ~50 secs, but we're
> still trying to keep it from resignaling excessively.
>
I've never seen this advice before but I've certainly seen networks
operate fine with adjust-interval much greater than 3x statistics
interval.
> I'll gladly accept any clue anyone can offer on this one. :)
>
> --
> Richard A Steenbergen <ras at e-gerbil.net> http://www.e-gerbil.net/ras
> GPG Key ID: 0xF8B12CBC (7535 7F59 8204 ED1F CC1C 53AF 4C41 5ECA F8B1 2CBC)
> _______________________________________________
> juniper-nsp mailing list juniper-nsp at puck.nether.net
> https://puck.nether.net/mailman/listinfo/juniper-nsp
>
More information about the juniper-nsp
mailing list