[c-nsp] High CPU load from loose TE tunnels
Anton Smith
anton at huge.geek.nz
Wed Jan 24 05:26:29 EST 2007
Rodney Dunn wrote:
> What code?
My bad - it is actually a 6500. Code s72033-pk9sv-mz.122-18.SXD7.bin
(12.2(18)SXD7).
The other end (exhibiting no problems) is 12.2(20)S6 (7300). It actually
has the CEF Scanner still though:
#show processes | i CEF
66 Lwe 40F4B92C 10196276 1711321 5958 4840/6000 0 CEF
Scanner
99 Lwe 40F4C260 780468 27792146 28 3704/6000 0 CEF
process
I forgot to note that there is a difference in the paths between the two
routers, probably easiest just to draw a small diagram:
R1(6500)--(loosely found hops)---(loose)--(strict)-x-(strict)--R2(7300)
The physical link down is where the x is above.
> The CEF scanner is gone in later code. We moved it to an entirely event
> driven architecture.
Could you perhaps tell me in which code version is the CEF scanner removed?
>
> A sh ip cef events might help explain what the scanner is trying to
> do. It's a bit more complicated with MPLS enabled because some of
> the MPLS code can make callbacks to the CEF scanner even though it's
> not CEF at all.
sh ip cef events doesn't seem to show anything interesting w.r.t. the
tunnel in question.
>
> debug mpls lfib cef|enc|adj might tell the answer if that's what it
> is.
The debugs don't seem to show anything useful either.
>
> Rodney
>
Cheers,
Anton
>
> On Tue, Jan 23, 2007 at 05:01:53PM +0000, Anton Smith wrote:
>> Hi all,
>>
>> I am having a problem with high CPU load on a 7600 because of a downed
>> TE tunnel that is configured to use a path that starts with a loose hop,
>> followed by strict hops. The tunnel is down because one of the strict
>> hops is not reachable (physical link down). The tunnel and path config
>> are as follows:
>>
>> interface Tunnel111
>> description Tunnel111
>> ip unnumbered Loopback1
>> tunnel destination x.x.x.4
>> tunnel mode mpls traffic-eng
>> tunnel mpls traffic-eng autoroute announce
>> tunnel mpls traffic-eng priority 2 2
>> tunnel mpls traffic-eng path-option 1 explicit name path1
>> tunnel mpls traffic-eng load-share 155
>>
>> ip explicit-path name path1 enable
>> index 2 next-address loose x.x.x.1
>> next-address x.x.x.2
>> next-address x.x.x.3
>> next-address x.x.x.4
>>
>>
>> Other tunnels that are configured with paths that are very similar (and
>> are fully up) do not cause this kind of CPU load. But when they go down,
>> they also create the same kind of load.
>>
>> The CPU histogram looks as follows:
>>
>> 44444 44444 44444 444444444455555 44444
>> 233333 66666 7777788886666655555444441111166666
>> 100
>> 90
>> 80
>> 70
>> 60
>> 50 ***** ***** *************** *****
>> 40 ***** ***** ***** *************** *****
>> 30 ***** ***** ***** *************** *****
>> 20 ***** ***** ***** *************** *****
>> 10 ***** ***** ************************ *****
>> 0....5....1....1....2....2....3....3....4....4....5....5....
>> 0 5 0 5 0 5 0 5 0 5
>>
>> CPU% per second (last 60 seconds)
>>
>> And the process that seems to be causing the load is CEF scanner:
>>
>> CPU utilization for five seconds: 44%/0%; one minute: 29%; five minutes: 18%
>> PID Runtime(ms) Invoked uSecs 5Sec 1Min 5Min TTY Process
>> 114 43044320 1823442 23606 37.11% 21.28% 11.57% 0 CEF
>> Scanner
>> ....
>>
>> If I admin shut the tunnel interface in question, the CPU load drops
>> back to near zero.
>>
>> I notice that when I run show mpls traffic-eng tunnels summary, the
>> activations and deactivations numbers increment steadily every few
>> seconds, as though the box is continually trying to bring up the tunnel
>> (more often than it should?):
>>
>> R1#show mpls traffic-eng tunnels summary
>> Signalling Summary:
>> LSP Tunnels Process: running
>> Passive LSP Listener: running
>> RSVP Process: running
>> Forwarding: enabled
>> Head: 15 interfaces, 13 active signalling attempts, 13 established
>> 23356 activations, 23343 deactivations
>> Midpoints: 0, Tails: 13
>> Periodic reoptimization: every 300 seconds, next in 213 seconds
>> Periodic FRR Promotion: Not Running
>> Periodic auto-bw collection: every 300 seconds, next in 269 seconds
>>
>> I have another router (the tail), which has return tunnels built (also
>> using a combination of loose and strict hops). This router also has a
>> downed tunnel interface (for the same reason that the first router does
>> - i.e. a downed physical link on a strict hop), but it does not exhibit
>> high CPU load nor does it seem to be periodic. In addition, the
>> 'activations' and 'deactivations' counters do not increment. However,
>> this other router is a 7301. The tunnels are not administratively shut
>> on the 7301.
>>
>> Does anybody have any ideas? How frequently does a 7600 attempt to bring
>> up a tunnel interface? I imagine that the CPU load is coming from the
>> CSPF calculation being run every few seconds in an attempt to find a
>> path to the first (loose) hop. Is it possible to change this frequency?
>> (I have tried changing the reoptimisation timers but I do not believe
>> this is the problem, since they are by default 300 seconds - and I see
>> high CPU load every few seconds).
>>
>> Any help much appreciated :).
>>
>> Regards,
>> Anton
>> _______________________________________________
>> cisco-nsp mailing list cisco-nsp at puck.nether.net
>> https://puck.nether.net/mailman/listinfo/cisco-nsp
>> archive at http://puck.nether.net/pipermail/cisco-nsp/
>
More information about the cisco-nsp
mailing list