[c-nsp] MPLS "Tag Control" process - what does this do?

Rodney Dunn rodunn at cisco.com
Fri Aug 3 07:42:48 EDT 2007


I'm pretty sure Reuben that it's controlling the allocation of
labels to prefixes. I'd need to dig a bit more on it.

How many routes do you have in the global routing table?

'sh ip route summ'.

> Aug  3 04:42:37.479: %TIB-3-REMOTETAG: 150.82.0.0/255.255.0.0, peer
> 61.84.96.254:0; tag 53643; Failure with LDP Label Release; prefix not in TIB
> Aug  3 04:43:08.171: %TIB-3-REMOTETAG: 192.232.71.0/255.255.255.0, peer
> 61.84.96.254:0; tag 54898; Failure with LDP Label Release; prefix not in TIB

Are you running TDP or LDP?

What is the peer and code?

I looked around a bit to try and understand that messge.
Seems it has to do with getting a label for a prefix we don't have.

Problem Description:
====================
Problem: LDP does not withdraw label before announcing a new label for 
	 the same FEC.

Solution:
To fix this problem a new config command is introduced:
  [no] mpls ldp neighbor A.B.C.D implicit-withdraw-label

default Behavior:
It will follow the LDP standard i.e. LDP will withdraw previously
advertised label before advertising a new label for a FEC. When
"mpls ldp neighbor A.B.C.D implicit-withdraw-label" is configured
LDP will not withdraw the previous label before advertising a new 
label.

default behavior is changed. Now when there is a need to change
label for a FEC:
1. LDP will send a Label withdraw and then after receiving Label
Release, it will advertise the new binding with a Label Mapping. If
after sending Label Withdraw, no Label Release is received from a peer(s)
and sufficient time (Currently set to 5 minutes) has passed, then LDP
will assume that peer(s) is not capable of sending a label release and it
will send Label Mapping to the peer.

2. LDP maintains a list of previous labels for which a Label Release is
awaited from any peer.
A new Label Mapping for a FEC is not announced to a peer if a Label
release for the same FEC is pending from the peer.

that was changes that went in under:

CSCdv74248
Externally found enhancement defect: Resolved (R)
LDP session drop after receiving a new label for the same FEC

that you would have in 12.3(19). But what about the peering router?

Rodney


On Fri, Aug 03, 2007 at 08:37:31PM +1000, Reuben Farrelly wrote:
> Hi,
> 
> We had a rather mysterious and puzzling event occur on one of our 7200s today, 
> running 12.3(19) and MPLS on an NPE-G1.
> 
> What happened was that a normal IPv4 BGP peer was added to the config for a new 
> iBGP session, and at or nearly the same time the router started chewing up 
> massive amounts of CPU, things like SNMP stopped working, radius authentication 
> started timing out and denying logins, although the router was still alive - and 
> (slowly) accepting telnet connections to it but subsequently failing 
> authentication on them.
> 
> A show proc cpu sorted looked like this (first few lines)
> 
> PID Runtime(ms)   Invoked      uSecs   5Sec   1Min   5Min TTY Process
>   164     4433204   8742181        507 63.44% 60.27% 40.98%   0 Tag Control
>   188     4424124  12026434        367  0.95%  0.80%  0.64%   0 LDP
>    52   383431108 451130068        849  0.87%  0.81%  1.59%   0 IP Input
>   176     1067504  37471947         28  0.31%  0.18%  0.11%   0 IP-EIGRP: PDM
> 
> 
> Quite a few messages like this were logged to the console during the problem 
> period and for s short while after it came right:
> 
> Aug  3 04:42:37.479: %TIB-3-REMOTETAG: 150.82.0.0/255.255.0.0, peer 
> 61.84.96.254:0; tag 53643; Failure with LDP Label Release; prefix not in TIB
> Aug  3 04:43:08.171: %TIB-3-REMOTETAG: 192.232.71.0/255.255.255.0, peer 
> 61.84.96.254:0; tag 54898; Failure with LDP Label Release; prefix not in TIB
> 
> At about that same time a 3550 switch in another part of the network suffered a 
> hardware failure/reload.  So I'm not entirely sure what caused the event - there 
> was another 7200 behind the broken switch but it wasn't a major path for any 
> sort of traffic pattern to drastically shift.  It may well be unrelated.
> 
> Traffic volumes across the 2 ATM PVCs that this router services were 10-15M 
> (only a 10M and a 8M PVC into it), so not extraordinarily high and certainly 
> could not have gone past 18M even if both PVCs were running full.
> 
> After 15 or so mins the router came right and has been OK ever since.
> 
> Can anyone tell me what this "Tag Control" process does, and what sort of 
> activity might cause it to burn so much CPU for such a length of time?  There 
> doesn't seem to be a whole lot about it on CCO and it's a bit worrying when 
> things like this happen and there is no real easily documented answer :-)
> 
> Thanks,
> Reuben
> _______________________________________________
> cisco-nsp mailing list  cisco-nsp at puck.nether.net
> https://puck.nether.net/mailman/listinfo/cisco-nsp
> archive at http://puck.nether.net/pipermail/cisco-nsp/


More information about the cisco-nsp mailing list