[c-nsp] VFI LDP transport signaled down (ME3600x)

Ihsan Junaidi Ibrahim ihsan.junaidi at gmail.com
Thu May 10 11:56:47 EDT 2012


Adam,

Shutting and unshutting both side of the VFIs resulting in the following:

PE1
---
May 10 23:52:29.485 MYT: %VFI-6-VFI_STATUS_CHANGED: Status of VFI ME002555 changed from DOWN to UP
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Circuit attributes, Receive update:
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Interface handle: 0x3E83
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Status: UP (0x1)
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Circuit directive: Go Active
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Payload encap: Ethernet
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Circuit Encap: VFI
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Segment type: 0x19
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Switch handle: 61469
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  MTU: 9178
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  I/F Str: pw100001
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .  Circuit string: vfi
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Process attrs
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Received Go Active service directive
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: . Receive status update
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... NMS: VC oper state:  DOWN
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... NMS:     err codes:  no-err
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... SYSLOG: VC is DOWN
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... Local ready
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... Local service is ready; send a label
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... Alloc local binding
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ..... Alloc label for dynamic
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... Populate local binding
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ........ Capability C000505, returned cap C000505, mask FFFEFFBF
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... Autosense enabled, no remote, Ethernet(5)
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... MTU set to 9178
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... FEC set to 129
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... Grouping off
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... Grouping ignored, set to 0
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ....... Control word on
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... PWID: already in use, reuse 14
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... Asking to reuse label 23
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... Requested label: any
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ...... Label request, label 0 pwid 0
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... Generate local event
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... No label
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: ... Check if can activate dataplane
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: .... Keep dataplane up
May 10 23:52:29.485 MYT: AToM: 277 cumulative msgs handled. rc=0
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Label response: label 23 pwid 14 reqid 27
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Generate local event
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Ready, label 23
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: Evt local ready, provisioned->local standby, AC-ready
May 10 23:52:29.485 MYT: AToM[200.28.0.120, 116]: . Take no action
May 10 23:52:29.485 MYT: AToM: 278 cumulative msgs handled. rc=0

PE2
---
May 10 23:51:20.404 MYT: %VFI-6-VFI_STATUS_CHANGED: Status of VFI ME002617 changed from ADMINDOWN to DOWN
May 10 23:51:20.404 MYT: %VFI-6-VFI_STATUS_CHANGED: Status of VFI ME002617 changed from DOWN to UP
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Circuit attributes, Receive update:
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Interface handle: 0x3E83
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Status: UP (0x1)
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Circuit directive: Go Active
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Payload encap: Ethernet
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Circuit Encap: VFI
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Segment type: 0x19
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Switch handle: 12297
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  MTU: 9178
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  I/F Str: pw100001
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .  Circuit string: vfi
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Process attrs
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Received Go Active service directive
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: . Receive status update
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... NMS: VC oper state:  DOWN
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... NMS:     err codes:  no-err
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... SYSLOG: VC is DOWN
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... Local ready
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... Local service is ready; send a label
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... Alloc local binding
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ..... Alloc label for dynamic
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... Populate local binding
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ........ Capability C000505, returned cap C000505, mask FFFEFFBF
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... Autosense enabled, no remote, Ethernet(5)
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... MTU set to 9178
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... FEC set to 129
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... Grouping off
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... Grouping ignored, set to 0
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ....... Control word on
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... PWID: already in use, reuse 2
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... Asking to reuse label 637
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... Requested label: any
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ...... Label request, label 0 pwid 0
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... Generate local event
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... No label
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: ... Check if can activate dataplane
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: .... Keep dataplane up
May 10 23:51:20.404 MYT: AToM: 3306 cumulative msgs handled. rc=0
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Label response: label 637 pwid 2 reqid 3
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Generate local event
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Ready, label 637
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: Evt local ready, ldp ready->local ready
May 10 23:51:20.404 MYT: AToM[200.28.0.15, 116]: . Advertise local vc label binding
May 10 23:51:20.404 MYT: AToM: 3307 cumulative msgs handled. rc=0
May 10 23:51:20.432 MYT: AToM[200.28.0.15, 116]: Start resend label timer
May 10 23:51:20.432 MYT: AToM LDP[200.28.0.15, 116]: Receive label release
May 10 23:51:20.432 MYT: AToM[200.28.0.15, 116]: Evt remote release, in local ready
May 10 23:51:20.432 MYT: AToM[200.28.0.15, 116]: . Take no action
May 10 23:51:20.432 MYT: AToM: 3308 cumulative msgs handled. rc=0
May 10 23:51:26.756 MYT: AToM[200.28.0.15, 116]: Stop resend label timer
May 10 23:51:26.756 MYT: AToM[200.28.0.15, 116]: Evt resend label timer expired, in local ready
May 10 23:51:26.756 MYT: AToM[200.28.0.15, 116]: . Resend label timer expired
May 10 23:51:27.100 MYT: AToM[200.28.0.15, 116]: Start resend label timer
May 10 23:51:27.100 MYT: AToM LDP[200.28.0.15, 116]: Receive label release
May 10 23:51:27.100 MYT: AToM[200.28.0.15, 116]: Evt remote release, in local ready
May 10 23:51:27.100 MYT: AToM[200.28.0.15, 116]: . Take no action
May 10 23:51:27.100 MYT: AToM: 3309 cumulative msgs handled. rc=0

ihsan
On May 10, 2012, at 11:19 PM, adam vitkovsky wrote:

> I was just thinking that maybe with the manual config one side gets selected
> as the session initiator while with bgp it's the other way around resulting
> in the fail -if indeed only one side is configured to accept the targeted
> sessions (but now that I think about it I guess it works the way the both
> ends initiate the session and than the session with lower id gets thorn
> down)
> And what does the debug says about the session not coming up please?
> 
> adam
> 
> -----Original Message-----
> From: Ihsan Junaidi Ibrahim [mailto:ihsan.junaidi at gmail.com] 
> Sent: Thursday, May 10, 2012 4:40 PM
> To: adam vitkovsky
> Cc: 'Pete Lumbis'; cisco-nsp at puck.nether.net
> Subject: Re: [c-nsp] VFI LDP transport signaled down (ME3600x)
> 
> Adam,
> 
> That's what I initially thought but an EoMPLS or a manual VPLS work just
> fine. Sample of EoMPLS,
> 
> es-103-glsfb#sh mpls l2transport vc 5070 detail Local interface: Gi0/19 up,
> line protocol up, Ethernet:1 up
>  Destination address: 200.28.0.120, VC ID: 5070, VC status: up
>    Output interface: Te0/2, imposed label stack {298 16}
>    Preferred path: not configured  
>    Default path: active
>    Next hop: 200.28.2.242
>  Create time: 1d02h, last status change time: 00:03:17
>  Signaling protocol: LDP, peer 200.28.0.120:0 up
>    Targeted Hello: 200.28.0.15(LDP Id) -> 200.28.0.120, LDP is UP
>    Status TLV support (local/remote)   : enabled/supported
>      LDP route watch                   : disabled
>      Label/status state machine        : established, LruRru
>      Last local dataplane   status rcvd: No fault
>      Last BFD dataplane     status rcvd: Not sent
>      Last BFD peer monitor  status rcvd: No fault
>      Last local AC  circuit status rcvd: No fault
>      Last local AC  circuit status sent: No fault
>      Last local LDP TLV     status sent: No fault
>      Last remote LDP TLV    status rcvd: No fault
>      Last remote LDP ADJ    status rcvd: No fault
>    MPLS VC labels: local 17, remote 16 
>    Group ID: local 0, remote 0
>    MTU: local 9178, remote 9178
>    Remote interface description: 
>  Sequencing: receive disabled, send disabled
>  Control Word: On (configured: autosense)
>  Dataplane:
>    SSM segment/switch IDs: 45083/8194 (used), PWID: 2
>  VC statistics:
>    transit packet totals: receive 1374, send 1374
>    transit byte totals:   receive 118164, send 87936
>    transit packet drops:  receive 0, seq error 0, send 0
> 
> Debugging did not turn up a whole lot of useful info that can be used to
> narrow down the problem.
> 
> Resetting the LDP neighbours to a clean state the proceeding logs only logs
> the EoMPLS targeted LDP session for VC 5070 (the above) but no information
> at all for VC 116 which is part of the VFI attached circuit.
> 
> ihsan
> 
> On May 10, 2012, at 9:42 PM, adam vitkovsky wrote:
> 
>> It almost appears like one of the routers doesn't accept targeted 
>> sessions Can you please check whether both ends are conf to accept ldp 
>> targeted sessions Or maybe the debug of targeted sessions would shed 
>> some light on why the session won't come up The bgp auto-discovery 
>> looks good
>> 
>> adam
>> 
>> -----Original Message-----
>> From: cisco-nsp-bounces at puck.nether.net 
>> [mailto:cisco-nsp-bounces at puck.nether.net] On Behalf Of Ihsan Junaidi 
>> Ibrahim
>> Sent: Thursday, May 10, 2012 6:00 AM
>> To: Pete Lumbis
>> Cc: cisco-nsp at puck.nether.net
>> Subject: Re: [c-nsp] VFI LDP transport signaled down (ME3600x)
>> 
>> On PE1,
>> 
>> es-103-glsfb#sh xconnect rib  detail
>> 
>> Local Router ID: 200.28.0.15
>> 
>> VPLS-ID: 9930:116, Target ID: 200.28.0.120  iBGP Peer
>> Next-Hop: 200.28.9.146
>> Hello-Source: 200.28.0.15
>> Route-Target: 9930:116
>> Incoming RD: 9930:116
>> Forwarder: VFI ME002555
>> Provisioned: Yes
>> NLRI handle: 69000001
>> 
>> PE2,
>> 
>> es-03-akhmw#sh xconnect rib detail
>> 
>> Local Router ID: 200.28.0.120
>> 
>> VPLS-ID: 9930:116, Target ID: 200.28.0.15  iBGP Peer
>> Next-Hop: 200.28.2.242
>> Hello-Source: 200.28.0.120
>> Route-Target: 9930:116
>> Incoming RD: 9930:116
>> Forwarder: VFI ME002617
>> Provisioned: Yes
>> NLRI handle: 77000001
>> 
>> On May 10, 2012, at 10:40 AM, Pete Lumbis wrote:
>> 
>>> What do you see in "show xconn rib"?
>>> 
>>> On Wed, May 9, 2012 at 10:36 AM, Ihsan Junaidi Ibrahim 
>>> <ihsan.junaidi at gmail.com> wrote:
>>>> Hi all,
>>>> 
>>>> My topology as follows:
>>>> 
>>>> PE1--P1--P2--P3--P4--P5--PE2
>>>> 
>>>> PE1 lo0 - 200.28.0.15 (15.2(2)S) loader 12.2(52r)EY1
>>>> PE2 lo0 - 200.28.0.120 (15.2(2)S) loader 12.2(52r)EY2
>>>> 
>>>> Are there specific nuances for an LDP signaled transport for EoMPLS 
>>>> and
>> VPLS in the Whales platform?
>>>> 
>>>> An xconnect from PE1 to PE2 is signaled successfully however a VPLS
>> instance based on BGP autodiscovery (manual VPLS works) is unable to 
>> bring up the LDP l2transport signal although the VFI is signaled up.
>>>> 
>>>> EoMPLS
>>>> ----
>>>> es-103-glsfb#sh xconnect peer 200.28.0.120 vc 5070
>>>> Legend:    XC ST=Xconnect State  S1=Segment1 State  S2=Segment2 State
>>>> UP=Up       DN=Down            AD=Admin Down      IA=Inactive
>>>> SB=Standby  HS=Hot Standby     RV=Recovering      NH=No Hardware
>>>> 
>>>> XC ST  Segment 1                         S1 Segment 2
>> S2
>>>> 
>> ------+---------------------------------+--+--------------------------
>> ------+---------------------------------+--+------
>> -+--
>>>> UP pri   ac Gi0/19:1(Ethernet)           UP mpls 200.28.0.120:5070
>> UP
>>>> 
>>>> es-103-glsfb#sh mpls l2transport vc 5070 detail Local interface: 
>>>> Gi0/19 up, line protocol up, Ethernet:1 up  Destination address: 
>>>> 200.28.0.120, VC ID: 5070, VC status: up
>>>>  Output interface: Te0/2, imposed label stack {298 16}
>>>>  Preferred path: not configured
>>>>  Default path: active
>>>>  Next hop: 200.28.2.242
>>>> Create time: 02:10:43, last status change time: 02:08:57  Signaling
>>>> protocol: LDP, peer 200.28.0.120:0 up
>>>>  Targeted Hello: 200.28.0.15(LDP Id) -> 200.28.0.120, LDP is UP
>>>>  Status TLV support (local/remote)   : enabled/supported
>>>>    LDP route watch                   : disabled
>>>>    Label/status state machine        : established, LruRru
>>>>    Last local dataplane   status rcvd: No fault
>>>>    Last BFD dataplane     status rcvd: Not sent
>>>>    Last BFD peer monitor  status rcvd: No fault
>>>>    Last local AC  circuit status rcvd: No fault
>>>>    Last local AC  circuit status sent: No fault
>>>>    Last local LDP TLV     status sent: No fault
>>>>    Last remote LDP TLV    status rcvd: No fault
>>>>    Last remote LDP ADJ    status rcvd: No fault
>>>>  MPLS VC labels: local 17, remote 16
>>>>  Group ID: local 0, remote 0
>>>>  MTU: local 9178, remote 9178
>>>>  Remote interface description:
>>>> Sequencing: receive disabled, send disabled  Control Word: On
>>>> (configured: autosense)
>>>> Dataplane:
>>>>  SSM segment/switch IDs: 45083/8194 (used), PWID: 2  VC statistics:
>>>>  transit packet totals: receive 24, send 21
>>>>  transit byte totals:   receive 2064, send 1344
>>>>  transit packet drops:  receive 0, seq error 0, send 0
>>>> 
>>>> VPLS
>>>> ----
>>>> es-103-glsfb#sh vfi
>>>> Legend: RT=Route-target, S=Split-horizon, Y=Yes, N=No
>>>> 
>>>> VFI name: ME002555, state: up, type: multipoint signaling: LDP  VPN
>>>> ID: 116, VPLS-ID: 9930:116
>>>> RD: 9930:116, RT: 9930:116
>>>> Bridge-Domain 116 attachment circuits:
>>>>  Vlan116
>>>> Neighbors connected via pseudowires:
>>>> Peer Address     VC ID        Discovered Router ID    S
>>>> 200.28.9.146     116          200.28.0.120            Y
>>>> 
>>>> es-103-glsfb#sh mpls l2transport vc 116 detail Local interface: VFI
>>>> ME002555 vfi up  Interworking type is Ethernet  Destination address: 
>>>> 200.28.0.120, VC ID: 116, VC status: down
>>>>  Last error: Local access circuit is not ready for label advertise 
>>>> Next hop PE address: 200.28.9.146
>>>>  Output interface: none, imposed label stack {}
>>>>  Preferred path: not configured
>>>>  Default path: no route
>>>>  No adjacency
>>>> Create time: 02:07:55, last status change time: 02:07:55  Signaling
>>>> protocol: LDP, peer unknown
>>>>  Targeted Hello: 200.28.0.15(LDP Id) -> 200.28.9.146, LDP is DOWN, 
>>>> no
>> binding
>>>>  Status TLV support (local/remote)   : enabled/None (no remote binding)
>>>>    LDP route watch                   : disabled
>>>>    Label/status state machine        : local standby, AC-ready, LnuRnd
>>>>    Last local dataplane   status rcvd: No fault
>>>>    Last BFD dataplane     status rcvd: Not sent
>>>>    Last BFD peer monitor  status rcvd: No fault
>>>>    Last local AC  circuit status rcvd: No fault
>>>>    Last local AC  circuit status sent: Not sent
>>>>    Last local LDP TLV     status sent: None
>>>>    Last remote LDP TLV    status rcvd: None (no remote binding)
>>>>    Last remote LDP ADJ    status rcvd: None (no remote binding)
>>>>  MPLS VC labels: local 23, remote unassigned
>>>>  AGI: type 1, len 8, 000A 26CA 0000 0074
>>>>  Local AII: type 1, len 4, DF1C 000F (200.28.0.15)
>>>>  Remote AII: type 1, len 4, DF1C 0078 (200.28.0.120)
>>>>  Group ID: local n/a, remote unknown
>>>>  MTU: local 9178, remote unknown
>>>>  Remote interface description:
>>>> Sequencing: receive disabled, send disabled  Control Word: On
>>>> (configured: autosense)
>>>> Dataplane:
>>>>  SSM segment/switch IDs: 0/0 (used), PWID: 14  VC statistics:
>>>>  transit packet totals: receive 0, send 0
>>>>  transit byte totals:   receive 0, send 0
>>>>  transit packet drops:  receive 0, seq error 0, send 0
>>>> 
>>>> I'm getting the account team into the loop but if anyone has 
>>>> encountered
>> this scenario before and managed to find the answer, that would be 
>> most helpful.
>>>> 
>>>> ihsan
>>>> _______________________________________________
>>>> cisco-nsp mailing list  cisco-nsp at puck.nether.net 
>>>> https://puck.nether.net/mailman/listinfo/cisco-nsp
>>>> archive at http://puck.nether.net/pipermail/cisco-nsp/
>> 
>> 
>> _______________________________________________
>> cisco-nsp mailing list  cisco-nsp at puck.nether.net 
>> https://puck.nether.net/mailman/listinfo/cisco-nsp
>> archive at http://puck.nether.net/pipermail/cisco-nsp/
>> 
> 
> 




More information about the cisco-nsp mailing list