MPLS Configuration Guide for Cisco NCS 5500 Series Routers, IOS XR Release 7.5.x
Bias-Free Language
The documentation set for this product strives to use bias-free language. For the purposes of this documentation set, bias-free is defined as language that does not imply discrimination based on age, disability, gender, racial identity, ethnic identity, sexual orientation, socioeconomic status, and intersectionality. Exceptions may be present in the documentation due to language that is hardcoded in the user interfaces of the product software, language used based on RFP documentation, or language that is used by a referenced third-party product. Learn more about how Cisco is using Inclusive Language.
MPLS (Multi Protocol Label Switching) is a forwarding mechanism based on label switching. In an MPLS network, data packets
are assigned labels and packet-forwarding decisions are taken based on the contents of the label. To switch labeled packets
across the MPLS network, predetermined paths are established for various source-destination pairs. These predetermined paths
are known as Label Switched Paths (LSPs). To establish LSPs, MPLS signaling protocols are used. Label Distribution Protocol
(LDP) is an MPLS signaling protocol used for establishing LSPs. This module provides information about how to configure MPLS
LDP.
IPv6 Support in MPLS LDP-MPLS LDPv6 makes the LDP control plane to run on IPv6 in order to setup LSPs for IPv6 prefixes. You can enable MPLS LDPv6
along with existing IPv4 services. LDPv6 feature support is explained in the IPv6 Support in MPLS LDP section.
Note
In the 7.2.1 release, LDPv6 is only supported for routers in an LSR role, and not the LDP Label Edge Router (LER) role.
To allow hashing for the Label Edge Router (LER) and Label Switched Routers (LSRs) with MPLS traffic or algorithm to use the
inner ethernet fields of the source MAC and destination MAC addresses, use the hw-module profile load-balance algorithm command with a suitable load-balancing profile.
Prerequisites for
Implementing MPLS Label Distribution Protocol
The following are the
prerequisites to implement MPLS LDP:
You must be in a user group associated with a task group that includes the proper task IDs. The command reference guides include
the task IDs required for each command. If you suspect user group assignment is preventing you from using a command, contact
your AAA administrator for assistance.
You must be
running
Cisco IOS XR software.
You must install a composite mini-image and the MPLS package.
You must activate
IGP.
We recommend to use a lower session holdtime bandwidth such as neighbors so that a session down occurs before an adjacency-down
on a neighbor. Therefore, the following default values for the hello times are listed:
Holdtime is 15 seconds.
Interval is 5 seconds.
For example, the
LDP session holdtime can be configured as 30 seconds by using the
holdtime
command.
Restrictions for MPLS LDP
When paths of different technologies are resolved over ECMP, it results in heterogeneous ECMP, leading to severe network traffic issues. Don’t use ECMP for any combination of the following technologies:
LDP.
BGP-LU, including services over BGP-LU loopback peering or recursive services at Level-3.
VPNv4.
6PE and 6VPE.
EVPN.
Recursive static routing.
Overview of Label
Distribution Protocol
Table 1. Feature History Table
Feature Name
Release Information
Feature Description
BFD, LACP Triggering TE FRR
Release 7.3.1
This feature is now supported on routers that have Cisco NC57 line
cards installed and operate in the native mode.
Table 2. Feature History Table
Feature Name
Release Information
Feature Description
Targeted LDP
Release 7.3.1
This feature is now supported on routers that have Cisco NC57 line
cards installed and operate in the native mode.
In IP forwarding, when
a packet arrives at a router the router looks at the destination address in the
IP header, performs a route lookup, and then forwards the packet to the next
hop. MPLS is a forwarding mechanism in which packets are forwarded based on
labels. Label Distribution Protocols assign, distribute, and install the labels
in an MPLS environment. It is the set of procedures and messages by which Label
Switched Routers (LSRs) establish LSPs through a network by mapping
network-layer routing information directly to data-link layer switched paths.
These LSPs may have an endpoint at a directly attached neighbor (comparable to
IP hop-by-hop forwarding), or may have an endpoint at a network egress node,
enabling switching via all intermediary nodes.
LSPs can be created statically, by RSVP traffic engineering (TE), or by LDP. LSPs created by LDP perform hop-by-hop path setup
instead of an end-to-end path. LDP enables LSRs to discover their potential peer routers and to establish LDP sessions with
those peers to exchange label binding information. Once label bindings are learned, the LDP is ready to set up the MPLS forwarding
plane.
Depending on the requirements, LDP requires some basic configuration tasks described in the following topics:
Configuring Label
Distribution Protocol
This section explains
the basic LDP configuration. LDP should be enabled on all interfaces that
connects the router to potential LDP peer routers. You can enable LDP on an
interface by specifying the interface under mpls ldp configuration mode.
Configuration
Example
This example shows
how to enable LDP over an interface.
Configuring Label
Distribution Protocol Discovery Parameters
LSRs that are running LDP send hello messages on all the LDP enabled interfaces to discover each other. So, the LSR that receives
the LDP hello message on an interface is aware of the presence of the LDP router on that interface. If LDP hello messages
are sent and received on an interface, there’s an LDP adjacency across the link between the two LSRs that are running LDP.
By default, hello messages are sent every 5 seconds with a hold time of 15 seconds. If the LSR doesn’t receive a discovery
hello from peer before the hold time expires, the LSR removes the peer LSR from the list of discovered LDP neighbors. The
LDP discovery parameters can be configured to change the default parameters.
LDP session between LSRs that aren’t directly connected is known as targeted LDP session. For targeted LDP sessions, LDP uses
targeted hello messages to discover the extended neighbors. By default, targeted hello messages are sent every 10 seconds
with a hold time of 90 seconds.
Configuration
Example
This example shows how to configure the following LDP discovery parameters:
This section
verifies the MPLS LDP discovery parameters configuration.
RP/0/RP0/CPU0:Router# show mpls ldp parameters
LDP Parameters:
Role: Active
Protocol Version: 1
Router ID: 192.168.70.1
Discovery:
Link Hellos: Holdtime:30 sec, Interval:10 sec
Targeted Hellos: Holdtime:120 sec, Interval:15 sec
Quick-start: Enabled (by default)
Transport address: IPv4: 192.168.70.1
Label Distribution
Protocol Discovery for Targeted Hellos
LDP session between LSRs that aren’t directly connected is known as targeted LDP session. For LDP neighbors which aren’t directly
connected, you should manually configure the LDP neighborship on both the routers.
Configuration
Example
This example shows how to configure LDP for non-directly connected routers, Router 1, and Router 2.
LDP allows you to
control the advertising and receiving of labels. You can control the exchange
of label binding information by using label advertisement control (outbound
filtering ) or label acceptance control (inbound filtering).
Label
Advertisement Control (Outbound Filtering)
Label Distribution
Protocol advertises labels for all the prefixes to all its neighbors. When this
is not desirable (for scalability and security reasons), you can configure LDP
to perform outbound filtering for local label advertisement for one or more
prefixes to one more peers. This feature is known as LDP outbound label
filtering, or local label advertisement control. You can control the exchange
of label binding information using the
mpls ldp
label advertise
command. Using the optional keywords, you can advertise
selective prefixes to all neighbors, advertise selective prefixes to defined
neighbors, or disable label advertisement to all peers for all prefixes.
Prefixes and peers advertised selectively are defined in the access list.
Configuration
Example: Label Advertisement Control
This example shows
how to configure outbound label advertisement control. In this example,
neighbors are specified to advertise and receive label advertisements. Also an
interface is specified for label advertisement.
RP/0/RP0/CPU0:Router(config)# mpls ldp
RP/0/RP0/CPU0:Router(config-ldp)# address-family ipv4
RP/0/RP0/CPU0:Router(config-ldp-af)# label local advertise to 10.0.0.1:0 for pfx_ac11
RP/0/RP0/CPU0:Router(config-ldp-af)# label local advertise interface TenGigE 0/0/0/5
RP/0/RP0/CPU0:Router(config-ldp-af)# commit
Label
Acceptance Control (Inbound Filtering)
LDP accepts labels
(as remote bindings) for all prefixes from all peers. LDP operates in liberal
label retention mode, which instructs LDP to keep remote bindings from all
peers for a given prefix. For security reasons, or to conserve memory, you can
override this behavior by configuring label binding acceptance for set of
prefixes from a given peer. The ability to filter remote bindings for a defined
set of prefixes is also referred to as LDP inbound label filtering or label
acceptance control.
Configuration
Example : Label Acceptance Control (Inbound Filtering)
This example shows
how to configure label acceptance control. In this example, an LSR is
configured to accept and retain label bindings from neighbors for prefixes
defined in access list .
RP/0/RP0/CPU0:Router(config)#mpls ldp
RP/0/RP0/CPU0:Router(config-ldp)#address-family ipv4
RP/0/RP0/CPU0:Router(config-ldp-af)#label remote accept from 192.168.1.1:0 for acl_1
RP/0/RP0/CPU0:Router(config-ldp-af)#label remote accept from 192.168.2.2:0 for acl_2
RP/0/RP0/CPU0:Router(config-ldp-af)#commit
Configuring Local
Label Allocation Control
LDP creates label bindings for all IGP prefixes and receives label
bindings for all IGP prefixes from all its peers. If an LSR receives label
bindings from several peers for thousands of IGP prefixes, it consumes
significant memory and CPU. In some scenarios, most of the LDP label bindings
may not useful for any application and you may required to limit the allocation
of local labels. This is accomplished using LDP local label allocation control,
where an access list can be used to limit allocation of local labels to a set
of prefixes. Limiting local label allocation provides several benefits,
including reduced memory usage requirements, fewer local forwarding updates,
and fewer network and peer updates.
Configuration Example
This example shows how to configure local label allocation using an IP
access list to specify a set of prefixes that local labels can allocate and
advertise.
RP/0/RP0/CPU0:Router(config)# mpls ldp
RP/0/RP0/CPU0:Router(config-ldp)# address-family ipv4
RP/0/RP0/CPU0:Router(config-ldp-af)# label local allocate for pfx_acl_1
RP/0/RP0/CPU0:Router(config-ldp-af)# commit
Configuring
Downstream on Demand
By default, LDP uses downstream unsolicited mode in which label
advertisements for all routes are received from all LDP peers. The downstream
on demand feature adds support for downstream-on-demand mode, where the label
is not advertised to a peer, unless the peer explicitly requests it. At the
same time, since the peer does not automatically advertise labels, the label
request is sent whenever the next-hop points out to a peer that no remote label
has been assigned.
In downstream on demand configuration, an ACL is used to specify the set
of peers for downstream on demand mode. For down stream on demand to be
enabled, it needs to be configured on both peers of the session. If only one
peer in the session has downstream-on-demand feature configured, then the
session does not use downstream-on-demand mode.
Configuration Example
This example shows how to configure LDP Downstream on Demand.
RP/0/RP0/CPU0:Router(config)# mpls ldp
RP/0/RP0/CPU0:Router(config-ldp)# session downstream-on-demand with ACL1
RP/0/RP0/CPU0:Router(config-ldp)# commit
Configuring Explicit
Null Label
Cisco MPLS LDP uses
implicit or explicit null label as local label for routes or prefixes that
terminate on the given LSR. These routes include all local, connected, and
attached networks. By default, the null label is
implicit-null
that allows LDP control plane to implement penultimate hop popping (PHP)
mechanism. When this is not desirable, you can configure
explicit-null
label that allows LDP control plane to implement ultimate hop popping (UHP)
mechanism. You can configure explicit-null feature on the ultimate hop LSR.
Access-lists can be used to specify the IP prefixes for which PHP is desired.
You can enforce
implicit-null local label for a specific prefix by using the
implicit-null-override command even if the prefix
requires a non-null label to be allocated by default. For example, by default,
an LSR allocates and advertises a non-null label for an IGP route. If you wish
to terminate LSP for this route on penultimate hop of the LSR, you can enforce
implicit-null label allocation and advertisement for this prefix using the
implicit-null-override
command.
Note
If the outgoing label is implicit-null on the penultimate hop (of the label switched path), the outermost label is removed, and the payload forwarded. The payload
is accounted as MPLS even if it is IP traffic, due to ASIC limitations in identifying the egress packet type correctly.
Configuration
Example: Explicit Null
This example shows
how to configure explicit null label.
This example shows
how to configure implicit null override for a set of prefixes.
RP/0/RP0/CPU0:Router(config)# mpls ldp
RP/0/RP0/CPU0:Router(config-ldp)# address-family ipv4
RP/0/RP0/CPU0:Router(config-ldp-af)# label local advertise implicit-null-override for acl-1
RP/0/RP0/CPU0:Router(config-ldp-af)# commit
Label Distribution
Protocol Auto-configuration
LDP auto-configuration allows you to automatically configure LDP on all
interfaces for which the IGP protocol is enabled. Typically, LDP assigns and
advertises labels for IGP routes and must often be enabled on all active
interfaces by an IGP. During LDP manual configuration, you must define the set
of interfaces under LDP which is a time-intensive procedure. LDP
auto-configuration eliminates the need to specify the same list of interfaces
under LDP and simplifies the configuration tasks.
Configuration Example: Enabling LDP Auto-Configuration for
OSPF
This example shows how to enable LDP auto-configuration for a
specified OSPF instance.
When a new link or
node comes up after a link failure, IP converges earlier and much faster than
MPLS LDP and may result in MPLS traffic loss until the MPLS convergence. If a
link flaps, the LDP session also flaps due to loss of link discovery. LDP
session protection minimizes traffic loss, provides faster convergence, and
protects existing LDP (link) sessions. When session protection is enabled for a
peer, LDP starts sending targeted hello (directed discovery) in addition to
basic discovery link hellos. When the direct link goes down, the targeted
hellos can still be forwarded to the peer LSR over an alternative path as long
as there is one. So, the LDP session stays up after the link goes down.
You can configure LDP
session protection to automatically protect sessions with all or a given set of
peers (as specified by peer-acl). When configured, LDP initiates backup
targeted hellos automatically for neighbors for which primary link adjacencies
already exist. These backup targeted hellos maintain LDP sessions when primary
link adjacencies go down.
Configuration
Example
This example shows
how to configure LDP session protection for peers specified by the access
control list peer-acl-1 for a maximum duration of 60 seconds.
Configuring Label
Distribution Protocol- Interior Gateway Protocol (IGP) Synchronization
Lack of
synchronization between LDP and Interior Gateway Protocol (IGP) can cause MPLS
traffic loss. Upon link up, for example, IGP can advertise and use a link
before LDP convergence has occurred or, a link may continue to be used in IGP
after an LDP session goes down.
LDP IGP
synchronization coordinates LDP and IGP so that IGP advertises links with
regular metrics only when MPLS LDP is converged on that link. LDP considers a
link converged when at least one LDP session is up and running on the link for
which LDP has sent its applicable label bindings and received at least one
label binding from the peer. LDP communicates this information to IGP upon link
up or session down events and IGP acts accordingly, depending on sync state.
LDP-IGP
synchronization is supported for both OSPF and ISIS protocols and is configured
under the corresponding IGP protocol configuration mode. Under certain
circumstances, it might be required to delay declaration of re-synchronization
to a configurable interval. LDP provides a configuration option to delay
declaring synchronization up for up to 60 seconds. LDP communicates this
information to IGP upon linkup or session down events.
From the 7.1.1 release, you can configure multiple MPLS-TE tunnel end points on an LER using the TLV 132 function in IS-IS.
You can configure a maximum of 63 IPv4 addresses or 15 IPv6 addresses on an LER.
Configuring LDP
IGP Synchronization: Open Shortest Path First (OSPF) Example
This example shows
how to configure LDP-IGP synchronization for an OSPF instance. The
synchronization delay is configured as 30 seconds.
Configuring Label
Distribution Protocol Graceful Restart
LDP Graceful Restart
provides a mechanism for LDP peers to preserve the MPLS forwarding state when
the LDP session goes down. Without LDP Graceful Restart, when an established
session fails, the corresponding forwarding states are cleaned immediately from
the restart and peer nodes. In this case, LDP forwarding has to restart from
the beginning, causing a potential loss of data and connectivity. If LDP
graceful restart is configured, traffic can continue to be forwarded without
interruption, even when the LDP session restarts. The LDP graceful restart
capability is negotiated between two peers during session initialization time.
During session initialization, a router advertises its ability to perform LDP
graceful restart by sending the graceful restart typed length value (TLV). This
TLV contains the reconnect time and recovery time. The values of the reconnect
and recovery times indicate the graceful restart capabilities supported by the
router. The reconnect time is the amount of time the peer router waits for the
restarting router to establish a connection. When a router discovers that a
neighboring router is restarting, it waits until the end of the recovery time
before attempting to reconnect. Recovery time is the amount of time that a
neighboring router maintains its information about the restarting router.
Configuration
Example
This example shows
how to configure LDP graceful restart. In this example, the amount of time that
a neighboring router maintains the forwarding state about the gracefully
restarting router is specified as 180 seconds. The reconnect time is configured
as 169 seconds.
Configuring Label
Distribution Protocol Nonstop Routing
LDP nonstop routing
(NSR) functionality makes failures, such as Route Processor (RP) or Distributed
Route Processor (DRP) fail over, invisible to routing peers with minimal to no
disruption of convergence performance. By default, NSR is globally enabled on
all LDP sessions except AToM.
A disruption in
service may include any of these events:
Route processor
(RP) or distributed route processor (DRP) failover
LDP process
restart
Minimum disruption
restart (MDR)
Note
Unlike graceful
restart functionality, LDP NSR does not require protocol extensions and does
not force software upgrades on other routers in the network, nor does LDP NSR
require peer routers to support NSR. L2VPN configuration is not supported on
NSR. Process failures of active LDP results in session loss and, as a result,
NSR cannot be provided unless RP switchover is configured as a recovery action.
Configuration
Example
This example shows
how to configure LDP Non-Stop Routing.
RP/0/RP0/CPU0:Router# show mpls ldp nsr summary
Mon Dec 7 04:02:16.259 UTC
Sessions:
Total: 1, NSR-eligible: 1, Sync-ed: 0
(1 Ready)
Configuring LDPv6
The LDP configuration model is extended to introduce IPv6 as an option under the address family submodes that reside under
LDP global and interface configurations. IPv6 address family is available under LDP global, LDP VRF global and interface configurations.
LDPv6 is supported only under default VRF. LDPv6 should be enabled on all interfaces that connects the router to potential
LDPv6 peer routers.
Note
In the 7.2.1 release, LDPv6 is only supported for routers in an LSR role, and not the LDP Label Edge Router (LER) role.
Restrictions for MPLS LDPv6
MPLS LDPv6 has the following restrictions:
IPv6 address family is supported only under default VRF.
Implicit enabling of IPv6 address family is not allowed. It needs explicit enabling.
It is recommended to configure a routable IPv6 discovery transport address when only LDP IPv6 is configured without explicitly
specifying a router-id.
IPv6 Support in MPLS LDP
MPLS LDPv6 makes the LDP control plane to run on IPv6 in order to setup LSPs for IPv6 prefixes. This support enables most
of the LDP functions supported on IPv4 to be extended to IPv6. In this context, support for native MPLS LDP over IPv6 is provided
in order to seamlessly continue providing existing services while enabling new ones.
LDP associates a forwarding equivalence class (FEC) with each label switched path (LSP) it creates. The FEC associated with
an LSP specifies which packets are mapped to that LSP. LDP establishes sessions with peers and exchanges FEC label bindings
with them to enable creation of LSPs to carry MPLS traffic destined to IP prefixes.
As per RFC 5036, LDP base specification defines procedures and messages for exchanging bindings for IPv4 and IPv6 addresses
and routing prefixes. LDPv6 related RFCs explain control plane and binding advertisement support for LDPv6.
The procedures of address bindings, label bindings, and forwarding setup are same for IPv4 and IPv6 address families in LDP.
The only difference is that a different address format is used according to the IP address family. While a single-stack IP
address family (IPv4-only or IPv6-only) enabled interfaces between a set of routers is the most typical deployment, scenarios
for LSR interconnections using both IPv4 and IPv6 interfaces are also supported.
LDP functionality can be broadly divided into two categories, control plane and LSR setup.
Control plane includes these functions - neighbor discovery (hello adjacencies), transport connection/endpoint (TCP connection),
session and peering, and bindings exchange.
LSP setup includes these functions - acquire FEC information through RIB, assign and advertise local label bindings for FEC,
advertise local (interface) IP address bindings and setup forwarding rewrites.
For the control plane, the underlying address family can be either IPv4-only, IPv6-only or both. Whereas for the LSP setup,
an LSP is setup for IPv4 or IPv6 FEC prefix.
This figure illustrates the main LDPv6 components.
LDP functions in an MPLS LDPv6 setup:
Receive routing updates from routing information base (RIB) for global IPv6 prefixes
Assign local labels for IPv6 prefixes
Receive IPv6 address or state notifications for local IPv6 enabled interfaces from IP Address Repository Manager (IP-ARM/IM)
and LAS for IPv6 link-local unicast addresses
Advertise/Accept IPv6 label bindings and address bindings to/from peers
Setup MPLS forwarding to create IPv6 LSPs
Provide IPv6 LSP information to MPLS OAM as and when requested
Service MIB requests for IPv6 control plane queries and generate MIB traps
Provide LDPv6 convergence status for a link to IGP for LDP-IGP Sync feature for IPv6
Topology Scenarios
A typical deployment scenario consists of single-stack IP address-family (IPv4-only or IPv6-only) enabled interfaces between
a set of routers. The following are some topology scenarios, and a description of the control plane and LSP setup scenarios.
Here, R2 is the reference router.
One dual-stack interface/same neighbor
R1_ _ _ _ _ _ _ _R2
IPv4+IPv6
Neighbor Discovery - IPv4 and IPv6 Hellos are sent on the interface to R1.
Transport Connection - IPv4 endpoints or IPv6 endpoints (as per user preference).
Label binding exchange - IPv4 and IPv6 prefixes.
Address binding exchange - IPv4 and IPv6 addresses.
LSPs - IPv4 and IPv6 over the same nexthop interface to R1.
Two single-stack interfaces/same neighbor:
1. (IPv4)
R1_ _ _ _ _ _ _ _ _R2
_ _ _ _ _ _ _ _ _
2. (IPv6)
Neighbor Discovery - IPv4 Hellos on interface-1 to R1, and IPv6 Hellos on interface-2 to R1.
Transport Connection - IPv4 endpoints or IPv6 endpoints (as per user preference).
Label binding exchange - IPv4 and IPv6 prefixes.
Address binding exchange - IPv4 and IPv6 addresses.
LSPs - IPv4 over nexthop interface-1 to R1, and IPv6 over nexthop interface-2 to R1.
Two single-stack interfaces/different neighbors with different address families:
1. (IPv4) 2. (IPv6)
R1_ _ _ _ _ _ _ _ R2_ _ _ _ _ _ _ _ R3
Neighbor Discovery - IPv4 Hellos on interface-1 to R1, and IPv6 Hellos on interface-2 to R3.
Transport Connection - IPv4 endpoints with R1 and IPv6 endpoints with R3.
Label binding exchange - IPv4 and IPv6 prefixes to R1 and R3.
Even if all three LSRs are dual-stack, traffic from R1 to R3 will not be completely labeled.
If there is IPv6 traffic, it is unlabeled from R1 to R2. Labels are imposed only at R2 (although in this specific case implicit
null imposition) to R3.
If there is IPv4 traffic, it is labeled from R1 to R2. But the traffic will go unlabeled between R2 and R3 given that no IPv4
adjacency exists between R2 and R3
Address binding exchange - IPv4 and IPv6 addresses to R1 and R3
LSPs - IPv4 over nexthop interface-1 to R1 and IPv6 over nexthop interface-2 to R3
Feature Support in LDPv6
The following features are supported in LDPv6:
Single-stack (native IPv6) and dual-stack (IPv4+IPv6) topologies.
New operating modes in LDP:
Native LDPv6
LDPv6 over IPv4 and LDPv4 over IPv6 connection endpoints
LDP Hellos carry optional transport address type length value (TLV) to notify a peer about TCP or transport connection endpoint.
An LSR can include either IPv4 or IPv6 transport address TLV in an IPv4 or IPv6 Hello message. There is no difference in the
TLV format of transport address for IPv4 and IPv6.
Only one transport connection is established between two discovered peers, whether there be single address family Hello adjacencies
or multi-address family (both IPv4 and IPv6) Hello adjacencies.
In a dual-stack setup, when LDP has the option to establish transport connection either using IPv4 endpoints or IPv6 endpoints,
IPv6 connection is preferred over IPv4 connection. If LDP is locally enabled for both IPv4 and IPv6 address families, every
new session is treated as potential dual-stack connection. Under such circumstances, IPv6 preference is kept in place for
maximum fifteen seconds for the session to establish, after which the LDP tries to establish a connection with the peer using
IPv4. A user can override this default behavior by specifying the preference for a set of dual-stack peers to use IPv4 transport
for the connection. Furthermore, a user may also specify maximum wait time to wait to establish the preferred transport connection.
If the preferred transport establishment times out, LDP tries to establish connection with other non-preferred transport address
families. This applies to both the cases when an LSR acts as active side or passive side for the TCP connection.
To override default IPv6 transport preference for dual-stack cases, use the mpls ldp neighbor dual-stack transport-connection prefer ipv4 command. To specify the maximum time the preferred address family connection must wait to establish a connection before resorting
to a non-preferred address family, use the mpls ldp neighbor dual-stack transport-connection max-wait command.
Once a transport connection is established, it is not torn down depending on preferences. If the address family related to
established transport connection is disabled under LDP, the corresponding transport connection is reset to reestablish the
connection.
For a single-stack setup, there is no contention; the transport connection uses the given address family.
LDP Control Plane is IPv6 aware
LDPv6 LSP forwarding setup - LDP interacts with LSD in order to setup IPv6 LSP forwarding. The steps involved in this interaction
are:
Label allocation for an IPv6 prefix is learnt from RIB.
Setup imposition and label switching forwarding path for given IPv6 prefix by creating IPv6 forwarding rewrites.
Like LDPv4, rewrite delete and label free operations are performed when a route disappears or is disallowed under LDP due
to label policy.
There is no new requirement related to MPLS enabling or disabling. LDP also MPLS-enables in LSD (if not already) any LDP enabled
interface, which is in the UP state for IP4 and/or IPv6 and has IPv4 and/or IPv6 addresses assigned.
In case of dual-stack LDP, a single Resource-Complete is sent by LDP to LSD once RIB-Converged notification is received for
both IPv4 and IPv6 redistribute tables.
Distribution of IPv4 and IPv6 bindings over a single LDP session established over IPv4 or IPv6
LDP Downstream on Demand
LDP session protection
LDP session protection is a feature to protect an LDPv6 session. In case of dual-stack hello adjacencies with a peer, there
is only a single targeted hello adjacency to protect the session. Session protection forms targeted adjacency of address family
same as the transport connection. For IPv6, the target of the session protection is the remote transport connection endpoint.
For IPv4, the target of the session protection is remote LSR ID.
LDP IGPv6 sync on IPv6 interface
This feature lets IGP support LDP IGP Sync feature for IPv6 address family. This means that Intermediate System-to-Intermediate
System (IS-IS) allows IGP under an interface’s IPv6 address family, whereas OSPFv3 implements it just like existing support
in OSPF for IPv4. When the IGP Sync feature is enabled, LDP convergence status on an interface is considered by the IGP under
the context of a given address family. This behavior applies to IGP Sync for both non-TE as well as TE tunnel interfaces.
LDP Typed Wildcard for IPv6 prefix FEC
This feature adds support for Typed Wildcard for IPv6 Prefix FEC. The support includes:
Being able to send or receive IPv6 Prefix Typed Wildcard FEC element in label messages.
Respond to Typed Wildcard Label Requests received from peer by replaying its label database for IPv6 prefixes.
Make use of Typed Wildcard Label Requests towards peers to request replay of peer label database for IPv6 prefixes. For example,
on local inbound policy changes.
Label allocation, advertisement and accept policies for IPv6 prefixes
Local label assignment and advertisement for IPv6 default-route (::/0)
Session MD5 authentication for IPv6 transport
IPv6 Explicit-Null label
IPv6 explicit null label feature support includes:
Advertisement and receipt of IPv6 explicit-null label to and from peers.
IPv6 explicit-null outgoing label in forwarding setup.
Explicit-null advertisement policy for a set of IPv6 prefixes and/or set of peers.
Explicit-null configuration change. Change in explicit-null configuration is handled by first transferring a wildcard withdraw
with null label to peer(s), followed by advertising the appropriate null (implicit or explicit) label to the peer(s) again.
This works without any issue as long as a single IP address family is enabled. In case of a dual-stack LSR peer, a change
of configuration related to explicit-null advertisement for a given address family may cause unnecessary mix-up in the other
address family.
LDPv6 LFA FRR
Local LFA FRR for IPv6 is supported. However, it is required that the primary and backup paths are of the same address family
type, that is, an IPv4 primary path must not have an IPv6 backup path.
NSF for LDPv6 traffic
Non-stop forwarding (NSF) support is either provided through LDP NSR or graceful restart mechanisms.
IGP/LDP NSR for IPv6
IGP/LDP Graceful Restart for IPv6
MPLS OAM: New FECs
LSPV supports two new FECs.
LDPv6 Prefix FEC Encoding/Decoding
Label Switched Path Verification (LSPV) encodes/decodes the LDP IPv6 Prefix FEC. Prefix is in the network byte order and the
trailing bits are to be set to zero when prefix length is shorter than 128 bits.
Generic IPv6 Prefix FEC Encoding/Decoding
LSPV encodes/decodes the generic IPv6 Prefix FEC. Prefix is in the network byte order and the trailing bits are to be set
to zero when prefix length is shorter than 128 bits.
Generic IPv6 FEC is used in addition to the LDPv6 FEC. This serves the following primary purposes:
Allows user to perform LSP ping and traceroute to verify data plane without involving control plane of the FEC in echo request
and response.
If support for a new FEC is preferred in the future, the generic FEC can be used until corresponding control plane is explicitly
supported by LSPV.
IPv6 LSR MIB
MPLS OAM LDP MIBS is extended to support IPv6. All LSR MIB objects that reference an InSegment prefix and OutSegment next
hop address are modified to support IPv6.
LSP ping support for LDPv6
LSP trace-route support for LDPv6
LSP tree-trace support for LDPv6
Scale
The same support as LDPv4 native is provided for LDPv6 native scale
Dual-stack–The aggregate scale of LDPv4 and LDPv6 is the same as the currently support for LDPv4 native scale
Unsupported Features in LDPv6
LDPv6 over TEv4 (traffic engineering)
Interfaces
LDP auto-config for IPv6
LDPv6 over TEv6
LDPv6 over GREv6
LDP auto-config for IPv6 IGP
LDP Label Edge Router (LER) function
Remote LFA FRR
L2VPN
L2VPN over IPv6 LSPs
L2VPN signaling with LDP when the nexthop address is IPv6
IPv6 BGP Redistribution
Applications with native LDPv6
Multicast extension to LDP (mLDP) for IPv6 FEC with label binding through IPv4 and IPv6 transport
ICCP - ICCP and LDP ICCP with IPv6 neighbor node
PW
L3VPN
Native IPv6 MPLS L3VPNs
4PE
4vPE
LDPv6 CSC
IPv6 Label Bindings
LDP stores label bindings associated with FEC prefix in its Label Information Base (LIB) [TIB in Cisco LDP]. An entry in LIB
corresponds to a prefix and holds the following bindings:
Local binding: Local label assigned for this prefix (which is learnt through local RIB)
Remote bindings: Array of peer labels (prefix-label bindings received in label mapping message from peers)
An entry in LIB can exist due to local binding presence, or due to remote binding(s) presence, or due to both local and remote
bindings presence. The forwarding setup, however, mandates that local binding be present for a prefix.
Extensions have been implemented to support IPv6 prefixes for LIB in LDP. For per-address family convergence or preference
reasons, separate or new LIB is implemented to keep and maintain IPv6 prefixes. In case of dual-stack LDP, LIBv4 is preferred
over LIBv6 wherever possible. For example, during background housekeeping function, LIBv4 is processed before LIBv6.
IPv6 Address Bindings
LDP needs to maintain IPv6 address database for local and peer interface addresses. The IPv4 address module for local/peer
addresses is extended to keep IPv4/IPv6 addresses in their respective databases, much like LIB database. In case of a dual-stack
LDP, IPv4 local address database function is preferred over IPv6 local address database function where ever possible.
LDP Control Plane: Bindings Advertisement
LDP base specification allows exchange of IPv4/IPv6 bindings (address/label) on an established session. When both IPv4 and
IPv6 address families are enabled under LDP, LDP distributes address/label bindings for both address families to its established
peer according to local policies. Following are a few significant points pertaining to bindings support for IPv6:
LDP allocates/advertises local label bindings for link-local IPv6 address prefixes. If received, such FEC bindings are ignored
LDP sends only the Prefix FEC of the single address family type in a FEC TLV and not include both. If such a FEC binding is
received, the entire message is ignored
LDP sends only the addresses belonging to same address family in a single address list TLV (in address or address withdraw
message)
If an address family is not enabled on receiving LSR, LDP discards any bindings received from peer(s) for the address family.
This means that when address family is enabled, LDP needs to reset existing sessions with the peers in order to re-learn the
discarded bindings. The implementation is optimized to reset only those sessions which were previously known to be dual-stack
and had sent bindings for both address families.
LSP Mapping
LDP uses IPv6 adjacency information instead of IP address to map an IPv6 link-local nexthop to an LDP peer.
In addition to other usual checks before using a label from nexthop LDP peer, LDP uses the nexthop label for a prefix of a
given address family, if there are one or more LDP hello adjacencies of the same address family type established with the
peer.
Label Policies
LDP allows a user to configure label policies for allocation, acceptance, receipt, and advertisement of labels for the given
prefixes.
Following are the significant points pertaining to the IPv6 support for label policies:
Label policies and their configurations are allowed under address family IPv6
Any policy that specifies prefix or a set of prefixes through an ACL, supports both IPv4 and IPv6 variants for address(s)
or ACLs
Any policy that specifies peer address or set of peer addresses through an ACL, supports both IPv4 and IPv6 variant for peer
address(s) or ACL
Any policy that specifies the peer’s LSR ID in a peer ACL continues to take IPv4 ACL based policy irrespective of the feature
configuration
Dual-Stack Capability TLV
Clear rules are specified in RFC 5036 to determine transport connection roles in setting up a TCP connection for single-stack
LDP. But RFC 5036 is not clear about dual-stack LDP, in which an LSR may assume different roles for different address families,
causing issues in establishing LDP sessions.
To ensure a deterministic transport connection role for the dual-stack LDP, the dual-stack LSR conveys its transport connection
preference in every LDP Hello message. This preference is encoded in a new TLV (Type Length Value) called the Dual-Stack Capability
TLV. Dual-stack LSR always checks for the presence of the dual-stack capability TLV in the received LDP Hello messages and
takes appropriate action for establishing or maintaining sessions.
RFC 7552 specifies more details about updates to LDPv6.
This field is reserved. It must be set to zero on transmission and ignored on receipt
MBZ
Must be zero
Compliance Check
The compliance check prevents sessions being formed with prior RFC 7552 implementation of LDPv6.
If the dual-stack capability TLV is not present in the received Hellos and the compliance check is configured, the local and
remote preferences must match to establish a session. If the preferences do not match, the LDP Hellos are dropped and the
session is not established. Compliance check has therefore been disabled by default.
Use the neighbor dual-stack tlv-compliance command in MPLS LDP configuration to enable the compliance check.
Configuring ISIS for IPv6 and LDPv6
Intermediate System-to-Intermediate System (IS-IS) is an Interior Gateway Protocol (IGP) that advertises link-state information
throughout the network to create a picture of the network topology. IPv6 IS-IS extends the address families supported by IS-IS
to include IPv6, in addition to IPv4.
Previously, IS-IS supported registration of only LDP IPv4 sync status change. This has now been enhanced to support registration
of notifications of LDP IPv6 sync status change. IS-IS determines the link-metrics to be advertised based on the LDP-IGP sync
status on the IPv4 and IPv6 address families.
IS-IS supports non-stop forwarding (NSF) by preserving the LDPv6-IGP sync status across high availability (HA) events of IS-IS
process restarts and failover.
IS-IS also supports LDPv6-IGP sync for LFA-FRR by checking the sync status of the backup interface (if it is configured with
LDP IPv6 sync).
LDP computes default local transport address for IPv6 from its IPv6 interface or address database by picking the lowest operational
loopback interface with global unicast IPv6 address. This means that any change in this loopback state or address, flaps or
changes the default transport address for IPv6 and may cause session flaps using such an address as transport endpoint. For
example, if a session is currently active on Loopback2 as during it's inception it was the lowest loopback with an IPv6 address,
and a lower loopback, Loopback0, is configured with an IPv6 address, the session does not flap. However, if it does flap,
the next time the session is attempted, Loopback0 is used.
The session flaps when configuring discovery transport address explicitly.
Use the discovery transport-address command under the LDP address family submode to specify the global transport address for IPv4 or IPv6.
Enable a global transport-address for the IPv6 address family.
It is recommended to configure global transport-address for IPv6 address family to avoid a potentially unstable default transport
address.
The LDP configuration model was changed with the introduction of explicit address family enabling under LDP (VRF) global and
LDP (VRF) interfaces. However, in order to support backward compatibility, the old configuration model was still supported
for default VRF. There was, however, no option to disable the implicitly enabled IPv4 address family under default VRF's global
or interface level.
A new configuration mpls ldp default-vrf implicit-ipv4 disable is now available to the user to disable the implicitly enabled IPv4 address family for the default VRF. The new configuration
provides a step towards migration to new configuration model for the default VRF that mandates enabling address family explicitly.
This means that if the new option is configured, the user has to explicitly enable IPv4 address family for default VRF global
and interface levels. It is recommended to migrate to this explicitly enabled IPv4 configuration model.
IPv4 is implicitly enabled under default VRF and any LDP interface under default VRF. To operate as an IPv6-only LSR, disable
the IPv4 address family.
Configure IPv4 as the preferred transport (overriding the default setting of IPv6 as preferred transport) to establish connection
for a set of dual-stack peers. You can also configure the maximum time (max-wait, in seconds) the preferred address family connection must wait to establish the transport connection before resorting to
the non-preferred address family.
The following examples shows how to configure IPv6-only LSR.
IPv4 is implicitly enabled under default VRF and any LDP interfaces under default VRF. In order to operate as an IPv6-only
LSR, the user must also explicitly disable IPv4 address family.
LDPv6 Configuration Without Explicit IPv6 Export Address
In this example, there is no explicit IPv6 export address. The loopback’s IPv6 address is used as the export address (6:6:6::6/128).
The router ID configured in MPLS LDP is not used in anyway for export. It is used only for LDP LSR identification.
LDPv6 Configuration With Explicit IPv6 Export Address
In this example, there is an explicit IPv6 export address. However, there is no IPv6 loopback. There is no router-id configured,
but the loopback IPv4 address is used.
This section provides
detailed conceptual information about setting up LSPs, LDP graceful restart,
and LDP session protection.
Setting Up Label Switched Paths
MPLS packets are forwarded between the nodes on the MPLS network using Label Switched Paths(LSPs). LSPs can be created statically
or by using a label distribution protocol like LDP. Label Switched Paths created by LDP performs hop-by-hop path setup instead
of an end-to-end path. LDP enables label switched routers (LSRs) to discover their potential peer routers and to establish
LDP sessions with those peers to exchange label binding information.
The following figure illustrates the process of label binding exchange for setting up LSPs.
For a given network (10.0.0.0), hop-by-hop LSPs are set up between each of the adjacent routers (or, nodes) and each node
allocates a local label and passes it to its neighbor as a binding:
R4 allocates local label L4 for prefix 10.0.0.0 and advertises it to its neighbors (R3).
R3 allocates local label L3 for prefix 10.0.0.0 and advertises it to its neighbors (R1, R2, R4).
R1 allocates local label L1 for prefix 10.0.0.0 and advertises it to its neighbors (R2, R3).
R2 allocates local label L2 for prefix 10.0.0.0 and advertises it to its neighbors (R1, R3).
R1’s label information base (LIB) keeps local and remote labels bindings from its neighbors.
R2’s LIB keeps local and remote labels bindings from its neighbors.
R3’s LIB keeps local and remote labels bindings from its neighbors.
R4’s LIB keeps local and remote labels bindings from its neighbors.
MPLS Forwarding
Once the label bindings are learned, MPLS forwarding plane is setup and packets are forwarded as shown in the following figure.
Because R3 is next hop for 10.0.0.0 as notified by the FIB, R1 selects label binding from R3 and installs forwarding entry
(Layer 1, Layer 3).
Because R3 is next hop for 10.0.0.0 (as notified by FIB), R2 selects label binding from R3 and installs forwarding entry (Layer
2, Layer 3).
Because R4 is next hop for 10.0.0.0 (as notified by FIB), R3 selects label binding from R4 and installs forwarding entry (Layer
3, Layer 4).
Because next hop for 10.0.0.0 (as notified by FIB) is beyond R4, R4 uses NO-LABEL as the outbound and installs the forwarding
entry (Layer 4); the outbound packet is forwarded IP-only.
Incoming IP traffic on ingress LSR R1 gets label-imposed and is forwarded as an MPLS packet with label L3.
Incoming IP traffic on ingress LSR R2 gets label-imposed and is forwarded as an MPLS packet with label L3.
R3 receives an MPLS packet with label L3, looks up in the MPLS label forwarding table and switches this packet as an MPLS
packet with label L4.
R4 receives an MPLS packet with label L4, looks up in the MPLS label forwarding table and finds that it should be Unlabeled,
pops the top label, and passes it to the IP forwarding plane.
IP forwarding takes over and forwards the packet onward.
Details of Label
Distribution Protocol Graceful Restart
LDP (Label
Distribution Protocol) graceful restart provides a control plane mechanism to
ensure high availability and allows detection and recovery from failure
conditions while preserving Nonstop Forwarding (NSF) services. Graceful restart
is a way to recover from signaling and control plane failures without impacting
forwarding.
Without LDP graceful
restart, when an established session fails, the corresponding forwarding states
are cleaned immediately from the restarting and peer nodes. In this case LDP
forwarding restarts from the beginning, causing a potential loss of data and
connectivity.
The LDP graceful
restart capability is negotiated between two peers during session
initialization time, in FT SESSION TLV. In this typed length value (TLV), each
peer advertises the following information to its peers:
Reconnect time
Advertises the
maximum time that other peer will wait for this LSR to reconnect after control
channel failure.
Recovery time
Advertises the
maximum time that the other peer has on its side to reinstate or refresh its
states with this LSR. This time is used only during session reestablishment
after earlier session failure.
FT flag
Specifies
whether a restart could restore the preserved (local) node state for this flag.
Once the graceful
restart session parameters are conveyed and the session is up and running,
graceful restart procedures are activated.
When configuring the
LDP graceful restart process in a network with multiple links, targeted LDP
hello adjacencies with the same neighbor, or both, make sure that graceful
restart is activated on the session before any hello adjacency times out in
case of neighbor control plane failures. One way of achieving this is by
configuring a lower session hold time between neighbors such that session
timeout occurs before hello adjacency timeout. It is recommended to set LDP
session hold time using the following formula:
This means that for
default values of 15 seconds and 5 seconds for link Hello holdtime and interval
respectively, session hold time should be set to 30 seconds at most.
Phases in
Graceful Restart
The graceful restart
mechanism is divided into different phases:
Control
communication failure detection
Control
communication failure is detected when the system detects either:
Missed LDP
hello discovery messages
Missed LDP
keepalive protocol messages
Detection
of Transmission Control Protocol (TCP) disconnection a with a peer
Forwarding
state maintenance during failure
Persistent
forwarding states at each LSR are achieved through persistent storage
(checkpoint) by the LDP control plane. While the control plane is in the
process of recovering, the forwarding plane keeps the forwarding states, but
marks them as stale. Similarly, the peer control plane also keeps (and marks as
stale) the installed forwarding rewrites associated with the node that is
restarting. The combination of local node forwarding and remote node forwarding
plane states ensures NSF and no disruption in the traffic.
Control state
recovery
Recovery
occurs when the session is reestablished and label bindings are exchanged
again. This process allows the peer nodes to synchronize and to refresh stale
forwarding states.
Control Plane
Failure
When a control plane
failure occurs, connectivity can be affected. The forwarding states installed
by the router control planes are lost, and the in-transit packets could be
dropped, thus breaking NSF. The following figure illustrates control plane
failure and recovery with graceful restart and shows the process and results of
a control plane failure leading to loss of connectivity and recovery using
graceful restart.
Recovery with
Graceful Restart
The R4 LSR
control plane restarts.
LIB is lost when
the control plane restarts.
The forwarding
states installed by the R4 LDP control plane are immediately deleted.
Any in-transit
packets flowing from R3 to R4 (still labeled with L4) arrive at R4.
The MPLS
forwarding plane at R4 performs a lookup on local label L4 which fails. Because
of this failure, the packet is dropped and NSF is not met.
The R3 LDP peer
detects the failure of the control plane channel and deletes its label bindings
from R4.
The R3 control
plane stops using outgoing labels from R4 and deletes the corresponding
forwarding state (rewrites), which in turn causes forwarding disruption.
The established
LSPs connected to R4 are terminated at R3, resulting in broken end-to-end LSPs
from R1 to R4.
The established
LSPs connected to R4 are terminated at R3, resulting in broken LSPs end-to-end
from R2 to R4.
When the LDP control
plane recovers, the restarting LSR starts its forwarding state hold timer and
restores its forwarding state from the checkpointed data. This action
reinstates the forwarding state and entries and marks them as old.
The restarting LSR
reconnects to its peer, indicated in the FT Session TLV, that it either was or
was not able to restore its state successfully. If it was able to restore the
state, the bindings are resynchronized.
The peer LSR stops
the neighbor reconnect timer (started by the restarting LSR), when the
restarting peer connects and starts the neighbor recovery timer. The peer LSR
checks the FT Session TLV if the restarting peer was able to restore its state
successfully. It reinstates the corresponding forwarding state entries and
receives binding from the restarting peer. When the recovery timer expires, any
forwarding state that is still marked as stale is deleted.
If the restarting
LSR fails to recover (restart), the restarting LSR forwarding state and entries
will eventually timeout and is deleted, while neighbor-related forwarding
states or entries are removed by the Peer LSR on expiration of the reconnect or
recovery timers.
Details of Session
Protection
LDP session protection
lets you configure LDP to automatically protect sessions with all or a given
set of peers (as specified by peer-acl). When configured, LDP initiates backup
targeted hellos automatically for neighbors for which primary link adjacencies
already exist. These backup targeted hellos maintain LDP sessions when primary
link adjacencies go down.
The Session Protection
figure illustrates LDP session protection between neighbors R1 and R3. The
primary link adjacency between R1 and R3 is directly connected link and the
backup; targeted adjacency is maintained between R1 and R3. If the direct link
fails, LDP link adjacency is destroyed, but the session is kept up and running
using targeted hello adjacency (through R2). When the direct link comes back
up, there is no change in the LDP session state and LDP can converge quickly
and begin forwarding MPLS traffic.
Note
When LDP session
protection is activated (upon link failure), protection is maintained for an
unlimited period time.
Controlling State Advertisements In An mLDP-Only Setup
This function explains the controlling of state advertisements of non-negotiated Label Distribution Protocol (LDP) applications.
This implementation is in conformance with RFC 7473 (Controlling State Advertisements of Non-negotiated LDP Applications).
The main purpose of documenting this function is to use it in a Multipoint LDP (mLDP)-only environment, wherein participating
routers don’t need to exchange any unicast binding information.
Non-Negotiated LDP Applications
The LDP capabilities framework enables LDP applications’ capabilities exchange and negotiation, thereby enabling LSRs to send
necessary LDP state. However, for the applications that existed prior to the definition of the framework (called non-negotiated LDP applications), there is no capability negotiation done. When an LDP session comes up, an LDP speaker may unnecessarily
advertise its local state (without waiting for any capabilities exchange and negotiation). In other words, even when the peer
session is established for Multipoint LDP (mLDP), the LSR advertises the state for these early LDP applications.
One example is IPv4/IPv6 Prefix LSPs Setup (used to set up Label Switched Paths [LSPs] for IP prefixes). Another example is L2VPN P2P FEC 128 and FEC 129 PWs Signaling (an LDP application that signals point-to-point [P2P] Pseudowires [PWs] for Layer 2 Virtual Private Networks [L2VPNs]).
In an mLDP-only setup, you can disable these non-negotiated LDP applications and avoid unnecessary LDP state advertisement.
An LDP speaker that only runs mLDP announces to its peer(s) its disinterest (or non-support) in non-negotiated LDP applications.
That is, it announces to its peers its disinterest to set up IP Prefix LSPs or to signal L2VPN P2P PW, at the time of session
establishment.
Upon receipt of such a capability, the receiving LDP speaker, if supporting the capability, disables the advertisement of
the state related to the application towards the sender of the capability. This new capability can also be sent later in a
Capability message, either to disable a previously enabled application’s state advertisement, or to enable a previously disabled
application’s state advertisement.
As a result, the flow of LDP state information in an mLDP-only setup is faster. When routers come up after a network event,
the network convergence time is fast too.
IP Address Bindings In An mLDP Setup
An LSR typically uses peer IP address(es) to map an IP routing next hop to an LDP peer in order to implement its control plane
procedures. mLDP uses a peer’s IP address(es) to determine its upstream LSR to reach the root node, and to select the forwarding
interface towards its downstream LSR. Hence, in an mLDP-only network, while it is desirable to disable advertisement of label
bindings for IP (unicast) prefixes, disabling advertisement of IP address bindings will break mLDP functionality.
Uninteresting State - For the Prefix-LSP LDP application, uninteresting state refers to any state related to IP Prefix FEC, such as FEC label bindings and LDP Status. IP address bindings are not
considered as an uninteresting state.
For the P2P-PW application LDP application, uninteresting state refers to any state related to P2P PW FEC 128 or FEC 129, such as FEC label bindings, MAC address withdrawal, and LDP
PW status.
Control State Advertisement
To control advertisement of uninteresting state of non-negotiated LDP applications, the capability parameter TLV State Advertisement Control Capability is used. This TLV is only present in the Initialization and Capability messages, and the TLV can hold one or more State Advertisement
Control (SAC) Elements.
As an example, consider two LSRs, S (LDP speaker) and P (LDP peer), that support all non-negotiated applications. S is participating
(or set to participate) in an mLDP-only setup. Pointers for this scenario:
By default, the LSRs will advertise state for all LDP applications to their peers, as soon as an LDP session is established.
The capabilities sac mldp-only function is enabled on S.
P receives an update from S via a Capability message that specifies to disable all four non-negotiated applications states.
P’s outbound policy towards S blocks and disables state for the unneeded applications.
S only receives mLDP advertisements from specific mLDP-participating peers.
Use Cases For Controlling State Advertisements
Two use cases are explained, mLDP-Based MVPN and Disable Prefix-LSPs On An L2VPN/PW tLDP Session.
mLDP-Based MVPN
A sample topology and relevant configurations are noted below.
The topology represents an MVPN profile 1 where an mLDP-based MVPN service is deployed over a Segment Routing core setup
mLDP is required to signal MP2MP LSPs, whereas SR handles the transport.
SAC capabilities are used to signal mLDP-only capability, which blocks unrequired unicast IPv4, IPv6, FEC128, and FEC129 related label binding advertisements.
The mldp-only option is enabled on PE routers and P routers to remove unwanted advertisements.
Capabilities Sent shows that mldp-only option disables all other advertisements.
Capabilities Received shows that mldp-only is enabled on peer PE2 too.
Disable Prefix-LSPs On An L2VPN/PW tLDP Session
A sample topology and relevant configurations are noted below.
The topology represents an L2VPN Xconnect service over a Segment Routing core setup.
By default, Xconnect uses tLDP to signal service labels to remote PEs.
By default, tLDP not only signals the service label, but also known (IPv4 and IPv6) label bindings to the tLDP peer, which
is not required.
The LDP SAC capabilities is an optional configuration enabled under LDP, and users can block IPv4 and IPv6 label bindings
by applying configurations on PE1 and PE2.
Configuration
PE1 Configuration
Disable IPv4 prefix LSP binding advertisements on PE1:
Whenever you disable a non-negotiated LDP application state on a router, you must include previously disabled non-negotiated
LDP applications too, in the same command line. If not, the latest configuration overwrites the existing ones. You can see
that ipv4-disable is added again, though it was already disabled.
PE2 Configuration
Enable SAC capability awareness on PE2, and make PE2 stop sending IPv4 prefix LSP binding advertisements to PE1: