The documentation set for this product strives to use bias-free language. For the purposes of this documentation set, bias-free is defined as language that does not imply discrimination based on age, disability, gender, racial identity, ethnic identity, sexual orientation, socioeconomic status, and intersectionality. Exceptions may be present in the documentation due to language that is hardcoded in the user interfaces of the product software, language used based on RFP documentation, or language that is used by a referenced third-party product. Learn more about how Cisco is using Inclusive Language.
The following topics describe the monitoring policies used by Cisco EPN Manager . For information on the supported MIBs and MIB objects, see Cisco Evolved Programmable Network Manager Supported Devices.
The Device Health Monitoring Policy monitors device CPU utilization, memory pool utilization, environmental temperature, and device availability for all devices in the network. By default, the policy polls devices for this information every 5 minutes, and an alarm is generated if CPU utilization, memory pool utilization, or environmental temperature thresholds are surpassed.
This monitoring policy that is activated by default after installation.
Note | This policy does not monitor the device CPU utilization and memory pool utilization for supported Cisco ONS or Cisco NCS 2000 devices, but it does monitor memory utilization and device availability. For information on how to manage this policy, see Set Up Basic Device Health Monitoring. |
An Interface Health Monitoring Policy monitors over 30 attributes to check interface operational status and performance. It polls device interfaces every 5 minutes and generates an alarm when interface discard, error, utilization, or byte rate thresholds are exceeded.
To protect the performance of large deployments, this policy is not activated by default.
Note | This policy does not monitor optical interfaces. Use an optical policy to monitor that information. See Optical 1 Day and Optical 15 Mins Monitoring Policies. |
See these topics for information on how to manage this policy:
To check whether an Interface Health policy is actively monitoring interfaces, see Check What Cisco EPN Manager Is Monitoring.
To set up interface monitoring, see Set Up Basic Interface Monitoring.
To adjust an interface monitoring policy, see Adjust What Is Being Monitored.
The Custom MIB Polling monitoring policy is a customizable policy you can use to monitor unsupported parameters—that is, parameters that are not polled by any of the existing monitoring policy types. When you create a Custom MIB Polling policy, you can choose from an extensive list of Cisco and other MIBs, or import new MIBs into the policy. If a Custom MIB Polling policy is collecting device performance information, you can display that data in the Performance dashboard by creating a generic dashlet (see Add a Customized Dashlet to the Device Trends Dashboard) . For more information on managing Custom MIB Polling monitoring policies, see the following topics:
To check if a Custom MIB Polling policy is being used to monitoring information, see Check What Cisco EPN Manager Is Monitoring.
To create a new Custom MIB Polling policy, see Create a Monitoring Policy for Unsupported Parameters and Third-Party Devices.
To adjust an existing Custom MIB Polling policy, see Adjust What Is Being Monitored.
An IP SLA Y.1731 monitoring policy uses the Y.1731 ITU-T recommendation to monitor over 70 fault and performance attributes in Metro Ethernet networks. When you create an IP SLA Y.1731 monitoring policy, it polls the parameters every 15 minutes and generates an alarm when delay, jitter, frame loss, and other thresholds are exceeded.
For more information on how to configure and manage an IP SLA Y.1731 monitoring policy, see these topics:
To check if IP SLA Y.1731 parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new IP SLA Y.1731 monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing IP SLA Y.1731 monitoring policy, see Adjust What Is Being Monitored.
A Pseudowire Emulation Edge to Edge (PWE3) monitoring policy polls approximately 20 attributes that emulate edge-to-edge services over a Packet Switched Network (PSN). When you create and enable a monitoring policy that uses this policy type, attributes are polled every 15 minutes by default. In addition, Cisco EPN Manager generates a minor alarm when the thresholds for the following attributes are surpassed on pseudowire virtual circuits (PW VCs):
For more information on how to configure and manage a PWE3 monitoring policy, see these topics:
To check if PWE3 parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new PWE3 monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing PWE3 monitoring policy, see Adjust What Is Being Monitored.
A Quality of Service monitoring policy polls over 60 service parameters to validate the quality of services running on network devices. When you create a Quality of Service monitoring policy, it polls the parameters every 15 minutes and generates an alarm when certain thresholds are exceeded. The following is a partial list of parameters that can cause an alarm:
To view all Quality of Service parameters that can cause TCAs, see Check Which Parameters and Counters Are Polled By a Monitoring Policy.
For more information on how to configure and manage a Quality of Service monitoring policy, see these topics:
To check if Quality of Service parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new Quality of Service monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing Quality of Service monitoring policy, see Adjust What Is Being Monitored.
An IP SLA monitoring policy monitors approximately 20 parameters to provide real-time performance information. When you create an IP SLA monitoring policy, it polls the parameters every 15 minutes. This monitoring policy does not generate any alarms; if you want to generate IP SLA-based alarms, use the IP SLA Y.1731 monitoring policy.
For more information on how to configure and manage an IP SLA monitoring policy, see these topics:
To check if IP SLA parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new IP SLA monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing IP SLA monitoring policy, see Adjust What Is Being Monitored.
A ME1200 QoS monitoring policy polls over 20 service parameters to validate the quality of selected services running on ME1200 devices. When you create a ME1200 Quality of Service monitoring policy, it polls the parameters every 15 minutes but does not generate an alarm when certain thresholds are exceeded. The following is a partial list of parameters that are polled by ME1200 QoS monitoring policy:
To view all ME1200 QoS parameters that are polled, see Check Which Parameters and Counters Are Polled By a Monitoring Policy.
For more information on how to configure and manage a ME1200 QoS monitoring policy, see these topics:
To check if ME1200 QoS parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new ME1200 QoS monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing ME1200 QoS monitoring policy, see Adjust What Is Being Monitored.
This monitoring policy polls over 5 parameters to monitor the BNG sessions as well as the IP addresses leased from the IP pools. When you create a BNG Sessions and IP Pools monitoring policy, it polls the parameters every 15 minutes and generates an alarm when certain thresholds are exceeded. The following is a partial list of parameters that can cause an alarm:
Number of used or free IP addresses in the IP pools.
Number of sessions for authenticated and up subscribers.
To view all BNG Sessions and IP Pools parameters that can cause TCAs, see Check Which Parameters and Counters Are Polled By a Monitoring Policy.
For more information on how to configure and manage a BNG Sessions and IP Pools monitoring policy, see these topics:
To check if BNG Sessions and IP Pools parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new BNG Sessions and IP Pools monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing BNG Sessions and IP Pools monitoring policy, see Adjust What Is Being Monitored.
The TDM/Sonet Ports monitoring policy monitors approximately 20 circuit emulation (CEM) parameters. When you create a TDM/SONET Ports monitoring policy, it polls the CEM parameters every 15 minutes and generates an alarm when errored seconds (ES), unavailable sessions (UAS), and code violation (CV) thresholds are exceeded.
For more information on how to configure and manage a TDM/SONET Ports monitoring policy, see these topics:
To check if TDM/SONET Ports parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new TDM/SONET Ports monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing TDM/SONET Ports monitoring policy, see Adjust What Is Being Monitored.
An Optical SFP monitoring policy polls health and performance information for optical small form-factor (SFP) interfaces. This policy polls temperature, voltage, current, and optical TX/RX power. When you create an Optical SFP monitoring policy, it polls the parameters every 1 minute. This policy does not generate TCAs.
For more information on how to configure and manage an Optical SFP monitoring policy, see these topics:
To check if Optical SFP parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new Optical SFP monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing Optical SFP monitoring policy, see Adjust What Is Being Monitored.
The Optical 1 Day and Optical 15 Mins monitoring policies poll the following optical interfaces:
Device availability and environmental temperature is monitored by the Device Health monitoring policy. See Device Health Monitoring Policy for more information.
See Performance Counters for Optical Monitoring Policies for a list of the parameters that these policies poll.
For more information on how to configure and manage an Optical 1 Day and Optical 15 Mins monitoring policy, see these topics:
To check if Quality of Service parameters are being monitored , see Check What Cisco EPN Manager Is Monitoring.
To create a new Quality of Service monitoring policy, see Create a New Monitoring Policy Using Out-of-the-Box Policy Types.
To adjust an existing Quality of Service monitoring policy, see Adjust What Is Being Monitored.
The following table lists the performance counters used by the optical policy types to monitor physical interfaces.
Performance counters marked with an asterisk (*) are applicable for all Cisco Optical Networking Services (ONS) and Cisco Network Convergence System (NCS) 2000 series devices. Performance counters marked with a double asterisk (**) are applicable for Cisco Network Convergence System (NCS) 4000 series devices.
Physical Interface Performance Counter |
Description |
OPR-MIN |
Minimum output power received by the optical circuit. |
OPR-AVG |
Average output power received by the optical circuit. |
OPR-MAX |
Maximum output power received by the optical circuit. |
OPT-MIN |
Minimum output power transmitted from the optical circuit. |
OPT-AVG |
Average output power transmitted from the optical circuit. |
OPT-MAX |
Maximum output power transmitted from the optical circuit. |
LBC-MIN* LBCL-MIN |
Minimum laser bias current for the optical circuit. |
LBC-AVG* LBCL-AVG |
Average laser bias current for the optical circuit. |
LBC-MAX* LBCL-MAX |
Maximum laser bias current for the optical circuit. |
DGD-MIN** |
Minimum differential group delay for the optical circuit. |
DGD-AVG** |
Average differential group delay for the optical circuit. |
DGD-MAX** |
Maximum differential group delay for the optical circuit. |
SOPMD-MIN** |
Minimum second order polarization mode dispersion for the optical circuit. |
SOPMD-AVG** |
Average second order polarization mode dispersion for the optical circuit. |
SOPMD_MAX** |
Maximum second order polarization mode dispersion for the optical circuit. |
OSNR-MIN** |
Minimum optical signal to noise ratio for the optical circuit. |
OSNR-AVG** |
Average optical signal to noise ratio for the optical circuit. |
OSNR-MAX** |
Maximum optical signal to noise ratio for the optical circuit. |
PDL-MIN** |
Minimum polarization dependent loss for the optical circuit. |
PDL-AVG** |
Average polarization dependent loss for the optical circuit. |
PDL-MAX** |
Maximum polarization dependent loss for the optical circuit. |
PCR-MIN** |
Minimum polarization change rate for the optical circuit. |
PCR-AVG** |
Average polarization change rate for the optical circuit. |
PCR-MAX** |
Maximum polarization change rate for the optical circuit. |
PMD-AVG*,** |
Average polarization mode dispersion for the optical circuit. |
PMD-MIN*,** |
Minimum polarization mode dispersion for the optical circuit. |
PN-MIN** |
Minimum phase noise for the optical circuit. |
PN-AVG** |
Average phase noise for the optical circuit. |
PN-MAX** |
Maximum phase noise for the optical circuit. |
PREFEC-BER* |
Pre-forward error correction bit error rate for the optical circuit. |
CD-MIN** |
Minimum chromatic dispersion for the optical circuit. |
CD-AVG** |
Average chromatic dispersion for the optical circuit. |
CD-MAX** |
Maximum chromatic dispersion for the optical circuit. |
The following table lists the performance counters used by the optical policy types to monitor OTN-FEC interfaces.
Performance counters marked with an asterisk (*) are applicable for all Cisco Optical Networking Services (ONS) and Cisco Network Convergence System (NCS) 2000 series devices.
OTN-FEC Interface Performance Counter |
Description |
BIT-EC* BIEC |
Number of bit errors corrected. |
UNC-WORDS* UCW |
Number of uncorrectable words. |
The following table lists the performance counters used by the optical policy types to monitor OTN-ODU interfaces.
OTN-ODU Interface Performance Counter |
Description |
BBE-PM |
Number of background block errors in path monitoring. |
BBER-PM |
Background block errors ratio in path monitoring. |
ES-PM |
Number of errored seconds in path monitoring. |
ESR-PM |
Errored seconds ratio in path monitoring. |
SES-PM |
Number of severely errored seconds in path monitoring. |
SESR-PM |
Severely errored seconds ratio in path monitoring. |
UAS-PM |
Number of unavailable seconds in path monitoring. |
FC-PM |
Number of failure counts (AIS/RFI detected) in path monitoring. |
gfpStatsRxFrames |
Number of generic framing procedure (GFP) frames received. |
gfpStatsTxFrames |
Number of GFP frames transmitted. |
gfpStatsRxOctets |
Number of GFP bytes received. |
gfpStatsTxOctets |
Number of GFP bytes transmitted. |
gfpStatsRxCRCErrors |
Number of packets received with a payload frame check sequence (FCS) error. |
gfpStatsRxMBitErrors |
Number of multiple bit errors. In the GFP core header at the GFP-transparent (GFP-T) receiver, these are uncorrectable. |
gfpStatsRxSBitErrors |
Number of single bit errors. In the GFP core header at the GFP-T receiver, these are correctable. |
gfpStatsRxTypeInvalid |
Number of packets received with invalid GFP type. This includes unexpected user payload identifier (UPI) type and errors in core header error check (CHEC). |
gfpStatsRxCIDInvalid |
Number of packets received with invalid CID. |
gfpStatsRoundTripLatencyUSec |
Round trip delay for the end-to-end Fibre Channel transport in milliseconds. |
gfpStatsTxDistanceExtBuffers |
Number of buffer credit transmitted for GFP-T transmitter (valid only if distance extension is enabled). |
gfpStatsRxSblkCRCErrors |
Number of super block cyclic redundancy check (CRC) errors. |
gfpStatsCSFRaised |
Number of GFP client signal fail (CSF) frames detected at the GFP-T receiver. |
gfpStatsLFDRaised |
Number of GFP loss of frame delineation (LFD) detected. |
gfpRxCmfFrame |
Number of client management frames (CMF) received. |
gfpTxCmfFrame |
Number of client management frames (CMF) transmitted. |
gfpStatsCHecRxMBitErrors |
Number of core header error control (cHEC) CRC multiple bit errors. |
gfpStatsTHecRxMBitErrors |
Number of type header error control (tHEC) CRC multiple bit errors. |
The following table lists the performance counters used by the optical policy types to monitor OTN-OTU interfaces.
OTN-OTU Interface Performance Counter |
Description |
BBE-SM |
Number of background block errors in section monitoring. |
BBER-SM |
Background block error ratio in section monitoring. |
ES-SM |
Number of errored seconds in section monitoring. |
ESR-SM |
Errored seconds ratio in section monitoring. |
SES-SM |
Number of severely errored seconds in section monitoring. |
SESR-SM |
Severely errored seconds ratio in section monitoring. |
UAS-SM |
Number of unavailable seconds in section monitoring. |
FC-SM |
Number of failure counts (AIS/RFI detected) in section monitoring. |
The following table lists the performance counters used by the optical policy types to monitor Ethernet interfaces.
Ethernet Interface Performance Counter |
Description |
ifInOctets |
The total number of octets received on the interface, including framing octets. |
ifInErrors |
The total number of received packets that were discarded because of errors. |
ifOutOctets |
The total number of transmitted octets, including framing packets. |
ifInUcastPkts |
The total number of unicast packets received since the last counter reset. |
ifOutUcastPkts |
The total number of packets requested by the higher-level protocols to be transmitted, and which were not addressed to a multicast or broadcast address at this sub-layer, including those that were discarded or not sent. |
ifInMulticastPkts |
The total number of multicast packets received since the last counter reset. |
ifOutMulticastPkts |
The total number of multicast frames transmitted error free. |
ifInBroadcastPkts |
The total number of broadcast packets received since the last counter reset. |
ifOutBroadcastPkts |
The total number of packets requested by higher-level protocols and addressed to a broadcast address at this sublayer, including those that were not transmitted. |
txTotalPkts |
The total number of packets transmitted. |
rxTotalPkts |
The total number of packets received. |
etherStatsOctets |
The total number of octets of data (including those in bad packets) received on the network (excluding framing bits but including FCS octets). |
etherStatsOversizePkts |
The total number of packets received that were longer than 1518 octets (excluding framing bits but including FCS octets) and were otherwise well formed. Note that for tagged interfaces, this number becomes 1522 bytes. |
dot3StatsFCSErrors |
A count of frames received on a particular interface that are an integral number of octets in length but do not pass the FCS check. |
dot3StatsFrameTooLongs |
A count of frames received on a particular interface that exceed the maximum permitted frame size. |
etherStatsJabbers |
The total number of packets received that were longer than 1518 octets (excluding framing bits but including FCS octets), and had either a bad FCS with an integral number of octets (FCS Error) or a bad FCS with a non-integral number of octets (Alignment Error). |
etherStatsPkts64Octets |
The total number of packets (including bad packets) received that were 64 octets in length (excluding framing bits but including FCS octets). |
etherStatsPkts65to127 Octets |
The total number of packets (including bad packets) received that were between 65 and 127 octets in length inclusive (excluding framing bits but including FCS octets). |
etherStatsPkts128to255 Octets |
The total number of packets (including bad packets) received that were between 128 and 255 octets in length inclusive (excluding framing bits but including FCS octets). |
etherStatsPkts256to511 Octets |
The total number of packets (including bad packets) received that were between 256 and 511 octets in length inclusive (excluding framing bits but including FCS octets). |
etherStatsPkts512to1023Octets |
The total number of packets (including bad packets) received that were between 512 and 1023 octets in length inclusive (excluding framing bits but including FCS octets). |
etherStatsPkts1024to1518Octets |
The total number of packets (including bad packets) received that were between 1024 and 1518 octets in length inclusive (excluding framing bits but including FCS octets). |
etherStatsMulticastPkts |
The total number of good packets received that were directed to a multicast address. |
etherStatsBroadcastPkts |
The total number of good packets received that were directed to the broadcast address. |
etherStatsUndersizePkts |
The total number of packets received that were less than 64 octets long (excluding framing bits, but including FCS octets) and were otherwise well formed. |
The following table lists the performance counters used by the optical policy types to monitor SONET interfaces.
Performance counters marked with an asterisk (*) are applicable for all Cisco Optical Networking Services (ONS) and Cisco Network Convergence System (NCS) 2000 series devices.
SONET Interface Performance Counter |
Description |
CV-S* CVS |
Number of coding violations per section. |
ES-S* ESS |
Number of errored seconds per section. |
SES-S* SESS |
Number of severely errored seconds per section. |
SEFS-S* SEFSS |
Number of severely errored frame seconds per section. |
CV-L* CVL |
Number of coding violations per line for near end and far end devices. |
ES-L* ESL |
Number of errored seconds per line for near end and far end devices. |
SES-L* SESL |
Number of severely errored seconds per line for near end and far end devices. |
UAS-L* UASL |
Number of unavailable seconds per line for near end and far end devices. |
FC-L* FCL |
Number of failure counts per line for near end and far end devices. |
The following table lists the performance counters used by the optical policy types to monitor SDH interfaces.
SDH Interface Performance Counter |
Description |
MS-ES |
Number of errored seconds per multiplex section for near end and far end devices. |
MS-ESR |
Error seconds ratio per multiplex section for near end and far end devices. |
MS-SES |
Number of severely errored seconds per multiplex section for near end and far end devices. |
MS-SESR |
Severely errored seconds ratio per multiplex section for near end and far end devices. |
MS-BBE |
Number of background block errors per multiplex section for near end and far end devices. |
MS-BBER |
Background block error ratio per multiplex section for near end and far end devices. |
MS-UAS |
Number of unavailable seconds per multiplex section for near end and far end devices. |
MS-EB |
Number of errored block per multiplex section for near end and far end devices. |
MS-FC |
Number of failure counts per multiplex section for near end and far end devices. |
MS-PSC |
Protection switching count per multiplex section. PSC is the number of times the service switches from a working card to a protection card and back. |
MS-PSC-R |
Protection switching count ring per multiplex section. This count is incremented only if ring switching is used. |
MS-PSC-S |
Protection switching count span per multiplex section. This count is incremented only if span switching is used. |
MS-PSC-W |
Protection switching count working per multiplex section. It is the count of the number of times traffic switches away from the working capacity in the failed line and back to the working capacity after the failure is cleared. PSC-W increments on the failed working line. |
MS-PSD |
Protection switching duration applies to the length of time, in seconds, that service is carried on another line. |
MS-PSD-R |
Protection switching duration ring is a count of the seconds that the protection line was used to carry service. This count is incremented only if ring switching is used. |
MS-PSD-S |
Protection switching duration span is a count of the seconds that the protection line was used to carry service. This count is incremented only if span switching is used. |
MS-PSD-W |
Protection switching duration working per multiplex section. |
RS-ES |
Number of errored seconds per regenerator section. |
RS-ESR |
Errored seconds ratio per regenerator section. |
RS-SES |
Number of severely errored seconds per regenerator section. |
RS-SESR |
Severely errored seconds ratio per regenerator section. |
RS-BBE |
Number of background block errors per regenerator section. |
RS-BBER |
Background block errors ratio per regenerator section. |
RS-UAS |
Number of unavailable seconds per regenerator section. |
RS-EB |
Number of errored block per regenerator section. |
RS-OFS |
Number of out-of-frame seconds per regenerator section. |