Overview of Cisco’s IP Fabric for Media Solution

This chapter contains information about Cisco's IP fabric for media solution.

About the IP Fabric for Media Solution

Today, the broadcast industry uses a serial digital interface (SDI) router and SDI cables to transport video and audio traffic. The SDI cables can carry only a single unidirectional signal. As a result, many cables, frequently stretched over long distances, are required, making it difficult and time-consuming to expand or change an SDI-based infrastructure.

Cisco’s IP fabric for media solution helps transition from an SDI router to an IP-based infrastructure. In an IP-based infrastructure, a single cable can carry multiple bidirectional traffic flows and can support different flow sizes without requiring changes to the physical infrastructure.

The IP fabric for media solution consists of a flexible spine and leaf architecture or a single modular switch topology. The solution uses Cisco Nexus 9000 Series switches with the Cisco non-blocking multicast (NBM) algorithm (an intelligent traffic management algorithm) and with or without the Nexus Dashboard Fabric Controller (NDFC). Using open APIs, the Cisco Nexus Dashboard Fabric Controller (NDFC) can integrate with various broadcast controllers. The solution provides a highly reliable (zero drop multicast), highly visible, highly secure, and highly available network.

Deployment Types

Cisco’s IP fabric for media solution supports the following types of deployments:

  • Spine-leaf topology—Flexible architecture for large-scale deployments that are typically seen in an IP studio.

  • Single modular switch—Architecture suitable for fixed deployments, with the controller providing features such as flow visibility, security, and monitoring.

Spine-Leaf Topology

Cisco's IP fabric for media solution supports a spine-leaf topology that consists of multiple spine and leaf switches. The topology supports any combination of leaf switches, including using just one type of leaf switch.

Media sources and receivers connect to the leaf switches, and receivers initiate IGMP join requests to the leaf switches in order to receive the media traffic.

Single Modular Switch Topology

Cisco's IP fabric for media solution supports a single modular switch topology that consists of one Cisco Nexus 9500 Series switch.

IP Fabric for Media Solution Components

Cisco Nexus 9000 Series Switches

The following Cisco Nexus 9000 Series switches are used to transport video and audio traffic through the IP fabric:

Cisco Nexus 9000 Series Switch

Number and Size of Ports

Role in Topology*

Cisco Nexus 9236C switch

36 x 40/100-Gbps ports

Spine or leaf in spine-leaf topology

Cisco Nexus 9272Q switch

72 x 40-Gbps ports

Spine or leaf in spine-leaf topology

Cisco Nexus 92160YC-X switch

48 x 1/10/25-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 9336C-FX2 switch

36 x 40/100-Gbps ports

Spine or leaf in spine-leaf topology

Cisco Nexus 9348GC-FXP switch

48 x 100-Mbps/1-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 9364C switch

64 x 40/100-Gbps ports

Spine in spine-leaf topology

Cisco Nexus 93108TC-EX switch

48 x 1/10-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93108TC-FX switch

48 x 10-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93180LC-EX switch

32 x 40/100-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93180YC-EX switch

48 x 1/10/25-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93180YC-FX switch

48 x 10/25-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93216TC-FX2 switch

96 x 1/10-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93240YC-FX2 switch

48 x 10/25-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 93360YC-FX2 switch

96 x 10/25-Gbps ports

Leaf in spine-leaf topology

Cisco Nexus 9504 or 9508 switch with the following line cards:

  • N9K-X9636C-R

  • N9K-X9636C-RX

  • N9K-X9636Q-R

Note

 

The N9K-X96136YC-R line card is not supported.

36 x 40/100-Gbps ports (for N9K-X9636C-R line cards)

36 x 40/100-Gbps ports (for N9K-X9636C-RX line cards)

36 x 40-Gbps ports (for N9K-X9636Q-R line cards)

Spine in spine-leaf topology or single modular switch

Cisco Nexus 9316D-GX switch

16 x 400/100-Gbps QSFP-DD ports

Leaf in spine-leaf topology

Cisco Nexus 9364C-GX switch

64 x 100/40-Gbps Quad Small Form-Factor Pluggable (QSFP28) ports

Leaf in spine-leaf topology

Cisco Nexus 93600CD-GX switch

28 x 100/40-Gbps Quad Small Form-Factor Pluggable (QSFP28) and 8 x 400/100-Gbps QSFP-DD ports

Leaf in spine-leaf topology

Cisco Nexus 93180YC-FX3S switch

48 25/50/100-Gigabit Ethernet SFP28 ports (ports 1-48) and 6 10/25/40/50/100-Gigabit QSFP28 ports (ports 49-54)

Leaf in spine-leaf topology

Cisco Nexus 93180YC-FX3

48 x 1/10/25 Gbps fiber ports and 6 x 40/100 Gbps QSFP28 ports

Leaf in spine-leaf topology

Cisco Nexus 93108TC-FX3P

48 x 100M/1/2.5/5/10 Gbps BASE-T ports

6 x 40/100 Gbps Quad small form-factor pluggable 28 (QSFP28) ports

Leaf in spine-leaf topology

Cisco Nexus 9348GC-FX3

48 x 10M/100M/1 Gbps BASE-T ports

4x 10/25 Gbps SFP28 ports

2 x 40/100 Gbps Quad small form-factor pluggable 28 (QSFP28) ports

Leaf in spine-leaf topology

N9K-X9624D-R2 line card

Line card with 24 400G QSFP-DD ports (only to be used with 8-slot chassis)

Spine or leaf in spine-leaf topology

Cisco Nexus 9508-FM-R2 line card

Fabric module for 400G line card (only to be used with 8-slot chassis)

Spine or leaf in spine-leaf topology

Cisco Nexus 9364D-GX2A switch

64 x 40/100/400G QSFP-DD ports

2 x 1/10G SFP+ ports

Spine or leaf switch in spine-leaf topology

Cisco Nexus 9348D-GX2A switch

48 x 40/100/400G QSFP-DD ports

2 x 1/10G SFP+ ports

Spine or leaf switch in spine-leaf topology

Cisco Nexus 9332D-GX2B switch

32 x 40/100/400G QSFP-DD ports

2 x 1/10G SFP+ ports

Spine or leaf switch in spine-leaf topology

Cisco Nexus 9808 switch with the following line cards:

Cisco Nexus X9836DM-A

Cisco Nexus 9808-FM-A

36 x 40/100/400G QSFP-DD ports (only to be used with 8-slot chassis)

Fabric module for Nexus 9808

Spine in spine-leaf topology or single modular switch

Cisco Nexus 9332D-H2R

32-port 400G QSFP-DD ports

Leaf in spine-leaf topology

Cisco Nexus 9364C-HX switch

64 x 40/100-Gbps QSFP28 ports

Spine or leaf in spine-leaf topology

Cisco Nexus 93400LD-GX2A switch

32 x 400-Gbps and 16 x 100-Gbps QSFP-DD/QSFP28 ports

Spine or leaf in spine-leaf topology

Cisco Nexus 9804 switch

4-slot modular chassis (ports and capabilities vary by installed line cards)

Spine or leaf in spine-leaf topology or single modular switch

Cisco Nexus 9408 switch

8-slot modular chassis (ports and capabilities vary by installed line cards)

Spine or leaf in spine-leaf topology or single modular switch

Cisco Nexus 9364C-GX switch

64 x 100/40-Gbps QSFP28 ports

Leaf in spine-leaf topology

Cisco Nexus 9336C-SE1 switch

36 x 40/100-Gbps QSFP28 ports

Spine or leaf in spine-leaf topology

*The role indicates the place in the fabric that makes the most sense given the port speeds supported by each switch. There are no restrictions as such on the role for which a switch can be used.

NDFC with IPFM

Through open APIs, the Cisco Nexus Dashboard Fabric Controller (NDFC) with IP Fabric for Media (IPFM) seamlessly integrates with the broadcast controller and provides a similar operator workflow with all the benefits of an IP-based infrastructure. The DCNM Media Controller features an intuitive GUI that enables you to configure your IP fabric using predefined templates that are designed for media networks.

The NDFC with IPFM enables you to do the following:

  • Configure secure generic or multicast-specific policies for individual hosts and allow or deny hosts based on their role.

  • Configure secure multicast-specific policies for multiple hosts and flows.

  • View the traffic flow and bandwidth utilization to identify problem areas (such as link failures or oversubscriptions) in your fabric.

  • Use flow analytics to measure and store bit rates and to display the details for individual traffic flows.

  • View an audit log of actions that are performed on the fabric.

Enhanced Payload for IPFM Critical Event

Beginning with Cisco NX-OS Release 10.6(1)F, enhancements provide more granular and user-friendly fault and notification information for improved network visibility and operational insight. This allows for faster identification and resolution of network issues. Key improvements include:

  • Clearer fault reasons and resolutions.

  • Key information, previously embedded within the Distinguished Name (DN), is now included as individual attributes in the payload. These attributes include:

    • source: The source IP address.

    • group: The multicast group IP address.

    • faultCode: The code that identifies the fault.

    • vrf: The Virtual Routing and Forwarding instance.

    This change eliminates the need to parse the DN to extract these values, providing a more direct way to access the information.

  • Improved presentation of switch critical events in network management and monitoring interfaces.

The enhancements to flow analytics, with the new fault categories and payload structure, provide a more detailed and actionable view of network behavior. This allows for more precise troubleshooting and optimization of traffic flows.

  • The new fault categories provide a more specific classification of fault conditions. This allows administrators to quickly pinpoint the source and nature of a problem.

  • The faultReason and faultResolution attributes in the enhanced payload provide clear explanations of the fault and specific steps to resolve it.

  • The new notification categories provide more granular information about network events. This allows administrators to proactively identify potential issues and optimize traffic flows before they impact performance.

  • The new notification categories provide more granular information about network events. This enables administrators to proactively identify potential issues and take appropriate actions to optimize traffic flows.


Note


Beginning with Cisco NX-OS Release 10.6(1)F, the legacy fault and notification outputs are deprecated. Subscribers and integrations must now consume the enhanced JSON payload structure, which provides enhanced name-value attribute pairs for all Fault and Notification Managed Objects (MOs). Ensure that any automated integrations or monitoring tools are updated accordingly.


Payload Structure: Old Vs New

The following example highlights the changes to the payload structure for flow faults:

Before Enhancement

After Enhancement

"nbmFaults": {

    "attributes": {

         "dn": "sys/nbm/show/faults/dom-default/faults-[s-[47.20.20.9]-g-[233.1.4.255]]",

         "faultDn": "s-[47.20.20.9]-g-[233.1.4.255]",

         "faultReason": "No Policer Avail",

         "faultResolution": "Please consult documentation",

         "modTs": "2025-04-01T16:03:15.175+00:00",

         "tStamp": "1743523395174"

}
"nbmFlowFaults": {

    "attributes": {

        "dn": "sys/nbm/show/faults/dom-default/flowfaults-[s-[47.20.20.9]-g-[233.1.4.255]]",

        "faultCode": "2076",

         "faultDn": "s-[47.20.20.9]-g-[233.1.4.255]",

         "faultReason": "Policer resources exhausted. Configured TCAM max has been reached",

         "faultResolution": "Review TCAM configuration if needed",

         "group": "233.1.4.255",

        "modTs": "2025-04-01T14:35:04.081+00:00",

        "source": "47.20.20.9",

        "tStamp": "1743518104080",

         "vrf": "default"

}

Enhanced Fault and Notification Payload Structure

Starting with Cisco NX-OS Release 10.6(1)F, new fault and notification categories provide clearer and more useful information for troubleshooting and optimizing the network. The following sections detail the new Fault and Notification Managed Objects (MO), along with examples of their enhanced payload structures.

Fault MO

The enhanced Fault Managed Objects (MOs) provide detailed and categorized information about various network issues, enabling more precise troubleshooting and quicker resolution. These MOs are designed to offer granular insights into different types of faults. These are the Fault MOs:

  • Flow Faults

    Indicate issues related to specific multicast flows, such as bandwidth shortages or policer resource exhaustion.

  • Sender Faults

    Report problems originating from the source device or endpoint, such as policy denials or connectivity issues.

  • Receiver Faults

    Highlight conditions impacting the receiver, such as insufficient bandwidth or configuration issues that prevent flow delivery.

  • Pim Passive Ingress Faults

    Identify faults associated with the ingress (incoming) interface of a flow, including VRF context or interface misconfigurations.

  • Pim Passive Egress Faults

    Relate to faults on the egress (outgoing) interface of a flow, such as invalid interface IPs or VRF mismatches.

Notify MO

The new Notification Managed Objects (MOs) offer granular information about network events and operational states, helping administrators proactively identify potential issues and optimize traffic flows before they impact performance. These are the Notification MOs:

  • Interface Level Usage MOs:

    Provide notifications on bandwidth utilization for ingress and egress interfaces, helping to monitor interface usage health.

    • Egress Event

      Notifies when bandwidth usage on an egress interface reaches or exceeds critical thresholds.

    • Ingress Event

      Notifies when bandwidth usage on an ingress interface reaches or exceeds critical thresholds.

  • Event MO for Flow Rate

    Indicates when a flow’s rate falls below or rises above configured thresholds.

  • Flow Provisioned MO

    Confirms when a new flow has been successfully provisioned in the network.

  • NAT Event MOs:

    Provide events related to NAT-specific bandwidth or translation state issues.

Flow Faults

The nbmFlowFaults Managed Object provides information about faults related to a specific flow, such as bandwidth or policer issues.

"nbmFlowFaults": {

    "attributes": {

        "dn": "sys/nbm/show/faults/dom-default/flowfaults-[s-[47.20.20.9]-g-[233.1.4.255]]",

        "faultCode": "2076",

         "faultDn": "s-[47.20.20.9]-g-[233.1.4.255]",

         "faultReason": "Policer resources exhausted. Configured TCAM max has been reached",

         "faultResolution": "Review TCAM configuration if needed",

         "group": "233.1.4.255",

        "modTs": "2025-04-01T14:35:04.081+00:00",

        "source": "47.20.20.9",

        "tStamp": "1743518104080",

         "vrf": "default"

}

This table provides a list of flow fault codes, their reasons, and suggested resolutions for the flow fault type.

Fault Code

Fault Reason

Fault Resolution

3051

No bandwidth available for sender

Please review bandwidth configuration and modify if needed

3201

Flow denied remotely

Please revisit participating upstream switches

3202

Flow denied remotely (external link)

Please revisit participating upstream switches

3376

Impacted by higher priority flow due to bandwidth contention

Effects of granular flow priority configuration

3126

Sender is not reachable

Please check the unicast routing table for the sender IP

3176

PIM not enabled

Please configure PIM

3151

PIM/IGMP host proxy not enabled

Please configure PIM/IGMP host proxy

3152

PIM/IGMP host proxy not enabled (external link)

Please configure PIM/IGMP host proxy

Sender Faults

The nbmSenderFaults Managed Object provides information about faults originating from the media sender, such as host policy denials or resource limitations.

"nbmSenderFaults": {

    "attributes": {

        "dn": "sys/nbm/show/faults/dom-default/senderfaults-[sys/nbm/show/endpoints/dom-default/h-[47.20.20.9]-if-0/g-[227.10.10.1]]",

       "faultCode": "2001",

        "faultDn": "sys/nbm/show/endpoints/dom-default/h-[47.20.20.9]-if-0/g-[227.10.10.1]",

        "faultReason": "Denied by sender host policy",

        "faultResolution": "Review sender host policy configuration and modify if needed"

         "group": "227.10.10.1",

          "modTs": "2025-04-01T14:25:13.635+00:00",

          "senderEndpoint": "47.20.20.9",

          "tStamp": "1743517513635",

          "vrf": "default"

}

This table provides a list of sender fault codes, their reasons, and suggested resolutions for the sender fault type.

Fault Code

Fault Reason

Fault Resolution

2001

Denied by sender host policy

Review sender host policy configuration and modify if needed

2002

Denied by sender host policy (external link)

Review sender host policy config and modify if needed

2051

No bandwidth available for sender

Please review bandwidth configuration and modify if needed

2052

No bandwidth available for sender (external link)

Please review bandwidth configuration and modify if needed

2076

Policer resources exhausted. Configured TCAM max has been reached

Review TCAM configuration if needed

2077

Policer resources exhausted (external link), configured TCAM max has been reached

Review TCAM configuration if needed

2377

Impacted by a higher priority flow due to policer unavailability

Effects of granular flow priority configuration

2101

No matching flow policy found

Please define the flow policy and bandwidth for this group

2151

PIM/IGMP host proxy not enabled

Please configure PIM/IGMP host proxy

2152

PIM/IGMP host proxy not enabled (external link)

Please configure PIM/IGMP host proxy

2351

No TCAM allocated to ing-nbm region

Please revisit TCAM configuration

2352

No TCAM allocated to ing-nbm region for external ingress interface

Please revisit TCAM configuration

Receiver Faults

The nbmReceiverFaults Managed Object provides information about faults related to the media receiver, such as bandwidth limitations or connectivity issues.

"nbmReceiverFaults": {

     "attributes": {

          "dn": "sys/nbm/show/faults/dom-default/receiverfaults-[sys/nbm/show/endpoints/dom-default/h-[47.20.10.1]-if-436231169/s-[47.20.20.9]-g-[227.10.10.1]]",

          "faultCode": "1026",

          "faultDn": "sys/nbm/show/endpoints/dom-default/h-[47.20.10.1]-if-436231169/s-[47.20.20.9]-g-[227.10.10.1]",

          "faultReason": "No bandwidth currently available for receiver",
          
          "faultResolution": "Please review flow policy if receiver needs to be stitched"

          "group": "227.10.10.1",

          "modTs": "2025-04-01T14:27:46.801+00:00",

          "receiverEndpoint": "47.20.10.1",

          "receiverInterface": "Ethernet1/47.1",

          "source": "47.20.20.9",

          "tStamp": "1743517666801",

          "vrf": "default"

}

This table provides a list of receiver fault codes, their reasons, and suggested resolutions for the receiver fault type.

Fault Code

Fault Reason

Fault Resolution

1026

No bandwidth currently available for receiver

Please review flow policy, if receiver needs to be stitched

Pim Passive Ingress Faults

The nbmFlowIngressFaults Managed Object provides information about faults related to the ingress interface of a media flow, such as VRF context issues or invalid interface configurations.

"nbmFlowIngressFaults": {
                "attributes": {
                "dn": "sys/nbm/show/faults/dom-default/flowingressfaults-[sys/nbm/conf/flows/dom-default/s-[47.20.20.9]-g-[230.1.0.1]]",
                "faultCode": "4230",
                "faultDn": "sys/nbm/conf/flows/dom-default/s-[47.20.20.9]-g-[230.1.0.1]",
                "faultReason": "IIF is not part of valid VRF context",
                "faultResolution": "Update VRF context on IIF if needed, then delete and re-add DN in fault",
                "group": "230.1.0.1",
                "ingressif": "null0_iif",
                "modTs": "2025-04-01T15:07:15.248+00:00",
                "source": "47.20.20.9",
                "tStamp": "1743520035248",
                "vrf": "default"
                }

This table provides a list of PIM passive ingress fault codes, their reasons, and suggested resolutions for the PIM passive ingress fault type.

Fault Code

Fault Reason

Fault Resolution

4226

Invalid interface IP on IIF

Configure interface IP address, then set RPF to unspecified and validate

4230

IIF is not part of valid VRF context

Update VRF context on IIF if needed, then delete and re-add DN in fault

4251

VRF context is shut down for ingress interface

Enable VRF context, then delete and re-add DN in fault

4276

Ingress interface mroute clear command initiated

Delete and re-add DN in fault

4076

Policer resources exhausted. Configured TCAM max has been reached

Review TCAM configuration if needed

4228

Interface IP (IIF) VRF context was changed

Revert interface VRF configuration if needed, then delete and re-add DN in fault

4232

Missing PIM/IGMP host proxy config on RPF

Please configure PIM/IGMP host proxy on interface, then delete and re-add DN in fault

Pim Passive Egress Faults

The nbmFlowEgressFaults Managed Object provides information about faults related to the egress interface of a media flow, such as invalid interface IP addresses or configuration errors.

"nbmFlowEgressFaults": {
                "attributes": {
                "dn": "sys/nbm/show/faults/dom-default/flowegressfaults-[sys/nbm/conf/flows/dom-default/s-[47.20.20.9]-g-[230.1.0.1]/if-[eth1/47.1]]",
                "egressif": "Eth1/47.1",
                "faultCode": "4227",
                "faultDn": "sys/nbm/conf/flows/dom-default/s-[47.20.20.9]-g-[230.1.0.1]/if-[eth1/47.1]",
                "faultReason": "Invalid interface IP on OIF",
                "faultResolution": "Configure interface IP address, then delete and re-add DN in fault",
                "group": "230.1.0.1",
                "modTs": "2025-04-01T15:06:59.738+00:00",
                "source": "47.20.20.9",
                "tStamp": "1743520019738",
                "vrf": "default"
                }

This table provides a list of PIM passive egress fault codes, their reasons, and suggested resolutions for the PIM passive egress fault type.

Fault Code

Fault Reason

Fault Resolution

4227

Invalid interface IP on OIF

Configure interface IP address, then delete and re-add DN in fault

4229

OIF is not part of valid VRF context

Update VRF context on OIF if needed, then delete and re-add DN in fault

4231

Missing OIF-PIM config on egress interface

Configure PIM on interface, then delete and re-add DN in fault

4252

VRF context is shut down for outgoing interface

Enable VRF context, then delete and re-add DN in fault

4277

Egress interface mroute clear command initiated

Delete and re-add DN in fault

Interface Level Usage

Egress Event: nbmEgressEvent

The nbmEgressEvent Managed Object provides information about interface-level usage, such as bandwidth utilization on egress interfaces.

"nbmEgressEvent": {
                
                "attributes": {
                
                "dn": "sys/nbm/show/notify/dom-vrf_pmn1/egressevent-[vrf:vrf_pmn1-INTF:Eth1/47.2-EGRESS]",
                
                "egressinterface": "Eth1/47.2",
                
                "modTs": "2025-04-03T08:12:36.338+00:00",
                
                "notifyCode": "5301",
                
                "notifyDn": "vrf:vrf_pmn1-INTF:Eth1/47.2-EGRESS",
                
                "reason": "CRITICAL: egress bandwidth usage is at or above 90%",
                
                "tStamp": "1743667956338",
                
                "vrf": "vrf_pmn1"
                
                }

This table provides a list of notification codes and their corresponding reasons for the egress event type.

Notify Code

Reason

5301

CRITICAL: egress bandwidth usage is at or above 90%

Ingress Event: nbmIngressEvent

The nbmIngressEvent Managed Object provides information about interface-level usage, such as bandwidth utilization on ingress interfaces.

"nbmIngressEvent": {
                
                "attributes": {
                
                "dn": "sys/nbm/show/notify/dom-vrf_pmn1/ingressevent-[vrf:vrf_pmn1-INTF:Eth1/42.1-INGRESS]",
                
                "ingressinterface": "Eth1/42.1",
                
                "modTs": "2025-04-03T08:16:11.366+00:00",
                
                "notifyCode": "5302",
                
                "notifyDn": "vrf:vrf_pmn1-INTF:Eth1/42.1-INGRESS",
                
                "reason": "CRITICAL: ingress bandwidth usage is at or above 90%",
                
                "tStamp": "1743668171366",
                
                "vrf": "vrf_pmn1"
                
                }

This table provides a list of notification codes and their corresponding reasons for the ingress event type.

Notify Code

Reason

5302

CRITICAL: ingress bandwidth usage is at or above 90%

Event MO for Flow Rate

nbmEvent

The nbmEvent Managed Object provides information about flow rate events, such as when the flow rate falls below a configured threshold or exceeds a threshold.

"nbmEvent": {
                "attributes": {
                "dn": "sys/nbm/show/notify/dom-default/event-[vrf:default-BW:s-47.20.20.1-g-225.1.1.1]",
                "group": "225.1.1.1",
                "modTs": "2025-04-01T14:09:01.530+00:00",
                "notifyCode": "5304",
                "notifyDn": "vrf:default-BW:s-47.20.20.1-g-225.1.1.1",
                "reason": "Rate below 60% of the configured flow policy",
                "source": "47.20.20.1",
                "tStamp": "1743516541530",
                "vrf": "default"
                }
                }

This table provides a list of notification codes and their corresponding reasons for the flow rate event type.

Notify Code

Reason

5303

Rate is over 100% of the configured flow policy

5304

Rate is below 60% of the configured flow policy

Flow Provisioned Event

nbmFlowEvent

The nbmFlowEvent Managed Object provides information about flow provisioning events, such as when a flow is successfully provisioned.

"nbmFlowEvent": {
                "attributes": {
                "dn": "sys/nbm/show/notify/dom-default/flowevent-[vrf:default-FLOW:s-[47.20.20.1]-g-[226.1.1.1]/oif-[Lo1]]",
                "egressinterface": "Lo1",
                "group": "226.1.1.1",
                "modTs": "2025-04-01T14:02:43.330+00:00",
                "notifyCode": "5201",
                "notifyDn": "vrf:default-FLOW:s-[47.20.20.1]-g-[226.1.1.1]/oif-[Lo1]",
                "reason": "Flow provisioned successfully",
                "source": "47.20.20.1",
                "tStamp": "1743516163330",
                "vrf": "default"
                }
                }

This table provides a list of notification codes and their corresponding reasons for the flow provisioned event type.

Notify Code

Reason

5201

Flow provisioned successfully

NAT Event

Oversubscription Event

The nbmInatOversubscriptionEvent Managed Object provides information about NAT oversubscription events, such as when the cumulative pre-NAT bandwidth exceeds the post-NAT bandwidth.

"nbmInatOversubscriptionEvent": {
                "attributes": {
                "dn": "sys/nbm/show/notify/dom-default/inatoversubscriptionevent-[vrf:default-post_s-[51.51.51.51]-post_g-[226.1.1.1]-ingress]",
                "group": "226.1.1.1",
                "modTs": "2025-04-01T14:19:04.393+00:00",
                "notifyCode": "5305",
                "notifyDn": "vrf:default-post_s-[51.51.51.51]-post_g-[226.1.1.1]-ingress",
                "reason": "Oversubscription: cumulative pre-NAT bandwidth is higher than post-NAT bandwidth from the respective flow policies",
                "source": "51.51.51.51",
                "tStamp": "1743517144393",
                "vrf": "default"
                }
                }

This table provides a list of notification codes and their corresponding reasons for the NAT oversubscription event type.

Notify Code

Reason

5305

Oversubscription: cumulative pre-NAT bandwidth is higher than post-NAT bandwidth from the respective flow policies

Bandwidth Mismatch Event

The nbmEnatBandwidthmismatchEvent Managed Object provides information about NAT bandwidth mismatch events, such as when there is a mismatch between the pre-NAT and post-NAT flow bandwidth.

"nbmEnatBandwidthmismatchEvent": {
                "attributes": {
                "destPort": "0",
                "dn": "sys/nbm/show/notify/dom-default/enatbandwidthmismatchevent-[vrf:default-post_s-[100.1.1.1]-post_g-[226.1.2.1]-pre_s-[47.20.20.1]-pre_g-[226.1.1.1]-S[0]-D[0]-egress/if-[Eth1/47.1]]",
                "group": "226.1.2.1",
                "modTs": "2025-04-01T14:02:44.123+00:00",
                "notifyCode": "5307",
                "notifyDn": "vrf:default-post_s-[100.1.1.1]-post_g-[226.1.2.1]-pre_s-[47.20.20.1]-pre_g-[226.1.1.1]-S[0]-D[0]-egress/if-[Eth1/47.1]",
                "preGroup": "226.1.1.1",
                "preSource": "47.20.20.1",
                "reason": "Pre- and post-translation flow bandwidth mismatch",
                "source": "100.1.1.1",
                "sourcePort": "0",
                "tStamp": "1743516164123",
                "vrf": "default"
                }
                }

This table provides a list of notification codes and their corresponding reasons for the NAT bandwidth mismatch event type.

Notify Code

Reason

5307

Pre- and post-translation flow bandwidth mismatch

Failure Handling

Cisco's IP fabric for media solution supports deterministic failure handling.

During a link or switch failure, the affected flows are moved to alternate links, provided sufficient bandwidth is available. With SMPTE 2022-7, redundancy is built on the endpoints, which ensures that the link or switch failure does not affect production traffic.

Cisco NX-OS Release 10.6(1)F introduces enhancements to failure handling that provide more detailed and actionable fault information. This allows administrators to effectively understand fault information, identify the root cause of failures, and implement appropriate remediation steps using their preferred network management tools.

Benefits of the IP Fabric for Media Solution

Cisco's IP fabric for media solution provides the following benefits:

  • Replaces specialized hardware (SDI routers) with a general-purpose switching infrastructure.

  • Supports various types and sizes of broadcasting equipment endpoints with port speeds up to 100 Gbps.

  • Supports the latest video technologies, including 4K and 8K ultra HD.

  • Scales horizontally. When you need more capacity, you can add a leaf switch to support more endpoints.

  • Provides a deterministic network with zero packet loss, ultra low latency, and minimal jitter.

  • Capable of synchronizing all media sources and receivers.

  • Provides deterministic failure handling that sends traffic to the receiver when a link fails between a leaf and the spine.

  • Supports the coexistence of live and file-based traffic flows for postproduction work.

  • Offers increased network security.

  • Provides a non-blocking network design to prevent the oversubscription of links.

  • Requires no changes to the existing operator workflow.

Related Documentation

Related Topic

Document Title

Cisco NDFC

Cisco Nexus Dashboard Fabric Controller Installation and Upgrade Guide

Cisco DCNM online help

Cisco Nexus Dashboard

Cisco Nexus Dashboard

Cisco NX-OS release information

Cisco Nexus 9000 Series NX-OS IP Fabric for Media Release Notes

Cisco NX-OS software upgrades

Cisco Nexus 9000 Series NX-OS Software Upgrade and Downgrade Guide

IGMP snooping and PIM

Cisco Nexus 9000 Series NX-OS Multicast Routing Configuration Guide

IP fabric for media scalability numbers

Cisco Nexus 9000 Series NX-OS Verified Scalability Guide

NX-API REST

Cisco Nexus 3000 and 9000 Series NX-API REST SDK User Guide and API Reference

OSPF

Cisco Nexus 9000 Series NX-OS Unicast Routing Configuration Guide

PTP

Cisco Nexus 9000 Series NX-OS System Management Configuration Guide

QoS

Cisco Nexus 9000 Series NX-OS Quality of Service Configuration Guide

TCAM carving

Cisco Nexus 9000 Series NX-OS Security Configuration Guide

VLANs

Cisco Nexus 9000 Series NX-OS Layer 2 Switching Configuration Guide