Methodology
This chapter contains the following sections:
•Test Cycle
•Sample Test Case
Test Cycle
The test cycle consists of the following steps:
1. Network configuration and verification
2. Image upgrade and rollback with ISSU
3. Induce network disruptions at each layer (see Network Disruption Test Cases)
4. Scale control-plane peers and routes for both unicast and multicast routing at the core layer
5. Scale the number of access level switches and hosts to stress ARP and IGMP at the aggregation layer
6. Extended uptime monitoring to check for CPU and memory usage anomalies
Network Disruption Test Cases
The following sections describe the test disruptions and the verification criteria:
•System Level
•Edge Layer
•Core Layer
•Aggregation Layer
•Access Layer
System Level
|
|
Image upgrade and rollback with ISSU |
Hitless upgrade/rollback for all configured features with parallel enhancement |
Edge Layer
|
|
Router Link Failure/Recovery between Edge and Public Cloud |
•BGP reconvergence (control-plane & data-plane) •IGP and Multicast services reconvergence (control-plane & data plane) •MPLS/VPN and LDP reconvergence (control-plane & data-plane) •BFD peer detection and client notifications •GRE Tunnel re-route due to transport disruption |
Member of Port-channel Failure/Recovery between Edge and Public Cloud |
•Traffic load-sharing for port-channels •LACP interoperability •Unidirectional Link Detection (UDLD) |
Clear IGP Neighbors/Process at Edge |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Unicast Routes at Edge |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Multicast Routes at Edge |
Stress test for control-plane recovery |
Edge Switch System Failure/Recovery |
•BGP reconvergence (control-plane & data-plane) •IGP and Multicast services reconvergence (control-plane & data plane) •MPLS/VPN and LDP reconvergence (control-plane & data-plane) •BFD peer detection and client notifications •GRE Tunnel re-route due to transport disruption •VDC failure does not impact other VDCs |
Edge Switch Power Redundancy |
Partial Power loss causes no impact to control/data plane |
Edge Switch Supervisor High-Availability |
•SSO/NSF, in-chassis and on peers •SSO/NSF interoperability |
Edge Switch Fabric High-Availability |
Fabric module failure causes no impact to control/data plane |
Line Card OIR at Edge Switch |
•Hitless operation for non-affected ports •Traffic load-sharing for distributed port-channels •IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications •LACP interoperability for distributed port-channels •Unidirectional Link Detection (UDLD) |
Core Layer
|
|
Router Link Failure/Recovery between Core and Edge |
•IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications |
Member of Port-channel Failure/Recovery between Core and Edge |
•Traffic load-sharing for port-channels •LACP interoperability •Unidirectional Link Detection (UDLD) |
Clear IGP Neighbors/Process at Core |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Unicast Routes at Core |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Multicast Routes at Core |
Stress test for control-plane recovery |
Core Switch System Failure/Recovery |
•IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications •PIM Rendezvous Point redundancy & Back-up verification •VDC failure does not impact other VDCs |
Core Switch Power Redundancy |
Partial Power loss causes no impact to control/data plane |
Core Switch Supervisor High-Availability |
•SSO/NSF, in-chassis and on peers •SSO/NSF interoperability |
Core Switch Fabric High-Availability |
Fabric module failure causes no impact to control/data plane |
Line Card OIR at Core Switch |
•Hitless operation for non-affected ports •Traffic load-sharing for distributed port-channels •IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications •LACP interoperability for distributed port-channels •Unidirectional Link Detection (UDLD) |
Aggregation Layer
|
|
Router Link Failure/Recovery between Aggregation and Core |
•IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications |
Member of Port-channel Failure/Recovery between Aggregation and Core |
•Traffic load-sharing for port-channels •LACP interoperability •Unidirectional Link Detection (UDLD) |
Router Link Failure/Recovery between Aggregation and Access |
•STP reconvergence •IGMP reprogramming with snooping •MAC address re-learning •Security ACL & FNF reprogramming •No FHRP impact •No ARP/ND impact •No BFD impact •vPC functionality |
Member of Port-channel Failure/Recovery between Aggregation and Access |
•Traffic load-sharing for port-channels •LACP interoperability •Unidirectional Link Detection (UDLD) |
Clear IGP Neighbors/Process at Aggregation |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Unicast Routes at Aggregation |
Stress test for control-plane recovery |
Clear IPv4/IPv6 Multicast Routes at Aggregation |
Stress test for control-plane recovery |
Aggregation Switch System Failure/Recovery |
•STP reconvergence •IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications •PIM Rendezvous Point redundancy & Back-up verification •PIM DR/BDR functionality •IGMP Snooping & Querier functionality •VDC failure does not impact other VDCs •Security ACL & FNF reprogramming •FHRP redundancy •MAC address learning •ARP/ND re-learning •vPC functionality |
Aggregation Switch Power Redundancy |
Partial Power loss causes no impact to control/data plane |
Aggregation Switch Supervisor High-Availability |
•SSO/NSF, in-chassis and on peers •SSO/NSF interoperability •No impact to vPC peering status |
Aggregation Switch Fabric High-Availability |
Fabric module failure causes no impact to control/data plane |
Line Card OIR at Aggregation Switch |
•Hitless operation for non-affected ports •Traffic load-sharing for distributed port-channels •IGP and PIM reconvergence (control-plane & data plane) •BFD peer detection and client notifications •LACP interoperability for distributed port-channels •Unidirectional Link Detection (UDLD) |
vPC peer-link/keep-alive Failure/Recovery |
vPC functionality and peering status |
vPC Leg Failure/Recovery |
•No impact to STP overlay •IGMP reprogramming with snooping •MAC address re-learning •Security ACL & FNF reprogramming •No FHRP impact •No ARP/ND impact |
vPC Leg member Failure/Recovery |
•Traffic load-sharing for port-channels •LACP interoperability •Unidirectional Link Detection (UDLD) |
Access Layer
|
|
Access/ToR Switch System Failure/Recovery |
•STP reconvergence •IGMP snooping reprogramming •MAC address re-learning •No impact to other vPCs |
Sample Test Case
|
|
Link failure between aggregation and core layers |
|
Verify network control and data plane recovery after link flap |
|
•Reference topology •Reference network configuration setup test case •Reference test plan for control and data plane setup matrices |
|
1. Fail one of the links between the aggregation and core layers. 2. Recover the above link. 3. Repeat the same test at least 5 iterations to ensure consistent behavior for the devices and network. 4. Repeat the above procedures for the other links between the aggregation and core layers. |
|
•During the link failure, traffic should drop in proportion to the number of links and paths affected, and the traffic should be able to reconverge within the expected time relative to a previously-established Catalyst 6500 baseline. •Ensure that the unicast and multicast routing protocols have detected peer failure in order to start network reconvergence within the expected time. •Verify the convergence pattern is as expected. •Verify the CPU usage pattern is as expected. •Verify the memory usage is as expected. •Verify the route tables for both unicast and multicast routing are updated correctly on all switches in the network. Ensure that only affected switches show change in the forwarding tables. •Verify the hardware forwarding entries, line card programming, fabric programming, outgoing interface, forwarding engine entries, for both unicast and multicast routing are updated correctly on all switches in the network. •Verify Layer 2 forwarding tables on aggregation and access switches. They should not be affected by this failure. |