Cisco Nexus 1000V System Management Configuration Guide, Release 4.2(1)SV1(5.1)
Virtualized Workload Mobility (DC to DC vMotion)
Downloads: This chapterpdf (PDF - 113.0KB) The complete bookPDF (PDF - 4.17MB) | Feedback

Table Of Contents

Virtualized Workload Mobility (DC to DC vMotion)

Information About Virtualized Workload Mobility (DC to DC vMotion)

Stretched Cluster

Split Cluster

Prerequisites for Virtualized Workload Mobility (DC to DC vMotion)

Guidelines and Limitations

Physical Site Considerations

Handling Inter-Site Link Failures

Headless Mode of Operation

Handling Additional Distance/Latency Between the VSM and VEM

Migrating a VSM

Migrating a VSM Hosted on an ESX or ESXi Host

Verifying the Virtualized Workload Mobility (DC to DC vMotion) Configuration

Monitoring Virtualized Workload Mobility (DC to DC vMotion)

Configuration Limits

Feature History for Virtualized Workload Mobility (DC to DC vMotion)


Virtualized Workload Mobility (DC to DC vMotion)


This chapter describes the Virtualized Workload Mobility (DC to DC vMotion) feature of the Cisco Nexus 1000. This feature addresses Cisco Nexus 1000 across two physical data centers.

This chapter includes the following sections:

Information About Virtualized Workload Mobility (DC to DC vMotion)

Prerequisites for Virtualized Workload Mobility (DC to DC vMotion)

Guidelines and Limitations

Verifying the Virtualized Workload Mobility (DC to DC vMotion) Configuration

Monitoring Virtualized Workload Mobility (DC to DC vMotion)

Feature History for Virtualized Workload Mobility (DC to DC vMotion)

Information About Virtualized Workload Mobility (DC to DC vMotion)

This section describes the Virtualized Workload Mobility (DC to DC vMotion) configurations and includes the following topics:

Stretched Cluster

Split Cluster

Physical Site Considerations

Stretched Cluster


Note A stretched cluster is a cluster with ESX/ESXi hosts in different physical locations.


In an environment where the same Cisco Nexus 1000 instance spans two data centers, this configuration allows you to have Virtual Ethernet Modules (VEMs) in different data centers be part of the same vCenter Server cluster.

By choosing this configuration, you are ensure that the VEMs in either data center (in a two data center environment) are a part of the same Dynamic Resource Scheduling (DRS) / VMware High Availability (VMW HA) / Fault Tolerance (FT) domain that allows for multiple parallel virtual machine (VM) migration events.

Split Cluster

The Split Cluster configuration is an alternate to the Stretched Cluster deployment. With this configuration, the deployment consists of one or more clusters on either physical site with no cluster that contains VEMs in multiple data centers. While this configuration allows for VM migration between physical data centers, these events are not automatically scheduled by DRS.

Prerequisites for Virtualized Workload Mobility (DC to DC vMotion)

Virtualized Workload Mobility (DC to DC vMotion) has the following prerequisites:

You must set up your DRS affinity rules to ensure that the VSM pair is restricted to one site.

Layer 2 extension between the two physical data centers over the DCI link.

Guidelines and Limitations

Virtualized Workload Mobility (DC to DC vMotion) has the following guidelines and limitations:

The VSM HA pair must be located in the same site as their storage and the active vCenter Server.

Layer 3 control mode is preferred.

If you are using Link Aggregation Control Protocol (LACP) on the VEM, use LACP offload.

Quality of Service bandwidth guarantees for control traffic over the DCI link.

Limit the number of physical data centers to two.

A maximum latency of 5 ms is supported for VSM-VEM control traffic.

For configuration limits in a two data center environment, see the "Configuration Limits" section.

Physical Site Considerations

When you are designing a physical site, follow these guidelines:

Check the average and maximum latency between a Virtual Supervisor Module (VSM) and VEM.

Follow the procedures to perform actions you would intend to do in normal operation. For example, VSM migration.

Design the system to handle the high probability of VSM-VEM communication failures where a VEM must function in headless mode due to data center interconnect (DCI) link failures.

Handling Inter-Site Link Failures

If the DCI link or Layer 2 extension mechanism fails, a set of VEM modules might run with their last known configuration for a period of time.

Headless Mode of Operation

For the period of time that the VSM and VEM cannot communicate, the VEM continues to operate with its last known configuration. Once the DCI link connectivity is restored and the VSM-VEM communication is reestablished, the system should come back to its previous operational state.

This mode type is no different than the headless mode of operation within a data center and has the following limitations for the headless VEM:

No new ports can be brought up on the headless VEM (new VMs coming up or VMs coming up after vMotion).

No NetFlow data exports.

Ports shut down because DHCPS/DAI rate limits are not automatically brought up until the VSM reconnects.

Port security options, such as aging or learning secure MAC addresses and shutting down/recovering from port-security violations, are not available until the VSM reconnects.

The Cisco Discovery Protocol (CDP) does not function for the disconnected VEM.

IGMP joins/leaves are not processed until the VSM reconnects.

Queries on BRIDGE and IF-MIB processed at the VSM give the last known status for the hosts in headless mode.

Handling Additional Distance/Latency Between the VSM and VEM

In a network where there is a considerable distance between the VSM and VEM, latency becomes a critical factor.

Because the control traffic between the VSM and VEM faces a sub-millisecond latency within a data center, latency can increase to a few milliseconds depending on the distance.

With an increased round-trip time, communication between the VSM and VEM takes longer. As you add VEMs and vEthernet interfaces, the time it takes to perform actions such as configuration commands, module insertions, port bring-up, and show commands increase because that many tasks are serialized.

Migrating a VSM

This section describes how migrate a VSM from one physical site to another.


Note If you are migrating a VSM on a Cisco Nexus 1010, see the Cisco Nexus 1010 Software Configuration Guide, Release 4.2(1)SP1(3).


Migrating a VSM Hosted on an ESX or ESXi Host

Use the following procedure to migrate a VSM that is hosted on an ESX or ESXi host from the local data center to the remote data center.

BEFORE YOU BEGIN

Before beginning this procedure, you must know or do the following:

Reduce the amount of time where the VSM runs with remote storage in another data center.

Do not bring up any new VMs or vMotion VMs that are hosted on any VEMs corresponding to the VSM that is being migrated.


Note For information on vMotion or storage vMotion, see the VMware documentation.


PROCEDURE


Step 1 Migrate the standby VSM to the backup site.

Step 2 Perform a storage vMotion for the standby VSM storage.

Step 3 Enter the system switchover command.

switch # system switchover 
 
 

Step 4 Migrate the original active VSM to the backup site.

Step 5 Perform a storage vMotion for the original active VSM storage.


Verifying the Virtualized Workload Mobility (DC to DC vMotion) Configuration

switch# show module 
 
 
switch# show interface 

Monitoring Virtualized Workload Mobility (DC to DC vMotion)

switch# show module 
 
 
switch# show interface 

Configuration Limits

See

Feature History for Virtualized Workload Mobility (DC to DC vMotion)

Table 15-1 lists the release history for this feature.

Table 15-1 Feature History for Virtualized Workload Mobility (DC to DC vMotion)

Feature Name
Releases
Feature Information

Virtualized Workflow Mobility (DC to DC vMotion)

4.2(1)SV1(4a)

This feature was introduced.