Guest

Cisco MDS 9700 Series Multilayer Directors

Large SAN Design Best Practices Using Cisco MDS 9700 and MDS 9500 Multilayer Directors White Paper

  • Viewing Options

  • PDF (1.3 MB)
  • Feedback

What You Will Learn

As SANs continue to grow in size, many factors need to be considered to help scale and manage them. This document focuses on large SAN deployments within a data center and provides best practices and design considerations for you to apply when designing a large physical fabric. It does not address networks that implement Inter-VSAN Routing (IVR), Fibre Channel or Fibre Channel over IP (FCIP)-based SAN extension, or intelligent fabric applications (for example, Cisco® Data Mobility Manager, or Cisco I/O Acceleration [IOA]).

Design Parameters

In SAN environments, many design criteria need to be addressed, such as the number of servers that access a shared storage frame, the network topology, and fabric scaling. This document focuses on the following design parameters:

   Deployments with 2500 or more end devices (servers, storage, and tape devices)

   Deployments in which a majority of end devices have connection speeds of 8 Gbps and 16 Gbps

   Deployments with identical dual physical fabrics (fabric A and fabric B)

Network Considerations

When designing a large Cisco MDS fabric, the following should be taken into consideration:

   Ports and port groups

   Dedicated and shared rate mode

   Port speed

   Inter-Switch Link (ISL)

   PortChannels

   8 Gbps compared to 10 Gbps Fibre channel ISL

   Fan-in, fan-out, and oversubscription ratio

   Zone types and smart zoning

   Virtual SAN (VSAN)

   Fabric login

Ports and Port Groups

Each port in the Cisco MDS 9000 Family is a member of one port group that shares common resources from an assigned pool of allocated bandwidth. This allows the appropriate bandwidth allocation for high-bandwidth and low-bandwidth devices. Table 1 provides some details needed to better understand bandwidth allocation per port and port groups for switching modules that will be covered in this paper.

Table 1.       Bandwidth and Port Group Configurations for Fibre Channel (FC) Modules

Part Number

Product Name

No. of Port Groups

No. of Ports Per Port Group

Bandwidth Per Port Group (Gbps)

DS-X9248-256K9

48-port 8-Gbps Adv FC module

8

6

32.41

12.82

DS-X9232-256K9

32-port 8-Gbps Adv FC module

8

4

32.41

12.82

DS-X9248-96K9

48-port 8-Gbps FC module

8

6

12.83

DS-X9224-96K9

24-port 8-Gbps FC module

8

3

12.83

DS-X9448-768K9

48-port 16-Gbps

12

4

644

1 MDS 9513 with fabric 3 module installed
2 MDS 9506 (all) and MDS 9509 (all) or MDS 9513 with fabric 2 module is more oversubscribed
3 MDS 9506 (all), MDS 9509 (all), or MDS 9513 (all)
4 MDS 9700 (all)

Dedicated and Shared Rate Mode

Ports on MDS 9000 series line cards are grouped into port groups that have a fixed amount of bandwidth per port group (see Table 1). The MDS 9000 family allows for the bandwidth of ports within a port group to be allocated based on the requirements of individual ports. When planning port bandwidth requirements, allocating the bandwidth within the port group is important. Ports with the port group can have bandwidth dedicated to them or ports can share a pool of bandwidth. Ports that require high sustained bandwidth, for example, ISL ports, storage and tape array ports, or high-bandwidth servers, can have bandwidth dedicated to them within a port group using the switchport rate-mode dedicated command. Other ports, typically servers that access shared storage array ports (that is, storage ports that have higher fan-out ratios), can share the bandwidth within a port group by using the switchport rate-mode shared command. When configuring the ports, it is important not to exceed the available bandwidth within a port group.

As an example, a MDS 9513 with a fabric 3 module installed and using a 48-port 8-Gbps Advanced Fibre Channel module has eight port groups of six ports each. Each port group has 32.4 Gbps of bandwidth available. It would not be possible to configure all six ports of a port group at 8-Gbps dedicated rate since that would require 48 Gbps of bandwidth and the port group only has 32.4 Gbps of bandwidth. It is possible to configure all six ports in shared rate mode, meaning that the ports would run at eight Gbps and will be over-subscribed at a rate of 1.48:1 (6 ports multiplied by 8 Gbps = 48 Gbps/32.4 Gbps). This is an over-subscription rate well below the over-subscription rate of the typical storage array port (fan-out ratio) and does not impact performance.

It is possible to mix dedicated and shared rate ports within a port group. Using the same environment as before, one port in the port group can be configured for dedicated eight Gbps of bandwidth and be used as an ISL port or a storage target port. This will allocate eight Gbps of the port group bandwidth to the dedicated rate mode port, leaving 24.4 Gbps of bandwidth for the remaining five ports to be shared, giving them an oversubscription ratio of 1.64:1 (5 ports multiplied by 8 Gbps = 40 Gbps/24.4).

Port Speed

The speed of an interface, combined with rate mode, determines the amount of shared resources available to the ports in the port group. The speed of an interface can be configured to automatically detect the speed of the attached device or it can be explicitly configured. When a port is configured for auto speed detect, the switch assumes it is capable of the highest speed allowed by that port, dependent on the type of line card module being used. This means that port bandwidth may be allocated for a higher speed than the attached device requires. For maximum use of bandwidth within a port group, it is recommended to explicitly specify the port speed of the attached device.

Inter-Switch Link (ISL)

An ISL is a connection between Fibre Channel switches. The number of ISLs required between Cisco MDS switches depends on the desired end-to-end oversubscription ratio. The storage port oversubscription ratio from a single storage port to multiple servers can be used to help determine the number of ISLs needed for each edge-to-core connection. Figure 1 shows three examples of storage, server, and ISL combinations, all with the same oversubscription ratio of 8:1.

   The first example has one 16-Gbps storage port with eight 16-Gbps server ports traversing
one 16-Gbps ISL.

   The second example has one 16-Gbps storage port with sixteen 8-Gbps server ports traversing
one 16-Gbps ISL.

   The third example has eight 16-Gbps storage port with sixty-four 16-Gbps server ports traversing eight
16-Gbps ISLs.

A 1:1 ratio of storage bandwidth-to-ISL bandwidth is recommended for SAN design. ISL bandwidth can be added with additional ISLs to provide greater availability in the event of a link failure.

Figure 1.      Number of ISLs Needed to Maintain Oversubscription Ratio

PortChannel

A PortChannel is an aggregation of multiple physical interfaces into one logical interface to provide higher aggregated bandwidth, load balancing, and link redundancy while providing fabric stability in the event that a member fails. PortChannels can connect to interfaces across different switching modules, so a failure of a switching module does not bring down the PortChannel link.

A PortChannel has the following functions:

   It provides a single, logical, point-to-point connection between switches.

   It provides a single VSAN ISL (E port) or trunking of multiple VSANs over an EISL (TE port). EISL ports exist only between Cisco switches and carry traffic for multiple VSANs, unlike an ISL.

   It increases the aggregate bandwidth on an ISL by distributing traffic among all functional links in the channel. PortChannels can contain up to 16 physical links and can span multiple modules for added high availability. Multiple PortChannels can be used if more than 16 ISLs are required between switches.

   It load balances across multiple links and maintains optimum bandwidth utilization. Load balancing is performed based on per-VSAN configuration (source ID [SID] and destination ID (DID); or SID, DID, and Exchange ID [OXID]).

   It provides high availability on an ISL. If one link fails, traffic is redistributed to the remaining links. If a link goes down in a PortChannel, the upper protocol is not aware of it. To the upper protocol, the link is still there, although the bandwidth is diminished. The routing tables are not affected by link failure.

8-Gbps and 10-Gbps Fibre Channel ISL

The Fibre Channel protocol is typically associated with the 1/2/4/8/16-Gbps speeds of attached devices. However, the Fibre Channel protocol also supports 10 Gbps, which can be used for ISLs. The use of 8-Gbps or 10-Gbps ISLs is a significant design consideration when 16-Gbps interfaces are not available.

At first glance, 10-Gbps Fibre Channel appears to be only a 25 percent increase over 8-Gbps Fibre Channel. However, due to the differences of the physical layer, 10-Gbps Fibre Channel actually has a data rate 50 percent greater than 8-Gbps Fibre Channel. To understand this, one must look at the way data is transmitted over the two interfaces. All data is encoded to ensure data integrity when transmitted over an interface. For 8-Gbps Fibre Channel, for every 8 bits of data, 10 bits are transmitted, imposing a 25 percent overhead. For 10-Gbps Fibre Channel, for every 64 bits of data, 66 bits are transmitted, an overhead on only 3.125 percent. This encoding, combined with the physical clock rate, determines the actual data rate of the interface. Table 2 outlines the differences in 8- and 10-Gbps data rates.

Table 2.       Comparison of 8 and 10-Gbps Data Rates

Fibre Channel
Interface Speed

Clock Rate

Data Encoding

Data Rate (Gbps)

Data Rate (MB/s)

8 Gbps

8.5 Gbps

8 bits : 10 bits

6.8 Gbps

850 MB/s

10 Gbps

10.51875 Gbps

64 bits : 66 bits

10.2 Gbps

1275 MB/s

When designing ISL connectivity, 10-Gbps Fibre Channel interfaces can provide greater bandwidth per ISL or reduce the number of ISLs between switch, reducing the amount of cabling required.

Different line card modules have their own port group settings. Depending upon the port group configuration, we can configure that port for regular 1/2/4/8/16 Fibre Channel speeds or 10-Gbps Fibre Channel speed. Note that depending on the specific line card module, not all the ports can be configured for 10-Gbps Fibre Channel.

Figures 2 and 3 below show the specific ports of the individual port groups that can be configured for 10-Gbps Fibre Channel speed. The interfaces that can be configured out of the port groups are identified within the yellow border and the ones that will be disabled by the switch are marked with a red cross (X). Only the first two port groups are shown, however the groupings of ports are the same for the remainder of the port groups.

Figure 2.      10-Gbps Fibre Channel Port Selection in DS-X9248-256k9

Figure 3.      10-Gbps Fibre Channel Port Selection in DS-X9232-256k9

For the DS-X9448-768K9, 16-Gbps 48-port line card module, all ports in paired port groups can operate in either 2/4/8/16-Gbps mode or 10-Gbps mode. This means that 10-Gbps is enabled in 8-port increments at port group level (ports 1-8, 9-16, 17-24, 25-32, 33-40, and 41-48). If 10-Gbps speed is enabled, it will be enabled for all interfaces in those two pairs of port groups, as mentioned previously.

Fan-in, Fan-out and Oversubscription Ratio

To efficiently and optimally use resources and to save deployment time and reduce management costs, SANs are designed to share array ports and ISL and line-card bandwidth. The terms used to describe this sharing include fan-in ratio, fan-out ratio, and oversubscription ratio. The term used depends on the point of reference being described. In general, the fan-in ratio is calculated as the ratio of host port bandwidth to storage array port bandwidth, and the fan-out ratio is calculated as the ratio of storage array port bandwidth to host port bandwidth. Oversubscription is a networking term that is generally defined as the overall bandwidth ratio between host and storage array ports. See Figure 4 for more details.

Figure 4.      Fan-in, Fan-out, and Oversubscription Ratios

Note:    Prior to N_Port ID Virtualization (NPIV) and Cisco N_Port Virtualization (NPV), a single port was limited to one fabric login. With NPIV and Cisco NPV-enabled switches, a single port can now support multiple fabric logins. Figure 5 shows 24 logins for 24 hosts and one for the switch.

Zone Types and Smart Zoning

Each virtual SAN (VSAN) has only one active zone set, which contains one or more zones. Each zone consists of one or more members to allow communication between the members. Cisco MDS 9000 SAN-OS and NX-OS Software provide multiple ways to identify zone members, but the commonly used ones are:

   PWWN: Port worldwide name of the device (most commonly used)

   Device alias: An easy-to-read name associated with a single device’s PWWN

Depending on the requirements of the environment, the type of zone members is a matter of preference. A recommended best practice is to create a device alias for end devices when managing the network. The device alias provides an easy-to-read name for a particular end device. For example, a storage array with PWWN 50:06:04:82:bf:d0:54:52 can be given a device-alias name of Tier1-arrayX-ID542-Port2. In addition, with device alias, when the actual device moves from one VSAN (VSAN 10) to a new VSAN (VSAN 20) in the same physical fabric, the device-alias name will follow that device. You do not need to re-enter the device alias for each port of the moved array in the new VSAN.

Note:    As a best practice for large SAN deployments, you should have more zones with two members rather than a single zone with three or more members. This practice is not a concern in smaller environments.

Smart zoning supports zoning multiple devices in a single zone by reducing the number of zoning entries that need to be programmed. This feature allows multiple member zones consisting of multiple initiators and multiple zones to be zoned together without increasing the size of the zone set. Smart zoning can be enabled at the zone level, zone-set level, zone-member level, or VSAN level.

Virtual SAN (VSAN)

Cisco MDS switches offer VSAN technology, which provides a simple and highly secure way to consolidate many SAN islands into a single physical fabric. Separate fabric services (per-VSAN zoning, name services, domains, separate role-based management, etc.) are provided for each VSAN, providing separation of both the control plane and the data plane.

VSANs have multiple use cases. For example, you can create a VSAN for each type of operating system (such as a VSAN for Microsoft Windows or HP-UX), or you can use them on the basis of business functions (a VSAN for development, for production, or for a lab, for instance). VSAN 1 is created on the Cisco MDS switch by default and cannot be deleted. As a best practice, VSAN 1 should be used as a staging area for unprovisioned devices, and other VSANs should be created for the production environments. With each VSAN having its own zones and zone sets, Cisco MDS switches can enable highly secure, scalable, and robust networks.

Fabric Login

When a Fibre Channel switch wants to forward traffic to its neighboring switch or host connector, it needs to interchange the login parameters. These parameters are called fabric logins. Each switch has some limitation on how many fabric logins can be allowed at any single time. Generally, the number of actual physical ports in the fabric is larger than the number of end devices (server, storage, and tape ports) in the physical fabric. The Cisco MDS Family supports enough numbers of fabric logins in a physical fabric, independent of the number of VSANs in the network for today’s network.

Typically, when designing a SAN, the number of end devices determines the number of fabric logins. The increase in blade server deployments and the consolidation of servers due to server virtualization technologies will affect the design of the network. With features such as NPIV and Cisco NPV, the number of fabric logins has further increased (Figure 5). The proliferation of NPIV-capable end devices such as host bus adaptors (HBAs) and Cisco NPV-mode switches makes the number of fabric logins on a per-port, per-line-card, per-switch, and per-physical-fabric basis a critical consideration. These fabric login limits will determine the design of the current SAN, as well as future growth. The total number of hosts and NPV switches will determine the fabric logins required on the core switch. In recent times, it is difficult to control the virtual machines hosted on servers so it is important to plan ahead and keep some reservations on the resources side. The current scalability numbers for the Cisco MDS product line can be found under Configuration Guides. To estimate the number of Flogis requirements for the SAN network we can use the formula: Number of hosts X number of initiators per host.

Figure 5.      Cisco NPV-Enabled Switches and Fabric Logins

Prior to NPIV and Cisco NPV, a single port was limited to one fabric login. With NPIV and Cisco NPV-enabled switches, a single port can now support multiple fabric logins. Figure 5 shows 24 logins for 24 hosts and one for the switch.

Cisco MDS 9500 and MDS 9700 Components

In this white paper, we will specifically talk about the best options to install the Cisco MDS 9700 Multilayer Director-class switch at the core with an existing SAN network running MDS 9513 Multilayer Directors at the edge and core. The Cisco MDS 9513 provides multiple line card options. Four of the mostly used and deployed line cards supported under MDS 9513 will be discussed in this paper. In the example here, a fabric-3 module (DS-13SLT-FAB3) with MDS 9513 provides 256 Gbps of fabric switching throughput per slot. The Cisco MDS 9700 has a 48-port 16-Gbps line card module that will be used with a MDS 9700 chassis. This module offers hardware-based slow drain, real-time power consumption reporting, and improved diagnostics capabilities.

1.     Cisco MDS 9513 Multilayer Director

Cisco MDS 9513 is a director-class multilayer series switch to help large-scale enterprises and service provides design and deploy large-scale data centers and scalable enterprise clouds to promote business transformation. The Cisco MDS 9513 is a 13-slot director in a 14 rack-unit (RU) form factor. Two slots are reserved for the redundant supervisor modules, and the 11 remaining slots are available for line card and service modules. The MDS 9513 can support a maximum of 528 ports per chassis or 1152 ports per rack with total throughput of 8.4 Tbps per chassis. It can provide 1/2/4/8 and 10 Gbps-Fibre Channel, 10-Gbps Fibre Channel over Ethernet (FCoE), and 1/2/4/8/10 Gbps FICON interfaces.

2.     Cisco MDS 9700 Series Multilayer Director

The Cisco MDS 9700 Series Multilayer Directors are the newest directors in the Cisco storage networking portfolio. The Cisco MDS 9710 Multilayer Director supports up to 384 line-rate 16-Gbps Fibre Channel or 10-Gbps Fibre Channel over Ethernet (FCoE) ports, and the Cisco MDS 9706 Multilayer Director supports up to 192 line-rate 16-Gbps Fibre Channel or 10-Gbps FCoE ports. They each provide up to 1.5 Tbps of per-slot throughput when populated with six fabric modules. Both directors also provide redundant supervisors, power supplies, and fan modules.

The Cisco MDS 9700 48-Port 16-Gbps Fibre Channel Switching Module delivers line-rate nonblocking 16-Gbps Fibre Channel performance to support scalability in the virtualized data centers. Line-rate 16-Gbps performance provides high-bandwidth throughput for consolidation of workloads from thousands of virtual machines while reducing the number of SAN components, providing scalability for future SAN growth at the same time. These line-card modules are hot-swappable and continue to provide all previous Cisco MDS features such as predictable performance, high availability, advanced traffic management capabilities, integrated VSANs, high-performance ISLs, fault detection, isolation of errored packets, and sophisticated diagnostics. This module offers new hardware-based slow drain, real-time power consumption reporting, and improved diagnostics capabilities.

With the Cisco MDS 9700 48-Port 10-Gbps FCoE Module, the Cisco MDS 9700 Series offers 10-Gbps FCoE capabilities, providing multiprotocol flexibility for SANs. This module extends the benefits of FCoE beyond the access layer to the data center core with a full line-rate FCoE solution for the Cisco MDS 9700 Series.

Customers can save money, simplify management, reduce power and cooling requirements, and improve flexibility by deploying FCoE, while protecting their Fibre Channel SAN investment with the Cisco MDS 9700 10-Gbps 48-Port FCoE Module. FCoE allows an evolutionary approach to I/O consolidation by preserving all Fibre Channel constructs. It maintains the latency, security, and traffic management attributes of Fibre Channel, as well as your investment in Fibre Channel tools, training, and SANs. FCoE also extends Fibre Channel SAN connectivity; now 100 percent of your network servers can be attached to the SAN.

3.     MDS 9513 24-Port 8-Gbps Fibre Channel Switching Module: DS-X9224-96K9

This module (Figure 6) delivers the performance needed for high-end storage systems and ISL. The front panel delivers 96 Gbps of Fibre Channel bandwidth with 24 ports, divided in the eight groups of three ports per group. The total allocated bandwidth per port group is 12.8 Gbps, one of which can be a dedicated port with a maximum of 8-Gbps bandwidth. The rest of the 4.8 Gbps can be shared among other ports in the same port group. This means the switch can provide eight of 8-Gbps interfaces or twenty-four 4-Gbps interfaces.

Figure 6.      Port Group Definitions on 24-port, 1/2/4/8 Gbps Fibre Channel Line Card Module

4.     MDS 9513 48-Port 8-Gbps Fibre Channel Switching Module: DS-X9248-96K9

This module (Figure 7) helps us get the port density and performance at the same time with a virtualization server environment. The front panel delivers 96 Gbps of Fibre Channel bandwidth with 48 ports, divided into eight groups of six ports per group. The total allocated bandwidth per port group is 12.8 Gbps, one of which can be a dedicated port with a maximum of 8-Gbps bandwidth. The rest of the 4.8 Gbps can be shared among the other five ports in the same port group. This means the switch can provide eight of 8-Gbps interfaces or twenty-four 4-Gbps interfaces.

Figure 7.      Port Group Definitions on 48-port, 1/2/4/8 Gbps Fibre Channel Line Card Module

5.     MDS 9513 32-Port 8-Gbps Advanced Fibre Channel Switching Module: DS-X9232-256K9

This module (Figure 8) is more suitable for high-end storage systems as well as for ISL connectivity. This module delivers 256 Gbps of front-panel bandwidth with a total of thirty-two 8-Gbps interface connectivity to the backend storage systems. There are eight port groups with four ports in each group. This module has no oversubscription ratio and all 32 ports can run at 8 Gbps full speed simultaneously.

Figure 8.      Port Group Definitions on 32-Port, 1/2/4/8/10 Gbps Fibre Channel Line Card Module

6.     MDS 9513 48-Port 8-Gbps Advanced Fibre Channel Switching Module: DS-X9248-256K9

With an 8-Gbps Fibre Channel bandwidth option, this module (Figure 9) is more suitable for port density at high-speed performance. The front panel delivers 256 Gbps of Fibre Channel bandwidth with 48 ports, divided into eight groups of six ports per group. The total allocated bandwidth per port group is 32 Gbps with a maximum speed of 8 Gbps per port.

Figure 9.      Port Group Definitions on 48-port, 1/2/4/8/10 Gbps Fibre Channel Line Card Module

 

7.     MDS 9700 48-Port 16-Gbps Fibre Channel Switching Module: DS-X9448-768K9

The Cisco MDS 9700 48-Port 16-Gbps Fibre Channel Switching Module (Figure 10) is the best in its class with a new chassis to deliver line-rate 16-Gbps Fibre Channel performance to support scalability in virtualized data centers. These line card modules are hot-swappable and compatible with 2/4/8/10 and 16-Gbps Fibre Channel interfaces.

Figure 10.    Port Group Definitions on 48-port, 2/4/8/10/16 Gbps Fibre Channel Line Card Module for the MDS 9700

SAN Topology Considerations

It is common practice in SAN environments to build two separate, redundant physical fabrics (fabric A and fabric B) to protect against the failure of a single physical fabric. This document shows a single fabric in the topology diagrams; however, customers would deploy two identical fabrics for redundancy. When designing for large networks, most environments will be one of two types of topologies within a physical fabric:

   Two-tier: Core-edge design

   Three-tier: Edge-core-edge design

Within the two-tier design, servers connect to the edge switches, and storage devices connect to one or more core switches (Figure 11). This design allows the core switch to provide storage services to one or more edge switches, thus servicing more servers in the fabric.

Figure 11.    Sample Core-Edge Design

In environments in which future growth of the network will likely cause the number of storage devices to exceed the number of ports available at the core switch, a three-tier design may be the best approach (Figure 12). This type of topology still uses a set of edge switches for server connectivity, but it adds a set of edge switches for storage devices. Both sets of edge switches connect to a core switch through ISLs.

Figure 12.    Sample Edge-Core-Edge Design

Sample Use Case Deployments

Figures 13 and 14 show two sample deployments of large-scale Cisco MDS fabrics with more than 2000 devices in a single fabric.

Figure 13.    U se Ca se 1 Topology with 8-Gbps Hosts Connected to 16-Gbps Storage Ports

This deployment allows scaling to nearly 2500 devices in a single fabric. The actual production environment has approximately 160 storage ports running at 16 Gbps of storage devices and roughly 2560 host ports. The environment requires a minimum of 8:1 oversubscription within the network, which requires each host edge switch to have a 2560-Gbps port channel using ISLs. Storage ports will not grow quite as rapidly and the core switch has room to grow to add more host edge switches and connect using ISLs. In this environment, the following was used in managing the network:

   Total of four VSANs created

     VSAN 1 used for staging new SAN devices

     VSAN 100 for a development SAN

     VSAN 200 for a lab SAN

     VSAN 300 for a production SAN

   Use of TACACS+ for authorization and authentication of MDS switches

   Use of Role-Based Access Control (RBAC) to create separate administrative roles for VSANs

   Use of device-aliases logical device identification

   Use of two member zones with device alias

Sample Deployment 2

Figure 14.    U se Ca se 2 Topology with 8-Gbps Hosts Connected to 16-Gbps Storage Ports

The deployment in Figure 14 scales to nearly 4000 devices with 240 storage ports running at 16 Gbps in a single fabric. The environment requires a minimum oversubscription ratio of 9:1 within the network, which requires each host edge switch to have a 3840 Gbps of ISL bandwidth. Again, storage ports will not grow quite as rapidly, and the core switch has room to grow to add more host edge switches. The storage edge and core Cisco MDS 9710 switches can have more 16-Gbps Fibre Channel line cards to meet the demands of future growth in storage or host ports.

In this environment, the following were used in managing the network:

   Total of five VSANs created

     VSAN 1 used for staging new devices

     Four VSANs based on business operations

   TACACS+ used for authorization and auditing of Cisco MDS switches

   Separate administrative roles created for VSANs

   Device alias created for environment

   Dynamic Port VSAN Membership (DPVM) feature enabled: This feature dynamically assigns VSAN membership to ports by assigning VSANs based on the device WWN. DPVM eliminates the need to reconfigure the port VSAN membership to maintain the fabric topology when a host or storage device connection is moved between two Cisco SAN switches or two ports within a switch. It retains the configured VSAN regardless of where a device is connected or moved

   Mixture of two- and three-member zones

Summary

With data centers continually growing, SAN administrators can now design networks that meet their current needs and can scale for demanding growth. Cisco MDS 9710 and 9706 Multilayer Director class switches, along with MDS 9513, provide embedded features to help SAN administrators in these tasks. SAN administrators deploying large Cisco SAN fabrics can use the design parameters and best practices discussed in this paper to design optimized and scalable SANs.

Table 3.       Cisco Product IDs of MDS 9513 and MDS 9700 Components

Part Number

Product Description

MDS 9700 Component

DS-C9710

MDS 9710 chassis, no power supplies, fans Included

DS-C9706

MDS 9706 chassis, no power supplies, fans Included

DS-X97-SF1-K9

MDS 9700 Series Supervisor-1

DS-X9710-FAB1

MDS 9710 crossbar switching fabric-1 module

DS-X9448-768K9

48-port 16-Gbps Fibre Channel switching module

Licensed Software

M97ENTK9

Enterprise package license for 1 MDS9700 switch

DCNM-SAN-M97-K9

DCNM for SAN license for MDS 9700

MDS 9513 Component

DS-C9513

Cisco MDS 9513 chassis

DS-X9530-SF2AK9

Cisco MDS 9500 supervisor/fabric-2A

DS-X9232-256K9

Cisco MDS 9000 Family 32-port 8-Gbps advanced Fibre Channel switching module

DS-X9248-256K9

Cisco MDS 9000 Family 48-port 8-Gbps advanced Fibre Channel switching module

DS-X9224-96K9

Cisco MDS 9000 Family 1/2/4/8-Gbps 24-port Fibre Channel module

DS-X9248-96K9

Cisco MDS 9000 Family 1/2/4/8-Gbps 48-port Fibre Channel module

Licensed Software

M9500ENT1K9

Cisco MDS 9500 Series enterprise package

DCNM-SAN-M95-K9

DCNM for SAN Advanced Edition for MDS 9500

DCNM-S-PAK-M95-K9

DCNM for SAN Advanced Edition for MDS 9500 configurable PAK (part of DCNM-SAN-PAK=)

L-DCNM-S-M95-K9

E-delivery DCNM for SAN Advanced Edition for MDS 9500 PAK (part of L-DCNM-S-PAK=)

For More Information

Cisco MDS 9700 Series documents:

   Cisco MDS 9700 Series Multilayer Directors

   Cisco MDS 9710 Multilayer Director Data Sheet

   Cisco MDS 9706 Multilayer Director Data Sheet

   Cisco MDS 9700 Series Supervisor-1 Module Data Sheet

   Cisco MDS 9700 48-Port 16-Gbps Fibre Channel Switching Module Data Sheet

   Compare Models: Learn about the similarities and differences of the models within this product series.

   Data Sheets and Literature (4)

   At-a-Glance Sheets

   Data Sheets

   Presentations

   White Papers

Cisco MDS 9500 Series documents:

   Cisco MDS 9513 Multilayer Director

   http://www.cisco.com/web/fw/i/s.gifCisco MDS 9000 Family 8-Gbps Advanced Fibre Channel Switching Modules (PDF - 440 KB)

   Cisco MDS 9000 Family 8-Gbps Fibre Channel Switching Modules (PDF - 210 KB)

   Cisco MDS 9500 and 9700 Series Multilayer Directors (PDF - 340 KB)

   Cisco MDS 9000 Family 8-Gbps Advanced Fibre Channel Switching Modules

   Cisco MDS 9000 Family 8-Gbps Fibre Channel Switching Modules

   Cisco MDS 9513 Multilayer Director Data Sheet

   Cisco MDS 9000 Family Investment Protection