Collaboration application infrastructure requirements for hardware and VMware has moved.

Click here for the new page.

All prior content has been deprecated and does not reflect current support policy.
  


Introduction

(top)

Note: Unless specifically indicated otherwise, references to UCS B-Series includes both UCS Mini and regular UCS 5100 Series Blade Server Chassis. UCS Mini is allowed as long as the UCS B-Series and C-Series servers in UCS Mini conform to Collaboration support policies.

Note: Not all UC apps support all hardware options. For supported options, check the application's page in At a Glance.

This web page describes supported compute, storage and network hardware for Virtualization of Cisco Unified Communications, including UC on UCS (Cisco Unified Communications on Cisco Unified Computing System). See the "How to..." links for guidelines on how to design, quote and procure a virtualized UC solution that follows Cisco's support policy.

Cisco uses three different support models:

  • UC on UCS Tested Reference Configuration (TRC), some of which are available as packaged collaboration solutions like Cisco Business Edition 6000 or Cisco Business Edition 7000.
    • "TRC" used by itself means "UC on UCS Tested Reference Configuration (TRC)".
  • UC on UCS Specs-based
    • "UC on UCS" used by itself refers to both UC on UCS TRC and UC on UCS Specs-based.
  • Third-party Server Specs-based
    • "Specs-based" used by itself refers to the common rules of UC on UCS Specs-based and Third-party Server Specs-based.
Below is a comparison of the hardware support options. Note that the following are identical regardless of the support model chosen:
  • Virtual machine (OVA) definitions
  • VMware product, version and feature support
  • VMware configuration requirements for UC
  • Application/VM Co-residency policy (specifically regarding application mix, 3rd-party support, no reservations / oversubscription, virtual/physical sizing rules and max VM count per server).

UC on UCS TRC UC on UCS Specs-based Third-Party Server Specs-based Other hardware
Basic Approach Configuration-based Rules-based Rules-based Not supported- does not satisfy this page's policy
Allowed for which UC apps? See At a Glance See At a Glance See At a Glance Not Supported
UC-required Virtualization Software
  • Click here for general requirements.
  • VMware vCenter is optional.
  • VMware vSphere is mandatory.
    • Click here for supported versions, editions, features, capacities and purchase options.
  • Click here for general requirements.
  • VMware vCenter is mandatory. Also mandatory to capture Statistics Level 4 for maximum duration at each level.
  • One of the following is mandatory:
    • Cisco UC Virtualization Foundation
    • VMware vSphere
    • Click here for supported versions, editions, features, capacities and purchase options.
  • Click here for general requirements.
  • VMware vCenter is mandatory. Also mandatory to capture Statistics Level 4 for maximum duration at each level.
  • VMware vSphere is mandatory:
    • Click here for supported versions, editions, features, capacities and purchase options.
  • N/A - not supported
    Allowed Servers Select Cisco UCS listed in Table 1. Must follow all TRC rules in this policy. Any Cisco UCS that satisfies this page's policy Any 3rd-party server model that satisfies this page's policy None
    Required Level of Virtualization/Server Experience Low / Medium High High N/A
    Cisco-tested? Joint validation of apps and server hardware by UC and UCS teams Generic server hardware validation by UCS team. Not jointly validated with UC apps by Cisco. No server hardware validation by Cisco. Not jointly validated with UC apps by Cisco. No Cisco testing (unspported hardware)
    Server Model, CPU and Component Choices Less (customer accepts tradeoff of less hardware flexibility for more UC predictability). More (customer assumes more test/design ownership to get more hardware flexibility) More (customer assumes more test/design ownership to get more hardware flexibility) None (unsupported hardware)
    Does Cisco TAC support UC apps? Yes, when all TRC rules in this policy are followed.
    UC apps on C-Series DAS-only TRC: Supported with Guaranteed performance
    UC apps on C-Series FC SAN TRC or B-Series FC SAN TRC: Supported with Guaranteed performance provided all shared storage requirements in this policy are met.
    Yes, when all Specs-based rules in this policy are followed. Supported with performance Guidance only Yes, when all Specs-based rules in this policy are followed. Supported with performance Guidance only UC apps not supported when deployed on unsupported hardware.
    Does Cisco TAC suppport the server? Yes. If used with UC apps, then all TRC rules in this policy must be followed. Yes. If used with UC apps, then all UC on UCS Specs-based rules in this policy must be followed. No. Cisco TAC supports products purchased from Cisco with a valid, paid-up maintenance contract. No. Cisco TAC supports products purchased from Cisco with a valid, paid-up maintenance contract. Also note UC apps also not supported when deployed on unsupported hardware.
    Who designs/determines the server's BOM? Customer wants Cisco to own Customer wants to own, with assistance from Cisco Customer wants to own N/A

    For more details on Cisco UCS servers in general, see the following:

  • Cisco UCS B-Series Servers Documentation Roadmap
  • Cisco UCS C-Series Servers Documentation Roadmap
  • Cisco UCS C-Series Integrated Management Controller Documentation
  • Cisco UCS Manager Documentation
  • Cisco UCS home page
  • Cisco UC on UCS solution home page

  • Cisco Business Edition Appliance and UC on UCS Tested Reference Configurations (TRCs) Servers

    (top)

    Note: Definition of a Tested Reference Configuration(TRC):
    • Includes specification of server model and loal components (CPU, RAM, adapters, local storage) at the orderable part number level.
      • UCS B-Series servers, do NOT include specification of blade server chassis or switching.
      • UCS E-Series servers, do NOT include specification of Cisco Integrated Server Router (ISR) chassis unless purchased as part of a Cisco Business Edition 6000S appliance.
    • For TRCs using DAS storage, includes specification of required RAID configuration (e.g. RAID5, RAID10, etc) - including battery backup cache or SuperCap.
    • Includes guidance on hardware installation and basic setup (full Cisco UCS installation documentation is on http://www.cisco.com/go/ucs and is the authoritative source).
      • Does not include specification of virtual-to physical network interface mapping as this is design-dependent.
      • Does not include specification of adapter settings (such as Cisco VIC, 3rd-party CNA / NIC / HBA) as this is design-dependent.
      • Does not include configuration settings or step by step procedures for hardware BIOS, firmware, drivers, RAID setup outside the guidance in the Cisco Collaboration on Virtual Servers document. See the Cisco UCS documentation for details.
    • Does not include design, installation and configuration of external hardware such as:
      • Network routing and switching (e.g. routers, gateways, MCUs, ethernet/FC/FCoE switches, Cisco Catalyst/Nexus/MDS, etc.)
      • QoS configuration of route/switch network devices.
      • Cisco UCS B-Series chassis and switching components (e.g. Cisco UCS 6100/6200, Cisco UCS 2100/220, Cisco UCS 5100)
      • Storage arrays (such as those from EMC, NetApp or other vendors)
    • Does not include configuration settings, patch recommendations or step by step procedures for required VMware virtualization software.

    Click here for basic guidance on TRC hardware setup.

    If you have a UCS C-Series server and need assistance identifying its hardware specs for comparison with the configurations below, see the Cisco UCS C-Series Servers Integrated Management Controller GUI Configuration Guide.

    UConUCS Tested Reference Configurations


    End of Sale Hardware

    Click here to download End of Sale Servers

     
    BE6000S and Extra-Small UConUCS TRCs


    Hardware Configuration Name and
    Part Numbers / SKUs / BOM
    Form Factor and Physical Specs Capacity Available to VMs (using required Sizing Rules)
    Extra-Small
     UConUCS TRC 
    (XS TRC)
    UCS E160S M3 TRC#1
    (blade-only)
    Click here for BOM
    TRC =

  • Cisco ISR Single-wide Blade Server
  • Single D-1528 (6-core / 1.9 GHz)
  • 32 GB RAM
  • VMware + UC apps boot from DAS (2x 900GB SAS, RAID1)
  • Onboard internal and external ethernet ports

  • 6 total physical cores (note: Intel Xeon Processor-D only supported on this TRC and only for select deployment models)
  • 28 GB physical RAM
  • ~900 GB
  • External port for CIMC, internal and external 1Gb ports used for LAN access.
  • Click here to download specifications and BOM for End of Sale Servers.


    BE6000M and Small UConUCS TRCs


    Hardware Configuration Name and
    Part Numbers / SKUs / BOM
    Form Factor and Physical Specs Capacity Available to VMs (using required Sizing Rules)
    Cisco Business Edition 6000M

    or

    Small
    UConUCS TRC
    (S TRC)

    BE6000M (M5)

    UCS C220 M5SX TRC#1
    Click here for BOM


  • 1RU Rack-mount Server
  • Single Xeon 4114
  • 48 GB RAM
  • VMware + UC apps boot from DAS (single 6-disks-RAID5 of 300GB 12G SAS 10K SFF)
  • Ethernet ports on motherboard

  • 10 total physical cores ("Restricted UC Performance" CPU type)
  • 40 GB physical RAM
  • ~1TB
  • 1x 1GbE (CIMC), 2x 10GbE (LOM)
  • Click here to download specifications and BOM for End of Sale Servers.


    BE6000H and Small Plus UConUCS TRCs


    Hardware Configuration Name and
    Part Numbers / SKUs / BOM
    Form Factor and Physical Specs Capacity Available to VMs (using required Sizing Rules)
    Cisco Business Edition 6000H

    or

    Small Plus
    UConUCS TRC
    (S+ TRC)

    BE6000H (M5)

    UCS C220 M5SX TRC#2
    Click here for BOM


  • 1RU Rack-mount Server
  • Dual Xeon 4114
  • 64 GB RAM
  • VMware + UC apps boot from DAS (single 8-disks-RAID5 of 300GB 12G SAS 10K SFF)
  • Ethernet ports on motherboard and single NIC

  • 20 total physical cores ("Restricted UC Performance" CPU type)
  • 56 GB physical RAM
  • ~1TB
  • 1x 1GbE (CIMC), 2x 10GbE and 4x 1GbE (LOM)
  • Click here to download specifications and BOM for End of Sale Servers.


    BE7000M and Medium UConUCS TRCs


    Hardware Configuration Name and
    Part Numbers / SKUs / BOM
    Form Factor and Physical Specs Capacity Available to VMs (using required Sizing Rules)
    Cisco Business Edition 7000M

    or

    Medium
    UConUCS TRC
    (M TRC)

    BE7000M (M5)

    UCS C240 M5SX TRC#1
    Click here for BOM


  • 2RU Rack-mount Server
  • Single Xeon 6132
  • 96 GB RAM
  • VMware + UC apps boot from DAS (dual 7-disks-RAID5 of 300GB 12G SAS 10K SFF)
  • Ethernet ports on motherboard and dual NIC

  • 14 total physical cores ("Full UC Performance" CPU type)
  • 88 GB physical RAM
  • Two volumes, each with ~1TB usable disk space
  • 1x 1GbE (CIMC), 2x 10GbE and 8x 1GbE (LOM)
  • Click here to download specifications and BOM for End of Sale Servers.


    BE7000H and Large UConUCS TRCs


    Hardware Configuration Name and
    Part Numbers / SKUs / BOM
    Form Factor and Physical Specs Capacity Available to VMs (using required Sizing Rules)
    Cisco Business Edition 7000H

    or

    Large
    UConUCS TRC
    (L TRC)
    HyperFlex HX220c M5SX TRC#1

    Click here for BOM

  • 2RU Rack-mount HyperFlex node
  • Dual Xeon 6142 (16-core, 2.6 GHz)
  • 256 GB RAM
  • VMware + UC apps boot from HX Data Platform
  • Ethernet ports on motherboard + Cisco VIC

  • 24 total physical cores ("Full UC Performance" CPU type) after Storage Controller overhead
  • 200 GB RAM (i.e. 256GB less 48GB for HX Data Platform and 8GB for ESXi)
  • Usable storage space ~5.76 TB (3-node HX cluster with RF=2)
  • 1x 1GbE (CIMC), 2x 10GbE (LOM), 2x 40GbE (VIC)
  • BE7000H (M5)

    UCS C240 M5SX TRC#2
    Click here for BOM


  • 2RU Rack-mount Server
  • Dual Xeon 6132
  • 192 GB RAM
  • VMware + UC apps boot from DAS (quad 6-disks-RAID5 of 300GB 12G SAS 10K SFF)
  • Ethernet ports on motherboard and dual NIC

  • 28 total physical cores ("Full UC Performance" CPU type)
  • 184 GB physical RAM
  • Four volumes, each with ~1TB usable disk space
  • 1x 1GbE (CIMC), 2x 10GbE and 8x 1GbE (NIC)
  • Click here to download specifications and BOM for End of Sale Servers.



    Virtualization Software Requirements

    (top)

    VMware vCenter is required for UC on UCS Specs-based, Specs-based 3rd-party infrastructure and HX TRC (because HX requires vCenter).
    VMware vCenter is optional for other UC on UCS TRC deployments.

    Just like some of the UC applications, vCenter can be configured to save more performance data. The more historical data saved, the bigger disk space needed by the database used by vCenter. Note, this is one of the main areas where you need vCenter rather than going directly to the ESXi host for performance data. vCenter can save historical data that the ESXi host does not keep.

    The configurations to change the amount historical data saved by vCenter is located in the vSphere client under Administration > Server Settings. For each interval duration and save time the statistic level can be set. The statistics levels range from 1 to 4 with level 4 containing the most data. View the data size estimates to ensure there is enough space to keep all statistics.

    For a UC on UCS Specs-based or Specs-based 3rd-party infrastructure, Statistics Level 4 is required on all statistics. Configuring VMware vCenter to capture detailed logs, as shown in Figure 1 below, is strongly recommended. If not configured by default, Cisco TAC may request enabling these settings in order to troubleshoot problems.

    Figure 1

    VMware virtualization software is required for Cisco TAC support.

  • See the Introduction for basic virtualization software requirements, including what is optional and what is mandatory.
  • For Cisco UCS, no UC applications run or install directly on the server hardware; all applications run only as virtual machines. Cisco UC does not support a physical, bare-metal, or nonvirtualized installation on Cisco UCS server hardware.

    All UC virtualization deployments must align with the VMware Hardware Compatibility List (HCL).

    All UC virtualization deployments must follow UC rules for supported VMware products, versions, editions and features as described here.

    Note: For UC on UCS Specs-based and Third-party Server Specs-based, use of VMware vCenter is mandatory, and Statistics Level 4 logging is mandatory. See above for how to configure VMware vCenter to capture these logs. If not configured by default, Cisco TAC may request enabling these settings in order to troubleshoot problems.



  • "Can I use this server?"

    (top)

    UC virtualization hardware support is most dependent on the Intel CPU model and the VMware Hardware Compatibility List (HCL).

    The server model only matters in the context of:

  • whether or not it is on the VMware HCL
  • what Intel CPU models does it carry (and are those CPU models allowed for UC virtualization)
  • can its hardware component options satisfy all other requirements of this policy
  • For additional considerations, see TAC TechNote 115955.

    Note:

    • UC does not support every CPU model
    • A given server model may not carry every (or any) CPU model that UC supports.
    • Therefore your server model choices may be artificially limited by which CPUs the server models carry.


    UC on UCS TRC UC on UCS Specs-based Third-Party Server Specs-based Not Supported
    Allowed Servers:
  • Vendors
  • Models / Generations
  • Form Factors
  • Only Cisco Unified Computing System B-Series Blade Servers and C-Series Rack-mount Servers listed in Table 1 are supported.

    HyperFlex M5 TRC permits chassis substitution of 1RU, 2RU, hybrid storage and all-flash models. TRC may also be used with HyperFlex Edge with HXDP 4.0+ and 10/25/40 GE network connections (1 GE network connections not supported).


    Any Cisco Unified Computing System server is supported as long as:

  • it is on the VMware HCL for the version of VMware vSphere ESXi required by UC.

  • it carries a CPU model supported by UC (described later in this policy).

  • it satisfies all other requirements of this policy

  • Otherwise, any Cisco UCS model, generation, form factor (rack, blade) may be used.

  • Any 3rd-party server model is supported as long as:

  • it is on the VMware HCL for the version of VMware vSphere ESXi required by UC.

  • it carries a CPU model supported by UC (described later in this policy).

  • it satisfies all other requirements of this policy

  • Otherwise, any 3rd-party vendor, model, generation, form factor (rack, blade) may be used.
  • The following are NOT supported:

  • Cisco or 3rd-party server models that do not satisfy the rules of this policy.

  • Cisco 7800 Series Media Convergence Servers (MCS 7800) regardless of CPU model

  • Cisco UCS Express (SRE-V 9xx on ISR router hardware)

  • Cisco UCS E-Series Blade Servers (E14x/16x on ISR router hardware) with CPUs that do not meet Processor policy requirements.

  • For additional considerations, please see TAC TechNote 115955.
  • Server or Component "Embedded Software"
  • BIOS
  • Firmware
  • Drivers

  • Always enable "Intel Virtualization Technology" in the hardware BIOS if it is present to enable.
  • Always enable hyperthreading in the hardware BIOS if it is present to enable, but always deploy 1 vCPU per physical CPU core (regardless of whether hyperthreading is enabled or not). For more details, see Collaboration Virtualization Sizing.
  • Cisco apps will specify supported versions of VMware vSphere ESXi. Follow server vendor guidelines for versions, levels and settings of BIOS, firmware and drivers required for the version of VMware vSphere ESXi in use.
  • Mechanical and Environmental
    Note: Energy-saving features that cause reduction in CPU performance or real-time relocation/powering-down of virtual machines (such as CPU throttling or VMware Dynamic Power Management) are not supported.

    Otherwise, there are no UC-specific requirements for form factor, rack mounting hardware, cable management hardware, power supplies, fans or cooling systems. Follow server vendor guidelines for these components.

    If you use a Cisco UCS bundle SKU, note that the rail kit, cable management and power supply options may not match what is available with non-bundled Cisco UCS.

    Redundant power supplies are highly recommended, particularly for UC on UCS.

    For Cisco UCS, it is strongly recommended to use the Cisco default rail kit, unless you have different rack types such as telco racks or racks proprietary to another server vendor. Cisco does not sell any other types of rack-mounting hardware; you must purchase such hardware from a third party.


  • Processors / CPUs

    (top)

    Allowed vs. Supported CPU policies are different for UC on UCS Tested Reference Configuration, UC on UCS Specs-based and 3rd-party Server Specs-based. Required physical core speeds (base frequencies) are different for applications using 1vcpu:1pcore sizing approach (see Collaboration Virtualization Sizing) vs. applications that support CPU Reservations, such as UCM/IMP (see UCM and IMP Caveated Support for VMware CPU Reservations and Distributed Resource Scheduler). The rest of this section covers physical CPU requirements for 1vcpu:1pcore sizing approach.

    Read the notes and BOTH CPU tables below to understand these support policy differences before locking in CPU parts choices.

    Note:

    • If an application has application-specific rules on its page on www.cisco.com/go/uc-virtualized, that rule takes precedence over this policy. Check EVERY app that will share a physical server before finalizing CPU choice.
    • CPU architectures are NOT allowed for Specs-based until explicitly listed below. Until listed below, they are not allowed, even if believed to be "better" since "newer".
    • Some CPU architectures, or CPU models within an architecture/family, are not a good fit for Collaboration requirements and will NEVER be allowed or listed below. Do not assume that all CPU architectures, or all models with an architecture, will be allowed.
    • Collaboration application support for new CPU architectures/models may lag the release date from Intel and/or server vendors.
    • Collaboration applications require minimum physical core speeds on allowed CPU architectures (normal not turbo). Higher capacity VM configurations will usually require higher minimum physical core speeds. Expect these speeds to be higher than what is required for many traditional business applications. Speeds need to be what is shown at ark.intel.com in the CPU model's description, or for Cisco UCS in the description of the CPU SKU. This is normal speed not "turbo". This speed may not be what is actually returned by running software such as OS CLIs.
    • Slower-speed CPUs categorized as "Restricted UC Performance" are only allowed for certain VM configurations of certain Collaboration apps, and they are usually lower capacity points of those applications. Do not assume large capacity VMs are allowed on slow-speed CPUs.
    • Physical CPU choices for UC on UCS Tested Reference Configurations are stricter. They are a subset of what is allowed for Specs-based. See CPU Table 2.
    • Physical CPUs shipping as part of a Packaged Collaboration Solution like Cisco Business Edition may not be changed.


    CPU Table 1 - Allowed Shipping Specs-based CPUs

    Click here for Allowed Older (end of sale) CPUs

    CPU Architecture & Models Full UC Performance CPUs Restricted UC Performance CPUs
    Intel Xeon 4200, 5200
    (Intel Cascade Lake)

    Intel Xeon 4100, 5100
    (Intel Skylake)


    Not Supported
    Any model with physical core speed 2.00 GHz to 2.49 GHz (base frequency not turbo).

    Some applications, or VMs of an application, may not support or require higher speed; click here and check each application's page. Must also follow all rules in Table 2 below.

    Intel Xeon 6200, 8200, 9200
    (Intel Cascade Lake)

    Intel Xeon 6100, 8100
    (Intel Skylake)


    Any model with physical core speed 2.50 GHz or higher (base frequency not turbo).

    Some applications, or VMs of an application, may not support or require higher speed; click here and check each application's page. Must also follow all rules in Table 2 below.


    CPU Table 2 - Additional TRC and Specs-based rules for CPU


    UC on UCS TRC UC on UCS Specs-based Third-party Server Specs-based Not supported
    Physical Sockets / CPU Quantity Must exactly match what is listed in TRC Table 1. Customer choice (subject to what server model allows). The following CPUs are NOT supported for UC:

  • Intel CPUs that are IN one of the supported architectures/families, but do NOT meet minimum physical core speeds, are not supported for UC.

  • Unlisted Intel CPU architectures/families (such as Intel Xeon 6500, E5-16xx, Core-i7, etc.) are NOT supported for UC. An Intel CPU architecture is not supported for UC unless listed in CPU Table 1 or in the Allowed Older (end of sale) CPUs. Note that end of sale CPU hardware may be carried by a server that is no longer supported by the server vendor.

  • Other CPU vendors such as AMD are not supported for UC.

    Cisco TAC is not obligated to troubleshoot UC app issues when deployed on unsupported hardware.

  • Physical CPU Vendor and CPU model Must either exactly match the TRC BOM's CPU model in TRC Table 1 or use a CPU model that satisfies the following requirements:

    • Same physical CPU core count as the TRC BOM's CPU model in Table 1.

    • Same CPU architecture as the TRC BOM's CPU model in Table 1.

    • Physical CPU core speed same or higher than that of the TRC BOM's CPU model in Table 1.

    E.g. if the TRC BOM was tested with 2-socket Intel Xeon E5-2680 v1 (SandyBridge, 8-core, 2.7 GHz), then substition of a pair of any other E5-2600 v1 with 8 cores and speed 2.7 Ghz or higher would be allowed and the server is still a TRC. Subsitition with E5-2600 v2 or v3 would NOT be a TRC due to different architecture, and would make the server UC on UCS Specs-based.
    See app links in table on http://www.cisco.com/go/virtualized-collaboration or Supported Applications for which CPU Types are allowed for a given VM configuration of a UC app. E.g. the Unified Communications Manager 7500 user VM configuration is only allowed on a Full UC Performance CPU.
    Total Physical CPU Cores Total available is fixed based on the CPU models in TRC Table 1.

    Total available depends on which CPU model customer chooses vs. the physical server's socket count and the CPU model selected.

  • Total required is based on Collaboration Virtualization Sizing.
  • You can use https://www.cisco.com/go/vmpt or http://www.cisco.com/go/quotecollab for help with VM placement.

    Cisco TAC is not obligated to troubleshoot UC app issues in deployments with insufficient physical processor cores or speed.


  • Memory / RAM

    (top)

    Note: Virtualization software licenses such as Cisco UC Virtualization Foundation or VMware vSphere limit the amount of total vRAM that can be used (and therefore the amount of physical RAM that can be used for UC VMs, due to UC sizing rules). See Unified Communications VMware Requirements for these limits. In general larger deployments, or deployments with high VM counts, will require very high vRAM totals and will therefore need to use VMware vSphere instead of Cisco UC Virtualization Foundation. If using high-memory-capacity servers, use VMware vSphere instead to ensure use of all physical memory.


    UC on UCS TRC Specs-based (UCS or 3rd-party Server)
    Physical RAM Total available is listed in Table 1. Additional memory may be added. Total available depends on the server chosen.
    Total required is dependent on the virtual machine quantity/size mix deployed on the hardware:

  • 2-4 GB required for virtualization software, depending on version. See Memory/RAM section of General sizing rules.

  • plus the sum of UC virtual machines' vRAM.

  • while following co-residency support policy rules. Per these rules, recall that UC does not support physical memory oversubscription (1 GB of vRAM must equal 1 GB of physical RAM). Cisco TAC is not obligated to troubleshoot UC app issues if the deployment has insufficient physical RAM.

  • Memory Module/DIMM
    Speed and Population
    For what was tested in a TRC, see Table 1.

    Follow server vendor guidelines for optimum memory population for the memory capacity required by UC.

  • For Cisco UCS, use the Specs Sheets at UCS Quick Catalog. E.g. for a UCS B200 M3 with 96GB total RAM, optimal is 4x8GB DIMM + 4x4GB DIMM. Using 6x16GB DIMM is not optimal.

    Otherwise, there are no UC-specific requirements (primarily because UC does not support memory oversubscription).

  • UC allows any DIMM speed (e.g. 1333 MHz, 1600 MHz, etc.).

  • UC allows any memory hardware module size, density and quantity (including changing the DIMM population on a Tested Reference Configuration) as long as UC-required RAM capacity is met, and the server vendor supports the intended memory configuration.


  • Storage

    (top)

    To be supported for UC, all storage systems - whether TRC or specs-based - must meet the following requirements:

    • Compatible with the VMware HCL and compatible with the supported server model used
    • Host-level kernel disk command latency < 4ms (no spikes above) and physical device command latency < 20 ms (no spikes above). For NFS NAS, guest latency < 24 ms (no spikes above)
    • Published vDisk capacity requirements of UC VMs.
      • See VM specs of application pages in At a Glance table on http://www.cisco.com/go/virtualized-collaboration
      • For DAS-only TRCs (including Cisco Business Edition 6000, Thin Provisioning (either from VMware or from storage array) is not supported. Thick provisioning must be used.
      • For diskless TRCs and any Specs-based server, thin provisioning (either from VMware or from storage array) is allowed with the caveat that disk space must be available to the VM as needed. Running out of disk space due to thin provisioning will crash the application and corrupt the virtual disk (which may also prevent restore from backup on the virtual disk).
    • Published IOPS capacity requirements of UC VMs (including excess capacity provisioned to handle IOPS spikes such as during Cisco Unified Communications Manager upgrades).
    • Other storage system design requirements (click here).

    Note: UC on UCS TRCs using only DAS storage (such as C220 M3S TRC#1) have been pre-designed and tested to meet the above requirements for any UC with UC co-residency scenario that will fit on the TRC. Detailed capacity planning is not required unless deploying

    • non-UC/3rd-party apps
    • VM OVA templates created later than the TRC
    • VM OVA templates with very large vDisks (300GB+).
    Note: All of the above requirements must be met for Cisco UC to function properly. Except for UC on UCS TRCs using DAS only, it is the customer's responsibility to design a storage system that meets the above requirements. Cisco TAC is not obligated to troubleshoot UC app issues when customer-provided storage is insufficient, overloaded or otherwise not meeting the above requirements.

    See below for supported storage hardware options.


    UC on UCS TRC Specs-based (UCS or 3rd-party Server)
    Supported Storage Options TRCs are only defined for:

  • DAS-only with UC-specified configuration (C260 M2, C240 M3S, C220 M3S, C210 M1/M2, C200 M2)

  • FC SAN with VMware local boot from DAS (B200 M1/M2, C210 M1/M2)

  • Diskless / boot from FC SAN (B440 M2, B230 M2, B200 M3, C210 M2)

  • DAS with customer-defined configuration (including local disks, external SAS, etc.)

  • FC, iSCSI, FCoE or Infiniband SAN
  • Diskless / boot from SAN via above transport options (only supported with VMware vSphere ESXi 4.1+ and compatible UC app versions).
  • NFS NAS
  • DAS Support Details


    UC on UCS TRC Specs-based (UCS or 3rd-party Server)
    Disk Size and Speed B-Series TRC

    may use the disk size/speed listed in Table 1 BOMs, or any other orderable size/speed for the blade server (since local disks are only used to boot VMware).

    C-Series TRC

    Both must be same or higher than specs listed in Table 1.

    E.g. for a TRC tested with 300 GB 10K rpm disks, then:

  • 300GB 15K rpm is supported (faster)

  • 146GB 10K rpm not supported (too small)

  • 7.2K rpm disk of any size not supported (too slow)

    HyperFlex M5 TRC

    For Capacity Disks, may substitute other disk technologies (SFF/LFF SAS, SED, SSD) on HyperFlex HCL that provide same or higher space and speed. For Boot / System / Cache Disks, may substitute other disks on HyperFlex HCL.

  • DAS is supported with customer-determined disk size, speed, quantity, technology, form factor and RAID configuration as long as:

  • compatible with the VMware HCL and compatible with the server model used

  • all UC latency, performance and capacity requirements are met. To ensure optimum UC app performance, be sure to use Battery Backup cache or SuperCap on RAID controllers for DAS.
  • TRC BOMs are updated as orderable disk drive options change. E.g. UCS C210 M2 TRC#1 was tested with 146GB 15K rpm disks, but due to 146GB disk EOL, the BOM now specifies 300GB 15K rpm disks (still supported as TRC since both size and speed are "same or higher" than what was tested).
    Disk Quantity, Technology, Form Factor C-Series TRCs

    Must exactly match what is listed in Table 1. E.g. if the TRC was tested with ten 2.5" SAS drives, then that must be used regardless of disk size or speed.

    HyperFlex M5 TRC

    See previous row.

    RAID Configuration C-Series TRCs

    RAID configuration, including physical-to-logical volume mapping, must exactly match Table 1 and the RAID instructions in the document Cisco Collaboration on Virtual Servers here.

    N/A for HyperFlex M5 TRC (does not use RAID).

    SAN / NAS Support Details

    • Applies to any TRC or Specs-based configuration connecting to FC, iSCSI, FCoE or NFS storage.
    • No UC requirement to dedicate arrays or storage groups to UC (vs. non-UC), or to one UC app vs. other UC apps.
    • The storage solution must be compatible with the server model used. E.g. for Cisco Unified Computing System: Cisco UCS Interoperability
    • The storage solution must be compatible with the VMware HCL. For example, refer to the "SAN/Storage" tab at http://www.vmware.com/resources/compatibility/search.php?sourceid=ie7&rls=com.microsoft:en-us:IE-SearchBox&ie=&oe=
    • No UC requirements on disk size, speed, technology (SAS, SATA, FC disk), form factor or RAID configuration as long as requirements for compatibility, latency, performance and capacity are met. "Tier 1 Storage" is generally recommended for UC deployments. See the UC Virtualization Storage System Design Requirements for an illustration of a best practices storage array configuration for UC.
    • There is no UC-specific requirement for NFS version. Use what VMware and the server vendor recommend for the vSphere ESXi version required by UC.
    • Use of storage network and array "features" (such as thin provisioning or EMC Powerpath) is allowed.
    • Otherwise any shared storage configuration is allowed as long as UC requirements for VMware HCL, server compatibility, latency, capacity and performance are met.

    Removable Media


    UC on UCS TRC Specs-based (UCS or 3rd-party Server)
    Boot from SD cards, USB flash or other
  • Not allowed or supported by Cisco for apps or VMware vSphere ESXi. TRCs are only validated as either boot from DAS or diskless boot from FC SAN depending on Table 1.

  • Note all current TRCs are either diskless blades or C-Series DAS/HDD. SD cards in C-Series TRCs are used for convenience to get the UCS utilities (like SCU and HUU), in lieu of a DVD drive.

  • HyperFlex TRCs must boot from Cisco FlexFlash and are not user-changeable.


  • Not allowed or supported by Cisco for apps. Must boot from DAS, SAN or NAS per Specs-based Storage requirements.

  • Allowed for VMware vSphere ESXi as long as supported by server vendor. Cisco Collaboration apps do not test, validate or troubleshoot these boot methods - work with server vendor if boot issues.

  • Otherwise, there are no UC-specific requirements or restrictions. The different methods of installing UC apps into VMs can leverage the following distribution types of Cisco UC software:

    • Physical delivery of UC apps via ISO image file on DVD.
    • Cisco eDelivery of UC apps via email with link to ISO image file download.


    IO Adapters, Controllers and Devices for LAN Access and Storage Access

    (top)

    All adapters used (NIC, HBA, CNA, VIC, etc.) must be on the VMware Hardware Compatibility List for the version of vSphere ESXi required by UC.


    UC on UCS TRC Specs-based (UCS or 3rd-party Server)
    Physical Adapter Hardware (NIC, HBA, VIC, CNA)
    For UCS B-Series TRC
    • Adapter vendor/model/technology - either match the adapters listed in Table 1 BOMs or substitute with any other supported adapter for the blade server model (which adapter "should" be used is dependent on deployment, design and UC apps).

    • Adapter quantity may be one or two.

    For UCS C-Series TRC:

    • must exactly match adapter vendor/model/technology (e.g. Intel i350 for 1GbE or QLogic QLE2462 for FC) listed in Table 1 BOMs.

    • Allowed NIC quantity must be same or higher than what is listed in Table 1 BOMs.

    • Allowed HBA/VIC/CNA quantity must exactly match Table 1 BOMs.

    • Any other changes are not allowed for a UC on UCS TRC, but are allowed for UC on UCS Specs-based.

    For HyperFlex M5 TRC:

    • For VIC, may substitute quantity 1+ of any 10/25/40GE or faster VIC on HyperFlex HCL. If TRC being used with HyperFlex Edge, must use 10/25/40GE interfaces (1GE interface options not supported).
    • For NIC, may substitute quantity 1+ of any NIC 1GE or faster on HyperFlex HCL.

  • Only the following I/O Devices are supported:
    • HBA for storage access
      • Fibre Channel - 2Gbps or faster
      • InfiniBand
    • NIC for LAN and/or shared storage access
      • Ethernet - 1Gbps or faste. Includes NFS and iSCSI for storage access.
    • Cisco VIC or 3rd-party Converged Network Adapter for LAN and/or storage access
      • FCoE - 10Gbps or faster
    • RAID Controllers for DAS storage access
      • SAS
      • SAS SATA Combo
      • SAS-RAID
      • SAS/SATA-RAID
      • SATA
  • The customer is also responsible for configuring redundant devices on the server (e.g. redundant NIC, HBA, VIA or CNA adapters).
  • There are no UC restrictions on hardware vendors for I/O Devices other than that VMware HCL and the server vendor/model must be compatible with them and support them.
  • IO Capacity and Performance
    In most cases detailed capacity planning is not required for LAN IO or storage access IO. TRC adapter choices have been made to accommodate the IO of all UC on UCS app co-residency scenarios that will fit on the TRC. For guidance on active vs. standby network ports, see the Cisco UC Design Guide and QoS Design Considerations for Virtual UC with UCS

    It is the customer's responsibility to ensure the external LAN and storage access meet UC app design requirements.


  • LAN access adapters must be able to accommodate the LAN usage of UC VMs (described in UC app design guides).
  • Storage access adapters must be able to accommodate the storage IOPS (described in the Storage section of this policy). Cisco TAC is not obligated troubleshoot UC apps issues in a deployment with insufficient or overloaded I/O devices.



  • Hardware Bills of Material (BOMs)

    (top)

    Note: Do not assume that other UCS bundle SKUs on Cisco Commerce Build and Price can be used with UC on UCS. Before quoting one of these bundles, identify the BOM that it ships and see below:
    • If the bundle meets TRC requirements on this page, it may be quoted for UC on UCS TRC.
    • If the bundle does NOT meet TRC requirements but DOES meet Specs-based requirements, then it may be quoted for UC on UCS Specs-based only.
    • If the bundle does NOT meet TRC requirements and also does NOT meet Specs-based requirements, then it may NOT be quoted for UC on UCS at all without modification.

    End of Sale Hardware
    Click here to download EOS Bills of Material

    HX220c M5SX TRC#1
    Large HyperFlex node TRC. See details

    Below is hardware-only configuration for a single HyperFlex TRC node and does not include required virtualization software licensing or required HyperFlex HX Data Platform (HXDP) subscription (click here for more details).

    A minimum supported solution requires 3 HyperFlex TRC nodes and a pair of 6x00 Fabric Interconnection Switches. Click here for a sample deployment.

    This configuration is also quotable as part of bundle HX-UC-C2X0M5-TRC1.

    Quantity Cisco Part Number Description
    1 HX220C-M5SX Cisco HyperFlex HX220c M5 Node
    2 Either:
    HX-CPU-6142

    HX-CPU-I6242


    2.6 GHz 6142/150W 16C/22MB Cache/DDR4 2666MHz

    Intel 6242 2.8GHz/150W 16C/24.75MB 3DX DDR4 2933 MHz


    8

    4

    Either:
    HX-MR-X32G2RS-H

    HX-ML-X64G4RT-H


    32GB DDR4-2666-MHz RDIMM/PC4-21300/dual rank/x4/1.2v

    64GB DDR4-2933-MHz LRDIMM/4Rx4/1.2v

    6 HX-HD12TB10K12N 1.2 TB 12G SAS 10K RPM SFF HDD
    1 HX-SD480G63X-EP 480GB 2.5in Enterprise Performance 6GSATA SSD(3X endurance)
    1 Either:
    HX-SD240G61X-EV

    HX-SD240GM1X-EV


    240GB 2.5 inch Enterprise Value 6G SATA SSD

    240GB 2.5 inch Enterprise Value 6G SATA SSD

    1 HX-M2-240GB 240GB SATA M.2
    1 HX-SAS-M5 Cisco 12G Modular SAS HBA (max 16 drives)
    1 HX-MLOM-C40Q-03 Cisco VIC 1387 Dual Port 40Gb QSFP CNA MLOM
    2 CVR-QSFP-SFP10G QSFP to SFP10G adapter
    1
    (optional)
    HX-PCIE-IRJ45 Intel i350 Quad Port 1Gb Adapter
    1 HX-RAILF-M4 Friction Rail Kit for C220 M4 rack servers
    2 HX-PSU1-1050W Cisco UCS 1050W AC Power Supply for Rack Server
    1 HX-MSD-32G 32GB Micro SD Card for UCS M5 servers
    2
    (auto-include)
    UCSC-BBLKD-S2 UCS C-Series M5 SFF drive blanking panel
    1
    (auto-include)
    UCSC-BZL-C220M5 C220 M5 Security Bezel
    2
    (auto-include)
    UCSC-HS-C220M5 Heat sink for UCS C220 M5 rack servers 150W CPUs & below
    1
    (auto-include)
    UCS-MSTOR-M2 Mini Storage carrier for M.2 SATA/NVME (holds up to 2)
    Note
  • HXDP Software Subscription required - sold separately.
  • License required for VMware vSphere ESXi - sold separately.

  • C240 M5SX TRC#2

    Large TRC. See details

    This configuration available as Business Edition 7000H appliance BE7H-M5-K9 or BE7H-M5-XU.

    Note: The C240 M5L is only permitted under UC on UCS Specs-based.

    Note: BE7000 appliances with ESXi 6.5 require vmfs5 (vmfs6 is not supported). For non-appliance TRCs, if you want to use vmfs6, some Collaboration applications do not support (see their virtualization page on Collaboration Virtualization) and a VMware patch is required to ensure storage performance.

    Note: For Intel Spectre/Meltdown concerns, see cisco-sa-20180104-cpusidechannel.

    Quantity Cisco Part Number Description
    1 UCSC-C240-M5SX UCS C240 M5 24 SFF + 2 rear drives w/o CPU,mem,HD,PCIe,PS
    2 UCS-CPU-6132 2.6 GHz 6132/140W 14C/19.25MB Cache/DDR4 2666MHz
    12 UCS-MR-X16G1RS-H 16GB DDR4-2666-MHz RDIMM/PC4-21300/single rank/x4/1.2v
    24 UCS-HD300G10K12N 300GB 12G SAS 10K RPM SFF HDD
    1 UCSC-RAID-M5HD Cisco 12G Modular RAID controller with 4GB cache
    1 R2XX-RAID5 Enable RAID 5 Setting
    2 UCSC-PCIE-IRJ45 Intel i350 Quad Port 1Gb Adapter
    1 UCSC-PCI-1B-240M5 Riser 1B incl 3 PCIe slots (x8, x8, x8); all slots from CPU1
    1 UCSC-RAILB-M4 Ball Bearing Rail Kit for C220 & C240 M4 & M5 rack servers
    2 UCSC-PSU1-1050W Cisco UCS 1050W AC Power Supply for Rack Server
    2(auto-included) UCSC-BBLKD-S2 UCS C-Series M5 SFF drive blanking panel
    1(auto-included) UCSC-PCIF-240M5 C240 M5 PCIe Riser Blanking Panel
    1(auto-included) CBL-SC-MR12GM5P Super Cap cable for UCSC-RAID-M5HD
    1(auto-included) UCSC-SCAP-M5 Super Cap for UCSC-RAID-M5, UCSC-MRAID1GB-KIT
    2(auto-included) UCSC-HS-C240M5 Heat sink for UCS C240 M5 rack servers 150W CPUs & below
    1 CIMC-LATEST IMC SW (Recommended) latest release for C-Series Servers.
    1 C1UCS-OPT-OUT Cisco ONE Data Center Compute Opt Out Option

    C240 M5SX TRC#1

    Medium TRC. See details

    This configuration available as Business Edition 7000M appliance BE7M-M5-K9 or BE7M-M5-XU.

    Note: The C240 M5L is only permitted under UC on UCS Specs-based.

    Note: BE7000 appliances with ESXi 6.5 require vmfs5 (vmfs6 is not supported). For non-appliance TRCs, if you want to use vmfs6, some Collaboration applications do not support (see their virtualization page on Collaboration Virtualization) and a VMware patch is required to ensure storage performance.

    Note: For Intel Spectre/Meltdown concerns, see cisco-sa-20180104-cpusidechannel.

    Quantity Cisco Part Number Description
    1 UCSC-C240-M5SX UCS C240 M5 24 SFF + 2 rear drives w/o CPU,mem,HD,PCIe,PS
    1 UCS-CPU-6132 2.6 GHz 6132/140W 14C/19.25MB Cache/DDR4 2666MHz
    6 UCS-MR-X16G1RS-H 16GB DDR4-2666-MHz RDIMM/PC4-21300/single rank/x4/1.2v
    14 UCS-HD300G10K12N 300GB 12G SAS 10K RPM SFF HDD
    1 UCSC-RAID-M5HD Cisco 12G Modular RAID controller with 4GB cache
    1 R2XX-RAID5 Enable RAID 5 Setting
    2 UCSC-PCIE-IRJ45 Intel i350 Quad Port 1Gb Adapter
    1 UCSC-PCI-1B-240M5 Riser 1B incl 3 PCIe slots (x8, x8, x8); all slots from CPU1
    1 UCSC-RAILB-M4 Ball Bearing Rail Kit for C220 & C240 M4 & M5 rack servers
    2 UCSC-PSU1-1050W Cisco UCS 1050W AC Power Supply for Rack Server
    12(auto-included) UCSC-BBLKD-S2 UCS C-Series M5 SFF drive blanking panel
    1(auto-included) UCSC-PCIF-240M5 C240 M5 PCIe Riser Blanking Panel
    1(auto-included) CBL-SC-MR12GM5P Super Cap cable for UCSC-RAID-M5HD
    1(auto-included) UCSC-SCAP-M5 Super Cap for UCSC-RAID-M5, UCSC-MRAID1GB-KIT
    1(auto-included) UCSC-HS-C240M5 Heat sink for UCS C240 M5 rack servers 150W CPUs & below
    1 CIMC-LATEST IMC SW (Recommended) latest release for C-Series Servers.
    1 C1UCS-OPT-OUT Cisco ONE Data Center Compute Opt Out Option

    C220 M5SX TRC#2

    Small Plus TRC. See details

    This configuration available as Business Edition 6000H appliance BE6H-M5-K9 or BE6H-M5-XU.

    Note: The C220 M5L is only permitted under UC on UCS Specs-based.

    Note: BE6000 appliances with ESXi 6.5 require vmfs5 (vmfs6 is not supported). For non-appliance TRCs, if you want to use vmfs6, some Collaboration applications do not support (see their virtualization page on Collaboration Virtualization) and a VMware patch is required to ensure storage performance.

    Note: For Intel Spectre/Meltdown concerns, see cisco-sa-20180104-cpusidechannel.

    Quantity Cisco Part Number Description
    1 UCSC-C220-M5SX UCS C220 M5 SFF 10 HD w/o CPU, mem, HD, PCIe, PSU
    2 UCS-CPU-4114 2.2 GHz 4114/85W 10C/13.75MB Cache/DDR4 2400MHz
    4 UCS-MR-X16G1RS-H 16GB DDR4-2666-MHz RDIMM/PC4-21300/single rank/x4/1.2v
    8 UCS-HD300G10K12N 300GB 12G SAS 10K RPM SFF HDD
    1 UCSC-RAID-M5 Cisco 12G Modular RAID controller with 2GB cache
    1 R2XX-RAID5 Enable RAID 5 Setting
    1 UCSC-PCIE-IRJ45 Intel i350 Quad Port 1Gb Adapter
    1 UCSC-RAILB-M4 Ball Bearing Rail Kit for C220 & C240 M4 & M5 rack servers
    2 UCSC-PSU1-770W Cisco UCS 770W AC Power Supply for Rack Server
    2(auto-included) UCSC-BBLKD-S2 UCS C-Series M5 SFF drive blanking panel
    1(auto-included) CBL-SC-MR12GM52 Super Cap cable for UCSC-RAID-M5 on C240 M5 Servers
    1(auto-included) UCSC-SCAP-M5 Super Cap for UCSC-RAID-M5, UCSC-MRAID1GB-KIT
    2(auto-included) UCSC-HS-C220M5 Heat sink for UCS C220 M5 rack servers 150W CPUs & below
    1 CIMC-LATEST IMC SW (Recommended) latest release for C-Series Servers.
    1 C1UCS-OPT-OUT Cisco ONE Data Center Compute Opt Out Option

    C220 M5SX TRC#1

    Small TRC. See details

    This configuration available as Business Edition 6000M appliance BE6M-M5-K9 or BE6M-M5-XU.

    Note: The C220 M5L is only permitted under UC on UCS Specs-based.

    Note: BE6000 appliances with ESXi 6.5 require vmfs5 (vmfs6 is not supported). For non-appliance TRCs, if you want to use vmfs6, some Collaboration applications do not support (see their virtualization page on Collaboration Virtualization) and a VMware patch is required to ensure storage performance.

    Note: For Intel Spectre/Meltdown concerns, see cisco-sa-20180104-cpusidechannel.

    Quantity Cisco Part Number Description
    1 UCSC-C220-M5SX UCS C220 M5 SFF 10 HD w/o CPU, mem, HD, PCIe, PSU
    1 UCS-CPU-4114 2.2 GHz 4114/85W 10C/13.75MB Cache/DDR4 2400MHz
    3 UCS-MR-X16G1RS-H 16GB DDR4-2666-MHz RDIMM/PC4-21300/single rank/x4/1.2v
    6 UCS-HD300G10K12N 300GB 12G SAS 10K RPM SFF HDD
    1 UCSC-RAID-M5 Cisco 12G Modular RAID controller with 2GB cache
    1 R2XX-RAID5 Enable RAID 5 Setting
    1 UCSC-RAILB-M4 Ball Bearing Rail Kit for C220 & C240 M4 & M5 rack servers
    1 UCSC-PSU1-770W Cisco UCS 770W AC Power Supply for Rack Server
    4(auto-included) UCSC-BBLKD-S2 UCS C-Series M5 SFF drive blanking panel
    1(auto-included) UCSC-PSU-BLKP1U Power Supply Blanking Panel for C220 M4 servers
    1(auto-included) CBL-SC-MR12GM52 Super Cap cable for UCSC-RAID-M5 on C240 M5 Servers
    1(auto-included) UCSC-SCAP-M5 Super Cap for UCSC-RAID-M5, UCSC-MRAID1GB-KIT
    1(auto-included) UCSC-HS-C220M5 Heat sink for UCS C220 M5 rack servers 150W CPUs & below
    1 CIMC-LATEST IMC SW (Recommended) latest release for C-Series Servers.
    1 C1UCS-OPT-OUT Cisco ONE Data Center Compute Opt Out Option

    E160S M3 TRC#1

    Extra-Small TRC. See details

    For a-la-carte ordering of this blade server as an option of an ISR router toplevel SKU, note the blade server SKU will vary by ISR model. Non-exhaustive examples are provided below.

    Note this is a special server configuration with deployment model restrictions. See www.cisco.com/go/vmpt for allowed application virtual machine mixes.

    Quantity Cisco Part Number Description
    1 Either:
    Blade-server-only
    UCS-E160S-M3/K9=

    Blade server as option of ISR (example only, SKU varies by ISR model)
    UCS-E160S-M3/K9

     

    UCS-E, SingleWide, 6 Core CPU, 8 GB Flash, 8 GB RAM, 1-2 HDD
     

    UCS-E, SingleWide, 6 Core CPU, 8 GB Flash, 8 GB RAM, 1-2 HDD

    1 EM3-MEM-8U16G 8 to 16 GB upgrade, 1200MHz VLP RDIMM/PC4-2400 1R for UCSEM3
    1 EM3-MEM-16G 16 GB 1200MHz VLP RDIMM/PC4-2400 1R for UCS-E M3
    2 E100S-HDD-SAS900G 900 GB, SAS hard disk drive for SingleWide UCS-E
    1 DISK-MODE-RAID-1 Configure hard drives as RAID 1 (Mirror)