Caveated Support for VMware CPU Reservations and Distributed Resource Scheduler
Go back to: Cisco Collaboration Virtualization
|
|
NOTE: Caveated support for VMware CPU reservations and Distributed Resource Scheduler is limited to the following applications:
- Cisco Unified Communications Manager (CUCM) 11.5(1) or greater
- Cisco Unified Communications Manager - IM & Presence (CUCM IM&P) 11.5(1) or greater
- Cisco Unity Connection (CUC) 11.5(1) or greater
NOTE: Support for a deployment leveraging this CPU Reservations policy is granted only following the successful review of a detailed design, including the submission of sizing results from the Cisco Collaboration Sizing Tool and VM placement from Cisco VM Placement Tool. For more information about these tools and the sizing of Unified CM clusters, see the Cisco Collaboration Solution Reference Design Guides. Customers who wish to pursue such a deployment must engage either their Cisco Account Team, Cisco Advanced Services, or their certified Cisco Partner to submit the design review request. CPU Reservations design reviews leverage the same team and process used for UCM megaclusters.
|
|
General Support Caveats |
(top) |
- Recall Cisco Collaboration workloads are resource-intensive and latency-sensitive because they process real-time signaling and media for voice, video, messaging and other forms of live communication. Compared to traditional workloads this introduces restrictions on design, deployment, management and operations to ensure real-time software will perform properly and stably under load.
- Using VMware's terminology (see vSphere Resource Management Guide in Technical References), Cisco is defining "Reservations", not "Limits" or "Shares" (which are not used or supported).
- Caveated support for CPU Reservations is intended for virtualized Collaboration installed base customers with datacenter-centric deployments , skilled/experienced VMware/compute admins and extensive baselining of their deployment's CPU utilization (during both steady-state and spike situations such as group bootup, BHCA, upgrades, backups and CDR writes). Not recommended for greenfield deployments (where no baseline exists), deployments having admins with low VMware/compute skillsets or deployments of Business Edition 6000/7000 appliances (since not tested with CPU Reservations and if post-sales hardware changes ever needed, would not be supported).
- NOTE: Cisco does NOT test VMware CPU Reservations. Cisco Business Edition 6000/7000 appliances and UConUCS TRCs are only tested with the "1vcpu:1pcore" sizing approach as described in the Collaboration Virtualization Sizing. Customers own any required testing of CPU Reservations in their environment.
- NOTE: Cisco does NOT provide prescriptive guidance for CPU Reservations other than the Required CPU Reservation described later in this policy. Cisco Collaboration Sizing Tool and Virtual Machine Placement Tool are still expressed in terms of fixed-configuration VMs using the "1vcpu:1pcore" sizing approach described in the Collaboration Virtualization Sizing.
- For minimum required CPU pcore speeds, this new policy only affects supported applications. Required minimum CPU pcore speed for all other Collab apps is UNCHANGED.
- Caveated support for CPU Reservations does NOT relax sizing rules for other hardware such as Memory, Storage GB / latency / IOPS, network access, etc. Be careful not to use CPU Reservations to place so many VMs on a host that one of these other resources becomes overcommitted.
- Conservative designs are encouraged. Resist the temptation to over-consolidate VMs, which makes change management more complex, creates single points of failure and thwarts redundancy.
- When using CPU reservations with these applications, TAC support for committed real-time performance is only achieved when all VMs are at the Required CPU Reservation, Cisco Collaboration Sizing Tool rules are followed and hardware meets TRC requirements. If ANY are not met, the deployment is treated as Specs-based (see TAC support clarifications in TAC TechNote #115955).
|
|
Required Physical Hardware |
(top) |
Example 4 - Marginal Deployment using Overcommitted pcores and Reduced CPU Reservation: Here is the same deployment as above using Reduced CPU Reservation for each UCM and IM&P VM but not following recommendation of keeping #vcpu ≤ #pcores. Unity Connection VMs do not permit overcommitted pcores or Reduced CPU Reservations, so must run on different esxihosts than those shown.
- Physical CPUs are E5v1+/E7v2+ from "Full UC Performance CPUs", with pcores of 3.40 GHz (faster than requried).
- Moderately overcommitted pcores (8 vcpu to 6 pcores is 1.33 ratio, greater than 1:1 but less than 3:1 so marginal but allowed).
- Using CPU Reservation from customer's baseline (3.71 GHz) with deployment's CPU load previously vetted on Collaboration Sizing Tool.
Example 5 - Unsupported Deployment using Overcommitted pcores: Here is the same deployment as above but with overcommitting pcores beyond Marginal (and not following recommendation of redundant physical servers).
- Physical CPUs are E5v1+/E7v2+ from "Full UC Performance CPUs", with pcores of 3.50 GHz (faster than required).
- Excessively overcommitted pcores (14 vcpu to 4 pcores is 3.5 ratio, which is 3:1 or higher so not supported regardless of hardware or CPU reservations used).
Example 6 - Hypothetical situations that would not be supported.
- 2vcpu VM running on one pcore whose base frequency = CPU Reservation
- 1vcpu VM running on two pcores with base frequency = 50% of CPU Reservation
|
|
Caveated Support for VMware Distributed Resource Scheduler |
(top) |
General Caveats
- NOTE: Unless specifically indicated, "DRS" in this document refers to VMware Distributed Resource Scheduler and not Cisco Disaster Recovery Solution.
- Cisco does NOT test VMware DRS with its applications. Cisco Business Edition 6000/7000 appliances and UConUCS TRCs are only tested with static VM placement (see VM Placement Tool and Collaboration Virtualization Sizing). Customers own any required testing of DRS in their environment.
- Use of DRS with supported application VMs has the same caveats as using vMotion with the same VMs because DRS's live migration also temporarily freezes the live VM to move it to a new host. Freezing a live real-time communications VM can have variable, unpredictable and unpreventable adverse effects, so use of DRS for live migration is NOT advised during business hours (particularly during your busy-hour when BHCA occurs).
Use of VMware DRS for Caveated Applications Virtual Machines
- Supported application VMs must use CPU Reservations, and all caveats and pre-requisites for supported applications' CPU Reservations apply to use of DRS with applications. Including source and destination ESXi hosts following usable capacity rules in Max GHz capacity of an ESXi host / physical server.
- VMware licensing must entitle the DRS feature. NOTE: Collaboration embedded OEM virtualization licenses do NOT entitle DRS.
- Use of DRS must meet all of VMware's DRS+HA+vMotion pre-requisites (see vSphere Resource Mgmt Guide in Technical Resources).
- Customer owns their DRS configuration. Cisco is not responsible for consulting or debugging the DRS configuration with respect to Collaboration applications. Cisco TAC not obligated to root-cause these application issues isolated to DRS.
- VM migration via DRS must NOT result in the following:
- Allowed DRS settings:
- DRS Mode / Automation Level (at both host and VM levels) = Manual and Partially Automated for all supported applications. Fully Automated for UCM and IMP only.
- DRS Migration Threshold = all options allowed but "Conservative" is recommended.
- Dynamic Power Mgmt = not supported (frequently implies CPU throttling to conserve power).
- DRS Affinity & Anti-affinity rules = allowed to help customers comply with SRND rules, but customer's responsibility to setup and debug.
|
|
Technical References from VMware |
(top) |
|