FlexPod for Epic EHR

Available Languages

Download Options

  • PDF
    (13.3 MB)
    View with Adobe Reader on a variety of devices
  • ePub
    (19.3 MB)
    View in various apps on iPhone, iPad, Android, Sony Reader, or Windows Phone
  • Mobi (Kindle)
    (10.7 MB)
    View on Kindle device or Kindle app on multiple devices

Bias-Free Language

The documentation set for this product strives to use bias-free language. For the purposes of this documentation set, bias-free is defined as language that does not imply discrimination based on age, disability, gender, racial identity, ethnic identity, sexual orientation, socioeconomic status, and intersectionality. Exceptions may be present in the documentation due to language that is hardcoded in the user interfaces of the product software, language used based on RFP documentation, or language that is used by a referenced third-party product. Learn more about how Cisco is using Inclusive Language.

Available Languages

Download Options

  • PDF
    (13.3 MB)
    View with Adobe Reader on a variety of devices
  • ePub
    (19.3 MB)
    View in various apps on iPhone, iPad, Android, Sony Reader, or Windows Phone
  • Mobi (Kindle)
    (10.7 MB)
    View on Kindle device or Kindle app on multiple devices

Table of Contents

 

 

 

Published: November 2022

TextDescription automatically generated with low confidence 

A picture containing text, clipartDescription automatically generated

In partnership with:

LogoDescription automatically generated

About the Cisco Validated Design Program

The Cisco Validated Design (CVD) program consists of systems and solutions designed, tested, and documented to facilitate faster, more reliable, and more predictable customer deployments. For more information, go to: http://www.cisco.com/go/designzone.

Executive Summary

The FlexPod Datacenter solution is a validated approach for deploying Cisco® and NetApp technologies and products to build shared private and public cloud infrastructure. Cisco and NetApp have partnered to deliver a series of FlexPod solutions that enable strategic data-center platforms. The success of the FlexPod solution is driven through its ability to evolve and incorporate both technology and product innovations in the areas of management, compute, storage, and networking. This document provides a detailed overview of the FlexPod architecture for Epic and covers the setup and installation of FlexPod to deploy Epic for healthcare. This document explains the design details of incorporating the Cisco X-Series modular platform into the FlexPod Datacenter and the ability to manage and orchestrate FlexPod components from the cloud using Cisco Intersight. Some of the key advantages of integrating Cisco UCS X-Series into the FlexPod infrastructure are:

      Simpler and programmable infrastructure: infrastructure as a code delivered through a single partner integrable open API

      Power and cooling innovations: higher power headroom and lower energy loss due to a 54V DC power delivery to the chassis

      Better airflow: midplane-free design with fewer barriers, therefore lower impedance

      Fabric innovations: PCIe/Compute Express Link (CXL) topology for heterogeneous compute and memory composability

      Innovative cloud operations: continuous feature delivery and no need for maintaining on-premise virtual machines supporting management functions

      Built for investment protections: design ready for future technologies such as liquid cooling and high-Wattage CPUs; CXL-ready

In addition to the compute-specific hardware and software innovations, the integration of the Cisco Intersight cloud platform with VMware vCenter and NetApp Active IQ Unified Manager delivers monitoring, orchestration, and workload optimization capabilities for different layers (virtualization and storage) of the FlexPod infrastructure. The modular nature of the Cisco Intersight platform also provides an easy upgrade path to additional services, such as workload optimization and Kubernetes.

Customers interested in understanding the FlexPod design and deployment details, including the configuration of various elements of design and associated best practices, should refer to Cisco Validated Designs for FlexPod, here: https://www.cisco.com/c/en/us/solutions/design-zone/data-center-design-guides/flexpod-design-guides.html

Solution Overview

Introduction

The Cisco Unified Compute System (Cisco UCS) X-Series is a brand-new modular compute system, configured and managed from the cloud. It is designed to meet the needs of modern applications and to improve operational efficiency, agility, and scale through an adaptable, future-ready, modular design. The Cisco Intersight platform is a Software-as-a-Service (SaaS) infrastructure lifecycle management platform that delivers simplified configuration, deployment, maintenance, and support. FlexPod systems deployed to host Epic HyperSpace, InterSystems Iris database, Cogito Clarity analytics and reporting suite, and services servers hosting the Epic application layer provide an integrated platform for a dependable, high-performance infrastructure that can be deployed rapidly.

Powered by the Cisco Intersight cloud-operations platform, the Cisco UCS X-Series enables the next-generation cloud-operated FlexPod infrastructure that not only simplifies data-center management but also allows the infrastructure to adapt to the unpredictable needs of modern applications as well as traditional workloads. With the Cisco Intersight platform, customers get all the benefits of SaaS delivery and the full lifecycle management of Intersight-connected distributed servers and integrated NetApp storage systems across data centers, remote sites, branch offices, and edge environments.

Audience

The intended audience of this document includes but is not limited to IT architects, sales engineers, field consultants, professional services, IT managers, partner engineering, and customers who want to take advantage of an infrastructure built to deliver IT efficiency and enable IT innovation.

Purpose of this Document

This document provides design guidance around incorporating the Cisco Intersight—managed UCS X-Series platform within FlexPod Datacenter infrastructure for deploying Epic Electronic Health Records (EHR). The document introduces various design elements and covers various considerations and best practices for a successful deployment. The document also highlights the design and product requirements for integrating virtualization and storage systems to Cisco Intersight to deliver a true cloud-based integrated approach to infrastructure management.

What’s New in this Release?

The following design elements distinguish this version of FlexPod from previous models:

      Integration of Cisco UCS X-Series into FlexPod Datacenter

      Deploying and managing Cisco UCS X-Series from the cloud using Cisco Intersight

      Integration of Cisco Intersight with NetApp Active IQ Unified Manager for storage monitoring and orchestration

      Integration of Cisco Intersight with VMware vCenter for interacting with, monitoring, and orchestrating the virtual environment

      Deployment utilizing Ansible playbooks to automate deployment steps

      Deployment utilizing Terraform scripts to automate deployment steps

Solution Summary

By running an Epic environment on the FlexPod architectural foundation, healthcare organizations can expect to see an improvement in staff productivity and a decrease in capital and operating expenses. The FlexPod Datacenter solution with Cisco UCS X-Series, VMware 7.0 U3, and NetApp ONTAP 9.10.1P1 for Epic EHR delivers several benefits specific to the healthcare industry:

      Simplified operations and lowered costs: Eliminate the expense and complexity of legacy proprietary RISC/UNIX platforms by replacing them with a more efficient and scalable shared resource capable of supporting clinicians wherever they are. This solution delivers higher resource utilization for greater ROI.

      Quicker deployment of infrastructure: Whether it’s in an existing data center or a remote location, the integrated and tested design of FlexPod Datacenter with Epic enables customers to have the new infrastructure up and running in less time with less effort.

      Scale-out architecture: Scale SAN and NAS from terabytes to tens of petabytes without reconfiguring running applications.

      Nondisruptive operations: Perform storage maintenance, hardware lifecycle operations, and software upgrades without interrupting the business.

      Secure multitenancy: This benefit supports the increased needs of virtualized server and storage shared infrastructure, enabling secure multitenancy of facility-specific information, particularly if hosting multiple instances of databases and software.

      Pooled resource optimization: This benefit can help reduce physical server and storage controller counts, load balance workload demands, and boost utilization while improving performance.

      Quality of service (QoS): FlexPod offers QoS on the entire stack. Industry-leading QoS storage policies enable differentiated service levels in a shared environment. These policies enable optimal performance for workloads and help in isolating and controlling runaway applications.

      Storage efficiency: Reduce storage costs with the NetApp 7: 1 storage efficiency guarantee.

      Agility: The industry-leading workflow automation, orchestration, and management tools offered by FlexPod systems allow IT to be far more responsive to business requests. These business requests can range from Epic backup and provisioning of additional test and training environments to analytics database replications for population health management initiatives.

      Productivity: Quickly deploy and scale this solution for optimal clinician end- user experiences.

Like all other FlexPod solution designs, FlexPod Datacenter with Cisco UCS X-Series and Intersight is configurable according to demand and usage. Customers can purchase exactly the infrastructure they need for their current application requirements and can then scale up by adding more resources to the FlexPod system or scale out by adding more FlexPod instances. By moving the management from the fabric interconnects into the cloud, the solution can respond to the speed and scale of customer deployments with a constant stream of new capabilities delivered from Intersight software-as-a-service model at cloud-scale.

Technology Overview

FlexPod Datacenter

Epic FlexPod Datacenter architecture is built using the following infrastructure components for compute, network, and storage:

      Cisco Unified Computing System (Cisco UCS)

      Cisco Nexus® and Cisco MDS switches

      NetApp All Flash FAS (AFF) storage systems

Figure 1.   FlexPod Datacenter Components

A picture containing timelineDescription automatically generated

All the FlexPod components have been integrated so that customers can deploy the solution quickly and economically while eliminating many of the risks associated with researching, designing, building, and deploying similar solutions from the foundation. One of the main benefits of FlexPod is its ability to maintain consistency at scale. Each of the component families shown in Figure 1 (Cisco UCS, Cisco Nexus, Cisco MDS, and NetApp controllers) offers platform and resource options to scale up or scale out the infrastructure while supporting the same features.

The FlexPod Datacenter solution with Cisco UCS X-Series is built using following hardware components:

      Cisco UCS X9508 Chassis with up to eight Cisco UCS X210c M6 Compute Nodes

      Fourth-generation Cisco UCS 6454 Fabric Interconnects to support 10GbE, 25GbE, and 100GbE connectivity from various components

      High-speed Cisco NX-OS-based Nexus 93180YC-FX3 switching design to support up to 100GE connectivity

      NetApp AFF A400 end-to-end NVMe storage with Fibre Channel connectivity and NFS Storage with high-speed Ethernet.

The software components of the solution consist of:

      Cisco Intersight platform to deploy, maintain and support the FlexPod components

      Cisco Intersight Assist Virtual Appliance to help connect NetApp ONTAP and VMware vCenter with Cisco Intersight

      NetApp Active IQ Unified Manager to monitor and manage the storage and for NetApp ONTAP integration with Cisco Intersight

      VMware vCenter to set up and manage the virtual infrastructure as well as Cisco Intersight integration

These key product highlights and features are described in the following sections.

Cisco Unified Compute System X-Series

The Cisco UCS X-Series Modular System is designed to take the current generation of the Cisco UCS platform to the next level with its future-ready design and cloud-based management. Decoupling and moving the platform management to the cloud allows Cisco UCS to respond to customer feature and scalability requirements in a much faster and efficient manner. Cisco UCS X-Series state of the art hardware simplifies the data-center design by providing flexible server options. A single server type, supporting a broader range of workloads, results in fewer different data-center products to manage and maintain. The Cisco Intersight cloud-management platform manages Cisco UCS X-Series as well as integrating with third-party devices, including VMware vCenter and NetApp storage, to provide visibility, optimization, and orchestration from a single platform, thereby driving agility and deployment consistency.

Figure 2.   Cisco UCS X9508 Chassis

Cisco UCS X9508 Chassis

The various components of the Cisco UCS X-Series are described in the following sections.

Cisco UCS X9508 Chassis

The Cisco UCS X-Series chassis is engineered to be adaptable and flexible. As seen in Figure 3, Cisco UCS X9508 chassis has only a power-distribution midplane. This midplane-free design provides fewer obstructions for better airflow. For I/O connectivity, vertically oriented compute nodes intersect with horizontally oriented fabric modules, allowing the chassis to support future fabric innovations. Cisco UCS X9508 Chassis’ superior packaging enables larger compute nodes, thereby providing more space for actual compute components, such as memory, GPU, drives, and accelerators. Improved airflow through the chassis enables support for higher power components, and more space allows for future thermal solutions (such as liquid cooling) without limitations.

Figure 3.   Cisco UCS X9508 Chassis – midplane free design

A picture containing text, indoor, kitchen applianceDescription automatically generated

The Cisco UCS X9508 7-Rack-Unit (7RU) chassis has eight flexible slots. These slots can house a combination of compute nodes and a pool of future I/O resources that may include GPU accelerators, disk storage, and nonvolatile memory. At the top rear of the chassis are two Intelligent Fabric Modules (IFMs) that connect the chassis to upstream Cisco UCS 6400 Series Fabric Interconnects. At the bottom rear of the chassis are slots ready to house future X-Fabric modules that can flexibly connect the compute nodes with I/O devices. Six 2800W Power Supply Units (PSUs) provide 54V power to the chassis with N, N+1, and N+N redundancy. A higher voltage allows efficient power delivery with less copper and reduced power loss. Efficient, 100mm, dual counter-rotating fans deliver industry-leading airflow and power efficiency, and optimized thermal algorithms enable different cooling modes to best support the customer’s environment.

Cisco UCS X9108-25G Intelligent Fabric Modules

For the Cisco UCS X9508 Chassis, the network connectivity is provided by a pair of Cisco UCS X9108-25G Intelligent Fabric Modules (IFMs). Like the fabric extenders used in the Cisco UCS 5108 Blade Server Chassis, these modules carry all network traffic to a pair of Cisco UCS 6400 Series Fabric Interconnects (FIs). IFMs also host the Chassis Management Controller (CMC) for chassis management. In contrast to systems with fixed networking components, Cisco UCS X9508’s midplane-free design enables easy upgrades to new networking technologies as they emerge making it straightforward to accommodate new network speeds or technologies in the future.

Figure 4.   Cisco UCS X9108-25G Intelligent Fabric Module

 A picture containing textDescription automatically generated

Each IFM supports eight 25Gb uplink ports for connecting the Cisco UCS X9508 Chassis to the FIs and 32 25Gb server ports for the eight compute nodes. IFM server ports can provide up to 200 Gbps of unified fabric connectivity per compute node across the two IFMs. The uplink ports connect the chassis to the UCS FIs, providing up to 400Gbps connectivity across the two IFMs. The unified fabric carries management, VM, and Fibre Channel over Ethernet (FCoE) traffic to the FIs, where management traffic is routed to the Cisco Intersight cloud operations platform, FCoE traffic is forwarded to the native Fibre Channel interfaces through unified ports on the FI (to Cisco MDS switches), and data Ethernet traffic is forwarded upstream to the data center network (via Cisco Nexus switches).

Cisco UCS X210c M6 Compute Node

The Cisco UCS X9508 Chassis is designed to host up to 8 Cisco UCS X210c M6 Compute Nodes. The hardware details of the Cisco UCS X210c M6 Compute Nodes are shown in Figure 5:

Figure 5.   Cisco UCS X210c M6 Compute Node

Cisco UCS X210c M6 Compute Node

The Cisco UCS X210c M6 features:

      CPU: Up to 2x 3rd Gen Intel Xeon Scalable Processors with up to 40 cores per processor and 1.5 MB Level 3 cache per core

      Memory: Up to 32 x 256 GB DDR4-3200 DIMMs for a maximum of 8 TB of main memory. The Compute Node can also be configured for up to 16 x 512-GB Intel Optane persistent memory DIMMs for a maximum of 12 TB of memory

      Disk storage: Up to 6 SAS or SATA drives can be configured with an internal RAID controller, or customers can configure up to 6 NVMe drives. 2 M.2 memory cards can be added to the Compute Node with RAID 1 mirroring.

      Virtual Interface Card (VIC): Up to 2 VICs including an mLOM Cisco VIC 14425 and a mezzanine Cisco VIC card 14825 can be installed in a Compute Node.

      Security: The server supports an optional Trusted Platform Module (TPM). Additional security features include a secure boot FPGA and ACT2 anticounterfeit provisions.

Cisco UCS Virtual Interface Cards (VICs)

Cisco UCS X210c M6 Compute Nodes support the following two Cisco fourth-generation VIC cards:

Cisco VIC 14425

Cisco VIC 14425 fits the mLOM slot in the Cisco X210c Compute Node and enables up to 50 Gbps of unified fabric connectivity to each of the chassis IFMs for a total of 100 Gbps of connectivity per server. Cisco VIC 14425 connectivity to the IFM and up to the fabric interconnects is delivered through 4x 25-Gbps connections, which are configured automatically as 2x 50-Gbps port channels. Cisco VIC 14425 supports 256 virtual interfaces (both Fibre Channel and Ethernet) along with the latest networking innovations such as NVMeoF over RDMA (ROCEv2), VxLAN/NVGRE offload, and so on.

Figure 6.   Single Cisco VIC 14425 in Cisco UCS X210c M6

Single Cisco VIC 14425 in Cisco UCS X210c M6

The connections between the 4th generation Cisco VIC (Cisco UCS VIC 1) in the Cisco UCS B200 blades and the I/O modules in the Cisco UCS 5108 chassis comprise of multiple 10Gbps KR lanes. The same connections between Cisco VIC 14425 and IFMs in Cisco UCS X-Series comprise of multiple 25Gbps KR lanes resulting in 2.5x better connectivity in Cisco UCS X210c M6 Compute Nodes. The network interface speed comparison between VMware ESXi installed on Cisco UCS B200 M5 with VIC 1440 and Cisco UCS X210c M6 with VIC 14425 is shown in Figure 7.

Figure 7.   Network Interface Speed Comparison

TableDescription automatically generated

Cisco VIC 14825

The optional Cisco VIC 14825 fits the mezzanine slot on the server. A bridge card (UCSX-V4-BRIDGE) extends this VIC’s 2x 50 Gbps of network connections up to the mLOM slot and out through the mLOM’s IFM connectors, bringing the total bandwidth to 100 Gbps per fabric for a total bandwidth of 200 Gbps per server.

Figure 8.   Cisco VIC 14425 and 14825 in Cisco UCS X210c M6

DiagramDescription automatically generated

Cisco UCS 6400 Series Fabric Interconnects

The Cisco UCS Fabric Interconnects (FIs) provide a single point for connectivity and management for the entire Cisco UCS system. Typically deployed as an active/active pair, the system’s FIs integrate all components into a single, highly available management domain controlled by the Cisco UCS Manager or Cisco Intersight. Cisco UCS FIs provide a single unified fabric for the system, with low-latency, lossless, cut-through switching that supports LAN, SAN, and management traffic using a single set of cables.

Figure 9.   Cisco UCS 6454 Fabric Interconnect

Related image, diagram or screenshot

Cisco UCS 6454 utilized in the current design is a 54-port Fabric Interconnect. This single RU device includes 28 10/25 Gbps Ethernet ports, 4 1/10/25-Gbps Ethernet ports, 6 40/100-Gbps Ethernet uplink ports, and 16 unified ports that can support 10/25 Gigabit Ethernet or 8/16/32-Gbps Fibre Channel, depending on the SFP.

Note:   For supporting the Cisco UCS X-Series, the fabric interconnects must be configured in Intersight Managed Mode (IMM). This option replaces the local management with Cisco Intersight cloud- or appliance-based management.

Cisco Intersight

The Cisco Intersight platform is a Software-as-a-Service (SaaS) infrastructure lifecycle management platform that delivers simplified configuration, deployment, maintenance, and support. The Cisco Intersight platform is designed to be modular, so customers can adopt services based on their individual requirements. The platform significantly simplifies IT operations by bridging applications with infrastructure, providing visibility and management from bare-metal servers and hypervisors to serverless applications, thereby reducing costs and mitigating risk. This unified SaaS platform uses a unified Open API design that natively integrates with third-party platforms and tools.

Figure 10.       Cisco Intersight overview

Cisco Intersight overview

The main benefits of Cisco Intersight infrastructure services are as follows:

      Simplify daily operations by automating many daily manual tasks

      Combine the convenience of a SaaS platform with the capability to connect from anywhere and manage infrastructure through a browser or mobile app

      Stay ahead of problems and accelerate trouble resolution through advanced support capabilities

      Gain global visibility of infrastructure health and status along with advanced management and support capabilities

      Upgrade to add workload optimization and Kubernetes services when needed

Cisco Intersight Virtual Appliance and Private Virtual Appliance

In addition to the SaaS deployment model running on Intersight.com, on-premises options can be purchased separately. The Cisco Intersight Virtual Appliance and Cisco Intersight Private Virtual Appliance are available for organizations that have additional data locality or security requirements for managing systems. The Cisco Intersight Virtual Appliance delivers the management features of the Cisco Intersight platform in an easy-to-deploy VMware Open Virtualization Appliance (OVA) or Microsoft Hyper-V Server virtual machine that allows you to control the system details that leave your premises. The Cisco Intersight Private Virtual Appliance is provided in a form factor specifically designed for users who operate in disconnected (air gap) environments. The Private Virtual Appliance requires no connection to public networks or back to Cisco to operate.

Cisco Intersight Assist

Cisco Intersight Assist helps customers add endpoint devices to Cisco Intersight. A data center could have multiple devices that do not connect directly with Cisco Intersight. Any device that is supported by Cisco Intersight, but does not connect directly with it, will need a connection mechanism. Cisco Intersight Assist provides that connection mechanism. In FlexPod, VMware vCenter and NetApp Active IQ Unified Manager connect to Intersight with the help of Intersight Assist VM.

Cisco Intersight Assist is available within the Cisco Intersight Virtual Appliance, which is distributed as a deployable virtual machine contained within an Open Virtual Appliance (OVA) file format. More details about the Cisco Intersight Assist VM deployment configuration is covered in later sections.

Licensing Requirements

The Cisco Intersight platform uses a subscription-based license with multiple tiers. Customers can purchase a subscription duration of one, three, or five years and select the required Cisco UCS server volume tier for the selected subscription duration. Each Cisco endpoint automatically includes a Cisco Intersight Base license at no additional cost when customers access the Cisco Intersight portal and claim a device. Customers can purchase any of the following higher-tier Cisco Intersight licenses using the Cisco ordering tool:

      Cisco Intersight Essentials: Essentials includes all the functions of the Base license plus additional features, including Cisco UCS Central Software and Cisco Integrated Management Controller (IMC) supervisor entitlement, policy-based configuration with server profiles, firmware management, and evaluation of compatibility with the Cisco Hardware Compatibility List (HCL).

      Cisco Intersight Advantage: Advantage offers all the features and functions of the Base and Essentials tiers. It includes storage widgets and cross-domain inventory correlation across compute, storage, and virtual environments (VMWare ESXi). It also includes OS installation for supported Cisco UCS platforms.

      Cisco Intersight Premier: In addition to all of the functions provided in the Advantage tier, Premier includes full subscription entitlement for Intersight Orchestrator, which provides orchestration across Cisco UCS and third-party systems.

Servers in the Cisco Intersight managed mode require at least the Essentials license. For more information about the features provided in the various licensing tiers, see https://intersight.com/help/getting_started#licensing_requirements

Cisco Nexus Switching Fabric

The Cisco Nexus 9000 Series Switches offer both modular and fixed 1/10/25/40/100 Gigabit Ethernet switch configurations with scalability up to 60 Tbps of nonblocking performance with less than five-microsecond latency, wire speed VXLAN gateway, bridging, and routing support.

Figure 11.       Cisco Nexus 93180YC-FX3 Switch

Cisco Nexus 93180YC-FX3 Switch

The Cisco Nexus 9000 series switch featured in this design is the Cisco Nexus 93180YC-FX3 configured in NX-OS standalone mode. NX-OS is a purpose-built data-center operating system designed for performance, resiliency, scalability, manageability, and programmability at its foundation. It provides a robust and comprehensive feature set that meets the demanding requirements of virtualization and automation.

The Cisco Nexus 93180YC-FX3 Switch is a 1RU switch that supports 3.6 Tbps of bandwidth and 1.2 bpps. The 48 downlink ports on the 93180YC-FX3 can support 1-, 10-, or 25-Gbps Ethernet, offering deployment flexibility and investment protection. The six uplink ports can be configured as 40- or 100-Gbps Ethernet, offering flexible migration options.

Cisco MDS 9132T 32G Multilayer Fabric Switch

The Cisco MDS 9132T 32G Multilayer Fabric Switch is the next generation of the highly reliable, flexible, and low-cost Cisco MDS 9100 Series switches. It combines high performance with exceptional flexibility and cost effectiveness. This powerful, compact one Rack-Unit (1RU) switch scales from 8 to 32 line-rate 32 Gbps Fibre Channel ports.

Figure 12.       Cisco MDS 9132T 32G Multilayer Fabric Switch

Related image, diagram or screenshot

The Cisco MDS 9132T delivers advanced storage networking features and functions with ease of management and compatibility with the entire Cisco MDS 9000 family portfolio for reliable end-to-end connectivity. This switch also offers state-of-the-art SAN analytics and telemetry capabilities that have been built into this next-generation hardware platform. This new state-of-the-art technology couples the next-generation port ASIC with a fully dedicated network processing unit designed to complete analytics calculations in real time. The telemetry data extracted from the inspection of the frame headers are calculated on board (within the switch) and, using an industry-leading open format, can be streamed to any analytics-visualization platform. This switch also includes a dedicated 10/100/1000BASE-T telemetry port to maximize data delivery to any telemetry receiver, including Cisco Data Center Network Manager.

NetApp AFF A-Series Storage

NetApp AFF A-Series controller lineup provides industry leading performance while continuing to provide a full suite of enterprise-grade data management and data protection features. AFF A-Series systems support end-to-end NVMe technologies, from NVMe-attached SSDs to frontend NVMe over Fibre Channel (NVMe/FC) host connectivity. These systems deliver enterprise class performance, making them a superior choice for driving the most demanding workloads and applications. With a simple software upgrade to the modern NVMe/FC SAN infrastructure, you can drive more workloads with faster response times, without disruption or data migration. Additionally, more organizations are adopting a “cloud first” strategy, driving the need for enterprise-grade data services for a shared environment across on-premises data centers and the cloud. As a result, modern all-flash arrays must provide robust data services, integrated data protection, seamless scalability, and new levels of performance — plus deep application and cloud integration. These new workloads demand performance that first-generation flash systems cannot deliver.

For more information about the NetApp AFF A-series controllers, see the AFF product page: https://www.netapp.com/us/products/storage-systems/all-flash-array/aff-a-series.aspx

You can view or download more technical specifications of the AFF A-series controllers here: https://www.netapp.com/us/media/ds-3582.pdf

NetApp AFF A400

The NetApp AFF A400 offers full end-to-end NVMe support. The frontend NVMe/FC connectivity makes it possible to achieve optimal performance from an all-flash array for workloads that include artificial intelligence, machine learning, and real-time analytics as well as business-critical databases. On the back end, the A400 supports both serial-attached SCSI (SAS) and NVMe-attached SSDs, offering the versatility for current customers to move up from their legacy A-Series systems and satisfying the increasing interest that all customers have in NVMe-based storage.

The NetApp AFF A400 offers greater port availability, network connectivity, and expandability. The NetApp AFF A400 has 10 PCIe Gen3 slots per high availability pair. The NetApp AFF A400 offers 25GbE or 100GbE, as well as 32Gb/FC and NVMe/FC network connectivity. This model was created to keep up with changing business needs and performance and workload requirements by merging the latest technology for data acceleration and ultra-low latency in an end-to-end NVMe storage system.

Note:   Cisco UCS X-Series is supported with all NetApp AFF systems running NetApp ONTAP 9 release.

Figure 13.       NetApp AFF A400 Front View

Related image, diagram or screenshot

Figure 14.       NetApp AFF A400 Rear View

A close-up of a computerDescription automatically generated with low confidence

NetApp ONTAP 9

NetApp storage systems harness the power of ONTAP to simplify the data infrastructure from edge, core, and cloud with a common set of data services and 99.9999 percent availability. NetApp ONTAP 9 data management software from NetApp enables customers to modernize their infrastructure and transition to a cloud-ready data center. ONTAP 9 has a host of features to simplify deployment and data management, accelerate and protect critical data, and make infrastructure future-ready across hybrid-cloud architectures.

NetApp ONTAP 9 is the data management software that is used with the NetApp AFF A400 all-flash storage system in this solution design. ONTAP software offers secure unified storage for applications that read and write data over block- or file-access protocol storage configurations. These storage configurations range from high-speed flash to lower-priced spinning media or cloud-based object storage. ONTAP implementations can run on NetApp engineered FAS or AFF series arrays and in private, public, or hybrid clouds (NetApp Private Storage and NetApp Cloud Volumes ONTAP). Specialized implementations offer best-in-class converged infrastructure, featured here as part of the FlexPod Datacenter solution or with access to third-party storage arrays (NetApp FlexArray virtualization). Together these implementations form the basic framework of the NetApp Data Fabric, with a common software-defined approach to data management, and fast efficient replication across systems. FlexPod and ONTAP architectures can serve as the foundation for both hybrid cloud and private cloud designs.

Read more about all the capabilities of ONTAP data management software here: https://www.netapp.com/us/products/data-management-software/ontap.aspx.

ONTAP 9.10 brings additional enhancements in manageability, data protection, networking and security protocols, and SAN and object storage. It also includes updated hardware support, increased MetroCluster IP solution scale, and supports IP-routed MetroCluster IP backend connections. See the ONTAP 9.10 release note below for more details:

https://library.netapp.com/ecm/ecm_download_file/ECMLP2492508

NetApp Active IQ Unified Manager

NetApp Active IQ Unified Manager is a comprehensive monitoring and proactive management tool for NetApp ONTAP systems to help manage the availability, capacity, protection, and performance risks of your storage systems and virtual infrastructure. The Unified Manager can be deployed on a Linux server, on a Windows server, or as a virtual appliance on a VMware host.

Active IQ Unified Manager enables monitoring your ONTAP storage clusters from a single redesigned, intuitive interface that delivers intelligence from community wisdom and AI analytics. It provides comprehensive operational, performance, and proactive insights into the storage environment and the virtual machines running on it. When an issue occurs on the storage infrastructure, Unified Manager can notify you about the details of the issue to help with identifying the root cause. The virtual machine dashboard gives you a view into the performance statistics for the VM so that you can investigate the entire I/O path from the vSphere host down through the network and finally to the storage. Some events also provide remedial actions that can be taken to rectify the issue. You can configure custom alerts for events so that when issues occur, you are notified through email and SNMP Traps. Active IQ Unified Manager enables planning for the storage requirements of your users by forecasting capacity and usage trends to proactively act before issues arise, preventing reactive short-term decisions that can lead to additional problems in the long term.

VMware vSphere 7.0 U3

VMware vSphere is a virtualization platform for holistically managing large collections of infrastructures (resources including CPUs, storage, and networking) as a seamless, versatile, and dynamic operating environment. Unlike traditional operating systems that manage an individual machine, VMware vSphere aggregates the infrastructure of an entire data center to create a single powerhouse with resources that can be allocated quickly and dynamically to any application in need.

VMware vSphere 7.0 U3 has several improvements and simplifications including, but not limited to:

      Fully featured vSphere Client (HTML5) client. (The flash-based vSphere Web Client has been deprecated and is no longer available.)

      Improved Distributed Resource Scheduler (DRS) – a very different approach that results in a much more granular optimization of resources

      Assignable hardware – a new framework that was developed to extend support for vSphere features when customers utilize hardware accelerators

      vSphere Lifecycle Manager – a replacement for VMware Update Manager, bringing a suite of capabilities to make lifecycle operations better

      Refactored vMotion – improved to support today’s workloads

For more information about VMware vSphere and its components, see: https://www.vmware.com/products/vsphere.html.

VMware vSphere vCenter

VMware vCenter Server provides unified management of all hosts and VMs from a single console and aggregates performance monitoring of clusters, hosts, and VMs. VMware vCenter Server gives administrators a deep insight into the status and configuration of compute clusters, hosts, VMs, storage, the guest OS, and other critical components of a virtual infrastructure. VMware vCenter manages the rich set of features available in a VMware vSphere environment.

Cisco Intersight Assist device connector for VMware vCenter and NetApp ONTAP

Cisco Intersight integrates with VMware vCenter and NetApp storage as follows: - Cisco Intersight uses the device connector running within Cisco Intersight Assist virtual appliance to communicate with the VMware vCenter. - Cisco Intersight uses the device connector running within a Cisco Intersight Assist virtual appliance to integrate with NetApp Active IQ Unified Manager. The NetApp AFF A400 should be added to NetApp Active IQ Unified Manager.

Figure 15.       Cisco Intersight and vCenter/NetApp Integration

Cisco Intersight and vCenter/NetApp Integration

The device connector provides a secure way for connected targets to send information and receive control instructions from the Cisco Intersight portal using a secure internet connection. The integration brings the full value and simplicity of Cisco Intersight infrastructure management service to VMware hypervisor and ONTAP data storage environments.

Enterprise SAN and NAS workloads can benefit equally from the integrated management solution. The integration architecture enables FlexPod customers to use new management capabilities with no compromise in their existing VMware or ONTAP operations. IT users will be able to manage heterogeneous infrastructure from a centralized Cisco Intersight portal. At the same time, the IT staff can continue to use VMware vCenter and NetApp Active IQ Unified Manager for comprehensive analysis, diagnostics, and reporting of virtual and storage environments. The functionality provided through this integration is covered in the upcoming solution design section.

Solution Design

The FlexPod Datacenter with Cisco UCS X-Series and Intersight for Epic EHR solution delivers a cloud-managed infrastructure solution on the latest Cisco UCS hardware. VMware vSphere 7.0 U3 hypervisor is installed on the Cisco UCS X210c M6 Compute Nodes configured for stateless compute design using boot from SAN. Additionally, Local Boot (m.2 MRAID) options are provided. NetApp AFF A400 provides the storage infrastructure required for setting up the VMware environment. The Cisco Intersight cloud-management platform is utilized to configure and manage the infrastructure. The solution requirements and design details are explained in this section.

This document does not replace deployment guidance from Epic, Cisco, or NetApp. This CVD will give instructions on how to deploy the hardware to implement the infrastructure that Epic has developed for your specific requirements. In all cases, follow the guidance given from Epic, Cisco, and NetApp. 

Requirements

The FlexPod Datacenter with Cisco UCS X-Series and Intersight for Epic EHR design meets the following general design requirements:

      Resilient design across all layers of the infrastructure with no single point of failure

      Scalable design with the flexibility to add compute capacity, storage, or network bandwidth as needed

      Modular design that can be replicated to expand and grow as the needs of the business grow

      Simplified design with ability to integrate and automate with external automation tools

      Cloud-enabled design which can be configured, managed, and orchestrated from the cloud using GUI or APIs

Physical Topology

FlexPod Datacenter with Cisco UCS X-Series supports both IP and Fibre Channel (FC)—based storage access design. For the FC designs, NetApp AFF A400 and Cisco UCS X-Series are connected through Cisco MDS 9132T Fibre Channel Switches and boot from SAN uses the FC network. In this design, VMware ESXi hosts access the VM datastore volumes on NetApp using NFS. The Epic VMs access CIFs shares. The physical connectivity details FC designs are explained below.

FC-based Storage Access: FC, NFS, and CIFS

The physical topology for the FC, NFS, and CIFS based FlexPod Datacenter is shown in Figure 16.

Figure 16.       FlexPod Datacenter Physical Topology for FC, NFS, and CIFS

DiagramDescription automatically generated

To validate the FC-based storage access in a FlexPod configuration, the components are set up as follows:

      Cisco UCS 6454 Fabric Interconnects provide the chassis and network connectivity.

      The Cisco UCS X9508 Chassis connects to fabric interconnects using Cisco UCS X9108-25G Intelligent Fabric Modules (IFMs), where four 25 Gigabit Ethernet ports are used on each IFM to connect to the appropriate FI.

      Cisco UCS X210c M6 Compute Nodes contain fourth-generation Cisco 14425 virtual interface cards.

      Cisco Nexus 93180YC-FX3 Switches in Cisco NX-OS mode provide the switching fabric.

      Cisco UCS 6454 Fabric Interconnect 100 Gigabit Ethernet uplink ports connect to Cisco Nexus 93180YC-FX3 Switches in a vPC configuration.

      The NetApp AFF A400 controller connects to the Cisco Nexus 93180YC-FX3 Switches using four 25 GE ports from each controller configured as a vPC for NFS traffic.

      Cisco UCS 6454 Fabric Interconnects are connected to the Cisco MDS 9132T switches using 32-Gbps Fibre Channel connections configured as a single port channel for SAN connectivity.

      The NetApp AFF controller connects to the Cisco MDS 9132T switches using 32-Gbps Fibre Channel connections for SAN connectivity.

      VMware 7.0 U3 ESXi software is installed on Cisco UCS X210c M6 Compute Nodes to validate the infrastructure.

VLAN Configuration

Table 1 lists VLANs configured for setting up the FlexPod environment along with their usage.

Table 1.     VLAN Usage

VLAN ID

Name

Usage

2

Native-VLAN

Use VLAN 2 as native VLAN instead of default VLAN (1).

40

OOB-MGMT-VLAN

Out-of-band management VLAN to connect management ports for various devices

41

IB-MGMT-VLAN

In-band management VLAN utilized for all in-band management connectivity - for example, ESXi hosts, VM management, and so on.

44

VM-Traffic

VM data traffic VLAN

42

NFS-VLAN

NFS VLAN for mounting datastores in ESXi servers for VMs

43

vMotion

VMware vMotion traffic

45

CIFS

VLAN for CIFs shares

Some of the key highlights of VLAN usage are as follows:

      VLAN 40 allows customers to manage and access out-of-band management interfaces of various devices.

      VLAN 41 is used for in-band management of VMs, ESXi hosts, and other infrastructure services

      VLAN 42 provides ESXi hosts access to the NSF datastores hosted on the NetApp Controllers for deploying VMs.

      VLAN 45 provides virtual machines access to the CIFS shares

Logical Topology

In FlexPod Datacenter deployments, each Cisco UCS server equipped with a Cisco Virtual Interface Card (VIC) is configured for multiple virtual Network Interfaces (vNICs), which appear as standards-compliant PCIe endpoints to the OS. The end-to-end logical connectivity including VLAN/VSAN usage between the server profile for an ESXi host and the storage configuration on NetApp AFF A400 controllers is captured in the following subsections.

Logical Topology for IP-based Storage Access

Figure 17 illustrates the end-to-end connectivity design for IP-based storage access.

Figure 17.       Logical end-to-end connectivity for IP-based storage access design

ChartDescription automatically generated

Each ESXi service profile supports:

      Managing the ESXi hosts using a common management segment

      Four vNICs where:

      Two redundant vNICs (vSwitch0-A and vSwitch0-B) carry management and infrastructure NFS traffic. The MTU value for these vNICs is set as a Jumbo MTU (9000).

      Two redundant vNICs (VDS-A and VDS-B) are used by the vSphere Distributed switch and carry VMware vMotion traffic and customer application data traffic. The MTU for the vNICs is set to Jumbo MTU (9000).

      Each ESXi host (compute node) mounts VM datastores from NetApp AFF A400 controllers using NFS for deploying virtual machines.

Logical Topology for FC-based Storage Access

Figure 18 illustrates the end-to-end connectivity design for FC-based storage access.

Figure 18.       Logical End-to-End Connectivity for FC Design

DiagramDescription automatically generated

Each ESXi server profile supports:

      Managing the ESXi hosts using a common management segment

      Diskless SAN boot using FC with persistent operating system installation for true stateless computing

      Or: Local boot utilizing M.2 MRAID Adapters

      Four vNICs where:

  Two redundant vNICs (vSwitch0-A and vSwitch0-B) carry in-band management, and Infrastructure NFS VLANs. The MTU value for these vNICs is set as a Jumbo MTU (9000).

  Two redundant vNICs (VDS-A and VDS-B) are used by the vSphere Distributed switch and carry VMware vMotion traffic and customer application data traffic. The MTU for the vNICs is set to Jumbo MTU (9000).

  One vHBA defined on Fabric A to provide access to SAN-A path.

  One vHBA defined on Fabric B to provide access to SAN-B path.

      Each ESXi host (compute node) mounts VM datastores from NetApp AFF A400 controllers using NFS for deploying virtual machines.

Compute System Connectivity

The Cisco UCS X9508 Chassis is equipped with the Cisco UCS X9108-25G intelligent fabric modules (IFMs). The Cisco UCS X9508 Chassis connects to each Cisco UCS 6454 FI using four 25GE ports, as shown in Figure 19. If customers require more bandwidth, all eight ports on the IFMs can be connected to each FI.

Figure 19.       Cisco UCS X9508 Chassis Connectivity to Cisco UCS Fabric Interconnects

Cisco UCS X9508 Chassis Connectivity to Cisco UCS Fabric Interconnects

Cisco Nexus Ethernet Connectivity

The Cisco Nexus 93180YC-FX3 device configuration covers the core networking requirements for Layer 2 and Layer 3 communication. Some of the key NX-OS features implemented within the design are:

      Feature interface-vans – Allows for VLAN IP interfaces to be configured within the switch as gateways.

      Feature HSRP – Allows for Hot Standby Routing Protocol configuration for high availability.

      Feature LACP – Allows for the utilization of Link Aggregation Control Protocol (802.3ad) by the port channels configured on the switch.

      Feature VPC – Virtual Port-Channel (vPC) presents the two Nexus switches as a single “logical” port channel to the connecting upstream or downstream device.

      Feature LLDP - Link Layer Discovery Protocol (LLDP), a vendor-neutral device discovery protocol, allows the discovery of both Cisco devices and devices from other sources.

      Feature NX-API – NX-API improves the accessibility of CLI by making it available outside of the switch by using HTTP/HTTPS. This feature helps with configuring the Cisco Nexus switch remotely using the automation framework.

      Feature UDLD – Enables unidirectional link detection for various interfaces.

Cisco UCS Fabric Interconnect 6454 Ethernet Connectivity

Cisco UCS 6454 FIs are connected to Cisco Nexus 93180YC-FX3 switches using 100GE connections configured as virtual port channels. Each FI is connected to both Cisco Nexus switches using a 100G connection; additional links can easily be added to the port channel to increase the bandwidth as needed. Figure 20 illustrates the physical connectivity details.

Figure 20.       Cisco UCS 6454 FI Ethernet Connectivity

Cisco UCS 6454 FI Ethernet Connectivity

NetApp AFF A400 Ethernet Connectivity

NetApp AFF A400 controllers are connected to Cisco Nexus 93180YC-FX3 switches using 25GE connections configured as virtual port channels. The storage controllers are deployed in a switchless cluster configuration and are connected to each other using the 100GE ports e3a and e3b. Figure 21 illustrates the physical connectivity details.

In Figure 21, the two storage controllers in the high-availability pair are drawn separately for clarity. Physically, the two controllers exist within a single chassis.

Figure 21.       NetApp AFF A400 Ethernet Connectivity

A screenshot of a computerDescription automatically generated with medium confidence

Cisco MDS SAN Connectivity – Fibre Channel Design Only

The Cisco MDS 9132T is the key design component bringing together the 32Gbps Fibre Channel (FC) capabilities to the FlexPod design. A redundant 32 Gbps Fibre Channel SAN configuration is deployed utilizing two MDS 9132Ts switches. Some of the key MDS features implemented within the design are:

      Feature NPIV – N port identifier virtualization (NPIV) provides a means to assign multiple FC IDs to a single N port.

      Feature fport-channel-trunk – F-port-channel-trunks allow for the fabric logins from the NPV switch to be virtualized over the port channel. This provides nondisruptive redundancy should individual member links fail.

      Smart-Zoning – a feature that reduces the number of TCAM entries by identifying the initiators and targets in the environment.

Cisco UCS Fabric Interconnect 6454 SAN Connectivity

For SAN connectivity, each Cisco UCS 6454 Fabric Interconnect is connected to a Cisco MDS 9132T SAN switch using 2 x 32G Fibre Channel port-channel connection, as shown in Figure 22.

Figure 22.       Cisco UCS 6454 FI SAN Connectivity

Cisco UCS 6454 FI SAN Connectivity

NetApp AFF A400 SAN Connectivity

For SAN connectivity, each NetApp AFF A400 controller is connected to both of Cisco MDS 9132T SAN switches using 32G Fibre Channel connections, as shown in Figure 23.

In Figure 23, the two storage controllers in the high-availability pair are drawn separately for clarity. Physically, the two controllers exist within a single chassis.

Figure 23.       NetApp AFF A400 SAN Connectivity

NetApp AFF A400 SAN Connectivity

Cisco UCS X-Series Configuration - Cisco Intersight Managed Mode

Cisco Intersight Managed Mode standardizes policy and operation management for Cisco UCS X-Series. The compute nodes in Cisco UCS X-Series are configured using server profiles defined in Cisco Intersight. These server profiles derive all the server characteristics from various policies and templates. At a high level, configuring Cisco UCS using Intersight Managed Mode consists of the steps shown in Figure 24.

Figure 24.       Configuration Steps for Cisco Intersight Managed Mode

Configuration Steps for Cisco Intersight Managed Mode

Set up Cisco UCS Fabric Interconnect for Cisco Intersight Managed Mode

During the initial configuration, for the management mode the configuration wizard enables customers to choose whether to manage the fabric interconnect through Cisco UCS Manager or the Cisco Intersight platform. Customers can switch the management mode for the fabric interconnects between Cisco Intersight and Cisco UCS Manager at any time; however, Cisco UCS FIs must be set up in Intersight Managed Mode (IMM) for configuring the Cisco UCS X-Series system. Figure 25 shows the dialog during initial configuration of Cisco UCS FIs for setting up IMM.

Figure 25.       Fabric Interconnect Setup for Cisco Intersight Managed Mode

Fabric Interconnect setup for Cisco Intersight Managed Mode

Claim a Cisco UCS Fabric Interconnect in the Cisco Intersight Platform

After setting up the Cisco UCS fabric interconnect for Cisco Intersight Managed Mode, FIs can be claimed to a new or an existing Cisco Intersight account. When a Cisco UCS fabric interconnect is successfully added to the Cisco Intersight platform, all future configuration steps are completed in the Cisco Intersight portal.

Figure 26.       Cisco Intersight: Adding Fabric Interconnects

Cisco Intersight: adding fabric interconnects as targets

Customers can verify whether a Cisco UCS fabric interconnect is in Cisco UCS Manager managed mode or Cisco Intersight Managed Mode by clicking on the fabric interconnect name and looking at the detailed information screen for the FI, as shown in Figure 27.

Figure 27.       Cisco UCS FI in Intersight Managed Mode

Cisco UCS FI in Intersight Managed Mode

Cisco UCS Chassis Profile

A Cisco UCS Chassis profile configures and associate chassis policy to an IMM claimed chassis. The chassis profile feature is available in Intersight only if customers have installed the Intersight Essentials License. The chassis-related policies can be attached to the profile either at the time of creation or later.

The chassis profile in a FlexPod is used to set the power policy for the chassis. By default, Cisco UCS X-Series power supplies are configured in GRID mode, but the power policy can be utilized to set the power supplies in non-redundant or N+1/N+2 redundant modes.

Cisco UCS Domain Profile

A Cisco UCS domain profile configures a fabric interconnect pair through reusable policies, allows configuration of the ports and port channels, and configures the VLANs and VSANs to be used in the network. It defines the characteristics of and configures the ports on the fabric interconnects. One Cisco UCS domain profile can be assigned to one fabric interconnect domain, and the Cisco Intersight platform supports the attachment of one port policy per Cisco UCS domain profile.

Some of the characteristics of the Cisco UCS domain profile in the FlexPod environment are:

      A single domain profile is created for the pair of Cisco UCS fabric interconnects.

      Unique port policies are defined for the two fabric interconnects.

      The VLAN configuration policy is common to the fabric interconnect pair because both fabric interconnects are configured for same set of VLANs.

      The VSAN configuration policies (FC connectivity option) are unique for the two fabric interconnects because the VSANs are unique.

      The Network Time Protocol (NTP), network connectivity, and system Quality-of-Service (QoS) policies are common to the fabric interconnect pair.

After the Cisco UCS domain profile has been successfully created and deployed, the policies including the port policies are pushed to Cisco UCS fabric interconnects. Cisco UCS domain profile can easily be cloned to install additional Cisco UCS systems. When cloning the UCS domain profile, the new UCS domains utilize the existing policies for consistent deployment of additional Cisco UCS systems at scale.

Figure 28.       Cisco UCS Domain Profile

Cisco UCS domain profile

The Cisco UCS X9508 Chassis and Cisco UCS X210c M6 Compute Nodes are automatically discovered when the ports are successfully configured using the domain profile as shown in Figure 29, Figure 30, and Figure 31.

Figure 29.       Cisco UCS X9508 Chassis Front View

Cisco UCS X9508 Chassis front view

Figure 30.       Cisco UCS X9508 Chassis Rear View

Cisco UCS X9508 Chassis rear view

Figure 31.       Cisco UCS X210c M6 Compute Nodes

Cisco UCS X210c M6 Compute Nodes

Server Profile Template

A server profile template enables resource management by simplifying policy alignment and server configuration. A server profile template is created using the server profile template wizard. The server profile template wizard groups the server policies into the following four categories to provide a quick summary view of the policies that are attached to a profile:

      Compute policies: BIOS, boot order, and virtual media policies

      Network policies: adapter configuration, LAN connectivity, and SAN connectivity policies

  The LAN connectivity policy requires you to create Ethernet network policy, Ethernet adapter policy, and Ethernet QoS policy.

  The SAN connectivity policy requires you to create Fibre Channel (FC) network policy, Fibre Channel adapter policy, and Fibre Channel QoS policy. SAN connectivity policy is only required for the FC connectivity option.

      Storage policies: not used in FlexPod

      Management policies: device connector, Intelligent Platform Management Interface (IPMI) over LAN, Lightweight Directory Access Protocol (LDAP), local user, network connectivity, Simple Mail Transfer Protocol (SMTP), Simple Network Management Protocol (SNMP), Secure Shell (SSH), Serial over LAN (SOL), syslog, and virtual Keyboard, Video, and Mouse (KVM) policies

Some of the characteristics of the server profile template for FlexPod are as follows:

      BIOS policy is created to specify various server parameters in accordance with FlexPod best practices.

      Boot order policy defines virtual media (KVM mapper DVD), all SAN paths for NetApp iSCSI or Fibre Channel logical interfaces (LIFs), and UEFI Shell.

      IMC access policy defines the management IP address pool for KVM access.

      Local user policy is used to enable KVM-based user access.

      For the iSCSI boot from SAN configuration, LAN connectivity policy is used to create six virtual network interface cards (vNICs) — two for management virtual switch (vSwitch0), two for application Virtual Distributed Switch (VDS), and one each for iSCSI A/B vSwitches. Various policies and pools are also created for the vNIC configuration.

      For the FC boot from SAN configuration, LAN connectivity policy is used to create four virtual network interface cards (vNICs) — two for management virtual switches (vSwitch0) and two for application Virtual Distributed Switch (VDS) — along with various policies and pools.

      For the FC connectivity option, SAN connectivity policy is used to create two virtual host bus adapters (vHBAs) — one for SAN A and one for SAN B — along with various policies and pools. The SAN connectivity policy is not required for iSCSI setup.

Figure 32 shows various policies associated with the server profile template.

Figure 32.       Server Profile Template for Fibre Channel Boot from SAN

A screenshot of a computer screenDescription automatically generated with medium confidence

Derive and Deploy Server Profiles from the Cisco Intersight Server Profile Template

The Cisco Intersight server profile allows server configurations to be deployed directly on the compute nodes based on polices defined in the server profile template. After a server profile template has been successfully created, server profiles can be derived from the template and associated with the Cisco UCS X210c M6 Compute Nodes, as shown in Figure 33.

Figure 33.       Deriving a Server Profile from Templates

A screenshot of a computer screenDescription automatically generated with medium confidence

On successful deployment of the server profile, the Cisco UCS X210c M6 Compute Nodes are configured with parameters defined in the server profile and can boot from the storage LUN hosted on NetApp AFF A400.

NetApp AFF A400 – Storage Virtual Machine (SVM) Design

To provide the necessary data segregation and management, a dedicated SVM, Infra-SVM, is created for hosting the Epic environment. The SVM contains the following volumes and logical interfaces (LIFs):

      Volumes

  ESXi boot LUNs used to enable ESXi host boot from SAN functionality using FC

  Infrastructure datastores used by the vSphere environment to store the VMs and swap files

      Logical interfaces (LIFs)

  NFS LIFs to mount NFS datastores in the vSphere environment

  FC LIFs for supporting FC Boot-from-SAN traffic

  CIFS LIFs to mount CIFS shares in the vSphere environment

Storage Consolidation

Combining Epic NAS and SAN workloads on the same all-flash array is supported and recommended to simplify data management having a single storage platform. This allows you to use a single data management strategy for all workloads regardless of access protocol.

Performance and Scale

WebBLOB is typically made up of lots of small files with documents and images exported from the production Iris ODB database. Depending on what document management application is used, WebBLOB can grow to hundreds of millions of files. High file count can impact metadata processing and frequently causing outage.

FlexGroups scales performance, metadata processing and capacity using multiple volumes across the entire cluster to up to 24 nodes. A single names space that scales to >20PB and >400 billion files. FlexGroups illuminates any issues with scale with high file count resulting in higher availability. NetApp recommends using FlexGroups for WebBLOB and other critical shares like homedir and profiles that make up an Epic environment. FlexGroups is built into ONTAP and recommended by NetApp.

Figure 34.      FlexGroups within ONTAP

DiagramDescription automatically generated

Reduce Cost

WebBLOB is ~95% cold data and a low performing tier and does not have to be deployed on all flash. NetApp offers options to reduce the cost of NAS while delivering on performance and capability with several options:

      FAS using optimized read and write FlashPool SSD’s is an excellent solution for NAS. FAS and AFF systems can be part of the same cluster. FlashPool is built into ONTAP and recommended by NetApp when using FAS.

      NAS deployed on all-flash and automatically tier backups and cold data to object storage on-premises or in the cloud using the FabricPool feature. All of which can be accomplished without any noticeable performance effect. You can generate a cold data report to review how cold your data is before enabling automatic tiering. You can customize the age of the data to tier through a policy. Epic customers have realized significant savings with this feature. FabricPool is built into ONTAP and optionally used to reduce cost.

Figure 35.       NetApp FabricPool Feature

DiagramDescription automatically generated

Disaster Recovery Options

ONTAP offers many levels of RPO, and RTO as shown below. At a minimum, the NAS solution should use async replication to DR every 15 minutes and the ability to failover quickly. This can be done by enabling SnapMirror. Since ONTAP controls the writes to storage for NAS, each snapshot is application consistent. Breaking the SnapMirror makes the DR read only copy writable. Sync replication is available but not required and not recommended for WebBLOB because of higher cost of disk.

Figure 36.       ONTAP offering multiple levels of RPO and RTO

Chart, timelineDescription automatically generated

To understand the simplification of NAS DR failover, see Figure 37.

Figure 37.       Simplifying NAS DR failover

DiagramDescription automatically generated

Details on volumes, VLANs, and logical interfaces (LIFs) are shown in Figure 38 for FC connectivity.

Figure 38.       NetApp AFF A400 – Infra-SVM

Graphical user interface, applicationDescription automatically generated

VMware vSphere – ESXi Design

Multiple vNICs (and vHBAs) are created for the ESXi hosts using the Cisco Intersight server profile and are then assigned to specific virtual and distributed switches. The vNIC and vHBA distribution for the ESXi hosts is as follows:

      Two vNICs (one on each fabric) for vSwitch0 to support core services such as management and NFS traffic.

      Two vNICs (one on each fabric) for vSphere Virtual Distributed Switch (VDS) to support customer data traffic and vMotion traffic.

      One vHBA each for Fabric-A and Fabric-B for FC stateless boot and datastore access. Figure 39 shows the ESXi vNIC configurations in detail.

Figure 39.       VMware vSphere – ESXi Host Networking for FC Boot from SAN

VMware vSphere – ESXi Host Networking for FC Boot from SAN

Cisco Intersight Integration with VMware vCenter and NetApp Storage

Cisco Intersight works with NetApp’s ONTAP storage and VMware vCenter using third-party device connectors. Since third-party infrastructure does not contain any built-in Intersight device connector, Cisco Intersight Assist virtual appliance enables Cisco Intersight to communicate with non-Cisco devices.

Note:   A single Cisco Intersight Assist virtual appliance can support both NetApp ONTAP storage and VMware vCenter.

Cisco Intersight integration with VMware vCenter and NetApp ONTAP enables customers to perform following tasks right from the Intersight dashboard:

      Monitor the virtualization and storage environment.

      Add various dashboard widgets to obtain useful at-a-glance information.

      Perform common Virtual Machine tasks such as power on/off, remote console and so on.

      Orchestrate virtual and storage environment to perform common configuration tasks.

      Orchestrate NetApp ONTAP storage tasks to setup a Storage Virtual Machine and provide NAS and SAN services.

The following sections explain the details of these operations. Since Cisco Intersight is a SaaS platform, the monitoring and orchestration capabilities are constantly being added and delivered seamlessly from the cloud.

Note:   The monitoring capabilities and orchestration tasks and workflows listed below provide an in-time snapshot for your reference. For the most up to date list of capabilities and features, customers should use the help and search capabilities in Cisco Intersight.

Figure 40.       Managing NetApp and VMware vCenter through Cisco Intersight using Intersight Assist

Managing NetApp and VMware vCenter through Cisco Intersight using Intersight Assist

Licensing Requirement

To integrate and view various NetApp storage and VMware vCenter parameters from Cisco Intersight, a Cisco Intersight Advantage license is required. To use Cisco Intersight orchestration and workflows to provision the storage and virtual environments, an Intersight Premier license is required.

Integrate Cisco Intersight with NetApp ONTAP Storage

To integrate NetApp AFF A400 with Cisco Intersight, customers need to deploy:

      Cisco Intersight Assist virtual appliance

      NetApp Active IQ Unified Manager virtual appliance

Using the Cisco Intersight Assist, NetApp Active IQ Unified Manager is claimed as a target in Cisco Intersight, as shown in Figure 41.

Figure 41.       Claiming NetApp Active IQ Unified Manager as a Target in Cisco Intersight

Claiming NetApp Active IQ Unified Manager as a target in Cisco Intersight

Obtain Storage-level Information

After successfully claiming the NetApp Active IQ Unified Manager as a target, customers can view storage-level information in Cisco Intersight if they have already added NetApp AFF A400 to the NetApp Active IQ Unified Manager.

Figure 42.       NetApp AFF A400 Information in Cisco Intersight

Graphical user interface, application, TeamsDescription automatically generated

Table 2 lists some of the core NetApp AFF A400 information presented through Cisco Intersight.

Table 2.     NetApp Storage Information in Cisco Intersight

Category

Name

Detail

General

Name

Name of the controller

Vendor

NetApp

 Model

 NetApp AFF model information (for example, AFF-A400)

Version

Software version

Monitoring

 Capacity

Total, used, and available system capacity

 

Summary of Nodes, Storage VMs, Aggregates, disks and so on. in the system

Inventory

Volumes

Volumes defined in the system and their status, size, usage, and configured export policies

LUNs

LUNs defined in the system and their status, size, usage, and mapped iGroups

Aggregates

Configured aggregates and their status, size, usage, and space savings

Storage VMs

 Storage VM (SVM) information, state, allowed protocols, and logical ethernet and fibre channel interface details.

Export policies

Export policies defined in the system and the associated SVMs

SAN initiator groups

SAN initiator groups, their type, protocol, initiator information, and associated SVMs

Licenses

Licenses installed on the system

Nodes

Controller information, such as model, OS, serial number, and so on.

Disks

Disk information, including type, model, size, node information, status of the disks, and aggregate details

Ports

Ethernet and FC ports configured on the system

Storage Widget in the Dashboard

Customers can also add the storage dashboard widgets to Cisco Intersight for viewing NetApp AFF A400 at a glance information on the Cisco Intersight dashboard, as shown in Figure 43.

Figure 43.       Storage Widgets in Cisco Intersight Dashboard

A screenshot of a computerDescription automatically generated with medium confidence

These storage widgets provide useful information such as:

      Storage arrays and capacity utilization

      Top-five storage volumes by capacity utilization

      Storage versions summary, providing information about the software version and the number of storage systems running that version

Cisco Intersight Orchestrator – NetApp ONTAP Storage

Cisco Intersight Orchestrator provides various workflows that can be used to automate storage provisioning. Some of the sample storage workflows available for NetApp ONTAP storage are listed in Table 3.

Table 3.      NetApp ONTAP Storage Workflows in Cisco Intersight Orchestrator

Name

Details

New NAS datastore

Create a NFS storage volume and build NAS datastore on the volume

New storage export policy

Create a storage export policy and add the created policy to a NFS volume

New storage host

Create a new storage host or igroup to enable SAN mapping

New storage interface

Create a storage IP or FC interface

New storage virtual machine

Create a storage virtual machine

New VMFS datastore

Create a storage volume and build a Virtual Machine File System (VMFS) datastore on the volume

Remove NAS datastore

Remove the NAS datastore and the underlying NFS storage volume

Remove storage export policy

Remove the NFS volume and the export policy attached to the volume

Remove storage host

Remove a storage host. If a host group name is provided as input, the workflow will also remove the host from the host group.

Remove VMFS datastore

Remove a VMFS data store and remove the backing volume from the storage device

Update NAS datastore

Update NAS datastore by expanding capacity of the underlying NFS volume

Update storage host

Update the storage host details. If the inputs for a task are provided, then the task is run; otherwise, it is  skipped.

Update VMFS datastore

Expand a datastore on the hypervisor manager by extending the backing storage volume to specified capacity, and then expand the data store to use the additional capacity

In addition to the above workflows, Cisco Intersight Orchestrator also provides many storage and virtualization tasks for customers to create custom workflow based on their specific needs. A sample subset of these tasks is highlighted in Figure 44.

Figure 44.       Storage Tasks for NetApp ONTAP

Storage Tasks for NetApp ONTAP

Integrate Cisco Intersight with VMware vCenter

To integrate VMware vCenter with Cisco Intersight, VMware vCenter can be claimed as a target using Cisco Intersight Assist Virtual Appliance, as shown in Figure 45.

Figure 45.       Claim VMware vCenter in Cisco Intersight as a Target

Claiming VMware vCenter in Cisco Intersight as a target

Obtain Hypervisor-level Information

After successfully claiming the VMware vCenter as a target, customers can view hypervisor-level information in Cisco Intersight including hosts, VMs, clusters, datastores, and so on.

Figure 46.       VMware vCenter Information in Cisco Intersight

VMware vCenter Information in Cisco Intersight

Table 4 lists some of the main virtualization properties presented in Cisco Intersight.

Table 4.     Virtualization (VMware vCenter) Information in Cisco Intersight

Category

Name

Details

General

Name

Name of the data center

Hypervisor manager

Host name or IP address of the vCenter

Clusters

Name

Name of the cluster

Data center

Name of the data center

Hypervisor type

ESXi

Hypervisor manager

vCenter IP address or the host name

CPU capacity

CPU capacity in the cluster (GHz)

CPU consumed

CPU cycles consumed by workloads (percentage and GHz)

Memory capacity

Total memory in the cluster (GB)

Memory consumed

Memory consumed by workloads (percentage and GB)

Total cores

All the CPU cores across the CPUs in the cluster

VMware cluster information allows customers to access additional details about hosts and virtual machines associated with the cluster.

 

Hosts

Name

Host name or IP address

Server

Server profile associated with the ESXi host

Cluster

Cluster information if the host is part of a cluster

Data center

VMware data center

Hypervisor type

ESXi

Hypervisor manager

vCenter IP address of host name

Uptime

Host uptime

Virtual Machines

Number and state of VMs running on a host

CPU Information

CPU cores, sockets, vendor, speed, capacity, consumption, and other CPU related information

Memory Information

Memory capacity and consumption information

Hardware Information

Compute node hardware information such as serial number, model and so on.

Host information allows customers to access additional details about clusters, VMs, datastores, and networking related to the current ESXi host.

 

Virtual machines

Name

Name of the VM

Guest OS

Operating system, for example, RHEL, CentOS, and so on.

Hypervisor type

ESXi

Host

ESXi host information for the VM

Cluster

VMware cluster name

Data center

VMware data center name

IP address

IP address(s) assigned to the VM

Hypervisor manager

IP address of host name of the vCenter

Resource Information

CPU, memory, disk, and network information

Guest Information

Hostname, IP address and operating system information

VM information allows customers to access additional details about clusters, hosts, datastores, networking, and virtual disks related to the current VM.

 

Datastores

Name

Name of the datastore in VMware vCenter

Type

NFS or VMFS and so on.

Accessible

Yes, if datastore is accessible; No, if datastore is inaccessible

Thin provisioning

Yes, if thin provisioning is allowed; No if thin provisioning is not allowed

Multiple host access

Yes, if multiple hosts can mount the datastore; No, if the datastore only allows a single host

Storage capacity

Space in GB or TB

Storage consumes

Percentage and GB

Data center

Name of VMware vCenter data center

Hypervisor manager

vCenter hostname or IP address

Datastore Cluster

Datastore cluster information if datastore cluster is configured

Hosts and Virtual Machines

Number if hosts connected to a datastore and number of VM hosted on the datastore

Datastore information allows customers to access additional details about hosts and VMs associated with the datastore.

 

Interact with Virtual Machines

VMware vCenter integration with Cisco Intersight allows customers to directly interact with the virtual machines (VMs) from the Cisco Intersight dashboard. In addition to obtaining in-depth information about a VM, including the operating system, CPU, memory, host name, and IP addresses assigned to the virtual machines, customers can use Intersight to perform following actions on the virtual machines:

      Launch VM console

      Power off

      Reset

      Shutdown guest OS

      Restart guest OS

      Suspend

Figure 47.       Virtual Machine Actions in Cisco Intersight

Virtual machine actions in Cisco Intersight

Cisco Intersight Orchestrator – VMware vCenter

Cisco Intersight Orchestrator provides various workflows that can be used for the VM and hypervisor provisioning. Some of the sample workflows available for VMware vCenter are captured in Table 5.

Table 5.     VMware vCenter Workflows in Cisco Intersight Orchestrator

Name

Details

New NAS Datastore

Create a NFS storage volume and build NAS datastore on the volume.

New VMFS Datastore

Create a storage volume and build VMFS datastore on the volume.

New Virtual Machine

Create a new virtual machine on the hypervisor from an OVA or OVF file. Datastore, Host/Cluster, and Image URL fields are mandatory. All other inputs are optional.

Remove NAS Datastore

Remove the NAS datastore and the underlying NFS storage volume.

Remove VMFS Datastore

Remove VMFS datastore and remove the backing volume from the storage device.

Update NAS Datastore

Update NAS datastore by expanding capacity of the underlying NFS volume.

Update VMFS Datastore

Expand a datastore on hypervisor manager by extending the backing storage volume to specified capacity, and then grow the datastore to utilize the additional capacity.

In addition to the above workflows, Cisco Intersight Orchestrator provides many tasks for customers to create custom workflows depending on their specific requirements. A sample subset of these tasks is highlighted in Figure 48.

Figure 48.       VMware vCenter Tasks in Cisco Intersight Orchestrator

VMware vCenter Tasks in Cisco Intersight Orchestrator

Deployment Hardware and Software

Architecture

FlexPod is a defined set of hardware and software that serves as an integrated foundation for both virtualized and non-virtualized solutions. VMware vSphere® built on FlexPod includes NetApp AFF storage, Cisco Nexus® networking, Cisco MDS storage networking, the Cisco Unified Computing System (Cisco UCS®), and VMware vSphere software in a single package. The design is flexible enough that the networking, computing, and storage can fit in one data center rack or be deployed according to a customer’s data center design. Port density enables the networking components to accommodate multiple configurations of this kind.

One benefit of the FlexPod architecture is the ability to customize or “flex” the environment to suit a customer’s requirements. A FlexPod can easily be scaled as requirements and demand change. The unit can be scaled both up (adding resources to a FlexPod unit) and out (adding more FlexPod units). The reference architecture detailed in this document highlights the resiliency, cost benefit, and ease of deployment of a Fibre Channel and IP-based storage solution. A storage system capable of serving multiple protocols across a single interface allows for customer choice and investment protection because it truly is a wire-once architecture.

Epic is prescriptive as to its customers’ hardware requirements because of an overarching requirement for delivering predictable low-latency system performance and high availability. FlexPod, a validated, rigorously tested converged infrastructure from the strategic partnership of Cisco and NetApp, is engineered and designed specifically for delivering predictable low-latency system performance and high availability. This approach results in Epic high comfort levels and ultimately the best response time for users of the Epic EHR system. The FlexPod solution from Cisco and NetApp meets Epic system requirements with a high performing, modular, validated, converged, virtualized, efficient, scalable, and cost-effective platform.

Figure 49 shows the VMware vSphere built on FlexPod components and the network connections for a configuration with the Cisco UCS 6454 Fabric Interconnects. This design has port-channeled 25 Gb Ethernet connections between the Cisco UCS X9508 Blade Chassis and the Cisco UCS Fabric Interconnects via the Cisco UCS 9108 25G Intelligent Fabric Modules, and port-channeled 25 Gb Ethernet connections between the Cisco UCS Fabric modules and Cisco Nexus 9000s, and between the Cisco Nexus 9000s and NetApp AFF A400 storage array. This infrastructure option expanded with Cisco MDS switches sitting between the Cisco UCS Fabric Interconnects and the NetApp AFF A400 to provide FC-booted hosts with 32 Gb FC block-level access to shared storage. The reference architecture reinforces the “wire-once” strategy, because as additional storage is added to the architecture, no re-cabling is required from the hosts to the Cisco UCS fabric interconnects.

Topology

Figure 49.       FlexPod with Cisco UCS 6454 Fabric Interconnects and NetApp AFF A-Series

DiagramDescription automatically generated

The reference hardware configuration includes:

      Two Cisco Nexus 93180YC-FX3 switches

      Two Cisco UCS 6454 fabric interconnects

      One Cisco UCS X9508 Chassis

      Two Cisco MDS 9132T multilayer fabric switches

      One NetApp AFF A400 (HA pair) running ONTAP 9.10.1P1 with NVMe SSD disks

Hardware and Software Versions

Table 6 lists the hardware and software revisions for this solution. It is important to note that the validated FlexPod solution explained in this document adheres to Cisco, NetApp, and VMware interoperability matrix to determine support for various software and driver versions. Customers can use the same interoperability matrix to determine support for components that are different from the current validated design.

Click the following links for more information:

      NetApp Interoperability Matrix Tool: http://support.netapp.com/matrix/

      Cisco UCS Hardware and Software Interoperability Tool: http://www.cisco.com/web/techdoc/ucs/interoperability/matrix/matrix.html

      VMware Compatibility Guide: http://www.vmware.com/resources/compatibility/search.php

Table 6.     Hardware and Software Revisions

Compute

Device

Image

Comments

Compute

Cisco UCS X210c M6 Servers

5.0(1b)

 

Cisco UCS Fabric Interconnects 6454

9.3(5)I42(1g)

 

Network

Cisco Nexus 93180YC-FX3 NX-OS

9.3(8)

 

Cisco MDS 9132T

8.4(2c)

 

Storage

NetApp AFF A400

ONTAP 9.10.1P1

 

Software

Cisco Data Center Network Manager (SAN)

11.5(1)

 

Cisco Intersight Assist Appliance

1.0.9-342

 

VMware vCenter Appliance

7.0 Update 3

 

VMware ESXi

7.0 U3

 

VMware ESXi nfnic FC Driver

5.0.0.12

Supports FC-NVMe

VMware ESXi nenic Ethernet Driver

1.0.35.0

 

NetApp ONTAP Tools for VMware vSphere

9.10

formerly Virtual Storage Console (VSC)

NetApp NFS Plug-in for VMware VAAI

2.0-15

 

NetApp SnapCenter for VMware vSphere

4.6

Includes the vSphere plug-in for SnapCenter

NetApp Active IQ Unified Manager

9.10P1

 

Management

Cisco Intersight Assist Virtual Appliance

1.0.9-342

 

NetApp Active IQ

N/A

 

Configuration Guidelines

This document explains how to configure a fully redundant, highly available configuration for a FlexPod unit with ONTAP storage. Therefore, reference is made to which component is being configured with each step, either 01 or 02 or A and B. For example, node01 and node02 are used to identify the two NetApp storage controllers that are provisioned with this document, and Cisco Nexus A or Cisco Nexus B identifies the pair of Cisco Nexus switches that are configured. The Cisco UCS Fabric Interconnects are similarly configured. Additionally, this document details the steps for provisioning multiple Cisco UCS hosts, and these examples are identified as: VM-Host-Infra-01, VM-Host-Infra-02 to represent infrastructure hosts deployed to each of the fabric interconnects in this document. Finally, to indicate that you should include information pertinent to your environment in a given step, appears as part of the command structure. See the following example for the network port vlan create command:

Usage:

network port vlan create ?

 [-node] <nodename>                  Node

 { [-vlan-name] {<netport>|<ifgrp>}  VLAN Name

 |  -port {<netport>|<ifgrp>}        Associated Network Port

 [-vlan-id] <integer> }              Network Switch VLAN Identifier

Example:

network port vlan create -node <node01> -vlan-name a0a-<vlan id>

This document is intended to enable you to fully configure the customer environment. In this process, various steps require you to insert customer-specific naming conventions, IP addresses, and VLAN schemes, as well as to record appropriate MAC addresses. Table 7 lists the VLANs necessary for deployment as outlined in this guide.

Table 7.     Necessary VLANs

VLAN Name

VLAN Purpose

ID used in Validating this Document

Subnet Used in Validating This Document

Default Gateway used in Validating this Document

Native-Vlan

VLAN to which untagged frames are assigned

2

N/A

N/A

OOB-MGMT

VLAN for out-of-band management interfaces

40

10.10.40.0/24

10.10.40.1

IB-MGMT

VLAN for in-band management interfaces

41

10.10.41.0/24

10.10.41.1

Infra-NFS

VLAN for Infrastructure NFS traffic

42

10.10.42.0/24

10.10.42.1

vMotion

VLAN for VMware vMotion

43

10.10.43.0/24

10.10.43.1

VM-Traffic

VLAN for Production VM Interfaces

44

10.10.44.0/24

10.10.44.1

CIFS-Traffic

VLAN for CIFS/SMB VM traffic

45

10.10.45.0/24

10.10.45.1

FCoE-A

VLAN for FCoE encapsulation of VSAN-A

101

N/A

N/A

FCoE-B

VLAN for FCoE encapsulation of VSAN-B

102

N/A

N/A

Table 8.     Virtual Machines

Virtual Machine Description

Host Name

IP Address

vCenter Server

vcenter.hc.cvd

10.10.40.4

NetApp ONTAP Tools for VMware vSphere

ontap-tools.hc.cvd

10.10.40.5

NetApp SnapCenter for VMware vSphere

scv.hc.cvd

10.10.40.9

NetApp Active IQ Unified Manager

aiq.hc.cvd

10.10.40.7

Cisco Intersight Assist

intersight.hc.cvd

10.10.40.8

Physical Infrastructure

FlexPod Cabling

The information in this section is provided as a reference for cabling the physical equipment in a FlexPod environment. To simplify cabling requirements, a cabling diagram was used (see Figure 50).

Figure 50 illustrates the details for the prescribed and supported configuration of the NetApp AFF A400 running NetApp ONTAP 9.10.1.

Note:   For any modifications of this prescribed architecture, consult the NetApp Interoperability Matrix Tool (IMT).

This document assumes that out-of-band management ports are plugged into an existing management infrastructure at the deployment site. These interfaces will be used in various configuration steps.

Note:   Be sure to use the cabling directions in this section as a guide.

The NetApp storage controller and disk shelves should be connected according to best practices for the specific storage controller and disk shelves. For disk shelf cabling, refer to NetApp Support.

Figure 50 details the cable connections used in the validation lab for the FlexPod topology based on the Cisco UCS 6454 fabric interconnect. Two 32Gb uplinks connect as port-channels to each Cisco UCS Fabric Interconnect from the MDS switches, and a total of four 32Gb links connect the MDS switches to the NetApp AFF controllers. Also, 25Gb links connect the Cisco UCS Fabric Interconnects to the Cisco Nexus Switches and the NetApp AFF controllers to the Cisco Nexus Switches. Additional 1Gb management connections will be needed for an out-of-band network switch that sits apart from the FlexPod infrastructure. Each Cisco UCS fabric interconnect and Cisco Nexus switch is connected to the out-of-band network switch, and each AFF controller has a connection to the out-of-band network switch. Layer 3 network connectivity is required between the Out-of-Band (OOB) and In-Band (IB) Management Subnets.

Figure 50.       FlexPod Cabling with Cisco UCS 6454 Fabric Interconnect

DiagramDescription automatically generated

Automation Workflow and Solution Deployment

If using the published Automation scripts (Ansible playbooks and Terraform scripts) to configure the FlexPod infrastructure, complete this section of the document. If completing a manual configuration, skip to section FlexPod Cisco Nexus Switch Manual Configuration.

The Ansible automated FlexPod solution uses a management workstation (control machine) to run Ansible playbooks to configure Cisco Nexus, NetApp ONTAP Storage, Cisco MDS, VMware ESXi and vCenter, NetApp ONTAP Tools, and NetApp Active IQ Unified Manager.

The Terraform automated FlexPod solution uses a management workstation to run the Terraform configuration scripts to configure the Cisco UCS components (UCS Domain and UCS servers).

Figure 51 illustrates the FlexPod solution implementation workflow which is explained in the following sections.

Note:   The FlexPod infrastructure layers are configured in the order illustrated.

Figure 51.       FlexPod Automation Workflow

Related image, diagram or screenshot

Figure 52.       Automation Workflow

DiagramDescription automatically generated

Clone GitHub Collection

You need to use a GitHub repository from one public location; the first step in the process is to clone the GitHub collection named FlexPod-for-EHR (https://github.com/ucs-compute-solutions/FlexPod-for-EHR.git) to a new empty folder on the management workstation. Cloning the repository creates a local copy, which is then used to run the playbooks that have been created for this solution.

Procedure 1.     Clone the GitHub Repository

Step 1.    From the management workstation, create a new folder for the project. The GitHub collection will be cloned in a new folder inside this one, named FlexPod-for-EHR.

Step 2.    Open a command-line or console interface on the management workstation and change directories to the new folder just created.

Step 3.    Clone the GitHub collection using the following command:

git clone https://github.com/ucs-compute-solutions/FlexPod-for-EHR.git

Step 4.    Change directories to the new folder named FlexPod-for-EHR.

Ansible

Figure 53.       Ansible Automation Workflow

Ansible Automation Workflow

Prerequisites

Setup of the solution begins with a management workstation that has access to the Internet and with a working installation of Ansible. The management workstation commonly runs a variant of Linux or MacOS for ease of use with these command-line-based tools. Instructions for installing the workstation are not included in this document, but basic installation and configuration of Ansible is covered. A guide for getting started with Ansible can be found at the following link:

      Getting Started with Red Hat Ansible: https://www.ansible.com/resources/get-started

      To use the Ansible playbooks demonstrated in this document, the management workstation must also have a working installation of Git and access to the Cisco DevNet public GitHub repository. The Ansible playbooks used in this document are cloned from the public repositories, located at the following links:

  Cisco DevNet: https://developer.cisco.com/codeexchange/github/repo/ucs-compute-solutions/FlexPod-for-EHR

  GitHub repository: https://github.com/ucs-compute-solutions/FlexPod-for-EHR

      The Cisco Nexus and MDS Switches, NetApp Storage and Cisco UCS must be physically racked, cabled, powered, and configured with management IP addresses before the Ansible-based installation procedure can begin as shown in the cabling diagram Figure 50. If necessary, upgrade the Cisco Nexus Switches to release 9.3(8) and the Cisco UCS Interconnects to 9.3(5)I42(1g) the packages for blades servers set to 5.0(1b).

      Before running each Ansible Playbook to setup the Network, Storage controllers, and VMware ESXi, various variables must be updated based on the customers environment and specific implementation with values such as the VLANs, UCS node WWPNs, IP addresses for NFS interfaces and values needed for VMware ESXi.

Procedure 1.     Prepare Management Workstation (Control Machine)

Note:     These installation steps are performed on the CentOS management host to prepare the host for solution deployment to support the automation of Cisco Nexus, NetApp Storage, Cisco MDS and VMware ESXi using Ansible Playbooks.

Note:   The following steps were performed on a CentOS 8.4 Virtual Machine as the root user.

Step 1.    Install EPEL repository on the management host.

dnf install https://dl.fedoraproject.org/pub/epel/epel-release-latest-8.noarch.rpm

Step 2.    Install Ansible engine.

dnf install ansible

Step 3.    Verify Ansible version to make sure it is release 2.9 or later.

ansible --version

ansible 2.9.23

  config file = /etc/ansible/ansible.cfg

  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']

  ansible python module location =
/usr/lib/python3.6/site-packages/ansible

  executable location = /usr/bin/ansible

  python version = 3.6.8 (default, Mar 19 2021, 08:58:41) [GCC 8.4.1 20200928 (Red Hat 8.4.1-1)]

Step 4.    Install NetApp specific python modules.

pip3 install netapp-lib

Step 5.    Install ansible-galaxy collections for Cisco Nexus, NetApp ONTAP, and VMware as follows:

ansible-galaxy collection install cisco.nxos
ansible-galaxy collection install netapp.ontap
ansible-galaxy collection install community.vmware

Step 6.    Optional: These collections can also be automatically installed using the requirements.yml file in the root of the Ansible repository with the command:

ansible-galaxy collection install -r requirements.yml

Terraform

Terraform is an open-source infrastructure as code software tool that enables you to create, change, and improve infrastructure safely and predictably.

Terraform helps with:

      Increased agility with reduced time to provision from weeks to minutes with automated workflow

      Control costs systematically as users and applications scale

      Reduce risk and discover errors before they happen with code reviews and embed provisioning guardrails

Terraform Providers

Providers are plugins that implement resource types likes Intersight. Terraform CLI finds and installs providers when initializing a working directory. It can automatically download providers from a Terraform registry or load them from a local mirror or cache.

Why Terraform provider for the Cisco Intersight?

The Cisco Intersight platform supports the Terraform provider. The Terraform provider allows organizations to develop Cisco Intersight resources as self-service infrastructure using code rather than manual provisioning. This approach provides several benefits:

      You can more quickly and easily scale Cisco Intersight resources. You can provision infrastructure in minutes, with little effort, using the automated workflows, performing the same tasks that used to take days.

      The operating model of Terraform is well suited for the Cisco Intersight platform, because it accommodates the shift from static to dynamic infrastructure provisioning. For example, if a resource is deleted in the Terraform configuration, it will be reflected in the Cisco Intersight platform when the new configuration is applied.

      Terraform maintains a state file, which is a record of the currently provisioned resources. State files provide a version history of Cisco Intersight resources, enabling a detailed audit trail of changes.

      The provider enables idempotency, producing the same result and state with repeated API calls.

      The set of files used to describe infrastructure in Terraform is known as a Terraform configuration. The configuration is written using HashiCorp Configuration Language (HCL), a simple human-readable configuration language, to define a desired topology of infrastructure resources.

Prerequisites

Setup of the solution begins with a management workstation that has access to the Internet and with a working installation of Terraform. The management workstation commonly runs a variant of Windows, Linux or MacOS for ease of use with these command-line-based tools. Instructions for installing the workstation are not included in this document, but basic installation and configuration of Terraform is covered.

      Introduction to Terraform: https://www.terraform.io/intro/index.html

      To use the Terraform scripts demonstrated in this document, the management workstation must also have a working installation of Git and access to the Cisco DevNet public GitHub repository. The Terraform scripts used in this document are cloned from the public repositories, located at the following links:

  Cisco DevNet: https://developer.cisco.com/codeexchange/explore/#search=ucs-compute-solutions

  GitHub repository: https://github.com/ucs-compute-solutions/FlexPod-for-EHR

      The Cisco UCS must be physically racked, cabled, powered, and configured with management IP addresses before the Terraform-based installation procedure can begin as shown in the cabling diagram (Figure 50).

      If necessary, upgrade the Cisco 6454 Fabric Interconnects to release 9.3(5)I42(1g) and the Cisco UCS Fabric Interconnects to 4.2(1f) with the firmware packages for blades servers set to 5.0(1b).

      Before running each Terraform script to setup the Cisco UCS, various variables have to be updated based on the customers environment and specific implementation with values such as the VLANs, pools and ports on Cisco UCS, IP addresses for NFS and iSCSI interfaces and values needed for VMware ESXi.

Prepare Management Workstation (Control Machine)

Instructions for installing Terraform on Windows, Linux or MacOS can be found at this link; https://learn.hashicorp.com/tutorials/terraform/install-cli

Verify that the installation worked by opening a new terminal session and listing Terraform’s available subcommands:

> terraform -help
Usage: terraform [global options] <subcommand> [args]

The available commands for execution are listed below.
The primary workflow commands are given first, followed by
less common or more advanced commands.

Main commands:
  init          Prepare your working directory for other commands
  validate      Check whether the configuration is valid
  plan          Show changes required by the current configuration
  apply         Create or update infrastructure
  destroy       Destroy previously-created infrastructure

All other commands:
  console       Try Terraform expressions at an interactive command prompt
  fmt           Reformat your configuration in the standard style
  force-unlock  Release a stuck lock on the current workspace
  get           Install or upgrade remote Terraform modules
  graph         Generate a Graphviz graph of the steps in an operation
  import        Associate existing infrastructure with a Terraform resource
  login         Obtain and save credentials for a remote host
  logout        Remove locally-stored credentials for a remote host
  output        Show output values from your root module
  providers     Show the providers required for this configuration
  refresh       Update the state to match remote systems
  show          Show the current state or a saved plan
  state         Advanced state management
  taint         Mark a resource instance as not fully functional
  test          Experimental support for module integration testing
  untaint       Remove the 'tainted' state from a resource instance
  version       Show the current Terraform version
  workspace     Workspace management

Global options (use these before the subcommand, if any):
  -chdir=DIR    Switch to a different working directory before executing the
                given subcommand.
  -help         Show this help output, or the help for a specified subcommand.
  -version      An alias for the "version" subcommand.

Terraform Provider Configuration

Each of the subdirectories for these steps include a main.tf file which is used to configure the providers Terraform uses to authenticate to Cisco Intersight. This is accomplished using API keys that must be created and saved for use within all four steps of this procedure.

Procedure 1.     Create Cisco Intersight API Keys

Step 1.    From the Cisco Intersight window, click Related image, diagram or screenshot and then click Settings.

Step 2.    From the API section, click API Keys.

API_Keys1

Step 3.    Click Generate API Key and enter a description and select Open API schema version 2.

generate_api_key

Step 4.    Use the icons on the right side of the dialog to copy the Secret Key as a text document and save the file in a location accessible on the Terraform workstation.

IMPORTANT! This is the only one time that the secret key can be viewed or downloaded. You cannot recover them later. However, you can create new access keys at any time.

API_key_blacked

Step 5.    Use the icon on the right side of the dialog to copy the API Key ID to the clipboard. Save this file SecretKet.txt to a location the Terraform scripts have access to.

Network Switch Configuration

This section provides a detailed procedure for configuring the Cisco Nexus 93180YC-FX switches for use in a FlexPod environment. The Cisco Nexus 93180YC-FX will be used for LAN switching in this solution.

Note:   Follow these steps precisely because failure to do so could result in an improper configuration.

Physical Connectivity

Follow the physical connectivity guidelines for FlexPod as explained in section FlexPod Cabling.

FlexPod Cisco Nexus Base

Before the Ansible Nexus switch setup playbook can be run, the Cisco Nexus switches must be brought up with a management IP address. The following procedures describe the basic configuration of the Cisco Nexus switches for use in a base FlexPod environment. This procedure assumes the use of Cisco Nexus 9000 9.3(8), the Cisco suggested Nexus switch release at the time of this validation.

Note:   The following procedure includes the setup of NTP distribution on both the mgmt0 port and the in-band management VLAN. The interface-vlan feature and ntp commands are used to set this up. This procedure also assumes that the default VRF is used to route the in-band management VLAN.

Note:   In this validation, port speed and duplex are hard set at both ends of every 100GE connection.

Note:   This validation assumes that both switches have been reset to factory defaults by using the “write erase” command followed by the “reload” command.

Procedure 1.     Set Up Initial Configuration

Cisco Nexus A

Note:   Set up the initial configuration for the Cisco Nexus A switch, from a serial console.

Step 1.    Configure the switch.

Note:   On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.

Abort Power On Auto Provisioning [yes - continue with normal setup, skip – bypass password and basic configuration, no - continue with Power On Auto Provisioning] (yes/skip/no)[no]: yes

Disabling POAP.......Disabling POAP

poap: Rolling back, please wait... (This may take 5-15 minutes)



         ---- System Admin Account Setup ----


Do you want to enforce secure password standard (yes/no) [y]: Enter

Enter the password for "admin": <password>

Confirm the password for "admin": <password>

Would you like to enter the basic configuration dialog (yes/no): yes

Create another login account (yes/no) [n]: Enter

Configure read-only SNMP community string (yes/no) [n]: Enter

Configure read-write SNMP community string (yes/no) [n]: Enter

Enter the switch name: <nexus-A-hostname>

Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: Enter

Mgmt0 IPv4 address: <nexus-A-mgmt0-ip>

Mgmt0 IPv4 netmask: <nexus-A-mgmt0-netmask>

Configure the default gateway? (yes/no) [y]: Enter

IPv4 address of the default gateway: <nexus-A-mgmt0-gw>
Configure advanced IP options? (yes/no) [n]: Enter

Enable the telnet service? (yes/no) [n]: Enter

Enable the ssh service? (yes/no) [y]: Enter

Type of ssh key you would like to generate (dsa/rsa) [rsa]: Enter

Number of rsa key bits <1024-2048> [1024]: Enter

Configure the ntp server? (yes/no) [n]: Enter

Configure default interface layer (L3/L2) [L2]: Enter

Configure default switchport interface state (shut/noshut) [noshut]: shut


Enter basic FC configurations (yes/no) [n]: n


Configure CoPP system profile (strict/moderate/lenient/dense) [strict]: Enter

Would you like to edit the configuration? (yes/no) [n]: Enter

Step 2.    Review the configuration summary before enabling the configuration.

Use this configuration and save it? (yes/no) [y]: Enter

Cisco Nexus B

Note:   Set up the initial configuration for the Cisco Nexus B switch, from a serial console.

Step 1.    Configure the switch.

Note:   On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.

Abort Power On Auto Provisioning  [yes - continue with normal setup, skip - bypass password and basic configuration, no - continue with Power On Auto Provisioning] (yes/skip/no) [no]: yes

Disabling POAP.......Disabling POAP

poap: Rolling back, please wait... (This may take 5-15 minutes)


         ---- System Admin Account Setup ----
 

Do you want to enforce secure password standard (yes/no) [y]: Enter

Enter the password for "admin": <password>

Confirm the password for "admin": <password>

Would you like to enter the basic configuration dialog (yes/no): yes

Create another login account (yes/no) [n]: Enter

Configure read-only SNMP community string (yes/no) [n]: Enter

Configure read-write SNMP community string (yes/no) [n]: Enter

Enter the switch name: <nexus-B-hostname>

Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: Enter

Mgmt0 IPv4 address: <nexus-B-mgmt0-ip>

Mgmt0 IPv4 netmask: <nexus-B-mgmt0-netmask>

Configure the default gateway? (yes/no) [y]: Enter

IPv4 address of the default gateway: <nexus-B-mgmt0-gw>

Configure advanced IP options? (yes/no) [n]: Enter

Enable the telnet service? (yes/no) [n]: Enter

Enable the ssh service? (yes/no) [y]: Enter

Type of ssh key you would like to generate (dsa/rsa) [rsa]: Enter

Number of rsa key bits <1024-2048> [1024]: Enter

Configure the ntp server? (yes/no) [n]: Enter

Configure default interface layer (L3/L2) [L2]: Enter

Configure default switchport interface state (shut/noshut) [noshut]: shut

Enter basic FC configurations (yes/no) [n]: Enter

Configure CoPP system profile (strict/moderate/lenient/dense) [strict]: Enter

Would you like to edit the configuration? (yes/no) [n]: Enter

2.    Review the configuration summary before enabling the configuration.

Use this configuration and save it? (yes/no) [y]: Enter

Procedure 2.     Ansible Nexus Switch Configuration

Note:   Configure the Cisco Nexus switches from the management workstation.

Step 1.    Add Nexus switch ssh keys to /root/.ssh/known_hosts. Adjust known_hosts as necessary if errors occur.

ssh admin@<nexus-A-mgmt0-ip>
exit
ssh admin@<nexus-B-mgmt0-ip>
exit

Step 2.    Edit the following variable files to ensure proper Cisco Nexus variables are entered:

      FlexPod-for-EHR/Ansible//inventory

      FlexPod-for-EHR/Ansible/group_vars/all.yml

      FlexPod-for-EHR/Ansible/host_vars/n9kA.yml

      FlexPod-for-EHR/Ansible/host_vars/n9kB.yml

      FlexPod-for-EHR/Ansible/roles/NEXUSconfig/defaults/main.yml

Step 3.    From /root/ FlexPod-for-EHR/Ansible/, run the Setup_Nexus.yml Ansible playbook.

  ansible-playbook ./Setup_Nexus.yml -i inventory

Step 4.    Once the Ansible playbook has been run on both switches, it is important to configure the local time so that logging time alignment and any backup schedules are correct. For more information on configuring the timezone and daylight savings time or summertime, please see Cisco Nexus 9000 Series NX-OS Fundamentals Configuration Guide, Release 9.3(x). Sample clock commands for the United States Eastern timezone are:

clock timezone EST -5 0
clock summer-time EDT 2 Sunday March 02:00 1 Sunday November 02:00 60

Step 5.    ssh into each switch and execute the following commands:

clock timezone <timezone> <hour-offset> <minute-offset>

clock summer-time <timezone> <start-week> <start-day> <start-month> <start-time> <end-week> <end-day> <end-month> <end-time> <offset-minutes>

Procedure 3.     FlexPod Cisco Nexus Switch Manual Configuration

Note:   Enable features on Cisco Nexus A and Cisco Nexus B.

Step 1.    Log in as admin using ssh.

Step 2.    Run the following commands:

config t
feature nxapi
feature udld
feature interface-vlan
feature lacp
feature vpc
feature lldp

Procedure 4.     Set Global Configurations

Cisco Nexus A and Cisco Nexus B

Step 1.    Run the following commands to set global configurations:

spanning-tree port type network default
spanning-tree port type edge bpduguard default
spanning-tree port type edge bpdufilter default
port-channel load-balance src-dst l4port
ntp server <global-ntp-server-ip> use-vrf management
ntp master 3
clock timezone <timezone> <hour-offset> <minute-offset>

(For Example: clock timezone EST -5 0)
clock summer-time <timezone> <start-week> <start-day> <start-month> <start-time> <end-week> <end-day> <end-month> <end-time> <offset-minutes>

ip route 0.0.0.0/0 <ib-mgmt-vlan-gateway>
copy run start

Note:   It is important to configure the local time so that logging time alignment and any backup schedules are correct. For more information on configuring the timezone and daylight savings time or summer time, please see Cisco Nexus 9000 Series NX-OS Fundamentals Configuration Guide, Release 9.3(x). Sample clock commands for the United States Eastern timezone are:

clock timezone EST -5 0
clock summer-time EDT 2 Sunday March 02:00 1 Sunday November 02:00 60

Procedure 5.     Create VLANs

Cisco Nexus A and Cisco Nexus B

Note:   Create the necessary virtual local area networks (VLANs) on both switches.

Step 1.    From the global configuration mode, run the following commands:

vlan <oob-mgmt-vlan-id>
name OOB-MGMT
vlan <ib-mgmt-vlan-id>
name IB-MGMT
vlan <native-vlan-id>
name Native-Vlan
vlan <vmotion-vlan-id>
name vMotion
vlan <vm-traffic-vlan-id>
name VM-Traffic
vlan <infra-nfs-vlan-id>
name Infra-NFS

vlan <CIFS-vlan-id>

name CIFS
exit

Procedure 6.     Add NTP Distribution Interface

Cisco Nexus A

Step 1.    From the global configuration mode, run the following commands:

interface vlan <ib-mgmt-vlan-id>
ip address <switch-a-ntp-ip>/<ib-mgmt-vlan-netmask-length>
no shutdown
exit
ntp peer <nexus-B-mgmt0-ip> use-vrf management

Cisco Nexus B

Step 1.    From the global configuration mode, run the following commands:

interface vlan <ib-mgmt-vlan-id>
ip address <switch-b-ntp-ip>/<ib-mgmt-vlan-netmask-length>
no shutdown
exit
ntp peer <nexus-A-mgmt0-ip> use-vrf management

Procedure 7.     Add Individual Port Descriptions for Troubleshooting and Enable UDLD for Cisco UCS Interfaces

Cisco Nexus A

Note:   In this step and in the following sections, configure the AFF nodename <st-node> and Cisco UCS 6454 fabric interconnect clustername <ucs-clustername> interfaces as appropriate to your deployment.

Step 1.    From the global configuration mode, run the following commands:

interface Eth1/25
description <ucs-clustername>-A:1/45
udld enable

interface Eth1/26
description <ucs-clustername>-A:1/46
udld enable

interface Eth1/27
description <ucs-clustername>-B:1/45
udld enable

interface Eth1/28
description <ucs-clustername>-B:1/46
udld enable

Note:   For fibre optic connections to Cisco UCS systems (AOC or SFP-based), entering udld enable will result in a message stating that this command is not applicable to fiber ports. This message is expected. If you have fibre optic connections, do not enter the udld enable command.

interface Eth1/17
description <st-clustername>-01:e0e

interface Eth1/18
description <st-clustername>-01:e0f

interface Eth1/19
description <st-clustername>-02:e0e

interface Eth1/20
description <st-clustername>-02:e0f

interface Eth1/49
description <nexus-b-hostname>:1/49

interface Eth1/50
description <nexus-b-hostname>:1/50

interface Eth1/50
description Uplink-SW

exit

Cisco Nexus B

Note:   Add the individual port descriptions for troubleshooting activity and verification for switch B and to enable aggressive UDLD on copper interfaces connected to Cisco UCS systems.

Step 1.    From the global configuration mode, run the following commands:

interface Eth1/25
description <ucs-clustername>-A:1/47
udld enable

interface Eth1/26
description <ucs-clustername>-A:1/48
udld enable

interface Eth1/27
description <ucs-clustername>-B:1/47
udld enable

interface Eth1/28
description <ucs-clustername>-B:1/48
udld enable

Note:   For fibre optic connections to Cisco UCS systems (AOC or SFP-based), entering udld enable will result in a message stating that this command is not applicable to fiber ports. This message is expected.

interface Eth1/17
description <st-clustername>-01:e0g

interface Eth1/18
description <st-clustername>-01:e0h

interface Eth1/19
description <st-clustername>-02:e0g

interface Eth1/20
description <st-clustername>-02:e0h

interface Eth1/49
description <nexus-a-hostname>:1/49

interface Eth1/50
description <nexus-a-hostname>:1/50

interface Eth1/50
description Uplink-SW
exit

Procedure 8.     Create Port Channels

Cisco Nexus A and Cisco Nexus B

Note:   Create the necessary port channels between devices on both switches.

Step 1.    From the global configuration mode, run the following commands:

interface Po10
description vPC peer-link
interface Eth1/49-50
channel-group 10 mode active
no shutdown

interface Po17
description <st-clustername>-01
interface Eth1/17-18
channel-group 17 mode active
no shutdown

interface Po19
description <st-clustername>-02
interface Eth1/19-20
channel-group 19 mode active
no shutdown

interface Po25
description <ucs-clustername>-a
interface Eth1/25-26
channel-group 25 mode active
no shutdown

interface Po26
description <ucs-clustername>-b
interface Eth1/27-28
channel-group 123 mode active
no shutdown

interface Po154
description MGMT-Uplink
interface Eth1/54
channel-group 154 mode active
no shutdown

exit

copy run start

Procedure 9.     Configure Port Channel Parameters

Cisco Nexus A and Cisco Nexus B

Note:   Configure port channel parameters on both switches.

Step 1.    From the global configuration mode, run the following commands:

interface Po10
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <oob-mgmt-vlan-id>, <ib-mgmt-vlan-id>, <infra-nfs-vlan-id>, <vmotion-vlan-id>, <vm-traffic-vlan-id>, <CIFS-vlan-id>
spanning-tree port type network
speed 100000
duplex full
interface Po17
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <ib-mgmt-vlan-id>, <infra-nfs-vlan-id>,<CIFS-vlan-id>
spanning-tree port type edge trunk
mtu 9216

interface Po19
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <ib-mgmt-vlan-id>, <infra-nfs-vlan-id>, <CIFS-vlan-id>
spanning-tree port type edge trunk
mtu 9216

interface Po25
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <oob-mgmt-vlan-id>, <ib-mgmt-vlan-id>, <infra-nfs-vlan-id>, <vmotion-vlan-id>, <vm-traffic-vlan-id>,<CIFS-vlan-id>
spanning-tree port type edge trunk
mtu 9216

interface Po26
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <oob-mgmt-vlan-id>, <ib-mgmt-vlan-id>, <infra-nfs-vlan-id>, <vmotion-vlan-id>, <vm-traffic-vlan-id>, <CIFS-vlan-id>
spanning-tree port type edge trunk
mtu 9216

interface Po154
switchport mode trunk
switchport trunk native vlan <native-vlan-id>
switchport trunk allowed vlan <oob-mgmt-vlan-id>, <ib-mgmt-vlan-id>
spanning-tree port type network
mtu 9216

exit

copy run start

Procedure 10.  Configure Virtual Port Channels

Cisco Nexus A

Step 1.    From the global configuration mode, run the following commands:

vpc domain <nexus-vpc-domain-id>
role priority 10
peer-keepalive destination <nexus-B-mgmt0-ip> source <nexus-A-mgmt0-ip>
peer-switch
peer-gateway
auto-recovery
delay restore 150

ip arp synchronize

interface Po10
vpc peer-link

interface Po17
vpc 17
interface Po19
vpc 19
interface Po25
vpc 21
interface Po26
vpc 26
interface Po154
vpc 154
exit
copy run start

Cisco Nexus B

Step 1.    From the global configuration mode, run the following commands:

vpc domain <nexus-vpc-domain-id>
role priority 20
peer-keepalive destination <nexus-A-mgmt0-ip> source <nexus-B-mgmt0-ip>
peer-switch
peer-gateway
auto-recovery
delay restore 150
ip arp synchronize

interface Po10
vpc peer-link
interface Po7
vpc 17
interface Po19
vpc 19
interface Po25
vpc 25
interface Po26
vpc 26
interface Po154
vpc 154
exit
copy run start

Procedure 11.  Switch Testing Commands

Note:   These commands can be used to check for correct switch configuration. Some of these commands need to be run after further configuration of the FlexPod components are completed to see complete results.

show run

show vpc

show port-channel summary

show ntp peer-status

show cdp neighbors

show lldp neighbors

show run int

show int

show udld neighbors

show int status

Storage Configuration

NetApp AFF A400 Controllers

See section NetApp Hardware Universe for planning the physical location of the storage systems:

      Site Preparation

      System Connectivity Requirements

      Circuit Breaker, Power Outlet Balancing, System Cabinet Power Cord Plugs, and Console Pinout Requirements

      AFF Series Systems

NetApp Hardware Universe

The NetApp Hardware Universe (HWU) application provides supported hardware and software components for any specific ONTAP version. It also provides configuration information for all the NetApp storage appliances currently supported by ONTAP software and a table of component compatibilities.

To confirm that the hardware and software components that you would like to use are supported with the version of ONTAP that you plan to install, follow these steps at the NetApp Support site.

1.     Access the HWU application to view the System Configuration guides. Click the Platforms menu to view the compatibility between different version of the ONTAP software and the NetApp storage appliances with your desired specifications.

2.     Alternatively, to compare components by storage appliance, click Compare Storage Systems.

Controllers

Follow the physical installation procedures for the controllers found here: https://docs.netapp.com/us-en/ontap-systems/index.html.

Disk Shelves

NetApp storage systems support a wide variety of disk shelves and disk drives. The complete list of disk shelves that are supported by the AFF A400 and AFF A800 is available at the NetApp Support site.

When using SAS disk shelves with NetApp storage controllers, refer to: https://docs.netapp.com/us-en/ontap-systems/sas3/index.html for proper cabling guidelines.

When using NVMe drive shelves with NetApp storage controllers, refer to: https://docs.netapp.com/us-en/ontap-systems/ns224/index.html for installation and servicing guidelines.

NetApp ONTAP 9.10

Complete Configuration Worksheet

Note:   Before running the setup script, complete the Cluster setup worksheet in the ONTAP 9 Documentation Center. You must have access to the NetApp Support site to open the cluster setup worksheet.

Procedure 1.     Ansible Storage Configuration

End-to-End ONTAP Storage Configuration for a FlexPod is automated with Ansible. ONTAP Storage can be deployed via Ansible after the ONTAP Cluster setup is complete and the Cluster management network is configured.

A playbook by name ‘Setup_ONTAP.yml’ is available at the root of this repository. It calls all the required roles to complete the setup of the ONTAP storage system.

The ONTAP setup is split into three sections, use the tags - ontap_config_part_1, ontap_config_part_2 and ontap_config_part_3 to execute parts of the playbook at the appropriate stage of setup.

Step 1.    Execute the playbook from the Ansible Control machine as an admin/ root user using the following commands:

Step 2.    After setup of Cisco Nexus switches:

ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_1

Step 3.    After setup of Cisco UCS:

ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_2

Step 4.    After setup of VMware vSphere 7.0 Setup:

ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_3

Step 5.    Optional: If you would like to run a part of the deployment, you may use the appropriate tag that accompanies each task in the role and run the playbook by executing the following command:

ansible-playbook -i inventory Setup_ONTAP.yml -t <tag_name>

Configure ONTAP Nodes

Before running the setup script, review the configuration worksheets in the Software setup section of the ONTAP 9 Documentation Center to learn about configuring ONTAP. Table 9 lists the information needed to configure two ONTAP nodes. Customize the cluster detail values with the information applicable to your deployment.

Table 9.     ONTAP Software Installation Prerequisites

Cluster Detail

Cluster Detail Value

Cluster node 01 IP address

<node01-mgmt-ip>

Cluster node 01 netmask

<node01-mgmt-mask>

Cluster node 01 gateway

<node01-mgmt-gateway>

Cluster node 02 IP address

<node02-mgmt-ip>

Cluster node 02 netmask

<node02-mgmt-mask>

Cluster node 02 gateway

<node02-mgmt-gateway>

ONTAP 9.10 URL (http server hosting ONTAP software)

<url-boot-software>

Procedure 1.     Configure Node 01

Step 1.    Connect to the storage system console port. You should see a Loader-A prompt. However, if the storage system is in a reboot loop, press Ctrl-C to exit the autoboot loop when the following message displays:   

Starting AUTOBOOT press Ctrl-C to abort...

Step 2.    Allow the system to boot up.

autoboot

Step 3.    Press Ctrl-C when prompted.

Note:   If ONTAP 9.10.1P1 is not the version of software being booted, continue with the following steps to install new software. If ONTAP 9.10.1P1 is the version being booted, select option 8 and y to reboot the node. Then continue with section Set Up Node.

Step 4.    To install new software, select option 7 from the menu.

Step 5.    Enter y to continue the installation.

Step 6.    Select e0M for the network port for the download.

Step 7.    Enter n to skip the reboot.

Step 8.    Select option 7 from the menu: Install new software first

Step 9.    Enter y to continue the installation

Step 10.                       Enter the IP address, netmask, and default gateway for e0M.

   Enter the IP address for port e0M: <node01-mgmt-ip>
   Enter the netmask for port e0M: <node01-mgmt-mask>
   Enter the IP address of the default gateway: <node01-mgmt-gateway>

Step 11.                       Enter the URL where the software can be found.

Note:   The e0M interface should be connected to management network and the web server must be reachable (using ping) from node 01.

<url-boot-software>

Step 12.                       Press Enter for the user name, indicating no user name.

Step 13.                       Enter y to set the newly installed software as the default to be used for subsequent reboots.

Step 14.                       Enter yes to reboot the node.

TextDescription automatically generated

Note:   When installing new software, the system might perform firmware upgrades to the BIOS and adapter cards, causing reboots and possible stops at the Loader-A prompt. If these actions occur, the system might deviate from this procedure.

Note:   During the ONTAP installation a prompt to reboot the node requests a Y/N response. The prompt requires the entire Yes or No response to reboot the node and continue the installation.

Step 15.                       Press Ctrl-C when the following message displays:

Press Ctrl-C for Boot Menu

Step 16.                       Select option 4 for Clean Configuration and Initialize All Disks.

Step 17.                       Enter y to zero disks, reset config, and install a new file system.

Step 18.                       Enter yes to erase all the data on the disks.

Note:   The initialization and creation of the root aggregate can take 90 minutes or more to complete, depending on the number and type of disks attached. When initialization is complete, the storage system reboots. Note that SSDs take considerably less time to initialize. You can continue with the configuration of node 02 while the disks for node 01 are zeroing.

Procedure 2.     Configure Node 02

Step 1.    Connect to the storage system console port. You should see a Loader-B prompt. However, if the storage system is in a reboot loop, press Ctrl-C to exit the autoboot loop when the following message displays:

Starting AUTOBOOT press Ctrl-C to abort...

Step 2.    Allow the system to boot up.

autoboot

Step 3.    Press Ctrl-C when prompted.

Note:   If ONTAP 9.10.1P1 is not the version of software being booted, continue with the following steps to install new software. If ONTAP 9.10.1P1 is the version being booted, select option 8 and y to reboot the node, then continue with section Set Up Node.

Step 4.    To install new software, select option 7.

Step 5.    Enter y to continue the installation.

Step 6.    Select e0M for the network port you want to use for the download.

Step 7.    Enter n to skip the reboot.

Step 8.    Select option 7: Install new software first

Step 9.    Enter y to continue the installation.

Step 10.                       Enter the IP address, netmask, and default gateway for e0M.

   Enter the IP address for port e0M: <node02-mgmt-ip>
   Enter the netmask for port e0M: <node02-mgmt-mask>
   Enter the IP address of the default gateway: <node02-mgmt-gateway>

Step 11.                       Enter the URL where the software can be found.

Note:   The web server must be reachable (ping) from node 02.

<url-boot-software>

Step 12.                       Press Enter for the username, indicating no user name.

Step 13.                       Enter y to set the newly installed software as the default to be used for subsequent reboots.

Step 14.                       Enter yes to reboot the node.

TextDescription automatically generated

Note:   When installing new software, the system might perform firmware upgrades to the BIOS and adapter cards, causing reboots and possible stops at the Loader-A prompt. If these actions occur, the system might deviate from this procedure.

Note:   During the ONTAP installation a prompt to reboot the node requests a Y/N response. The prompt requires the entire Yes or No response to reboot the node and continue the installation.

Step 15.                       Press Ctrl-C when you see this message:

Press Ctrl-C for Boot Menu

Step 16.                       Select option 4 for Clean Configuration and Initialize All Disks.

Step 17.                       Enter y to zero disks, reset config, and install a new file system.

Step 18.                       Enter yes to erase all the data on the disks.

Note:   The initialization and creation of the root aggregate can take 90 minutes or more to complete, depending on the number and type of disks attached. When initialization is complete, the storage system reboots. Note that SSDs take considerably less time to initialize.

Procedure 3.     Set Up Node

Note:   From a console port program attached to the storage controller A (node 01) console port, run the node setup script. This script appears when ONTAP 9.10.1P1 boots on the node for the first time.

Step 1.    Follow the prompts to set up node 01.

  Welcome to node setup.

  You can enter the following commands at any time:

    "help" or "?" - if you want to have a question clarified,

    "back" - if you want to change previously answered questions, and

    "exit" or "quit" - if you want to quit the setup wizard.

       Any changes you made before quitting will be saved.


   You can return to cluster setup at any time by typing "cluster setup".

   To accept a default or omit a question, do not enter a value.


   This system will send event messages and weekly reports to NetApp Technical

   Support.

   To disable this feature, enter "autosupport modify -support disable" within 24

   hours.

   Enabling AutoSupport can significantly speed problem determination and resolution

   should a problem occur on your system.
  
   For further information on AutoSupport, see:

   http://support.netapp.com/autosupport/


   Type yes to confirm and continue {yes}: yes



   Enter the node management interface port [e0M]: Enter

   Enter the node management interface IP address: <node01-mgmt-ip>

   Enter the node management interface netmask: <node01-mgmt-mask>

   Enter the node management interface default gateway: <node01-mgmt-gateway>

   A node management interface on port e0M with IP address <node01-mgmt-ip> has been

   created



   Use your web browser to complete cluster setup by accessing https://<node01-mgmt-

   ip>


   Otherwise press Enter to complete cluster setup using the command line interface:

Step 2.    To complete cluster setup, open a web browser and navigate to https://node01-mgmt-ip.

Table 10.   Cluster Create in ONTAP Prerequisites

Cluster Detail

Cluster Detail Value

Cluster name

<clustername>

Cluster Admin SVM

<cluster-adm-svm>

Infrastructure Data SVM

<infra-data-svm>

ONTAP base license

<cluster-base-license-key>

Cluster management IP address

<clustermgmt-ip>

Cluster management netmask

<clustermgmt-mask>

Cluster management gateway

<clustermgmt-gateway>

Cluster node 01 IP address

<node01-mgmt-ip>

Cluster node 01 netmask

<node01-mgmt-mask>

Cluster node 01 gateway

<node01-mgmt-gateway>

Cluster node 02 IP address

<node02-mgmt-ip>

Cluster node 02 netmask

<node02-mgmt-mask>

Cluster node 02 gateway

<node02-mgmt-gateway>

Node 01 service processor IP address

<node01-sp-ip>

Node 01 service processor network mask

<node01-sp-mask>

Node 01 service processor gateway

<node01-sp-gateway>

Node 02 service processor IP address

<node02-sp-ip>

Node 02 service processor network mask

<node02-sp-mask>

Node 02 service processor gateway

<node02-sp-gateway>

Node 01 node name

<st-node01>

Node 02 node name

<st-node02>

DNS domain name

<dns-domain-name>

DNS server IP address

<dns-ip>

NTP server A IP address

<switch-a-ntp-ip>

NTP server B IP address

<switch-b-ntp-ip>

SNMPv3 User

<snmp-v3-usr>

SNMPv3 Authentication Protocol

<snmp-v3-auth-proto>

SNMPv3 Privacy Protocol

<snmpv3-priv-proto>

Note:   Cluster setup can also be performed using the CLI. This document describes the cluster setup using the NetApp ONTAP System Manager guided setup.

Step 3.    Complete the required information on the Initialize Storage System screen:

Graphical user interface, websiteDescription automatically generated

Step 4.    In the Cluster screen, follow these steps:

a.     Enter the cluster name and administrator password.

b.     Complete the Networking information for the cluster and each node.

c.     Check the box for Use time services (NTP) and enter the IP addresses of the time servers in a comma separated list.

Note:   The nodes should be discovered automatically; if they are not, Refresh the browser page. By default, the cluster interfaces are created on all the new factory shipping storage controllers.

Note:   If all the nodes are not discovered, then configure the cluster using the command line.

Note:   The node management interface can be on the same subnet as the cluster management interface, or it can be on a different subnet. In this document, we assume that it is on the same subnet.

Graphical user interface, applicationDescription automatically generated

Step 5.    Click Submit.

Graphical user interface, applicationDescription automatically generated

Step 6.    A few minutes will pass while the cluster is configured. When prompted, login to ONTAP System Manager to                  continue the cluster configuration.

Note:   You can use Ansible scripts at this point to configure the remaining part of the Storage Configurations.

Step 7.    Edit the following variable files to ensure proper ONTAP Storage variables are entered:

      FlexPod-for-EHR/Ansible/inventory

      FlexPod-for-EHR/Ansible/group_vars/all.yml

      FlexPod-for-EHR/Ansible/group_vars/ontap

      FlexPod-for-EHR/Ansible/vars/ontap_main.yml

Step 8.    From /root/ FlexPod-for-EHR/Ansible, run the Setup_ONTAP.yml Ansible playbook.

  ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_1

Note:   Use the -vvv tag to see detailed execution output log.

Note:   Ansible will implement all the Storage configurations tasks in this section. Skip the below steps and move to the Cisco UCS Configuration section.

Step 9.    From the Dashboard click the Cluster menu on the left and select Overview.

Step 10.                       Click the More ellipsis button in the Overview pane at the top right of the screen and select Edit.

Graphical user interface, application, TeamsDescription automatically generated

Step 11.                       Add additional cluster configuration details and click Save to make the changes persistent:

a.     Cluster location

b.     DNS domain name

c.     DNS server IP addresses

Note:   DNS server IP addresses can be added individually or with a comma separated list on a single line.

Graphical user interface, text, application, TeamsDescription automatically generated

Step 12.                       Click Save to make the changes persistent.

Step 13.                       Select the Settings menu under the Cluster menu.

Graphical user interface, application, TeamsDescription automatically generated

Step 14.                       If AutoSupport was not configured during the initial setup, click the ellipsis in the AutoSupport tile and select More options.

Step 15.                       To enable AutoSupport click the slider.

Step 16.                       Click Edit to change the transport protocol, add a proxy server address and a mail host as needed.

Step 17.                       Click Save to enable the changes.

Step 18.                       In the Email tile to the right, click Edit and enter the desired email information:

a.     Email send from address

b.     Email recipient addresses

c.     Recipient Category

Step 19.                       Click Save when complete.

Graphical user interface, applicationDescription automatically generated

Step 20.                       Select CLUSTER > Settings at the top left of the page to return to the cluster settings page.

Step 21.                       Locate the Licenses tile on the right and click the detail arrow.

Step 22.                       Add the desired licenses to the cluster by clicking Add and entering the license keys in a comma separated list.

Graphical user interface, applicationDescription automatically generated

A screenshot of a computerDescription automatically generated

Step 23.                       Configure storage aggregates by selecting the Storage menu on the left and selecting Tiers.

Step 24.                       Click Add Local Tier and allow ONTAP System Manager to recommend a storage aggregate configuration.

Graphical user interface, application, TeamsDescription automatically generated

Step 25.                       ONTAP will use best practices to recommend an aggregate layout. Click the Recommended details link to view the aggregate information.

Step 26.                       Optionally, enable NetApp Aggregate Encryption (NAE) by checking the box for Configure Onboard Key Manager for encryption.

Step 27.                       Enter and confirm the passphrase and save it in a secure location for future use.

Step 28.                       Click Save to make the configuration persistent.

Graphical user interface, applicationDescription automatically generated

Note:   Careful consideration should be taken before enabling aggregate encryption. Aggregate encryption may not be supported for all deployments. Please review the NetApp Encryption Power Guide and the Security Hardening Guide for NetApp ONTAP 9 (TR-4569) to help determine if aggregate encryption is right for your environment.

Procedure 4.     Log into the Cluster

Step 1.    Open an SSH connection to either the cluster IP or the host name.

Step 2.    Log into the admin user with the password you provided earlier.

Procedure 5.     Verify Storage Failover

Step 1.    Verify the status of the storage failover.

         storage failover show

Note:   Both <st-node01> and <st-node02> must be capable of performing a takeover. Continue with step 2 if the nodes can perform a takeover.

Step 2.    Enable failover on one of the two nodes if it was not completed during the installation.

         storage failover modify -node <st-node01> -enabled true

Note:   Enabling failover on one node enables it for both nodes.

Step 3.    Verify the HA status for a two-node cluster.

Note:   This step is not applicable for clusters with more than two nodes.

         cluster ha show

Step 4.    If HA is not configured use the below commands. Only enable HA mode for two-node clusters. Do not run this command for clusters with more than two nodes because it causes problems with failover.

         cluster ha modify -configured true

         Do you want to continue? {y|n}: y

Step 5.    Verify that hardware assist is correctly configured.

         storage failover hwassist show

Step 6.    If hwassist storage failover is not enabled, enable using the following commands.

storage failover modify --hwassist-partner-ip <node02-mgmt-ip> -node <st-node01>

storage failover modify --hwassist-partner-ip <node01-mgmt-ip> -node <st-node02>

Procedure 6.     Set Auto-Revert on Cluster Management

Note:   A storage virtual machine (SVM) is referred to as a Vserver or vserver in the GUI and CLI.

Step 1.    Run the following command:

net interface modify -vserver <clustername> -lif cluster_mgmt_lif_1 -auto-revert true

Procedure 7.     Zero All Spare Disks

Step 2.    To zero all spare disks in the cluster, run the following command:

disk zerospares

Note:   Advanced Data Partitioning creates a root partition and two data partitions on each SSD drive in an AFF configuration. Disk autoassign should have assigned one data partition to each node in an HA pair. If a different disk assignment is required, disk autoassignment must be disabled on both nodes in the HA pair by running the disk option modify command. Spare partitions can then be moved from one node to another by running the disk removeowner and disk assign commands.

Procedure 8.     Set Up Service Processor Network Interface

Step 1.    To assign a static IPv4 address to the Service Processor on each node, run the following commands:

system service-processor network modify -node <st-node01> -address-family IPv4 -enable true -dhcp none -ip-address <node01-sp-ip> -netmask <node01-sp-mask> -gateway <node01-sp-gateway>

system service-processor network modify -node <st-node02> -address-family IPv4 -enable true -dhcp none -ip-address <node02-sp-ip> -netmask <node02-sp-mask> -gateway <node02-sp-gateway>

Note:   The Service Processor IP addresses should be in the same subnet as the node management IP addresses.

Procedure 9.     Create Manual Provisioned Aggregates (Optional)

Note:   An aggregate containing the root volume is created during the ONTAP setup process. To manually create additional aggregates, determine the aggregate name, the node on which to create it, and the number of disks it should contain.

Step 1.    To create new aggregates, run the following commands:

storage aggregate create -aggregate <aggr1_node01> -node <st-node01> -diskcount <num-disks> -disktype SSD-NVM

storage aggregate create -aggregate <aggr1_node02> -node <st-node02> -diskcount <num-disks> -disktype SSD-NVM

Note:   Customers should have the minimum number of hot spare disks for the recommended hot spare disk partitions for their aggregate.

Note:   For all-flash aggregates, you should have a minimum of one hot spare disk or disk partition. For non-flash homogenous aggregates, you should have a minimum of two hot spare disks or disk partitions. For Flash Pool aggregates, you should have a minimum of two hot spare disks or disk partitions for each disk type.

Note:   In an AFF configuration with a small number of SSDs, you might want to create an aggregate with all, but one remaining disk (spare) assigned to the controller.

Note:   The aggregate cannot be created until disk zeroing completes. Run the storage aggregate show command to display the aggregate creation status. Do not proceed until both aggr1_node01 and aggr1_node02 are online.

Procedure 10.  Remove Default Broadcast Domains

Note:   By default, all network ports are included in separate default broadcast domain. Network ports used for data services (for example, e0e, e0f, and so on) should be removed from their default broadcast domain and that broadcast domain should be deleted.

Step 1.    To perform this task, run the following commands:

network port broadcast-domain delete -broadcast-domain <Default-N> -ipspace Default

network port broadcast-domain show

Note:   Delete the Default broadcast domains with Network ports (Default-1, Default-2, and so on).

Procedure 11.  Disable Flow Control on 25/100GbE Data Ports

Step 1.    Run the following command to configure the ports on node 01:

network port modify -node <st-node01> -port e4a,e4b -flowcontrol-admin none
network port modify -node <st-node01> -port e0e,e0f,e0g,e0h -flowcontrol-admin            none

Step 2.    Run the following command to configure the ports on node 02:

network port modify -node <st-node02> -port e4a,e4b -flowcontrol-admin none
network port modify -node <st-node02> -port e0e,e0f,e0g,e0h -flowcontrol-admin none

aa16-a400::> net port show -node * -port e0e,e0f,e0g,e0h -fields speed-    admin,duplex-admin,flowcontrol-admin
(network port show)

         node         port duplex-admin speed-admin flowcontrol-admin
         ------------ ---- ------------ ----------- -----------------
         aa16-a400-01 e0e  auto         auto        none
         aa16-a400-01 e0f  auto         auto        none
         aa16-a400-01 e0g  auto         auto        none
         aa16-a400-01 e0h  auto         auto        none
         aa16-a400-02 e0e  auto         auto        none
         aa16-a400-02 e0f  auto         auto        none
         aa16-a400-02 e0g  auto         auto        none
         aa16-a400-02 e0h  auto         auto        none
         8 entries were displayed.

 

aa16-a400::> net port show -node * -port e4a,e4b -fields speed-admin,duplex-  admin,flowcontrol-admin
(network port show)

         node         port duplex-admin speed-admin flowcontrol-admin
         ------------ ---- ------------ ----------- -----------------
         aa16-a400-01 e4a  auto         auto        none
         aa16-a400-01 e4b  auto         auto        none
         aa16-a400-02 e4a  auto         auto        none
         aa16-a400-02 e4b  auto         auto        none
         4 entries were displayed.

Procedure 12.  Disable Auto-Negotiate on Fibre Channel Ports

Step 1.    Disable each FC adapter in the controllers with the fcp adapter modify command.

         fcp adapter modify -node <st-node01> -adapter 5a --status-admin down
         fcp adapter modify -node <st-node01> -adapter 5b --status-admin down
         fcp adapter modify -node <st-node02> -adapter 5a --status-admin down
         fcp adapter modify -node <st-node02> -adapter 5b --status-admin down

Step 2.    Set the desired speed on the adapter and return it to the online state.

         fcp adapter modify -node <st-node01> -adapter 5a -speed 32 -status-admin up
         fcp adapter modify -node <st-node01> -adapter 5b -speed 32 -status-admin up
         fcp adapter modify -node <st-node02> -adapter 5a -speed 32 -status-admin up
         fcp adapter modify -node <st-node02> -adapter 5b -speed 32 -status-admin up

Procedure 13.  Enable Cisco Discovery Protocol

Step 1.    To enable the Cisco Discovery Protocol (CDP) on the NetApp storage controllers, run the following command to enable CDP in ONTAP:

node run -node * options cdpd.enable on

Note:   To be effective, CDP must also be enabled on directly connected networking equipment such as switches and routers.

Procedure 14.  Enable Link-layer Discovery Protocol on all Ethernet Ports

Note:   You need to enable the exchange of Link-layer Discovery Protocol (LLDP) neighbor information between the storage and network switches.

Step 1.    Enable LLDP on all ports of all nodes in the cluster.

node run * options lldp.enable on

Procedure 15.  Create Management Broadcast Domain

Step 1.    If the management interfaces are required to be on a separate VLAN, create a new broadcast domain for those interfaces by running the following command:

  network port broadcast-domain create -broadcast-domain IB-MGMT -mtu 1500

  network port broadcast-domain show

Procedure 16.  Create NFS Broadcast Domain

Step 1.    To create an NFS data broadcast domain with a maximum transmission unit (MTU) of 9000, run the following commands to create a broadcast domain for NFS in ONTAP:

  network port broadcast-domain create -broadcast-domain Infra_NFS -mtu 9000

  network port broadcast-domain show

Procedure 17.  Create CIFS Broadcast Domain

Step 1.    To create an CIFs data broadcast domain with a maximum transmission unit (MTU) of 9000, run the following commands to create a broadcast domain for NFS in ONTAP:

network port broadcast-domain create -broadcast-domain Infra_CIFS -mtu 9000

network port broadcast-domain show

Procedure 18.  Create Interface Groups

Step 1.    To create the LACP interface groups for the 25GbE data interfaces, run the following commands:

network port ifgrp create -node <st-node01> -ifgrp a0a -distr-func port -mode multimode_lacp
network port ifgrp add-port -node <st-node01> -ifgrp a0a -port e0e
network port ifgrp add-port -node <st-node01> -ifgrp a0a -port e0f
network port ifgrp add-port -node <st-node01> -ifgrp a0a -port e0g
network port ifgrp add-port -node <st-node01> -ifgrp a0a -port e0h
network port ifgrp create -node <st-node02> -ifgrp a0a -distr-func port -mode multimode_lacp
network port ifgrp add-port -node <st-node02> -ifgrp a0a -port e0e
network port ifgrp add-port -node <st-node02> -ifgrp a0a -port e0f
network port ifgrp add-port -node <st-node02> -ifgrp a0a -port e0g
network port ifgrp add-port -node <st-node02> -ifgrp a0a -port e0h

network port ifgrp show

Procedure 19.  Change MTU on Interface Groups

Step 1.    To change the MTU size on the base interface-group ports before creating the VLAN ports, run the following commands:

network port modify -node <st-node01> -port a0a -mtu 9000

network port modify -node <st-node02> -port a0a -mtu 9000

Procedure 20.  Create VLANs

Step 1.    Create the management VLAN ports and add them to the management broadcast domain.

network port vlan create -node <st-node01> -vlan-name a0a-<ib-mgmt-vlan-id>     

network port vlan create -node <st-node02> -vlan-name a0a-<ib-mgmt-vlan-id>

network port broadcast-domain add-ports -broadcast-domain IB-MGMT -ports <st-node01>:a0a-<ib-mgmt-vlan-id>,<st-node02>:a0a-<ib-mgmt-vlan-id>

network port vlan show

Step 2.    Create the NFS VLAN ports and add them to the Infra_NFS broadcast domain.

network port vlan create -node <st-node01> -vlan-name a0a-<infra-nfs-vlan-id>     

network port vlan create -node <st-node02> -vlan-name a0a-<infra-nfs-vlan-id>

network port broadcast-domain add-ports -broadcast-domain Infra_NFS -ports <st-node01>:a0a-<infra-nfs-vlan-id>,<st-node02>:a0a-<infra-nfs-vlan-id>

Step 3.    Create the CIFs VLAN ports and add them to the Infra_CIFS broadcast domain.

network port vlan create -node <st-node01> -vlan-name a0a-<infra-cifs-vlan-id>     

network port vlan create -node <st-node02> -vlan-name a0a-<infra-cifs-vlan-id>

network port broadcast-domain add-ports -broadcast-domain Infra_CIFS -ports <st-node01>:a0a-<infra-cifs-vlan-id>,<st-node02>:a0a-<infra-cifs-vlan-id>

Procedure 21.  Configure Timezone

Step 1.    Set the time zone for the cluster.

timezone -timezone <timezone>

Note:   For example, in the eastern United States, the time zone is America/New_York.

Procedure 22.  Configure Simple Network Management Protocol

Step 1.    Configure basic SNMP information, such as the location and contact. When polled, this information is visible as the sysLocation and sysContact variables in SNMP.

snmp contact <snmp-contact>

         snmp location "<snmp-location>"

         snmp init 1

         options snmp.enable on

Step 2.    Configure SNMP traps to send to remote hosts, such as an Active IQ Unified Manager server or another fault management system.

         snmp traphost add <oncommand-um-server-fqdn>

Procedure 23.  Configure SNMPv3 Access

Step 1.    SNMPv3 offers advanced security by using encryption and passphrases. The SNMPv3 user can run SNMP utilities from the traphost using the authentication and privacy settings that you specify. To configure SNMPv3 access, run the following commands:

security login create -user-or-group-name <<snmp-v3-usr>> -application snmp -authentication-method usm



Enter the authoritative entity's EngineID [local EngineID]:


Which authentication protocol do you want to choose (none, md5, sha, sha2-256) [none]: <<snmp-v3-auth-proto>>


Enter the authentication protocol password (minimum 8 characters long):


Enter the authentication protocol password again:


Which privacy protocol do you want to choose (none, des, aes128) [none]: <<snmpv3-priv-proto>>


Enter privacy protocol password (minimum 8 characters long):


Enter privacy protocol password again:

Note:   Refer to the SNMP Configuration Express Guide for additional information when configuring SNMPv3 security users.

Procedure 24.  Create SVM

Step 1.    Run the vserver create command.

vserver create -vserver Infra-SVM -rootvolume infra_svm_root -aggregate aggr1_node01 -rootvolume-security-style unix

Step 2.    Remove the unused data protocols from the SVM: iSCSI.

vserver remove-protocols -vserver Infra-SVM -protocols iscsi

Step 3.    Add the two data aggregates to the Infra-SVM aggregate list for the NetApp ONTAP Tools.

vserver modify -vserver Infra-SVM -aggr-list <aggr1_node01>,<aggr1_node02>

Step 4.    Enable and run the NFS protocol in the Infra-SVM.

vserver nfs create -vserver Infra-SVM -udp disabled -v3 enabled -v4.1 enabled -vstorage enabled

Note:   If the NFS license was not installed during the cluster configuration, make sure to install the license before starting the NFS service.

Step 5.    Verify the NFS vstorage parameter for the NetApp NFS VAAI plug-in was enabled.

vserver nfs show -fields vstorage

Procedure 25.  Create CIFS service

Note:   You can enable and configure CIFS servers on storage virtual machines (SVMs) with NetApp FlexVol®️ volumes to let SMB clients access files on your cluster. Each data SVM in the cluster can be bound to exactly one Active Directory domain. However, the data SVMs do not need to be bound to the same domain. Each data SVM can be bound to a unique Active Directory domain.

Note:   Before configuring the CIFS service on your SVM, the DNS must be configured.

Step 1.    Configure the DNS for your SVM.

dns create -vserver Infra-SVM -domains <domain_name> -name-servers <dns_server_ip>

Note:   The node management network interfaces should be able to route to the Active Directory domain controller to which you want to join the CIFS server. Alternatively, a data network interface must exist on the SVM that can route to the Active Directory domain controller.

Step 2.    Create a network interface on the IB-MGMT VLAN.

network interface create -vserver Infra-SVM -lif <<svm_mgmt_lif_name>> -role data -data-protocol none -home-node <<st-node-01>> -home-port a0a-<IB-MGMT-VLAN> -address <svm-mgmt-ip> -netmask <svm-mgmt-mask> -failover-policy broadcast-domain-wide -firewall-policy mgmt -auto-revert true

Step 3.    Create the CIFS service.

vserver cifs create -vserver Infra-SVM -cifs-server Infra-CIFS -domain <domain.com>

 

In order to create an Active Directory machine account for the CIFS server, you must supply the name and  password of a Windows account with sufficient privileges to add computers to the "CN=Computers" container  within the"DOMAIN.COM" domain.

Enter the user name: Administrator@active_diectory.local

Enter the password: 

Procedure 26.  Modify Storage Virtual Machine Options

Note:   NetApp ONTAP can use automatic node referrals to increase SMB client performance on SVMs with FlexVol volumes. This feature allows the SVM to automatically redirect a client request to a network interface on the node where the FlexVol volume resides.

Step 1.    To enable automatic node referrals on your SVM, run the following command:

set -privilege advanced

vserver cifs options modify -vserver Infra-SVM -is-referral-enabled true

Procedure 27.  Create Load-Sharing Mirrors of SVM Root Volume

Step 1.    Create a volume to be the load-sharing mirror of the infrastructure SVM root volume on each node.

volume create -vserver Infra-SVM -volume infra_svm_root_m01 -aggregate <aggr1_node01> -size 1GB -type DP

volume create -vserver Infra-SVM -volume infra_svm_root_m02 -aggregate <aggr1_node02> -size 1GB -type DP

Step 2.    Create a job schedule to update the root volume mirror relationships every 15 minutes.

job schedule interval create -name 15min -minutes 15

Step 3.    Create the mirroring relationships.

snapmirror create -source-path Infra-SVM:infra_svm_root -destination-path Infra-SVM:infra_svm_root_m01 -type LS -schedule 15min

snapmirror create -source-path Infra-SVM:infra_svm_root -destination-path Infra-SVM:infra_svm_root_m02 -type LS -schedule 15min

Step 4.    Initialize the mirroring relationship.

snapmirror initialize-ls-set -source-path Infra-SVM:infra_svm_root

snapmirror show -type ls

Procedure 28.  Create Block Protocol (FC) Service

Step 1.    Run the following command to create the FCP service on each SVM. This command also starts the FCP service and sets the worldwide name (WWN) for the SVM:

vserver fcp create -vserver Infra-SVM -status-admin up

vserver fcp show

Note:   If the FC license was not installed during the cluster configuration, make sure to install the license before creating the FC service.

Procedure 29.  Configure HTTPS Access

Step 1.    Increase the privilege level to access the certificate commands.

set -privilege diag

Do you want to continue? {y|n}: y

Step 2.    Generally, a self-signed certificate is already in place. Verify the certificate and obtain parameters (for example, the ) by running the following command:

security certificate show

Step 3.    For each SVM shown, the certificate common name should match the DNS fully qualified domain name (FQDN) of the SVM. Delete the two default certificates and replace them with either self-signed certificates or certificates from a certificate authority (CA). To delete the default certificates, run the following commands:

security certificate delete -vserver Infra-SVM -common-name Infra-SVM -ca Infra- SVM -type server -serial <serial-number>

Note:   Deleting expired certificates before creating new certificates is a best practice. Run the security certificate delete command to delete the expired certificates. In the following command, use TAB completion to select and delete each default certificate.

Step 4.    To generate and install self-signed certificates, run the following commands as one-time commands. Generate a server certificate for the Infra-SVM and the cluster SVM. Use TAB completion to aid in the completion of these commands.

security certificate create -common-name <cert-common-name> -type server -size 2048 -country <cert-country> -state <cert-state> -locality <cert-locality> -organization <cert-org> -unit <cert-unit> -email-addr <cert-email> -expire-days <cert-days> -protocol SSL -hash-function SHA256 -vserver Infra-SVM

Step 5.    To obtain the values for the parameters required in step 5 (<cert-ca> and <cert-serial>), run the security certificate show command.

Step 6.    Enable each certificate that was just created by using the -server-enabled true and -client-enabled false parameters. Use TAB completion to aid in the completion of these commands.

security ssl modify -vserver <clustername> -server-enabled true -client-enabled false -ca <cert-ca> -serial <cert-serial> -common-name <cert-common-name>

Step 7.    Disable HTTP cluster management access.

system services firewall policy delete -policy mgmt -service http -vserver <clustername>

Note:   It is normal for some of these commands to return an error message stating that the entry does not exist.

Step 8.    Change back to the normal admin privilege level and verify that the system logs are available in a web browser.

set -privilege admin

https://<node01-mgmt-ip>/spi
https://<node02-mgmt-ip>/spi

Procedure 30.  Configure NFSv3 and NFSv4.1 on SVM

Step 1.    Create a new rule for the infrastructure NFS subnet in the default export policy.

vserver export-policy rule create -vserver Infra-SVM -policyname default -ruleindex 1 -protocol nfs -clientmatch <infra-nfs-subnet-cidr> -rorule sys -rwrule sys -superuser sys -allow-suid true

Step 2.    Assign the FlexPod export policy to the infrastructure SVM root volume.

volume modify -vserver Infra-SVM -volume infra_svm_root -policy default

Procedure 31.  Create CIFS Export Policy

Note:   Optionally, you can use export policies to restrict CIFS access to files and folders on CIFS volumes. You can use export policies in combination with share level and file level permissions to determine effective access rights.

Step 3.    Create an export policy that limits access to devices in the domain by running the following command:

export-policy create -vserver Infra-SVM -policyname cifs

export-policy rule create -vserver Infra-SVM -policyname cifs -clientmatch <domain_name> -rorule

krb5i,krb5p -rwrule krb5i,krb5p

Procedure 32.  Create FlexVol Volumes

The following information is required to create a NetApp FlexVol® volume:

      The volume name

      The volume size

      The aggregate on which the volume exists

Step 1.    Create a FlexVol volume by running the following commands:

volume create -vserver Infra-SVM -volume infra_datastore_01 -aggregate <aggr1_node01> -size 1TB -state online -policy default -junction-path /infra_datastore_01 -space-guarantee none -percent-snapshot-space 0

volume create -vserver Infra-SVM -volume infra_datastore_02 -aggregate <aggr1_node02> -size 1TB -state online -policy default -junction-path /infra_datastore_02 -space-guarantee none -percent-snapshot-space 0

volume create -vserver Infra-SVM -volume infra_swap -aggregate <aggr1_node01> -size 100GB -state online -policy default -junction-path /infra_swap -space-guarantee none -percent-snapshot-space 0 -snapshot-policy none

volume create -vserver Infra-SVM -volume esxi_boot -aggregate <aggr1_node01> -size 320GB -state online -policy default -space-guarantee none -percent-snapshot-space 0

snapmirror update-ls-set -source-path Infra-SVM:infra_svm_root

Note:   If you are going to setup and use SnapCenter to backup the infra_datastore volume, add “-snapshot-policy none” to the end of the volume create command for the infra_datastore volume.

Procedure 33.  Modify Volume Efficiency

Step 1.    On NetApp AFF systems, deduplication is enabled by default. To disable the efficiency policy on the infra_swap volume, run the following command:

   volume efficiency off -vserver Infra-SVM -volume infra_swap

Procedure 34.  Create FlexGroup Volumes

A FlexGroup volume is a scale-out NAS container that provides high performance along with automatic load distribution and scalability. A FlexGroup volume contains several constituents that automatically and transparently share the traffic. A FlexGroup volume is a single namespace container that can be managed in a similar way as FlexVol volumes.

Step 1.    Run the following commands to create FlexGroup volumes:

volume create -vserver Infra-SVM -volume flexgroup_vol_01 -aggr-list aggr01_node01,aggr01_node02 -aggr-list-multiplier 4 -state online -policy cifs -size 800GB -junction-path /flexgroup_vol_01 -space-guarantee none -percent-snapshot-space 0

volume create -vserver Infra-SVM -volume flexgroup_vol_02 -aggr-list aggr01_node01,aggr01_node02 -aggr-list-multiplier 4 -state online -policy cifs -size 800GB -junction-path /flexgroup_vol_02 -space-guarantee none -percent-snapshot-space 0

Procedure 35.  Create CIFS Shares

A CIFS share is a named access point in a volume that enables CIFS clients to view, browse, and manipulate files on a file server.

Step 1.    Run the following commands to create CIFS share on the infrastructure SVM:

cifs share create -vserver Infra-SVM -share-name <CIFS_share_1> -path /flexgroup_vol_01 -share properties oplocks,browsable,continuously-available,showsnapshot

cifs share create -vserver Infra-SVM -share-name <CIFS_share_2> -path /flexgroup_vol_02 -share properties oplocks,browsable,continuously-available,showsnapshot

Procedure 36.  Create FC LIFs

Step 1.    Run the following commands to create four FC LIFs (two on each node):

network interface create -vserver Infra-SVM -lif fcp-lif-01a -role data -data-protocol fcp -home-node <st-node01> -home-port 5a -status-admin up

network interface create -vserver Infra-SVM -lif fcp-lif-01b -role data -data- protocol fcp -home-node <st-node01> -home-port 5b -status-admin up

network interface create -vserver Infra-SVM -lif fcp-lif-02a -role data -data-protocol fcp -home-node <st-node02> -home-port 5a -status-admin up

network interface create -vserver Infra-SVM -lif fcp-lif-02b -role data -data-protocol fcp -home-node <st-node02> -home-port 5b -status-admin up

network interface show

Procedure 37.  Create NFS LIFs

Step 1.    Run the following commands to create NFS LIFs, run the following commands:

network interface create -vserver Infra-SVM -lif nfs-lif-01 -role data -data-protocol nfs -home-node <st-node01> -home-port a0a-<infra-nfs-vlan-id> -address <node01-nfs-lif-01-ip> -netmask <node01-nfs-lif-01-mask> -status-admin up -failover-policy broadcast-domain-wide -firewall-policy data -auto-revert true

network interface create -vserver Infra-SVM -lif nfs-lif-02 -role data -data-protocol nfs -home-node <st-node02> -home-port a0a-<infra-nfs-vlan-id> -address <node02-nfs-lif-02-ip> -netmask <node02-nfs-lif-02-mask> -status-admin up -failover-policy broadcast-domain-wide -firewall-policy data -auto-revert true

network interface show

Procedure 38.  Create CIFS LIFs

Step 1.    Run the following commands to create CIFS LIFs:

network interface create -vserver Infra-SVM -lif cifs_lif01 -role data -data-protocol cifs -home-node <st-node01> -home-port a0a-<infra-cifs-vlan-id> –address <node01-cifs_lif01-ip> -netmask <node01-cifs_lif01-mask> -status-admin up –failover-policy broadcast-domain-wide –firewall-policy data –auto-revert true

network interface create -vserver Infra-SVM -lif cifs_lif02 -role data -data-protocol cifs -home-node <st-node02> -home-port a0a-<infra-cifs-vlan-id> –address <node02-cifs_lif02-ip> -netmask <node02-cifs_lif02-mask>> -status-admin up –failover-policy broadcast-domain-wide –firewall-policy data –auto-revert true

network interface show

Procedure 39.  Configure FC-NVMe Datastore for vSphere 7.0U3

Note:   You need to configure FC-NVMe Datastores for vSphere 7.0U3 enable the FC-NVMe protocol on an existing SVM or create a separate SVM for FC-NVMe workloads.

Step 1.    Verify that you have NVMe Capable adapters installed in your cluster.

network fcp adapter show -data-protocols-supported fc-nvme

Step 2.    Add the NVMe protocol to the SVM and list it.

         vserver add-protocols -vserver Infra-SVM -protocols nvme

         vserver show -vserver Infra-SVM -fields allowed-protocols

         aa16-a400::> vserver show -vserver Infra-SVM -fields allowed-protocols

         vserver   allowed-protocols

         --------- -----------------------

         Infra-SVM nfs,fcp,ndmp,nvme

Step 3.    Create NVMe service.

         vserver nvme create -vserver Infra-SVM

         vserver nvme show -vserver Infra-SVM

         aa16-a400::> vserver nvme show -vserver Infra-SVM 
         Vserver Name: Infra-SVM

         Administrative Status: up

Step 4.    Create NVMe FC LIFs.

network interface create -vserver <SVM_name> -lif <lif_name> -role data -data-protocol fc-nvme -home-node <home_node> -home-port <home_port>

network interface create -vserver Infra-SVM -lif fc-nvme-lif-01a -role data -data-protocol fc-nvme -home-node <st-node01> -home-port 5a -status-admin up

network interface create -vserver Infra-SVM -lif fc-nvme-lif-01b -role data -data-protocol fc-nvme -home-node <st-node01> -home-port 5b -status-admin up

network interface create -vserver Infra-SVM -lif fc-nvme-lif-02a -role data -data-protocol fc-nvme -home-node <st-node02> -home-port 5a -status-admin up

network interface create -vserver Infra-SVM -lif fc-nvme-lif-02b -role data -data-protocol fc-nvme -home-node <st-node02> -home-port 5b -status-admin up

 network interface show

Note:   You can only configure two NVMe LIFs per node on a maximum of four nodes.

net int show -vserver <vserver name> -data-protocol fc-nvme

aa16-a400::> net int show -vserver Infra-SVM -data-protocol fc-nvme

  (network interface show)

            Logical    Status     Network            Current       Current Is

Vserver     Interface  Admin/Oper Address/Mask       Node          Port    Home

----------- ---------- ---------- ------------------ ------------- ------- ----

Infra-SVM

            fc-nvme-lif-01a

                         up/up    20:0f:d0:39:ea:17:12:9b

                                                     aa16-a400-01  5a      true

            fc-nvme-lif-01b

                         up/up    20:10:d0:39:ea:17:12:9b

                                                     aa16-a400-01  5b      true

            fcp-nvme-lif-02a

                         up/up    20:11:d0:39:ea:17:12:9b

                                                     aa16-a400-02  5a      true

            fcp-nvme-lif-02b

                         up/up    20:12:d0:39:ea:17:12:9b

                                                     aa16-a400-02  5b      true

4 entries were displayed.

Step 5.    Create Volume.

volume create -vserver SVM-name -volume vol_name -aggregate aggregate_name -size volume_size -state online -space-guarantee none -percent-snapshot-space 0


aa16-a400::> vol create -vserver Infra-SVM -volume NVMe_datastore_01 -aggregate aa16_a400_01_NVME_SSD_1 -size 100G -state online -space-guarantee none -percent-snapshot-space 0


[Job 162] Job succeeded: Successful

Procedure 40.  Add Infrastructure SVM Administrator

Step 1.    Run the following commands:

network interface create -vserver Infra-SVM -lif svm-mgmt -role data -data-protocol none -home-node <st-node02> -home-port a0a-<ib-mgmt-vlan-id> -address <svm-mgmt-ip> -netmask <svm-mgmt-mask> -status-admin up -failover-policy broadcast-domain-wide -firewall-policy mgmt -auto-revert true

Step 2.    Create a default route that enables the SVM management interface to reach the outside world.

network route create -vserver Infra-SVM -destination 0.0.0.0/0 -gateway <svm-mgmt-gateway>

network route show

Step 3.    Set a password for the SVM vsadmin user and unlock the user.

security login password -username vsadmin -vserver Infra-SVM
Enter a new password: <password>
Enter it again: <password>

security login unlock -username vsadmin -vserver Infra-SVM

Note:   A cluster serves data through at least one and possibly several SVMs. These steps have created a single data SVM. If you would like to configure your environment with multiple SVMs, this is a good time to create them.

Procedure 41.  Configure AutoSupport

Note:   NetApp AutoSupport® sends support summary information to NetApp through HTTPS.

Step 1.    Run the following command to configure AutoSupport:

system node autosupport modify -node * -state enable -mail-hosts <mailhost> -transport https -support enable -noteto <storage-admin-email>

Cisco UCS Configuration

This FlexPod deployment explains the configuration steps for the Cisco UCS 6454 Fabric Interconnects (FI). The same base configuration should be done whether you are performing an automated or manual configuration.

Perform Initial Setup of Cisco UCS 6454 Fabric Interconnects for FlexPod Environments

This section provides the detailed procedures for configuring the Cisco Unified Computing System (Cisco UCS) for use in a FlexPod environment. The steps are necessary to provision the Cisco UCS X-Series servers and should be followed precisely to avoid improper configuration.

Procedure 1.     Cisco UCS Fabric Interconnect A

Step 1.    Connect to the console port on the first Cisco UCS fabric interconnect.

            ---- Basic System Configuration Dialog ----

  This setup utility will guide you through the basic configuration of
  the system. Only minimal configuration including IP connectivity to
  the Fabric interconnect and its clustering mode is performed through these steps.

  Type Ctrl-C at any time to abort configuration and reboot system.
  To back track or make modifications to already entered values,
  complete input till end of section and answer no when prompted
  to apply configuration.


<control+C>  

  Type 'reboot' to abort configuration and reboot system
  or Type 'X' to cancel GUI configuratuion and go back to console  or Press any other key to see the installation progress from GUI (reboot/X) ? x
 
  Enter the configuration method. (console/gui) ? console
 
  Enter the management mode. (ucsm/intersight)? intersight

  You have chosen to setup a new Fabric interconnect in "intersight" managed mode. Continue? (y/n): y

  Enforce strong password? (y/n) [y]: y

  Enter the password for "admin": <password>
  Confirm the password for "admin": <password>

  Enter the switch fabric (A/B) []: A

  Enter the system name:  <system name>

  Physical Switch Mgmt0 IP address : XX.XX.XX.XX

  Physical Switch Mgmt0 IPv4 netmask : XXX.XXX.XXX.XXX

  IPv4 address of the default gateway : XX.XX.XX.XX

    DNS IP address : <dns server ip>

  Configure the default domain name? (yes/no) [n]: y

    Default domain name : <dns domain name>

  Following configurations will be applied:

    Management Mode=intersight
    Switch Fabric=A
    System Name=<system_name>
    Enforced Strong Password=yes
    Physical Switch Mgmt0 IP Address=<XX.XX.XX.XX>
    Physical Switch Mgmt0 IP Netmask=<XXX.XXX.XXX.XXX>
    Default Gateway=XX.XX.XX.XX
    DNS Server=XX.XX.XX.XX
    Domain Name=<domain_name>

  Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): y
  Applying configuration. Please wait.

  Configuration file - Ok

Step 2.    Wait for the login prompt for UCS Fabric Interconnect A before proceeding to the next section.

Procedure 2.     Configure Cisco UCS Fabric Interconnect B in Cisco Intersight Managed Mode

Step 1.    Connect to the console port on the second Cisco UCS fabric interconnect.

             ---- Basic System Configuration Dialog ----

  This setup utility will guide you through the basic configuration of
  the system. Only minimal configuration including IP connectivity to
  the Fabric interconnect and its clustering mode is performed through these steps.

  Type Ctrl-C at any time to abort configuration and reboot system.
  To back track or make modifications to already entered values,
  complete input till end of section and answer no when prompted
  to apply configuration.


<control+C>

  Type 'reboot' to abort configuration and reboot system or hit enter to continue. (reboot/<CR>) ? <enter>

   Starting GUI for initial setup
   Starting GUI for initial setup.


   Switch can now be configured from GUI. Use https://XX.XX.XX.XX and click
   on 'Express Setup' link. If you want to cancel the configuration from GUI and go back,
   press the 'ctrl+c' key and choose 'X'. Press any other key to see the installation progress from GUI
   Note: Intersight management mode setup available through console based configuration method alone.


<control+C>

  Type 'reboot' to abort configuration and reboot system
  or Type 'X' to cancel GUI configuratuion and go back to console  or Press any other key to see the installation progress from GUI (reboot/X) ? x
 
  Enter the configuration method. (console/gui) ? console

  Installer has detected the presence of a peer Fabric interconnect. This Fabric interconnect will be added to the cluster. Continue (y/n) ? y

  Enter the admin password of the peer Fabric interconnect:
    Connecting to peer Fabric interconnect... done
    Retrieving config from peer Fabric interconnect... done
    Management Mode is  : intersight
    Peer Fabric interconnect management mode   : intersight
    Peer Fabric interconnect Mgmt0 IPv4 Address: XX.XX.XX.XX
    Peer Fabric interconnect Mgmt0 IPv4 Netmask: XXX.XXX.XXX.XXX
 
    Peer FI is IPv4 Cluster enabled. Please Provide Local Fabric Interconnect Mgmt0 IPv4 Address 

  Physical Switch Mgmt0 IP address : <FI-B management-address>

  Local fabric interconnect model(UCS-FI-6454)
  Peer fabric interconnect is compatible with the local fabric interconnect. Continuing with the installer...

  Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): yes
  Applying configuration. Please wait.

  Completing basic configuration setup

Return from NXOS
Cisco UCS 6400 Series Fabric Interconnect

Step 2.    Wait for the login prompt for UCS Fabric Interconnect B before proceeding to the next section.

Procedure 3.     Cisco Intersight Setup for Fabric Interconnects

Note:   If you have an existing Intersight account, connect to https://intersight.com and sign in with your Cisco ID, and select the appropriate account.

Note:   If you do not already have a Cisco Intersight account on Cisco Intersight, connect to https://intersight.com.

Step 1.    Click Create an account.

Step 2.    Sign-in with your Cisco ID.

Step 3.    Read, scroll through, and accept the End User License Agreement and click Next.

Step 4.    Enter an Account Name and click Create.

Procedure 4.     Log into Cisco Fabric Interconnect Device Console

Step 1.    Open a web browser and navigate to the Cisco UCS fabric interconnect address.

Graphical user interface, applicationDescription automatically generated

Note:   If performing the UCS installation using Terraform, note the serial numbers of both Fabric Interconnects listed on the System Information page. These will be required in the Create_DomainProfile/terraform.tfvars file.

FI_console_SNs

Step 2.    Click the Device Connector tab in the console.

FI_Device_Connector1

Note:   if a Proxy server is needed, the configuration is located under the Settings Related image, diagram or screenshot menu.

FI_Device_Connector2

Step 3.    Open Cisco Intersight, click Admin > Targets

Step 4.    Select Cisco UCS Domain (Intersight Managed) and click Start.

FI_select_target

Step 5.    Copy the Device ID from the Fabric Interconnect Console and copy it to the Device ID field in Intersight.

Step 6.    Copy the Claim Code from the Fabric Interconnect Console and copy it to the Claim Code field in Intersight.

Step 7.    Click Claim.

Step 8.    From the Cisco Intersight window, click Related image, diagram or screenshot and then click Licensing. If this is a new account, all servers connected to the UCS Domain will appear under the Base license Tier. If you have purchased Cisco Intersight licenses and have them in your Cisco Smart Account, click Register and follow the prompts to register this Cisco Intersight account to your Cisco Smart Account. Cisco Intersight also offers a one-time 90-day trial of Premier licensing for new accounts. Click Start Trial and then Start to begin this evaluation. The remainder of this section will assume Premier licensing.

Step 9.    From the Licensing Window, click Set Default Tier. From the drop-down list select Premier for Tier and click Set.

IMM_Default_Tier

Step 10.                        9. Click Refresh to refresh the Intersight window with Premier, Advantage, and Essentials features added.

Cisco Intersight-based Cisco UCS Infrastructure Upgrade

Cisco Intersight can be used to upgrade Cisco UCS Managed Infrastructure. This procedure demonstrates an Intersight-based upgrade of a Cisco UCS Infrastructure (Cisco UCS Manager, fabric interconnects, and IOM modules) from software release 4.1(3b) to 4.2(1f).

Procedure 1.     Upgrade Cisco UCS infrastructure using Cisco Intersight

Step 1.    To start the upgrade, select OPERATE > Fabric Interconnects. To the right of one of the fabric interconnects that you want to upgrade, click … and select Upgrade Firmware.

FI_Select_FIs

Step 2.    Click Start.

Step 3.    Ensure the UCS Domain you want to upgrade is shown and selected and click Next.

Note:   If you receive this error message go back in the process and Enable “Advanced Mode” and disable “Fabric Evacuation.”

FI_upgrade_error

FI_upgrade_domain

Step 4.    Select Version 4.2(1i), select “Advanced Mode, uncheck ”Fabric Interconnect Traffic Evacuation” and click Next.

FI_upgrade_version

Step 5.    Click Upgrade then click Upgrade again to bring up the window allowing you to log in to retrieve software.

FI_Upgrade_summary

Note:   It may be necessary to click Upgrade again to bring up the window allowing you to log in to retrieve software. Login with Cisco ID and password and click Submit.

Step 6.    Click Requests, then click the Upgrade Firmware request.

FI_upgrade_status

Step 7.    While the Cisco UCS Manager and the subordinate fabric interconnect are in the process of upgrading, click Continue and then click Continue again to continue the upgrade.

FI_upgrade_continue

Step 8.    When the upgrade completes, the Status will show Success. M6 servers can now be inserted in the chassis or attached to the FIs. Other servers will be upgraded by the Host Firmware Package setting.

FI_Upgrade_success

Step 9.    From here, proceed to Terraform Automation Workflow if you plan to utilize Terraform to deploy the Cisco UCS resources, or skip to Cisco UCS Manual Configuration if you intend to perform the configuration manually.

Terraform Automation Workflow and Solution Deployment

If using the published Terraform scripts to configure the FlexPod infrastructure, complete this section of the document. If completing a manual configuration, skip to the next section of the document. The Terraform automated FlexPod solution uses a management workstation (control machine) to run Terraform scripts to configure Cisco UCS servers into Cisco Intersight.

Figure 54 illustrates the FlexPod solution implementation workflow which is explained in the following sections. The FlexPod infrastructure layers are first configured in the order illustrated.

Figure 54.       Terraform Automation Workflow

TerraformAutomation Workflow

Terraform Cisco UCS Domain Configuration

The configuration of the Cisco UCS Domain using Terraform is divided into two sections:

1.     Create Domain Profile

2.     Deploy Domain Profile

These two steps require configuration of various variables in files located within each subdirectory of the Terraform repository.

Procedure 1.     Create UCS Domain Profile from the Terraform Management Workstation

Step 1.    Move to the Create_DomainProfile directory of the repository FlexPod-for-EHR/Terraform/

Step 2.    Edit the following variable file to ensure proper variables are entered:  terraform.tfvars

Note:   It is critical when entering the variable files that consistency is maintained between the Terraform scripts and Ansible scripts.

Step 3.    From Create_DomainProfile directory, run the following Terraform commands;

terraform init

Note:   This command will initialize the directory structure and download the Intersight provider information from the Terraform Registry.

Step 4.    Once the initialization is complete, run the following command;

terraform plan

Note:   Terraform plan creates an execution plan based on the configuration files. It does not actually carry out the instructions to Intersight. Once the plan stage has completed, review the output to determine if any errors need to be corrected prior to applying the configuration.

Step 5.    If the plan appears to be correct, run the following command;

terraform apply

Step 6.    Respond “yes” when Terraform asks; “Do you want to perform these actions?”

Note:   Terraform apply performs the actual configuration of the infrastructure in Cisco Intersight. This can be verified by clicking Configure > Profiles > UCS Domain Profiles. At this point, the profile is created by not Deployed.

TF_Domain_Prof_not_applied

You will deploy the profile in the next section.

Procedure 2.     Deploy Domain Profile

Step 1.    Move to the Deploy_DomainProfile directory of the repository

Step 2.    Edit the following variable file to ensure proper variables are entered:

terraform.tfvars

Note:   This file can be used to “Deploy” or “Unassign” the Domain profile by simply changing the variable and running terraform plan and terraform apply.

Step 3.    From Create_DomainProfile directory, run the following Terraform commands:

terraform init

Note:   This command will initialize the directory structure and download the Intersight provider information from the Terraform Registry.

Step 4.    Once the initialization is complete, run the following command:

terraform plan

Note:   Terraform plan creates an execution plan based on the configuration files. It does not actually carry out the instructions to Intersight. Once the plan stage has completed, review the output to determine if any errors need to be corrected prior to applying the configuration.

Step 5.    If the plan appears to be correct, run the following command:

terraform apply

Step 6.    Respond “yes” when Terraform asks; “Do you want to perform these actions?”

Note:   Terraform apply performs the actual configuration of the infrastructure in Cisco Intersight. This can be verified by clicking on Configure > Profiles > UCS Domain Profiles. At this point, the profile is Deployed.

TF_Deploy_DomainProfile_applied.png

Note:   This step can take time to complete.

Cisco UCS Server Configuration

The creation of Cisco UCS Server profiles is divided into two boot modes, local-boot (M.2) or Boot-from-SAN (FC). Each mode offers the ability to create Server Templates which can be used to derive profiles, or to create Server Profiles (without the creating templates.) Refer to Table 11 for clarification of required scripts. The choice is left to the customer. In either case, the unused directories can be deleted from the local automation repository.

Table 11.   Terraform Scripts for Cisco UCS Server Configuration

 

SAN (fc) Boot

Local (m.2) Boot

Profiles

Create SAN-Boot_Profiles

Create Local-Boot_Profiles

 

Deploy SAN-Boot_Profiles

Deploy Local-Boot_Profiles

Templates

Create SAN-Boot_Templates

Create_Local-Boot_Templates

Procedure 1.     Create UCS Server Profiles from the Terraform Management Workstation

Step 1.    Move to the Create_SAN-Boot_Server_Profiles directory of the repository FlexPod-for-EHR/Terraform/

Step 2.    Edit the following variable file to ensure proper variables are entered:

terraform.tfvars

Note:   Obtain the WWPN and Boot device names from the NetApp Storage Array.

Step 3.    From Create_SAN-Boot_Server_Profiles directory, run the following Terraform commands:

terraform init

Note:   This command will initialize the directory structure and download the Intersight provider information from the Terraform Registry.

Step 4.    Once the initialization is complete, run the following command:

terraform plan

Note:   Terraform plan creates an execution plan based on the configuration files. It does not actually carry out the instructions to Intersight. Once the plan stage has completed, review the output to determine if any errors need to be corrected prior to applying the configuration.

Step 5.    If the plan appears to be correct, run the following command:

terraform apply

Step 6.    Respond “yes” when Terraform asks; “Do you want to perform these actions?”

Note:   Terraform apply performs the actual configuration of the infrastructure in Cisco Intersight. This can be verified by clicking on Configure > Profiles > UCS Domain Profiles. At this point, the profile is created by not Deployed.

UCS_Server_Prof_not_applied

You will deploy the profile in the next section.

Procedure 2.     Deploy UCS Server Profiles from the Terraform Management Workstation

Step 1.    Move to the Deploy_SAN-Boot_Server_Profiles directory of the repository FlexPod-for-EHR/Terraform/

Step 2.    Edit the following variable file to ensure proper variables are entered:

terraform.tfvars

Note:   This file can be used to “Deploy” or “Unassign” the Domain profile by simply changing the variable and running terraform plan and terraform apply.

Step 3.    From Deploy_SAN-Boot_Server_Profiles directory, run the following Terraform commands:

terraform init

Note:   This command will initialize the directory structure and download the Intersight provider information from the Terraform Registry.

Step 4.    Once the initialization is complete, run the following command:

terraform plan

Note:   Terraform plan creates an execution plan based on the configuration files. It does not actually carry out the instructions to Intersight. Once the plan stage has completed, review the output to determine if any errors need to be corrected prior to applying the configuration.

Step 5.    If the plan appears to be correct, run the following command:

terraform apply

Step 6.    Respond “yes” when Terraform asks; “Do you want to perform these actions?”

Note:   Terraform apply performs the actual configuration of the infrastructure in Cisco Intersight. This can be verified by clicking on Configure > Profiles > UCS Server Profiles. At this point, the profile is Deployed.

TF_Deploy_DomainProfile_applied.png

Note:   This step can take time to complete.

Procedure 3.     Create UCS Server Templates from the Terraform Management Workstation

Step 1.    Move to the Create_SAN-Boot_Server_Templates directory of the repository FlexPod-for-EHR/Terraform

Step 2.    Edit the following variable file to ensure proper variables are entered:

terraform.tfvars

Note:   Obtain the WWPN and Boot device names from the NetApp Storage Array.

Step 3.    From Create_SAN-Boot_Server_Templates directory, run the following Terraform commands:

terraform init

Note:   This command will initialize the directory structure and download the Intersight provider information from the Terraform Registry.

Step 4.    Once the initialization is complete, run the following command:

terraform plan

Note:   Terraform plan creates an execution plan based on the configuration files. It does not actually carry out the instructions to Intersight. Once the plan stage has completed, review the output to determine if any errors need to be corrected prior to applying the configuration.

Step 5.    If the plan appears to be correct, run the following command:

terraform apply

Step 6.    Respond “yes” when Terraform asks; “Do you want to perform these actions?”

Note:   Terraform apply performs the actual configuration of the infrastructure in Cisco Intersight. This can be verified by clicking on Configure > Profiles > UCS Domain Profiles. At this point, the profile is created by not Deployed.

UCS_Server_Prof_not_applied

Procedure 4.     Derive UCS Server Profiles from Templates

Note:   Currently, there is no Terraform module to automate creation of Profiles from Templates.

Step 1.    From the Server profile template Summary screen, click Derive Profiles.

Note:   This action can also be performed later by navigating to Templates, clicking “…” next to the template name and selecting Derive Profiles.

Step 2.    Under the Server Assignment, select Assign Now and select Cisco UCS X210c M6 servers. Customers can select one or more servers depending on the number of profiles to be deployed.

A screenshot of a computerDescription automatically generated

Note:   Alternately, you can derive multiple profiles and assign servers later.

A screenshot of a computerDescription automatically generated with medium confidence

Step 3.    Click Next.

A screenshot of a computerDescription automatically generated

Step 4.    Adjust the names and Start Index for Suffix as desired.

Step 5.    Click Next.

Step 6.    Review the resulting Profiles and Server assignments and click Derive.

Cisco UCS Manual Configuration

Create a Cisco UCS Domain Profile

First you need to create a Cisco UCS domain profile to configure the fabric interconnect ports and discover connected chassis. A domain profile is composed of several policies. Table 12 lists the policies required for the solution described in this document.

Table 12.   Policies required for domain profile

Policy

Description

Multicast policy for each VLAN policy

allow/disallow IGMP snooping

VLAN Policy

Allowed VLANs, Native VLAN

VSAN configuration policy for fabric A

VSAN Name, ID

VSAN configuration policy for fabric B

VSAN Name, ID

Port configuration policy for fabric A

Server links, Uplinks and Port-Channels

Port configuration policy for fabric B

Server links, Uplinks and Port-Channels

System QoS Policy

Jumbo Frame support, QoS settings

Optional Policies

Network Time Protocol (NTP) policy

Time zone and NTP server configuration

Syslog

Syslog Sev. And server settings

Network Connectivity

DNS settings

SNMP

Simple Network Management Protocol settings

Switch Control

Switching modes, MAC aging time, and so on.

 

A screenshot of a computerDescription automatically generated

Procedure 1.     Create a Cisco UCS Domain Profile

Step 1.    Create the Multicast policy:

a.     Under Policies, select Create Policy. Then click Multicast.

b.     Select the Organization (if one is in use)

c.     Give the policy a name (for example, Multicast_Policy)

Note:   Each VLAN can have a separate Multicast Policy, or you can use the same policy on all VLANs.

A screenshot of a computerDescription automatically generated with medium confidence

d.     Move the selectors to indicate the policy desired

Step 2.    Create the VLAN configuration policy:

a.     Under Policies, select Create Policy. Then select VLAN and give the VLAN a name (for example, VLAN_Config_Policy).

b.     Click Add VLANs to add your required VLANs.

c.     Click Multicast Policy to add the multicast policy you just created for your VLAN policy.

d.     Click Create.

A screenshot of a computerDescription automatically generated

Note:   If you will be using the same VLANs on fabric interconnect A and fabric interconnect B, you can use the same policy for both.

Step 3.    Create the VSAN configuration policy:

a.     Under Policies, select Create Policy. Then select VSAN and give the VSAN a name (for example, VSAN_Config_Policy_A).

b.     Click Add VSAN and add the required VSAN names and IDs.

c.     Add the corresponding FCoE VLAN ID.

Note:   According to best practices, the VSAN IDs should be different for each fabric, so you should create two separate policies.

Step 4.    Create the Port Configuration policy:

a.     Under Policies, select Create Policy.

b.     Select Port and Start.

c.     Give the port policy a name.

d.     Click Next.

e.     To configure Converged Ports, use the slider to define Fibre Channel ports. Click Next.

A screenshot of a computerDescription automatically generated with medium confidence

f.      Define the port roles: server ports for chassis and server connections, Fibre Channel ports for SAN connections, or network uplink ports.

g.     Select ports 1 through 4 and click Configure.

A screenshot of a computerDescription automatically generated with medium confidence

h.     From the drop-down list, select FC Uplink and define the admin speed and VSAN ID (the solution described here uses 32 GBps for the speed and VSAN ID 111 for fabric interconnect A and 112 for fabric interconnect B).

Graphical user interface, applicationDescription automatically generated

i.       To configure server ports, select the ports that have chassis or rack-mounted servers plugged into them and click Configure. From the drop-down list, select Server.

Graphical user interface, applicationDescription automatically generated

j.       To configure network uplink ports, select the ports connected to the upstream network switches and click Configure. From the drop-down list, select Ethernet Uplink.

Graphical user interface, applicationDescription automatically generated

k.     To configure Port-Channels, select unconfigured ports, select Create Port Channel.

Graphical user interfaceDescription automatically generated with medium confidence

l.       Select “Ethernet Uplink Port Channel.

m.   Enter the Port-Channel ID and admin speed.

n.     Additional policies can be created and applied as needed.

Graphical user interface, applicationDescription automatically generated

o.     Click Save to save the port policy.

Step 5.    Repeat steps 1 – 4 to create a port policy for Fabric Interconnect B.

Step 6.    Create the NTP policy:

a.     Under Policies, select Create Policy.

b.     Select NTP and Start.

c.     Give the NTP policy a name.

d.     Click Next.

e.     Define the name or IP address for the NTP servers.

f.      Define the correct time zone.

g.     Click Create.

Step 7.    (Optional) Create the syslog policy:

a.     Under Policies, select Create Policy.

b.     Select Syslog and Start.

c.     Give the syslog policy a name.

d.     Click Next.

e.     Define the syslog severity level that triggers a report.

f.      Define the name or IP address for the syslog servers.

g.     Click Create.

Note:   You do not need to enable the syslog server.

Step 8.    Create the Domain Profile:

a.     Click Profiles in the side menu.

b.     Select UCS Domain Profiles at the top of the screen.

c.     Click Create UCS Domain Profile.

A blue screen with white textDescription automatically generated with low confidence

d.     Give the profile a name (for example, HC-UCS) and click Next.

e.     Select the fabric interconnect domain pair created when you claimed your fabric interconnects.

A screenshot of a computerDescription automatically generated with medium confidence

f.      Under VLAN & VSAN Configuration, click Select Policy to select the policies created earlier. (Be sure that you select the appropriate policy for each side of the fabric.)

g.     Under Ports Configuration, select the port configuration policies created earlier. (Be sure to assign the correct fabric policy to the appropriate side: that is, assign port_FIA to fabric interconnect A and assign port_FIB to fabric interconnect B)

h.     Under UCS Domain Configuration, select all the policies you created earlier.

A screenshot of a computerDescription automatically generated

i.       Click Next.

j.       Click Deploy.

Graphical user interface, application, websiteDescription automatically generated

Create a Cisco UCS Server Profile

Much like Cisco UCS Manager, Cisco Intersight managed mode lets you to deploy server profiles that consist of a series of pools and policies. This capability allows detailed management of the entire environment.

Pools used in a Server Profile

Some Cisco UCS Server Profiles are made up of values derived from pools that are assigned to devices and used to build a server profile. Table 13 lists the Pools required for UCS servers in this CVD.

Table 13.   List of Required Pools

Pool

Description

WWNN Pool

A collection of WWN addresses that can be allocated to VHBAs of a Server Profile

WWPN Pool

A collection of WW Port Name addresses that can be allocated to VHBAs of a Server Profile

NVME Pool

A collection of WWPN addresses used for NMVe Initiators of the VHBA of a server Profile

MAC Pool

A collection of MAC addresses that can be allocated to VNICs of a server profile.

IP Pool

A collection of IPv4 and/or IPv6 addresses that can be allocated to other configuration entities like server profiles

Procedure 1.     Create Required Pools

Step 1.    Create two MAC address pools: one for fabric A and one for fabric B.

Graphical user interface, application, TeamsDescription automatically generated

Step 2.    Create a WWNN Pool.

UCSS-WWNN-Pool

Step 3.    For Boot from FC, create two FC WWPN Pools, one for fabric A, one for fabric B.

Note:   If you’re not utilizing Boot-from-SAN, skip this step.

Graphical user interface, applicationDescription automatically generated

Step 4.    For Shared Storage access (NVMe), create two WWPN Pools, one for Fabric A, one for Fabric B

Related image, diagram or screenshot

Step 5.    Create two IP Pools for IMC Access. One pool uses IP addresses from the [IB-MGMT] range, the second pool uses IP addresses from the [OOB-MGMT] range.

A screenshot of a computerDescription automatically generated

Polices in a Server Profile

Table 14 lists the policies that make up a server profile.

Table 14.   Server Profile Policies

Policy

 

 Compute Configuration

 

BIOS

 

Boot Order

 

Power

 

Virtual Media

Management Configuration

 

IMC Access

 

IPMI Over LAN

 

Local User

 

Serial Over LAN

 

SNMP

 

Syslog

 

Virtual KVM

Storage Configuration

 

SD Card

 

Storage

Network Configuration Policies

 

LAN Connectivity

 

SAN Connectivity

Policies Needed by Other Policies

 

Fibre Channel Network (A/B)

 

Fibre Channel Adapter

 

Fibre Channel QoS

 

Ethernet Network Group Policy

 

Ethernet Network Control Policy

 

Ethernet QoS Policy

 

Ethernet Adapter Policy

Procedure 2.     Create Supporting Policies

Step 1.    Create two Fibre Channel Network Policies. One for Fabric A and one for Fabric B. Be sure to select UCS Server (FI-Attached) and enter the VSAN ID for each fabric.

Graphical user interface, application, TeamsDescription automatically generated

Step 2.    Create a Fibre Channel Adapter Policy:

A screenshot of a computerDescription automatically generated with medium confidence

a.     Based on Epic performance guidance, increase the I/O Throttle Count to 1024

b.     Set the LUN Queue Depth to 254

c.     For NVME initiator, set the SCSI I/O queues to 16

A screenshot of a computerDescription automatically generated with medium confidence

Step 3.    Create a Fibre Channel QoS Policy:

A screenshot of a computer screenDescription automatically generated

a.     Ethernet Network Group Policy. Ensure the listed VLANs match the VLANs in the UCS domain profile.

Graphical user interface, applicationDescription automatically generated

b.     Ethernet Network Control Policy:

Graphical user interface, applicationDescription automatically generated

c.     Ethernet QoS Policy (ensure the MTU is set to 9000):

A screenshot of a computerDescription automatically generated with medium confidence

d.     Ethernet Adapter Policy:

A screenshot of a computerDescription automatically generated with medium confidence

Step 4.    For local-boot, create an SD Card Policy.

A screenshot of a computer screenDescription automatically generated

Procedure 3.     Configure BIOS Policy

Step 1.    Click Select Policy next to BIOS and in the pane on the right, click Create New.

Step 2.    Verify correct organization is selected from the drop-down list  and provide a name for the policy (for example, HC-BIOS-Policy).

Step 3.    Click Next.

Step 4.    On the Policy Details screen, select the appropriate values for the BIOS settings. In this deployment, the BIOS values were selected based on recommendations from Epic and in the performance tuning guide for Cisco UCS M6 BIOS: https://www.cisco.com/c/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/performance-tuning-guide-ucs-m6-servers.html.

Graphical user interface, application, websiteDescription automatically generated

a.     LOM and PCIe Slot > CDN Support for LOM: Enabled

b.     Processor > Enhanced CPU performance: Auto

c.     Memory > NVM Performance Setting: Balanced Profile

Step 5.    Define a boot order Policy:

a.     Add the ability to boot from KVM-mapped ISO

Note:   If using FC Boot, values must be obtained from the NetApp Controller.

A screenshot of a computerDescription automatically generated

Note:   If using local storage (M.2):

A screenshot of a computerDescription automatically generated

Procedure 4.     Create Local User Policy

Step 1.    Create a local user policy:

A screenshot of a computerDescription automatically generated with medium confidence

Step 2.    Create a Virtual KVM policy:

 Graphical user interface, text, application, TeamsDescription automatically generated

Step 3.    Define a boot order policy:

Graphical user interfaceDescription automatically generated

Step 4.    Define the Cisco Integrated Management Controller (IMC) Access policy and IP address pool for the keyboard, video, and mouse (KVM):

a.     Enter the In-Band Mgmt VLAN ID.

b.     Use the In-Band Mgmt and OOB-Mgmt IP pools in the respective slots.

A screenshot of a computerDescription automatically generated

Procedure 5.     LAN Connectivity Policy

Note:   Epic Best practices suggest creating 2 vNICs for public and management traffic, and 2 vNICs for vMotion traffic.

A screenshot of a computerDescription automatically generated

Step 1.    For vNIC Configuration, select Auto vNICs Placement.

Step 2.    Click Add vNIC. Use the MAC Pools and Policies created earlier as needed.

A screenshot of a computerDescription automatically generated with medium confidence

Procedure 6.     Create a SAN Connectivity Policy

Note:   The required vHBAs will depend on if Boot-From-SAN is in use.

A screenshot of a computerDescription automatically generated with medium confidence

A screenshot of a computerDescription automatically generated

Step 1.    Select Auto vHBA Placement then Pool under WWNN Address.

Step 2.    Select the WWNN Pool created earlier

Step 3.    Click Add vHBA.

Step 4.    If using FC Boot-from SAN, create two FC HBAs, one in Fabric A and One in Fabric B.

Graphical user interface, applicationDescription automatically generated

Step 5.    Create two NVME HBAs for attaching to the A400 NVME SVMs.

Graphical user interface, applicationDescription automatically generated

The final SAN Connectivity Policy is shown below:

A screenshot of a computerDescription automatically generated

Finalize Cisco UCS Server Profile

After you have created all these polices and pools, step through the Create Server Profile wizard and select each policy or pool as appropriate. After this process is complete, the service profile can be cloned, converted to a template, or directly assigned to a compute node.

SAN Switch Configuration

This section explains how to configure the Cisco MDS 9000s for use in a FlexPod environment. Follow the steps precisely because failure to do so could result in an improper configuration.

Physical Connectivity

Follow the physical connectivity guidelines for FlexPod as explained in section FlexPod Cabling.

FlexPod Cisco MDS Base

The following procedures describe how to configure the Cisco Nexus switches for use in a base FlexPod environment. This procedure assumes you are using the Cisco MDS 9132T with NX-OS 8.4(2c).

Procedure 1.     Configure Cisco MDS 9132T A Switch

Note:   On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning. Enter y to get to the System Admin Account Setup.

Step 1.    Configure the switch using the command line.

         ---- System Admin Account Setup ----
 

Do you want to enforce secure password standard (yes/no) [y]: Enter


Enter the password for "admin": <password>

Confirm the password for "admin": <password>
 

Would you like to enter the basic configuration dialog (yes/no): yes
 

Create another login account (yes/no) [n]: Enter
 

Configure read-only SNMP community string (yes/no) [n]: Enter
 

Configure read-write SNMP community string (yes/no) [n]: Enter
 

Enter the switch name : <mds-A-hostname>
 

Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: Enter
 

Mgmt0 IPv4 address : <mds-A-mgmt0-ip>
 

Mgmt0 IPv4 netmask : <mds-A-mgmt0-netmask>
 

Configure the default gateway? (yes/no) [y]: Enter
 

IPv4 address of the default gateway : <mds-A-mgmt0-gw>
 

Configure advanced IP options? (yes/no) [n]: Enter
 

Enable the ssh service? (yes/no) [y]: Enter
 

Type of ssh key you would like to generate (dsa/rsa) [rsa]: Enter
 

Number of rsa key bits <1024-2048> [1024]: Enter
 

Enable the telnet service? (yes/no) [n]: Enter
 

Configure congestion/no_credit drop for fc interfaces? (yes/no) [y]: Enter



Enter the type of drop to configure congestion/no_credit drop? (con/no) [c]: Enter
 

Enter milliseconds in multiples of 10 for congestion-drop for logical-type edge
in range (<200-500>/default), where default is 500.  [d]: Enter
 

Enable the http-server? (yes/no) [y]: Enter
 

Configure clock? (yes/no) [n]: Enter
 

Configure timezone? (yes/no) [n]: Enter
 

Configure summertime? (yes/no) [n]: Enter
 

Configure the ntp server? (yes/no) [n]: Enter
 

Configure default switchport interface state (shut/noshut) [shut]: Enter
 

Configure default switchport trunk mode (on/off/auto) [on]: auto
 

Configure default switchport port mode F (yes/no) [n]: yes
 

Configure default zone policy (permit/deny) [deny]: Enter
 

Enable full zoneset distribution? (yes/no) [n]: Enter
 

Configure default zone mode (basic/enhanced) [basic]: Enter

Step 2.    Review the configuration.

Would you like to edit the configuration? (yes/no) [n]: Enter

Use this configuration and save it? (yes/no) [y]: Enter

Procedure 2.     Configure Cisco MDS 9132T B Switch

To set up the initial configuration for the Cisco MDS B switch, follow these steps:

Note:   On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning. Enter y to get to the System Admin Account Setup.

Step 1.    Configure the switch using the command line.

         ---- System Admin Account Setup ----
 

Do you want to enforce secure password standard (yes/no) [y]: Enter
 

Enter the password for "admin": <password>

Confirm the password for "admin": <password>
 

Would you like to enter the basic configuration dialog (yes/no): yes
 

Create another login account (yes/no) [n]: Enter
 

Configure read-only SNMP community string (yes/no) [n]: Enter
 

Configure read-write SNMP community string (yes/no) [n]: Enter
 

Enter the switch name : <mds-B-hostname>
 

Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: Enter
 

Mgmt0 IPv4 address : <mds-B-mgmt0-ip>
 

Mgmt0 IPv4 netmask : <mds-B-mgmt0-netmask>
 

Configure the default gateway? (yes/no) [y]: Enter
 

IPv4 address of the default gateway : <mds-B-mgmt0-gw>
 

Configure advanced IP options? (yes/no) [n]: Enter
 

Enable the ssh service? (yes/no) [y]: Enter
 

Type of ssh key you would like to generate (dsa/rsa) [rsa]: Enter


Number of rsa key bits <1024-2048> [1024]: Enter



Enable the telnet service? (yes/no) [n]: Enter
 

Configure congestion/no_credit drop for fc interfaces? (yes/no)  [y]: Enter
 

Enter the type of drop to configure congestion/no_credit drop? (con/no) [c]: Enter
 

Enter milliseconds in multiples of 10 for congestion-drop for logical-type edge
in range (<200-500>/default), where default is 500.  [d]: Enter
 

Enable the http-server? (yes/no) [y]: Enter
 

Configure clock? (yes/no) [n]: Enter
 

Configure timezone? (yes/no) [n]: Enter


Configure summertime? (yes/no) [n]: Enter
 

Configure the ntp server? (yes/no) [n]: Enter
 

Configure default switchport interface state (shut/noshut) [shut]: Enter
 

Configure default switchport trunk mode (on/off/auto) [on]: auto
 

Configure default switchport port mode F (yes/no) [n]: yes
 

Configure default zone policy (permit/deny) [deny]: Enter
 

Enable full zoneset distribution? (yes/no) [n]: Enter
 

Configure default zone mode (basic/enhanced) [basic]: Enter

Step 2.    Review the configuration.

Would you like to edit the configuration? (yes/no) [n]: Enter

Use this configuration and save it? (yes/no) [y]: Enter

Procedure 3.     Configure FlexPod Cisco MDS Ansible Switch

Step 1.    Add MDS switch ssh keys to /root/.ssh/known_hosts. Adjust known_hosts as necessary if errors occur.

ssh admin@<mds-A-mgmt0-ip>
exit

ssh admin@<mds-B-mgmt0-ip>
exit

Step 2.    Edit the following variable files to ensure proper MDS variables are entered:

      FlexPod-for-EHR/Ansible/inventory

      FlexPod-for-EHR/Ansible/group_vars/all.yml

      FlexPod-for-EHR/Ansible/host_vars/mdsA.yml

      FlexPod-for-EHR/Ansible/host_vars/mdsB.yml

      FlexPod-for-EHR/Ansible/roles/MDSconfig/defaults/main.yml

Note:   The FC and FC-NVMe NetApp LIF WWPNs should have already been entered into the all.yml file so that MDS device alias can be properly configured. The Cisco UCS server initiator WWPNs for both FC and FC-NVMe should also be entered into all.yml. To query these WWPNs, log into the Cisco Intersight web interface and select “Configure > Pools >.

Procedure 4.     Obtain the WWPNs for NetApp FC LIFs

Step 1.    Run the following commands on NetApp cluster management console:

   network interface show -vserver Infra-SVM -data-protocol fcp     

Procedure 5.     Obtain the WWPNs for Cisco UCS Server Profiles

Step 1.    From the Intersight GUI, go to: CONFIGURE > Profiles. Select UCS Server Profiles and click [Server Profile Name]. Under General, click SAN Connectivity and find the WWPN information for various vHBAs in the window on the right.

Graphical user interface, text, applicationDescription automatically generated

Step 2.    Alternately, a list of assigned WWPNs can be found by exporting the values from the WWPN Pools. Select each of the pools,  click the Export button to download a .csv file containing the port names and server assignments.

Graphical user interface, applicationDescription automatically generated

Step 3.    From /root/ FlexPod-for-EHR/Ansible, run the Setup_MDS.yml Ansible playbook.

ansible-playbook ./Setup_MDS.yml -i inventory

Step 4.    Once the Ansible playbook has been run and configured both switches, it is important to configure the local time so that logging time alignment and any backup schedules are correct. For more information on configuring the timezone and daylight savings time or summertime, please see Cisco MDS 9000 Series Fundamentals Configuration Guide, Release 8.x. Sample clock commands for the United States Pacific timezone are:

clock timezone PST -8 0

clock summer-time PDT 2 Sunday March 02:00 1 Sunday November 02:00 60

Step 5.    ssh into each switch and run the following commands:

clock timezone <timezone> <hour-offset> <minute-offset>

clock summer-time <timezone> <start-week> <start-day> <start-month> <start-time> <end-week> <end-day> <end-month> <end-time> <offset-minutes>

FlexPod Cisco MDS Switch Manual Configuration

Procedure 1.     Enable Licenses on Cisco MDS 9132T A and Cisco MDS 9132T B

Step 1.    Log in as admin.

Step 2.    Run the following commands:

configure terminal
feature npiv
feature fport-channel-trunk

Procedure 2.     Add NTP Servers and Local Time Configuration on Cisco MDS 9132T A and Cisco MDS 9132T B

Step 1.    From the global configuration mode, run the following command:

ntp server <nexus-A-mgmt0-ip>

ntp server <nexus-B-mgmt0-ip>

clock timezone <timezone> <hour-offset> <minute-offset>

clock summer-time <timezone> <start-week> <start-day> <start-month> <start-time> <end-week> <end-day> <end-month> <end-time> <offset-minutes>

Note:   It is important to configure the local time so that logging time alignment, any backup schedules, and SAN Analytics forwarding are correct. For more information on configuring the timezone and daylight savings time or summer time, please see Cisco MDS 9000 Series Fundamentals Configuration Guide, Release 8.x. Sample clock commands for the United States Pacific timezone are:
clock timezone PST -8 0
clock summer-time PDT 2 Sunday March 02:00 1 Sunday November 02:00 60

Procedure 3.     Configure Individual Ports for Cisco MDS 9132T A

Step 1.    From the global configuration mode, run the following commands:

interface fc1/1
switchport description <st-clustername>-1:5a
switchport speed 32000
switchport trunk mode off
no shutdown
exit

interface fc1/2
switchport description \<st-clustername>-2:5a
switchport speed 32000
switchport trunk mode off
no shutdown
exit

interface fc1/5
switchport description <ucs-clustername>-a:1/1
channel-group 15
no shutdown
exit

interface fc1/6
switchport description <ucs-clustername>-a:1/2
channel-group 15
no shutdown
exit

interface port-channel15
channel mode active
switchport trunk allowed vsan <vsan-a-id>
switchport description <ucs-clustername>-a
switchport speed 32000
no shutdown
exit

Note:   If VSAN trunking is not being used between the Cisco UCS Fabric Interconnects and the MDS switches, do not enter “switchport trunk allowed vsan ” for interface port-channel15. Also, the default setting of the switchport trunk mode auto is being used for the port channel.

Procedure 4.     Configure Individual Ports for Cisco MDS 9132T B

Step 1.    From the global configuration mode, run the following commands:

interface fc1/1
switchport description <st-clustername>-1:5b
switchport speed 32000
switchport trunk mode off
no shutdown
exit

interface fc1/2
switchport description <st-clustername>-2:5b
switchport speed 32000
switchport trunk mode off
no shutdown
exit

interface fc1/5
switchport description <ucs-clustername>-b:1/1
channel-group 15
no shutdown
exit

interface fc1/6
switchport description <ucs-clustername>-b:1/2
channel-group 15
no shutdown
exit

interface port-channel15
channel mode active
switchport trunk allowed vsan <vsan-b-id>
switchport description <ucs-clustername>-b
switchport speed 32000
no shutdown
exit

Note:   If VSAN trunking is not being used between the Cisco UCS Fabric Interconnects and the MDS switches, do not enter “switchport trunk allowed vsan ” for interface port-channel15. Also, the default setting of the switchport trunk mode auto is being used for the port channel.

Procedure 5.     Create VSANs for Cisco MDS 9132T A

Step 1.    From the global configuration mode, run the following commands:

vsan database
vsan <vsan-a-id>
vsan <vsan-a-id> name Fabric-A
exit

zone smart-zoning enable vsan <vsan-a-id>

vsan database
vsan <vsan-a-id> interface fc1/1
vsan <vsan-a-id> interface fc1/2
vsan <vsan-a-id> interface port-channel15
exit

Procedure 6.     Create VSANs for Cisco MDS 9132T B

Step 1.    From the global configuration mode, run the following commands:

vsan database
vsan <vsan-b-id>
vsan <vsan-b-id> name Fabric-B
exit

zone smart-zoning enable vsan <vsan-b-id>

vsan database
vsan <vsan-b-id> interface fc1/1
vsan <vsan-b-id> interface fc1/2
vsan <vsan-b-id> interface port-channel15

exit

Procedure 7.     Create Device Aliases for Cisco MDS 9132T A

Step 1.    From the global configuration mode, run the following commands:

device-alias mode enhanced
device-alias database
device-alias name Infra_SVM-fcp-lif-01a pwwn <fcp-lif-01a-wwpn>
device-alias name Infra_SVM-fcp-lif-02a pwwn <fcp-lif-02a-wwpn>
device-alias name Infra_SVM-fc-nvme-lif-01a pwwn <fc-nvme-lif-01a-wwpn>
device-alias name Infra_SVM-fc-nvme-lif-02a pwwn <fc-nvme-lif-02a-wwpn>
device-alias name VM-Host-Infra-FCP-01-A pwwn <vm-host-infra-fcp-01-wwpna>
device-alias name VM-Host-Infra-FCP-02-A pwwn <vm-host-infra-fcp-02-wwpna>
device-alias name VM-Host-Infra-FCP-03-A pwwn <vm-host-infra-fcp-03-wwpna>

device-alias name VM-Host-Infra-FCP-04-A pwwn <vm-host-infra-fcp-04-wwpna>
device-alias name VM-Host-Infra-FC-NVMe-01-A pwwn <vm-host-infra-fc-nvme-01-wwpna>
device-alias name VM-Host-Infra-FC-NVMe-02-A pwwn <vm-host-infra-fc-nvme-02-wwpna>
device-alias name VM-Host-Infra-FC-NVMe-03-A pwwn <vm-host-infra-fc-nvme-03-wwpna>
device-alias name VM-Host-Infra-FC-NVMe-04-A pwwn <vm-host-infra-fc-nvme-04-wwpna>
device-alias commit

Procedure 8.     Create Device Aliases for Cisco MDS 9132T B

Step 1.    From the global configuration mode, run the following commands:

device-alias mode enhanced
device-alias database
device-alias name Infra_SVM-fcp-lif-01b pwwn <fcp-lif-01b-wwpn>
device-alias name Infra_SVM-fcp-lif-02b pwwn <fcp-lif-02b-wwpn>
device-alias name Infra_SVM-fc-nvme-lif-01b pwwn <fc-nvme-lif-01b-wwpn>
device-alias name Infra_SVM-fc-nvme-lif-02b pwwn <fc-nvme-lif-02b-wwpn>
device-alias name VM-Host-Infra-FCP-01-B pwwn <vm-host-infra-fcp-01-wwpnb>
device-alias name VM-Host-Infra-FCP-02-B pwwn <vm-host-infra-fcp-02-wwpnb>
device-alias name VM-Host-Infra-FCP-03-B pwwn <vm-host-infra-fcp-03-wwpnb>
device-alias name VM-Host-Infra-FCP-04-B pwwn <vm-host-infra-fcp-04-wwpnb>

device-alias name VM-Host-Infra-FC-NVMe-01-B pwwn <vm-host-infra-fc-nvme-01-wwpnb>
device-alias name VM-Host-Infra-FC-NVMe-02-B pwwn <vm-host-infra-fc-nvme-02-wwpnb>
device-alias name VM-Host-Infra-FC-NVMe-03-B pwwn <vm-host-infra-fc-nvme-03-wwpnb>

device-alias name VM-Host-Infra-FC-NVMe-04-B pwwn <vm-host-infra-fc-nvme-04-wwpnb>
device-alias commit

Procedure 9.     Create Zones and Zonesets on Cisco MDS 9132T A

Step 1.    To create the required zones and zonesets on Fabric A, run the following commands:

configure terminal
zone name FCP-Infra_SVM-Fabric-A vsan <vsan-a-id>
member device-alias VM-Host-Infra-FCP-01-A init
member device-alias VM-Host-Infra-FCP-02-A init
member device-alias VM-Host-Infra-FCP-03-A init

member device-alias VM-Host-Infra-FCP-04-A init
member device-alias Infra_SVM-fcp-lif-01a target
member device-alias Infra_SVM-fcp-lif-02a target
exit

zone name FC-NVMe-Infra_SVM-Fabric-A vsan <vsan-a-id>
member device-alias VM-Host-Infra-FC-NVMe-01-A init
member device-alias VM-Host-Infra-FC-NVMe-02-A init
member device-alias VM-Host-Infra-FC-NVMe-03-A init

member device-alias VM-Host-Infra-FC-NVMe-04-A init
member device-alias Infra_SVM-fc-nvme-lif-01a target
member device-alias Infra_SVM-fc-nvme-lif-02a target
exit

zoneset name FlexPod-Fabric-A vsan <vsan-a-id>
member FCP-Infra_SVM-Fabric-A
member FC-NVMe-Infra_SVM-Fabric-A
exit

zoneset activate name FlexPod-Fabric-A vsan \<vsan-a-id>

show zoneset active
copy r s

Note:   Since Smart Zoning is enabled, a single zone for each storage protocol (FCP and FC-NVMe) is created with all host initiators and targets for the Infra_SVM instead of creating separate zones for each host with the host initiator and targets. If a new host is added, its initiator can simply be added to each single zone in each MDS switch and then the zoneset reactivated. If another SVM is added to the FlexPod with FC and/or FC-NVMe targets, new zones can be added for that SVM.

Procedure 10.  Create Zones and Zonesets on Cisco MDS 9132T B

Step 1.    To create the required zones and zoneset on Fabric B, run the following commands:

configure terminal

zone name FCP-Infra_SVM-Fabric-B vsan <vsan-b-id>
member device-alias VM-Host-Infra-FCP-01-B init
member device-alias VM-Host-Infra-FCP-02-B init
member device-alias VM-Host-Infra-FCP-03-B init

member device-alias VM-Host-Infra-FCP-04-B init
member device-alias Infra_SVM-fcp-lif-01b target
member device-alias Infra_SVM-fcp-lif-02b target
exit

zone name FC-NVMe-Infra_SVM-Fabric-B vsan <vsan-b-id>
member device-alias VM-Host-Infra-FC-NVMe-01-B init
member device-alias VM-Host-Infra-FC-NVMe-02-B init
member device-alias VM-Host-Infra-FC-NVMe-03-B init

member device-alias VM-Host-Infra-FC-NVMe-04-B init
member device-alias Infra_SVM-fc-nvme-lif-01b target
member device-alias Infra_SVM-fc-nvme-lif-02b target
exit

zoneset name FlexPod-Fabric-B vsan <vsan-b-id>
member FCP-Infra_SVM-Fabric-B
member FC-NVMe-Infra_SVM-Fabric-B
exit

zoneset activate name FlexPod-Fabric-B vsan <vsan-b-id>

show zoneset active

copy r s

Storage Configuration – ONTAP Boot Storage Setup

This configuration requires information from both the server profiles and NetApp storage system. After creating the boot LUNs, initiator groups and appropriate mappings between the two, Cisco UCS server profiles will be able to see the boot disks hosted on NetApp controllers.

Ansible Configuration

This section provides details about the Ansible Scripts used to configure ONTAP Boot storage.

Procedure 1.     Configure the Storage for ONTAP Boot

Step 1.    Edit the following variable file and update the fcp_igroups variables:

FlexPod-for-EHR/Ansible/vars/ontap_main.yml

Note:   Update the initiator WWPNs for FC igroups.

Step 2.    From /root/ FlexPod-for-EHR/Ansible, invoke the ansible scripts for this section using the following command:

ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_2

Note:   Use the -vvv tag to see detailed execution output log.

Manual Configuration

Procedure 1.     Create Boot LUNs for ESXi Servers using NetApp Cluster Management Console

Step 1.    Run the following commands to create boot LUNs for all the ESXi servers:

lun create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-01 -size 32GB -ostype  vmware -space-reserve disabled

lun create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-02 -size 32GB -ostype vmware -space-reserve disabled

lun create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-03 -size 32GB -ostype vmware -space-reserve disabled

Create igroups

Refer to section Obtain the WWPNs for Cisco UCS Server Profiles to obtain host WWPNs.

Procedure 1.     Create Initiator Groups for FC Storage Access using NetApp Cluster Management Console

Step 1.    To access boot LUNs, following igroups for individual hosts are created:

lun igroup create -vserver Infra-SVM -igroup Epic-Host-01 -protocol fcp -ostype vmware -initiator <Epic-Host-01-wwpna>, <Epic-Host-01-wwpnb>

lun igroup create -vserver Infra-SVM -igroup Epic-Host-02 -protocol fcp -ostype vmware -initiator <Epic-Host-02-wwpna>, <Epic-Host-02-wwpnb>

lun igroup create -vserver Infra-SVM -igroup Epic-Host-03 -protocol fcp -ostype vmware -initiator <Epic-Host-03-wwpna>, <Epic-Host-03-wwpnb>

Step 2.    To view and verify the FC igroups just created, use the following command:

lun igroup show -vserver Infra-SVM -protocol fcp

Step 3.    (Optional) To access a common datastore from all the hosts, a common igroup for all the servers can be created as follows:

lun igroup create –vserver Infra-SVM –igroup MGMT-Hosts –protocol fcp –ostype vmware –initiator <Epic-Host-01-wwpna>, <Epic-Host-01-wwpnb>, <Epic-Host-02-wwpna>, <Epic-Host-02-wwpnb>, <Epic-Host-03-wwpna>, <Epic-Host-03-wwpnb>

Procedure 2.     Map Boot LUNs to igroups

Step 1.    Map the boot LUNs to FC igroups, by entering the following commands on NetApp cluster management console:

lun mapping create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-01 -igroup Epic-Host-01 -lun-id 0

lun mapping create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-02 -igroup Epic-Host-02 -lun-id 0

lun mapping create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-03 -igroup Epic-Host-03 -lun-id 0

lun mapping create -vserver Infra-SVM -path /vol/esxi_boot/Epic-Host-04 -igroup Epic-Host-04 -lun-id 0

Step 2.    To verify the mapping was setup correctly, issue the following command:

lun mapping show -vserver Infra-SVM -protocol fcp

VMware vSphere 7.0U3 Setup

VMware ESXi 7.0U3

This section provides detailed instructions for installing VMware ESXi 7.0 in a FlexPod environment. On successful completion of these steps, multiple ESXi hosts will be provisioned and ready to be added to VMware vCenter.

Several methods exist for installing ESXi in a VMware environment. These procedures focus on how to use the built-in keyboard, video, mouse (KVM) console and virtual media features in Cisco UCS Manager to map remote installation media to individual servers and connect to their boot logical unit numbers (LUNs).

Procedure 1.     Download ESXi 7.0U3 from VMware

Step 1.    Click here: Cisco Custom Image for ESXi 7.0 U3 Install ISO.

Step 2.    You will need a user id and password on vmware.com to download this software.

Step 3.    Download the .iso file.

Procedure 2.     Log into Cisco Intersight and Launch KVM

Note:   The Cisco Intersight KVM enables the administrators to begin the installation of the operating system (OS) through remote media. It is necessary to log into the Cisco Intersight to access KVM.

Step 1.    Log into Cisco Intersight.

Step 2.    From the main menu, click Servers.

Step 3.    Find the Server and click “…” to see more options.

Step 4.    Click Launch vKVM.

Graphical user interface, applicationDescription automatically generated

Step 5.    Follow the prompts to ignore certificate workings (if any) and launch the HTML5 KVM console.

Step 6.    Repeat steps 1 - 5 to launch the HTML5 KVM console for all the servers.

Procedure 3.     Set Up VMware ESXi Installation on each ESXi Host

Step 1.    In the KVM window, click Virtual Media > vKVM-Mapped vDVD.

Step 2.    Browse and select the ESXi installer ISO image file downloaded in the last step.

Step 3.    Click Map Drive.

Step 4.    Select Macros > Static Macros > Ctrl + Alt + Delete to reboot the Server if the server is showing shell prompt. If the server is shutdown, from Intersight, select “…” next to server and click Power On.

Step 5.    Monitor the server boot process in the KVM. Server should find the boot LUNs and begin to load the ESXi installer.

Note:   If the ESXi installer fails to load because the software certificates cannot be validated, reset the server, and when prompted, press F2 to go into BIOS and set the system time and date to current. The ESXi installer should load properly.

Procedure 4.     Install VMware ESXi

Step 1.    After the ESXi installer is finished loading (from the last step), press Enter to continue with the installation.

Step 2.    Read and accept the end-user license agreement (EULA). Press F11 to accept and continue.

Note:   It may be necessary to map function keys as User Defined Macros under the Macros menu in the KVM console.

Step 3.    Select the NetApp boot LUN that was previously set up as the installation disk for ESXi and press Enter to continue with the installation.

Step 4.    Select the appropriate keyboard layout and press Enter.

Step 5.    Enter and confirm the root password and press Enter.

Step 6.    The installer issues a warning that the selected disk will be repartitioned. Press F11 to continue with the installation.

Step 7.    After the installation is complete, click Virtual Media to unmap the installer ISO. Press Enter to reboot the server.

Procedure 5.     Set Up Management Networking for ESXi Hosts

Note:   Adding a management network for each VMware host is required for managing the host.

Step 1.    After the server has finished rebooting, in the UCS KVM console, press F2 to customize VMware ESXi.

Step 2.    Log in as root, enter the password set during installation, and press Enter to log in.

Step 3.    Use the down arrow key to select Troubleshooting Options and press Enter.

Step 4.    Select Enable SSH and press Enter.

Step 5.    Press Esc to exit the Troubleshooting Options menu.

Step 6.    Select the Configure Management Network option and press Enter.

Step 7.    Select Network Adapters and press Enter.

Step 8.    Using the spacebar, select vmnic1 in addition to the pre-selected vmnic0.

TextDescription automatically generated

Step 9.    Press Enter.

Step 10.                       Under VLAN (optional) enter the IB-MGMT VLAN and press Enter.

Graphical user interface, textDescription automatically generated

Step 11.                       Select IPv4 Configuration and press Enter.

Note:   When using DHCP to set the ESXi host networking configuration, setting up a manual IP address is not required.

Step 12.                       Select the Set static IPv4 address and network configuration option by using the arrow keys and space bar.

Step 13.                       Under IPv4 Address, enter the IP address for managing the ESXi host.

Step 14.                       Under Subnet Mask, enter the subnet mask.

Step 15.                       Under Default Gateway, enter the default gateway.

Step 16.                       Press Enter to accept the changes to the IP configuration.

Step 17.                       Select the IPv6 Configuration option and press Enter.

Step 18.                       Using the spacebar, select Disable IPv6 (restart required) and press Enter.

Step 19.                       Select the DNS Configuration option and press Enter.

Note:   If the IP address is configured manually, the DNS information must be provided.

Step 20.                       Using the spacebar, select the option Use the following DNS server addresses and hostname:

a.     Under Primary DNS Server, enter the IP address of the primary DNS server.

b.     Optional: Under Alternate DNS Server, enter the IP address of the secondary DNS server.

c.     Under Hostname, enter the fully qualified domain name (FQDN) for the ESXi host.

d.     Press Enter to accept the changes to the DNS configuration.

e.     Press Esc to exit the Configure Management Network submenu.

f.      Press Y to confirm the changes and reboot the ESXi host.

Procedure 6.     Configure FlexPod VMware ESXi Ansible Hosts

Step 1.    Edit the following variable files to ensure proper VMware specific variables are entered:

      FlexPod-for-EHR/Ansible/inventory

      FlexPod-for-EHR/Ansible/group_vars/all.yml

      FlexPod-for-EHR/Ansible/roles/ESXIhosts/defaults/main.yml

Step 2.    From /root/ FlexPod-for-EHR/Ansible, run the Setup_ESXi.yml Ansible playbook:

ansible-playbook ./Setup_ESXi.yml -i inventory

FlexPod VMware ESXi Manual Configuration

Although the VMware ESXi Ansible configuration configures all three ESXi hosts, the manual configuration, after the installation of necessary drivers, configures only the first host using the ESXi web interface then adds the second and third hosts after vCenter is installed.

Procedure 1.     Install Cisco VIC Drivers and NetApp NFS Plug-in for VAAI

Step 1.    Download the offline bundle for the Cisco VIC nfnic driver and the NetApp NFS Plug-in for VMware VAAI to the Management workstation:

      Download the ISO Image of Cisco UCS related VMware drivers only from: https://software.cisco.com/download/home/286329080/type/283853158/release/5.0(1a).

      Mount ISO image and extract file Cisco-nfnic_5.0.0.15-1OEM.700.1.0.15843807_18697950.zip.

      Follow Storage > Cisco > VIC > ESXi_7.0U2 > nfnic_5.0.0.15 and copy the Cisco-nfnic_5.0.0.15-1OEM.700.1.0.15843807_18697950.zip file.

      NetApp NFS Plug-in for VMware VAAI 2.0 (NetAppNasPluginV2.0.zip) from https://mysupport.netapp.com/site/products/all/details/nfsplugin-vmware-vaai/downloads-tab/download/61278/2.0.

Note:   Cisco VIC nenic version 1.0.4.0 is already included in the Cisco Custom ISO for VMware vSphere version 7.0.2.

Note:   Consult the Cisco UCS Hardware Compatibility List and the NetApp Interoperability Matrix Tool to determine latest supported combinations of firmware and software.

Step 2.    Install VMware VIC Drivers and the NetApp NFS Plug-in for VMware VAAI on the ESXi hosts:

A screenshot of a computerDescription automatically generated

Step 3.    Using an SCP program, copy the two bundles referenced above to the /tmp directory on each ESXi host.

Step 4.    SSH to each VMware ESXi host and log in as root.

Step 5.    Run the following commands on each host:

esxcli software component apply -d /tmp/Cisco-nfnic_5.0.0.15-1OEM.700.1.0.15843807_18697950.zip

esxcli software vib install -d /tmp/NetAppNasPluginV2.0.zip
 
reboot

Step 6.    After reboot, SSH back into each host and use the following commands to ensure the correct version are installed:

esxcli software component list | grep nfnic
esxcli software vib list | grep NetApp

Procedure 2.     Configure FlexPod VMware ESXi for First ESXi Host

Note:   In this procedure, you’re only setting up the first ESXi host. The remaining hosts will be added to vCenter and setup from the vCenter.

Step 1.    Log into the first ESXi host using the VMware Host Client.

Step 2.    Open a web browser and navigate to the first ESXi server’s management IP address.

Step 3.    Enter root as the User name.

Step 4.    Enter the <root password>.

Step 5.    Click Log in to connect.

Step 6.    Decide whether to join the VMware Customer Experience Improvement Program or not and click OK.

Procedure 3.     Set Up VMkernel Ports and Virtual Switch on the First ESXi Host

Step 1.    From the Host Client Navigator, select Networking.

Step 2.    In the center pane, select the Virtual switches tab.

Step 3.    Highlight the vSwitch0 line.

Step 4.    Click Edit settings.

Step 5.    Change the MTU to 9000.

Step 6.    Expand Link Discovery

Step 7.    Set Mode to Both

Step 8.    Expand NIC teaming.

Step 9.    In the Failover order section, select vmnic1 and click Mark active.

Step 10.                       Verify that vmnic1 now has a status of Active.

Step 11.                       Click Save.

Step 12.                       Select Networking, then select the Port groups tab.

Step 13.                       In the center pane, right-click VM Network and select Edit settings.

Step 14.                       Name the port group IB_MGMT. Set the VLAN ID to <IB-MGMT-VLAN> (for example, 41).

Note:    (Optional) The IB-MGMT VLAN can be set as the native VLAN for the vSwitch0 vNIC templates and the port group’s VLAN ID can be set to 0.

Step 15.                       Click Save to finalize the edits for the IB-MGMT Network port group.

Step 16.                       At the top, select the VMkernel NICs tab.

Step 17.                       Click Add VMkernel NIC.

Step 18.                       For New port group, enter VMkernel-Infra-NFS.

Step 19.                       For Virtual switch, select vSwitch0.

Step 20.                       Enter <infra-nfs-vlan-id> (for example, 42) for the VLAN ID.

Step 21.                       Change the MTU to 9000.

Step 22.                       Select Static IPv4 settings and expand IPv4 settings.

Step 23.                       Enter the NFS IP address and netmask.

Step 24.                       Leave TCP/IP stack set at Default TCP/IP stack and do not choose any of the Services.

Step 25.                       Click Create.

Step 26.                       Select the Virtual Switches tab, then vSwitch0. The properties for vSwitch0 should be similar to the following screenshot:

Graphical user interface, applicationDescription automatically generated

Procedure 4.     Mount Datastores on the First ESXi Host

Step 1.    From the Web Navigator, click Storage.

Step 2.    In the center pane, click the Datastores tab.

Step 3.    In the center pane, select New Datastore to add a new datastore.

Step 4.    In the New datastore popup, select Mount NFS datastore and click Next.

Step 5.    Enter infra_datastore_01 for the datastore name and IP address of NetApp nfs-lif-01 LIF for the NFS server. Enter /infra_datastore_01 for the NFS share. Select the NFS version. Click Next.

Graphical user interface, applicationDescription automatically generated

Step 6.    Review information and click Finish.

Step 7.    The datastore should now appear in the datastore list.

Step 8.    In the center pane, select New Datastore to add a new datastore.

Step 9.    In the New datastore popup, select Mount NFS datastore and click Next.

Step 10.                       Enter infra_swap for the datastore name and IP address of NetApp nfs-lif-01 LIF for the NFS server. Enter /infra_swap for the NFS share. Select the NFS version. Click Next.

Step 11.                       Click Finish. The datastore should now appear in the datastore list.

Procedure 5.     Configure NTP Server on the First ESXi Host

Step 1.    From the Web Navigator, click Manage.

Step 2.    In the center pane, click System > Time & date.

Step 3.    Click Edit NTP Settings.

Step 4.    Select Use Network Time Protocol (enable NTP client).

Step 5.    Use the drop-down list to select Start and stop with host.

Step 6.    Enter the NTP server IP address(es) in the NTP servers.

Graphical user interface, text, application, emailDescription automatically generated 

Step 7.    Click Save to save the configuration changes.

Step 8.    Select the Services tab.

Step 9.    Right-click ntpd and click Start.

Step 10.                       System > Time & date should now show “Running” for the NTP service status.

Graphical user interfaceDescription automatically generated with medium confidence 

Procedure 6.     Configure ESXi Host Swap on the First ESXi Host

Step 1.    From the Web Navigator, click Manage.

Step 2.    In the center pane, click System > Swap.

Step 3.    Click Edit settings.

Step 4.    Use the drop-down list to select infra_swap. Leave all other settings unchanged.

Graphical user interface, text, applicationDescription automatically generated

Step 5.    Click Save to save the configuration changes.

Procedure 7.     Configure Host Power Policy on the First ESXi Host

Note:   Implementation of this policy is recommended in Performance Tuning Guide for Cisco UCS M6 Servers: for maximum VMware ESXi performance. This policy can be adjusted based on customer requirements.

Step 1.    From the Web Navigator, click Manage.

Step 2.    In the center pane, click Hardware > Power Management.

Step 3.    Click Change policy.

Step 4.    Select High performance and click OK.

Graphical user interface, applicationDescription automatically generated 

VMware vCenter 7.0U3

The procedures in the following subsections provide detailed instructions for installing the VMware vCenter 7.0U3c Server Appliance in a FlexPod environment.

Procedure 1.     Download vCenter 7.0U3c from VMware

Step 1.    Click the following link: https://customerconnect.vmware.com/downloads/details?downloadGroup=VC70U3C&productId=974&rPId=78220and download the VMware-VCSA-all-7.0.3-19234570.iso.

Step 2.    You will need a VMware user id and password on vmware.com to download this software.

Procedure 2.     Install the VMware vCenter Server Appliance

Note:   The VCSA deployment consists of 2 stages: installation and configuration.

Step 1.    Locate and copy the VMware-VCSA-all-7.0.3-19234570.iso file to the desktop of the management workstation. This ISO is for the VMware vSphere 7.0U3c vCenter Server Appliance.

Step 2.    Mount the ISO image as a disk on the management workstation. (For example, with the Mount command in Windows Server 2012 and above).

Step 3.    In the mounted disk directory, navigate to the vcsa-ui-installer > win32 directory and double-click installer.exe. The vCenter Server Appliance Installer wizard appears.

Graphical user interface, applicationDescription automatically generated

Step 4.    Click Install to start the vCenter Server Appliance deployment wizard.

Step 5.    Click NEXT in the Introduction section.

Step 6.    Read and accept the license agreement and click NEXT.

Step 7.    In the “vCenter Server deployment target” window, enter the FQDN or IP address of the destination host, User name (root) and Password. Click NEXT.

Note:   Installation of vCenter on a separate existing management infrastructure is recommended. If a separate management infrastructure is not available, customers can choose the recently configured first ESXi host as an installation target.

Step 8.    Click YES to accept the certificate.

Step 9.    Enter the Appliance VM name and password details shown in the “Set up vCenter Server VM” section. Click NEXT.

Step 10.                       In the “Select deployment size” section, select the Deployment size and Storage size. For example, select “Small” and “Default.” Click NEXT.

Step 11.                       Select the datastore (for example, infra_datastore_02) for storage. Click NEXT.

Step 12.                       In the Network Settings section, configure the following settings:

a.     Select a Network: IB-MGMT Network

Note:   It is important that the vCenter VM stay on the IB-MGMT Network on vSwitch0 and not moved to a vDS. If vCenter is moved to a vDS and the virtual environment is completely shut down and then brought back up, trying to bring up vCenter on a different host than the one it was running on before the shutdown, will cause problems with the network connectivity. With the vDS, for a virtual machine to move from one host to another, vCenter must be up and running to coordinate the move of the virtual ports on the vDS. If vCenter is down, the port move on the vDS cannot occur correctly. Moving vCenter to a different host on vSwitch0 does not require vCenter to already be up and running.

b.     IP version: IPV4

c.     IP assignment: static

d.     FQDN: <vcenter-fqdn>

e.     IP address: <vcenter-ip>

f.      Subnet mask or prefix length: <vcenter-subnet-mask>

g.     Default gateway: <vcenter-gateway>

h.     DNS Servers: <dns-server1>,<dns-server2>

Step 13.                       Click NEXT.

Step 14.                       Review all values and click FINISH to complete the installation.

Note:   The vCenter Server appliance installation will take a few minutes to complete.

Step 15.                       When Stage 1, Deploy vCenter Server, is complete, click CONTINUE to proceed with stage 2.

Step 16.                       Click NEXT.

Step 17.                       In the vCenter Server configuration window, configure these settings:

a.     Time Synchronization Mode: Synchronize time with NTP servers.

b.     NTP Servers: NTP server IP addresses.

c.     SSH access: Enabled.

Step 18.                       Click NEXT.

Step 19.                       Complete the SSO configuration as shown below (or according to your organization’s security policies):

Graphical user interface, applicationDescription automatically generated

Verify vCenter CPU Settings

If a vCenter deployment size of Small or larger was selected in the vCenter setup, it is possible that the VCSA’s CPU setup does not match the Cisco UCS server CPU hardware configuration. Cisco UCS X210c M6 and B200 M6 servers are 2-socket servers. During this validation, the Small deployment size was selected and vCenter was setup for a 4-socket server.  This setup can cause issues in the VMware ESXi cluster Admission Control.

Procedure 1.     Resolve vCenter CPU Settings

Step 1.    Open a web browser on the management workstation and navigate to vCenter or ESXi server where vCenter appliance was deployed and login.

Step 2.    Click the vCenter VMright-click and select Edit settings.

Step 3.    In the Edit settings window, expand CPU and check the value of Sockets.

Graphical user interface, text, applicationDescription automatically generated

Step 4.    If the number of Sockets match the server configuration, click Cancel.

Step 5.    If the number of Sockets does not match the server configuration, it will need to be adjusted:

a.     Right-click the vCenter VM and click Guest OS > Shut down. Click Yes on the confirmation.

b.     When vCenter is shut down, right-click the vCenter VM and select Edit settings.

c.     In the Edit settings window, expand CPU and change the Cores per Socket value to make the Sockets value equal to the server configuration.

Step 6.    Click Save.

Step 7.    Right-click the vCenter VM and click Power > Power on. Wait approximately 10 minutes for vCenter to come up.

Procedure 2.     Setup VMware vCenter Server

Step 1.    Using a web browser, navigate to https://<vcenter-ip-address>:5480. Navigate security screens.

Step 2.    Log into the VMware vCenter Server Management interface as root with the root password set in the vCenter installation.

Step 3.    In the menu on the left, click Time.

Step 4.    Click EDIT to the right of Time zone.

Step 5.    Select the appropriate Time zone and click SAVE.

Step 6.    In the menu on the left select Administration.

Step 7.    According to your Security Policy, adjust the settings for the root user and password.

Step 8.    In the menu on the left click Update.

Step 9.    Follow the prompts to stage and install any available vCenter updates.

Step 10.                       In the upper right-hand corner of the screen, click root > Logout to logout of the Appliance Management interface.

Step 11.                       Using a web browser, navigate to https://<vcenter-ip-address> and navigate through security screens.

Note:   With VMware vCenter 7.0 and above, the use of the vCenter FQDN is required.

Step 12.                       Click LAUNCH VSPHERE CLIENT (HTML5).

Note:   The VMware vSphere HTML5 Client is the only option in vSphere 7. All the old clients have been deprecated.

Step 13.                       Log in using the Single Sign-On username (administrator@vsphere.local) and password created during the vCenter installation. Dismiss the Licensing warning.

Procedure 3.     Add AD User Authentication to vCenter (Optional)

Step 1.    In the AD Infrastructure, using the Active Directory Users and Computers tool, setup a Domain Administrator user with a user name such as flexadmin (FlexPod Admin).

Step 2.    Connect to https://<vcenter-ip> and select LAUNCH VSPHERE CLIENT (HTML5).

Step 3.    Log in as Administrator@vsphere.local (or the SSO user set up in vCenter installation) with the corresponding password.

Step 4.    Under Menu, click Administration. In the list on the left, under Single Sign On, click Configuration.

Step 5.    In the center pane, under Configuration, click the Identity Provider tab.

Step 6.    In the list under Type, select Active Directory Domain.

Step 7.    Click JOIN AD.

Step 8.    Fill in the AD domain name, the Administrator user, and the domain Administrator password.  Do not fill in an Organizational unit. Click JOIN.

Step 9.    Click Acknowledge.

Step 10.                       In the list on the left under Deployment, select System Configuration. Click the radio button to select the vCenter, then click REBOOT NODE.

Step 11.                       Input a reboot reason and click REBOOT.  The reboot will take approximately 10 minutes for full vCenter initialization. 

Step 12.                       Log back into the vCenter vSphere HTML5 Client as Administrator@vsphere.local.

Step 13.                       Under Menu, click Administration. In the list on the left, under Single Sign On, click Configuration.

Step 14.                       In the center pane, under Configuration, click the Identity Provider tab. Under Type, select Identity Sources. Click ADD.

Step 15.                       Make sure Active Directory (Integrated Windows Authentication) is selected, your Windows Domain name is listed, and Use machine account is selected. Click ADD.

Step 16.                       In the list select the Active Directory (Integrated Windows Authentication) Identity source type. If desired, select SET AS DEFAULT and click OK.

Step 17.                       On the left under Access Control, select Global Permissions.

Step 18.                       In the center pane, click the + sign to add a Global Permission.

Step 19.                       In the Add Permission window, select your AD domain for the Domain.

Step 20.                       On the User/Group line, enter either the FlexPod Admin username or the Domain Admins group. Leave the Role set to Administrator. Check the box for Propagate to children.

Note:   The FlexPod Admin user was created in the Domain Admins group. The selection here depends on whether the FlexPod Admin user will be the only user used in this FlexPod If additional users will be added later.  By selecting the Domain Admins group, any user placed in that AD Domain group will be able to login to vCenter as an Administrator.

Step 21.                       Click OK to add the selected User or Group. The user or group should now appear in the Global Permissions list with the Administrator role.

Step 22.                       Log out and log back into the vCenter HTML5 Client as the FlexPod Admin user.  You will need to add the domain name to the user, for example, flexadmin@domain.

Procedure 4.     vCenter - Initial Configuration

Step 1.    In the center pane, click ACTIONS > New Datacenter.

Step 2.    Type Epic-DC in the Datacenter name field.

Step 3.    Click OK.

Step 4.    Expand the vCenter.

Step 5.    Right-click the datacenter Epic-DC in the list in the left pane. Click New Cluster…

Step 6.    Provide a name for the cluster (for example, Epic-Cluster).

Step 7.    Turn on DRS and vSphere HA. Do not turn on vSAN.

Graphical user interface, applicationDescription automatically generated

Step 8.    Click NEXT and then click FINISH to create the new cluster.

Step 9.    Right-click the cluster and click Settings.

Step 10.                       Click Configuration > General in the list located and click EDIT.

Step 11.                       Click Datastore specified by host for the Swap file location and click OK.

Step 12.                       Right-click the cluster and select Add Hosts.

Step 13.                       In the IP address or FQDN field, enter either the IP address or the FQDN of the first VMware ESXi host. Enter root as the Username and the root password. Click NEXT.

Step 14.                       In the Security Alert window, select the host and click OK.

Step 15.                       Verify the Host summary information and click NEXT.

Step 16.                       Ignore warnings about the host being moved to Maintenance Mode and click FINISH to complete adding the host to the cluster.

Note:   The added ESXi host will have Warnings that the ESXi Shell and SSH have been enabled. These warnings can be suppressed. The host will also have a TPM Encryption Key Recovery alert that can be reset to green.

Step 17.                       In the center pane under Storage, click Storage Devices. Make sure the NETAPP Fibre Channel Disk LUN 0 is selected.

Step 18.                       Click the Paths tab.

Step 19.                       Ensure that 4 paths appear, two of which should have the status Active (I/O). The output below shows the paths for the boot LUN.

Graphical user interface, text, application, emailDescription automatically generated

Add and Configure VMware ESXi Hosts in vCenter

Procedure 1.     Add the ESXi Hosts to vCenter

Step 1.    From the Home screen in the VMware vCenter HTML5 Interface, click Menu > Hosts and Clusters.

Step 2.    Right-click the cluster and click Add Hosts.

Step 3.    In the IP address or FQDN field, enter either the IP address or the FQDN name of the configured VMware ESXi host. Also enter the user id (root) and associated password. If more than one host is being added, add the corresponding host information, optionally selecting “Use the same credentials for all hosts.” Click NEXT.

Step 4.    Select all hosts being added and click OK to accept the thumbprint(s).

Step 5.    Review the host details and click NEXT to continue.

Step 6.    Review the configuration parameters and click FINISH to add the host(s).

Note:   The added ESXi host(s) will be placed in Maintenance Mode and will have Warnings that the ESXi Shell and SSH have been enabled. These warnings can be suppressed. The TPM Encryption Recovery Key Backup Alarm can also be Reset to Green.

Procedure 2.     Set Up VMkernel Ports and Virtual Switch

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    In the center pane, click the Configure tab.

Step 3.    In the list, select Virtual switches under Networking.

Step 4.    Expand Standard Switch: vSwitch0.

Step 5.    Select EDIT to Edit settings.

Step 6.    Change the MTU to 9000.

Step 7.    Select Teaming and failover located on the left.

Step 8.    In the Failover order section, use the arrow icons to move the vmnics until both are Active adapters.

Graphical user interface, application, emailDescription automatically generated

Step 9.    Click OK.

Step 10.                       In the center pane, to the right of VM Network click “…” > Remove to remove the port group. Click YES on the confirmation.

Step 11.                       Click ADD NETWORKING to add a new VM port group.

Step 12.                       Select Virtual Machine Port Group for a Standard Switch and click NEXT.

Step 13.                       Ensure vSwitch0 is shown for Select an existing standard switch and click NEXT.

Step 14.                       Name the port group “IB_MGMT” and set the VLAN to IB-MGMT-VLAN (for example, 41). Click NEXT.

Note:   (Optional) The IB-MGMT VLAN can be set as the native VLAN for the vSwitch0 vNIC templates on Cisco Intersight. The port group’s VLAN ID can then be set to 0.

Step 15.                       Click FINISH to complete adding the IB_MGMT VM port group.

Step 16.                       Under Networking, click VMkernel adapters.

Step 17.                       In the center pane, click Add Networking.

Step 18.                       Make sure VMkernel Network Adapter is selected and click NEXT.

Step 19.                       Select an existing standard switch and click BROWSE. Select vSwitch0 and click OK. Click NEXT.

Step 20.                       For Network label, enter VMkernel-Infra-NFS.

Step 21.                       Enter <infra-nfs-vlan-id for example, 42> for the VLAN ID.

Step 22.                       Select Custom for MTU and set the value to 9000.

Step 23.                       Leave the Default TCP/IP stack selected and do not select any of the Enabled services. Click NEXT.

Step 24.                       Select Use static IPv4 settings and enter the IPv4 address and subnet mask for the Infra-NFS VMkernel port for this ESXi host.

Step 25.                       Click NEXT.

Step 26.                       Review the settings and click FINISH to create the VMkernel port.

Step 27.                       To verify the vSwitch0 setting, under Networking, select Virtual switches, then expand vSwitch0. The properties for vSwitch0 should be similar to:

Graphical user interfaceDescription automatically generated

Step 28.                       Repeat steps 1 – 27 for all the ESXi hosts being added.

Procedure 3.     Mount Required Datastores

Step 1.    From the vCenter Home screen, click Menu > Storage.

Step 2.    Expand Epic-DC.

Step 3.    Right-click infra_datastore_01 and select Mount Datastore to Additional Hosts.

Step 4.    Select all the ESXi host(s) and click OK

Step 5.    Repeat steps 1 – 4  to mount the infra_datastore_2 and infra_swap datastores on all the ESXi host(s).

Step 6.    Select infra_datastore_01 and in the center pane, click Hosts. Verify that all the ESXi host(s) are listed. Repeat this process to verify that both infra_datastore_2 and infra_swap datastores are also mounted on all hosts.

Procedure 4.     Configure NTP on ESXi Host

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    In the center pane, select the Configure tab.

Step 3.    In the list under System, click Time Configuration.

Step 4.    Click EDIT next to Network Time Protocol.

Step 5.    Check the box for Enable.

Step 6.    Enter the NTP Server IP address(es) in the NTP servers box separated by a comma.

Step 7.    Check the box for Start NTP Service.

Step 8.    Use the drop-down list to select Start and stop with host.

Step 9.    Click OK to save the configuration changes.

Step 10.                       Verify that NTP service is now enabled and running, and the clock is now set to correct time.

Step 11.                       Repeat steps 1 – 10 for all the ESXi hosts.

Procedure 5.     Change ESXi Power Management Policy

Note:   Implementation of this policy is recommended in Performance Tuning Guide for Cisco UCS M6 Servers: https://www.cisco.com/c/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/performance-tuning-guide-ucs-m6-servers.html for maximum VMware ESXi performance. This policy can be adjusted based on customer requirements.

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    In the center pane, click the Configure tab.

Step 3.    Under Hardware, click Overview. Scroll to the bottom and next to Power Management, click EDIT POWER POLICY.

Step 4.    Select High performance and click OK.

FlexPod VMware vSphere Distributed Switch (vDS)

This section provides detailed procedures for setting up VMware vDS in vCenter. Based on the VLAN configuration in Intersight, a vMotion, and a VM-Traffic port group will be added to the vDS. Any additional VLAN-based port groups added to the vDS would require changes in Intersight and the Cisco Nexus 9K switches.

In this document, the infrastructure ESXi management VMkernel ports, the In-Band management interfaces including the vCenter management interface, and the infrastructure NFS VMkernel ports are left on vSwitch0 to facilitate bringing the virtual environment back up in the event it needs to be completely shut down. The vMotion VMkernel ports are moved to the vDS to allow for future QoS support. The vMotion port group is also pinned to Cisco UCS fabric B and pinning configuration in vDS ensures consistency across all ESXi hosts.

Procedure 1.     Configure the VMware vDS in vCenter

Step 1.    After logging into the VMware vSphere HTML5 Client, click Networking under Menu. 

Step 2.    Right-click the Epic-DC datacenter and click Distributed Switch > New Distributed Switch.

Step 3.    Give the Distributed Switch a descriptive name (for example, Epic_vDS) and click NEXT.

Step 4.    Make sure version 7.0.0 – ESXi 7.0.0 and later is selected and click NEXT.

Step 5.    Change the Number of uplinks to 2. If VMware Network I/O Control is to be used for Quality of Service, leave Network I/O Control Enabled. Otherwise, Disable Network I/O Control. Enter VM-Traffic for the Port group name. Click NEXT.

Step 6.    Review the information and click FINISH to complete creating the vDS.

Step 7.    Expand the Epic-DC datacenter and the newly created vDS. Click the newly created vDS.

Step 8.    Right-click the VM-Traffic port group and click Edit Settings.

Step 9.    Click VLAN.

Step 10.                       Click VLAN for VLAN type and enter the <VM-Traffic> VLAN ID (for example, 44). Click OK.

Step 11.                       Right-click the vDS and click Settings > Edit Settings.

Step 12.                       In the Edit Settings window, click the Advanced tab.

Step 13.                       Change the MTU to 9000. The Discovery Protocol can optionally be changed to Link Layer Discovery Protocol and the Operation to Both. Click OK.

Graphical user interface, text, application, emailDescription automatically generated

Step 14.                       To create the vMotion port group, right-click the vDS, select Distributed Port Group > New Distributed Port Group.

Step 15.                       Enter VMKernel-vMotion as the name and click NEXT.

Step 16.                       Set the VLAN type to VLAN, enter the VLAN ID used for vMotion (for example, 43), check the box for Customize default policies configuration, and click NEXT.

Step 17.                       Leave the Security options set to Reject and click NEXT.

Step 18.                       Leave the Ingress and Egress traffic shaping options as Disabled and click NEXT.

Step 19.                       Select Uplink 1 from the list of Active uplinks and click MOVE DOWN twice to place Uplink 1 in the list of Standby uplinks. This will pin all vMotion traffic to UCS Fabric Interconnect B except when a failure occurs.

Graphical user interface, applicationDescription automatically generated

Step 20.                       Click NEXT.

Step 21.                       Leave NetFlow disabled and click NEXT.

Step 22.                       Leave Block all ports set as No and click NEXT.

Step 23.                       Confirm the options and click FINISH to create the port group.

Step 24.                       Right-click the vDS and click Add and Manage Hosts.

Step 25.                       Make sure Add hosts is selected and click NEXT.

Step 26.                       Click the green + sign to add new hosts. Select the first ESXi host and click OK. Click NEXT.

Step 27.                       Select vmnic2 and click Assign uplink. Choose Uplink 1 and click OK.

Step 28.                       Select vmnic3 and click Assign uplink. Select Uplink 2 and click OK.

Note:   It is important to assign the uplinks as shown below. This allows the port groups to be pinned to the appropriate Cisco UCS fabric.

Graphical user interface, websiteDescription automatically generated

Step 29.                       Click NEXT.

Step 30.                       Do not migrate any VMkernel ports and click NEXT.

Step 31.                       Do not migrate any virtual machine networking ports. Click NEXT.

Step 32.                       Click FINISH to complete adding the ESXi host to the vDS.

Step 33.                       Select Hosts and Clusters under Menu and select the first ESXi host. In the center pane, select the Configure tab.

Step 34.                       In the list under Networking, select VMkernel adapters.

Step 35.                       Select ADD NETWORKING.

Step 36.                       In the Add Networking window, ensure that VMkernel Network Adapter is selected and click NEXT.

Step 37.                       Ensure that Select an existing network is selected and click BROWSE.

Step 38.                       Select VMKernel-vMotion and click OK.

Step 39.                       Click NEXT.

Step 40.                       From the MTU drop-down list, select Custom and ensure the MTU is set to 9000.

Step 41.                       From the TCP/IP stack drop-down list, select vMotion. Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 42.                       Select Use static IPv4 settings and fill in the IPv4 address and Subnet mask for the first ESXi host’s vMotion IP address. Click NEXT.

Step 43.                       Review the information and click FINISH to complete adding the vMotion VMkernel port.

Procedure 2.     Add the ESXi Host(s) to the VMware Virtual Distributed Switch

Step 1.    From the VMware vSphere HTML5 Client, click Networking under Menu.

Step 2.    Right-click the vDS and select Add and Manage Hosts.

Step 3.    Ensure that Add hosts is selected and click NEXT.

Step 4.    Click the green + sign to add New hosts.  Select the ESXi host(s) and click OK. Click NEXT.

Step 5.    Select vmnic2 on each host and click Assign uplink. Select Uplink 1 and click OK.

Step 6.    Select vmnic3 on each host and click Assign uplink. Select Uplink 2 and click OK.

Step 7.    If more than one host is being connected to the vDS, check the box for Apply this uplink assignment to the rest of the hosts.

Note:   It is important to assign the uplinks as defined in these steps. This allows the port groups to be pinned to the appropriate Cisco UCS fabric.

Step 8.    Click NEXT.

Step 9.    Do not migrate any VMkernel ports and click NEXT.

Step 10.                       Do not migrate any VM ports and click NEXT.

Step 11.                       Click FINISH to complete adding the ESXi host(s) to the vDS.

Procedure 3.     Add the vMotion VMkernel Port(s) to the ESXi Host

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    Click the Configure tab.

Step 3.    In the list under Networking, click VMkernel adapters.

Step 4.    Select Add Networking to add host networking.

Step 5.    Make sure VMkernel Network Adapter is selected and click NEXT.

Step 6.    Select BROWSE to the right of Select an existing network.

Step 7.    Select VMKernel-vMotion on the vDS and click OK.

Step 8.    Click NEXT.

Step 9.    Make sure the Network label is VMkernel-vMotion with the vDS in parenthesis. From the drop-down list, select Custom for MTU and make sure the MTU is set to 9000. Select the vMotion TCP/IP stack and click NEXT.

Step 10.                       Select Use static IPv4 settings and input the host’s vMotion IPv4 address and Subnet mask.

Step 11.                       Click NEXT.

Step 12.                       Review the parameters and click FINISH to add the vMotion VMkernel port.

Finalize the vCenter and ESXi Setup

This section details how to finalize the VMware installation.

Procedure 1.     Configure ESXi Host Swap

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    In the center pane, click the Configure tab.

Step 3.    In the list under System, click System Swap.

Step 4.    In the right pane, click EDIT.

Step 5.    Select Can use datastore and use the drop-down list to select infra_swap. Leave all other settings unchanged.

Graphical user interface, text, application, emailDescription automatically generated

Step 6.    Click OK to save the configuration changes.

Procedure 2.     Configure Virtual Machine Swap File Location

Step 1.    In Virtual Center, click on the Cluster name, and select the Configure menu.

Step 2.    Select the General option.

Step 3.    In the window on the right, click EDIT above the Datastore specified by host.

Step 4.    Select Datastore specified by host option.

Step 5.    Click OK.

Step 6.    For each server in the cluster, click on Configuration, then under Virtual Machines, select Swap File Location.

Step 7.    On the right, click EDIT.

Step 8.    Double-click the infra_swap datastore to select it.

Step 9.    Click OK.

A picture containing graphical user interfaceDescription automatically generated

Procedure 3.     Verify ESXi Host Fibre Channel Multi-Path Configuration

Step 1.    In the vCenter HTML5 Interface, under Hosts and Clusters select the ESXi host.

Step 2.    In the center pane, click the Configure tab.

Step 3.    In the list under Storage, click Storage Devices. Make sure the NETAPP Fibre Channel Disk is selected.

Step 4.    Select the Paths tab.

Step 5.    Ensure that 4 FC paths appear, two of which should have the status Active (I/O).

Graphical user interface, text, application, emailDescription automatically generated

Procedure 4.     VMware ESXi 7.0 U3 TPM Attestation

Note:   If your Cisco UCS servers have Trusted Platform Module (TPM) 2.0 modules installed, the TPM can provide assurance that ESXi has booted with UEFI Secure Boot enabled and using only digitally signed code. In the Create a Cisco UCS Server Profile section of this document, UEFI secure boot was enabled in the boot policy. A server can boot with UEFI Secure Boot with or without a TPM 2.0 module. If it has a TPM, VMware vCenter can attest that the server booted with UEFI Secure Boot.

Step 1.    For Cisco UCS servers that have TPM 2.0 modules installed, TPM Attestation can be verified in the vSphere HTML5 Client. 

Step 2.    In the vCenter HTML5 Interface, under Hosts and Clusters select the cluster.

Step 3.    In the center pane, click the Monitor tab.

Step 4.    Click Monitor > Security. The Attestation status will show the status of the TPM.

Storage Configuration – ONTAP NVMe Namespace Mapping and Finalizing ONTAP Storage

Procedure 1.     Ansible Configuration

Step 1.    Edit the following variable file and update the “namespaces” and “subsystem” variables under nvme_specs:

FlexPod-for-EHR/Ansible/vars/ontap_main.yml

Note:   Add the NQNs from each host to the subsystem variable. The NVME namespace will be shared by all the hosts in the nvme subsystem.

Step 2.    From /root/ FlexPod-for-EHR/Ansible, invoke the ansible scripts for this section use the following command:

ansible-playbook -i inventory Setup_ONTAP.yml -t ontap_config_part_3

Note:   Use the -vvv tag to see detailed execution output log.

Procedure 2.     Manual Configuration

Step 1.    Create NVMe namespace:

vserver nvme namespace create -vserver SVM_name -path path -size size_of_namespace -ostype OS_type

AA17-A400::> vserver nvme namespace create -vserver Infra-SVM -path /vol/NVMe_datastore_01/NVMe_namespace_01 -ostype vmware -size 50G

Created a namespace of size 50GB (53687091200).

Step 2.    Create NVMe subsystem:

vserver nvme subsystem create -vserver SVM_name -subsystem name_of_subsystem -ostype OS_type

AA17-A400::> vserver nvme subsystem create -vserver Infra-SVM -subsystem nvme_infra_hosts -ostype vmware

Step 3.    Verify the subsystem was created:

vserver nvme subsystem show -vserver SVM_name

AA17-A400::> vserver nvme subsystem show -vserver Infra-SVM

Vserver Subsystem    Target NQN

------- ------------ --------------------------------------------------------

Infra-SVM

        nvme_infra_hosts

                     nqn.1992-08.com.netapp:sn.e01bbb1de4f911ebac6fd039ea166b8c:subsystem. nvme_infra_host_01_02_03

Procedure 3.     Configure NVMe over FC on ESXi Host

Note:   Complete this procedure whether the Ansible configuration or manual configuration was used to set up the NVME namespace and subsystem.

Step 1.    Enable NVMe/FC with Asymmetric Namespace Access (ANA):

esxcfg-advcfg -s 0 /Misc/HppManageDegradedPaths

Step 2.    Reboot the Host. After reboot, verify that the HppManageDegradedPaths parameter is now disabled:

esxcfg-advcfg -g /Misc/HppManageDegradedPaths

The Value of HppManageDegradedPaths is 0

Step 3.    Get the ESXi host NQN string and add this to the corresponding subsystem on the ONTAP array:

esxcli nvme info get

Step 4.    Add the host NQN(s) obtained in the last step to the NetApp ONTAP subsystem one by one:

vserver nvme subsystem host add -vserver <SVM_name> -subsystem <Subsystem_name> -host-nqn
<Host_NQN:subsystem.Subsystem_name>


AA17-A400::> vserver nvme subsystem host add -vserver Infra-SVM  -subsystem nvme_infra_hosts  -host-nqn nqn.2014-08.com.vmware:nvme:ESXi-01

 

AA17-A400::> vserver nvme subsystem host add -vserver Infra-SVM  -subsystem nvme_infra_hosts  -host-nqn nqn.2014-08.com.vmware:nvme:ESXi-02

Note:   It is important to add the host NQNs using separate commands as shown. ONTAP will accept a comma separated list of host NQNs without generating an error message however the ESXi hosts will not be able to map the namespace.

Step 5.    Verify the host NQNs were added successfully:

AA17-A400::> vserver nvme subsystem host show

 

Vserver Subsystem Host NQN

 

------- --------- ----------------------------------------------------------

 

Infra-SVM

 

        nvme_infra_hosts

 

                  nqn.2014-08.com.vmware:nvme: ESXi-01

 

                  nqn.2014-08.com.vmware:nvme:ESXi-02

 

2 entries were displayed.

Note:   In the example above, host NQNs for two FC ESXi hosts in an ESXi cluster were added to the same subsystem to create a shared datastore.

Step 6.    Map the Namespace to the subsystem:

vserver nvme subsystem map add -vserver SVM_name -subsystem subsystem_name -path path

AA17-A400::> vserver nvme subsystem map add -vserver Infra_SVM -subsystem nvme_infra_hosts -path /vol/NVMe_datastore_01/NVMe_namespace_01

Step 7.    Verify the Namespace is mapped to the subsystem:

vserver nvme subsystem map show -vserver Infra-SVM -instance

 

AA17-A400::> vserver nvme subsystem map show -vserver Infra-SVM -instance

 Vserver Name: Infra-SVM

 Subsystem: nvme_infra_hosts

 NSID: 00000001h

Namespace Path: /vol/NVMe_datastore_01/NVMe_namespace_01

Namespace UUID: 01add6cf-d1c3-4d17-92f9-149f683a1e4d

 

Step 8.    Reboot each ESXi host and then verify that the ONTAP target NVMe/FC controllers are properly discovered on the ESXi Host:

[root@nx-esxi-1:~] esxcli nvme controller list

Name Controller Number Adapter Transport Type Is Online

nqn.1992-08.com.netapp:sn.50919001efe111ebb785d039ea166b8c:subsystem.nvme_infra_host_01#vmhba65#2005d039ea17129b:2006d039ea17129b 258 vmhba65 FC true

nqn.1992-08.com.netapp:sn.50919001efe111ebb785d039ea166b8c:subsystem.nvme_infra_host_01#vmhba64#2005d039ea17129b:2007d039ea17129b 261 vmhba64 FC true

nqn.1992-08.com.netapp:sn.50919001efe111ebb785d039ea166b8c:subsystem.nvme_infra_host_01#vmhba65#2005d039ea17129b:2008d039ea17129b 265 vmhba65 FC true

nqn.1992-08.com.netapp:sn.50919001efe111ebb785d039ea166b8c:subsystem.nvme_infra_host_01#vmhba64#2005d039ea17129b:2009d039ea17129b 266 vmhba64 FC true

Procedure 4.     Configure DNS

Step 1.    Run the following commands to configure DNS for the Infra_SVM:

dns create -vserver <vserver name> -domains <dns-domain> -nameservers <dns-servers>

dns create -vserver Infra_SVM -domains flexpod.cisco.com -nameservers 10.1.156.250,10.1.156.251

Procedure 5.     Delete Residual Default Broadcast Domains (Applicable for 2-node cluster only)

Step 1.    Run the following commands to delete the Default broadcast domains that are not in use:

broadcast-domain delete -broadcast-domain <broad-domain-name>

broadcast-domain delete -broadcast-domain Default-1

Procedure 6.     Test AutoSupport

Step 1.    Run the following commands to test the AutoSupport configuration by sending a message from all nodes of the cluster:

autosupport invoke -node * -type all -message "FlexPod storage configuration completed"

Procedure 7.     ESXi Host NVMe over FC Datastore Configuration

Step 1.    The remaining steps in the VMware vSphere Client are manual steps that should be completed whether Ansible configuration or manual configuration is being done. Verify that the NVMe Fibre Channel Disk is mounted on each ESXi host. Under Hosts and Clusters select the ESXi host. In the center pane, select Configure > Storage > Storage Devices. The NVMe Fibre Channel Disk should be listed under Storage Devices. Select the NVMe Fibre Channel Disk, then select Paths underneath. Verify 2 paths have a status of Active (I/O) and 2 paths have a status of Active.

Step 2.    Repeat this step for all 3 hosts.

Graphical user interface, text, application, emailDescription automatically generated

Step 3.    For any of the three hosts, right-click the host under Hosts and Clusters and select Storage > New Datastore. Leave VMFS selected and click NEXT.

Step 4.    Name the datastore and select the NVMe Fibre Channel Disk. Click NEXT.

NVME_datastore_name

Step 5.    Leave VMFS 6 selected and click NEXT.

Step 6.    Leave all Partition configuration values at the default values and click NEXT.

Step 7.    Review the information and click FINISH.

Step 8.    Select Storage and select the just-created NVMe datastore. In the center pane, select Hosts. Ensure all three hosts have the datastore mounted.

NVME_datastore_mount

FlexPod Management Tools Setup

Cisco Intersight Hardware Compatibility List (HCL) Status

Cisco Intersight evaluates the compatibility of customer’s UCS system to check if the hardware and software have been tested and validated by Cisco or Cisco partners. Intersight reports validation issues after checking the compatibility of the server model, processor, firmware, adapters, operating system, and drivers, and displays the compliance status with the Hardware Compatibility List (HCL).

To determine HCL compatibility for VMware ESXi, Cisco Intersight uses Cisco UCS Tools. The Cisco UCS Tools is part of VMware ESXi Cisco custom ISO, and no additional configuration is required.

Note:   For more details on Cisco UCS Tools manual deployment and troubleshooting, refer to: https://intersight.com/help/saas/resources/cisco_ucs_tools#about_cisco_ucs_tools  

To find detailed information about the hardware compatibility of a compute node, in Cisco Intersight select Servers in the left menu bar, click a server, select HCL.

A screenshot of a computerDescription automatically generated with medium confidence

NetApp ONTAP Tools 9.10 Deployment Procedure

The ONTAP tools for VMware vSphere provide end-to-end life cycle management for virtual machines in VMware environments that use NetApp storage systems. It simplifies storage and data management for VMware environments by enabling administrators to directly manage storage within the vCenter Server.

This section describes the deployment procedures for the NetApp ONTAP Tools for VMware vSphere.

NetApp ONTAP Tools for VMware vSphere 9.10 Pre-installation Considerations

The following licenses are required for ONTAP Tools on storage systems that run ONTAP 9.8 or above:

      Protocol licenses (NFS, FCP, and/or iSCSI)

      NetApp FlexClone® ((optional) Required for performing test failover operations for SRA and for vVols operations of VASA Provider.

      NetApp SnapRestore® (for backup and recovery)

      The NetApp SnapManager® Suite - NetApp SnapMirror® or NetApp SnapVault® ((optional) Required for performing failover operations for SRA and VASA Provider if using vVols replication.)

Note:   The Backup and Recovery capability has been integrated with SnapCenter and requires additional licenses for SnapCenter to perform backup and recovery of virtual machines and applications.

Table 15.   Port Requirements for NetApp ONTAP Tools

Port

Requirement

443 (HTTPS)

Secure communications between VMware vCenter Server and the storage systems

8143 (HTTPS)

VSC listens for secure communications

9083 (HTTPS)

VASA Provider uses this port to communicate with the vCenter Server and obtain TCP/IP settings

7

VSC sends an echo request to ONTAP to verify reachability and is required only when adding storage system and can be disabled later.

The requirements for deploying NetApp ONTAP Tools (VSC) are listed here.

Download NetApp ONTAP Tools OVA

Download the NetApp ONTAP Tools 9.10 OVA (netapp-ontap-tools-for-vmware-vsphere-9.10-8127.ova) from NetApp support site: https://mysupport.netapp.com/site/products/all/details/otv/downloads-tab/download/63792/9.10

Procedure 1.     Install NetApp ONTAP Tools utilizing Ansible

Step 1.    Clone the repository from https://github.com/NetApp-Automation/ONTAP-Tools-for-VMware-vSphere

Step 2.    Follow the instructions in the README file in the repository to ensure the Ansible environment is configured properly.

Step 3.    Update the following variable files:

hosts

vars/ontap_tools_main.yml

group_vars/vcenter

Step 4.    To invoke the ansible scripts use the following command:

ansible-playbook Setup_ONTAP_tools.yml -i hosts

Procedure 2.     Install NetApp ONTAP Tools Manually

Step 1.    Launch the vSphere Web Client and navigate to Hosts and Clusters.

Step 2.    Select ACTIONS for the Epic-DC datacenter and select Deploy OVF Template.

Graphical user interface, applicationDescription automatically generated

Step 3.    Browse to the ONTAP tools OVA file(downloaded from the NetApp Support site) and select the file.

Step 4.    Enter the VM name and select a datacenter or folder to deploy the VM and click NEXT.

Step 5.    Select a host cluster resource to deploy OVA and click NEXT.

Step 6.    Review the details and accept the license agreement.

Step 7.    Select an appropriate datastore volume and select the Thin Provision option for the virtual disk format.

Graphical user interface, applicationDescription automatically generated

Step 8.    From Select Networks, select a destination network and click NEXT.

Step 9.    From Customize Template, enter the ONTAP tools administrator password, vCenter name or IP address and other configuration details and click NEXT.

ONTAP_cust_template

Step 10.                       Review the configuration details entered and click FINISH to complete the deployment of NetApp-ONTAP-Tools VM.

Graphical user interface, applicationDescription automatically generated

Step 11.                       Power on the ONTAP-tools VM and open the VM console.

Step 12.                       During the ONTAP-tools VM boot process, you see a prompt to install VMware Tools. From vCenter, right-click the ONTAP-tools VM > Guest OS > Install VMware Tools.

Step 13.                       Networking configuration and vCenter registration information was provided during the OVF template customization, therefore after the VM is up and running, ONTAP-Tools and vSphere API for Storage Awareness (VASA) is registered with vCenter.

Step 14.                       Refresh the vCenter Home Screen and confirm that the ONTAP tools is installed.

Note:   The NetApp ONTAP tools vCenter plug-in is only available in the vSphere HTML5 Client and is not available in the vSphere Web Client.

ONTAP plug-in menu

Procedure 3.     Download the NetApp NFS Plug-in for VAAI

Step 1.    Download the NetApp NFS Plug-in 2.0 for VMware .vib file from the NFS Plugin Download page and save it to your local machine or admin host.

Graphical user interface, text, applicationDescription automatically generated

Procedure 4.     Install the NetApp NFS Plug-in for VAAI

Note:   If the NFS Plug-in for VAAI was previously installed on the ESXi hosts along with the Cisco UCS VIC drivers; it is not necessary to re-install.

Step 1.    Rename the .vib file that you downloaded from the NetApp Support Site to NetAppNasPlugin.vib to match the predefined name that ONTAP tools uses.

Step 2.    Click Settings in the ONTAP tool Getting Started page.

Step 3.    Click NFS VAAI Tools tab.

Step 4.    Click Change in the Existing version section.

Step 5.    Browse and select the renamed .vib file, and then click Upload to upload the file to the virtual appliance.

Step 6.    In the Install on ESXi Hosts section, select the ESXi host on which you want to install the NFS Plug-in for VAAI, and then click Install.

Graphical user interface, text, application, emailDescription automatically generated

Step 7.    Reboot the ESXi host after the installation finishes.

Procedure 5.     Verify the VASA Provider

Note:   The VASA provider for ONTAP is enabled by default during the installation of the NetApp ONTAP tools.

Step 1.    From the vSphere Client, click Menu > ONTAP tools.

Step 2.    Click Settings.

Step 3.    Click Manage Capabilities in the Administrative Settings tab.

Step 4.    In the Manage Capabilities dialog box, click Enable VASA Provider if it was not pre-enabled.

Step 5.    If Enable VASA Provider is not enabled, click to enable then enter the IP address of the virtual appliance for ONTAP tools, VASA Provider, and VMware Storage Replication Adapter (SRA) and the administrator password, and then click Apply.

Text, applicationDescription automatically generated

Procedure 6.     Discover and Add Storage Resources

Step 1.    Using the vSphere Web Client, log in to the vCenter. If the vSphere Web Client was previously opened, close the tab, and then reopen it.

Step 2.    In the Home screen, click the Home tab and click ONTAP tools.

Note:   When using the cluster admin account, add storage from the cluster level.

Note:   You can modify the storage credentials with the vsadmin account or another SVM level account with role-based access control (RBAC) privileges. Refer to the ONTAP 9 Administrator Authentication and RBAC Power Guide for additional information.

Step 3.    Click Storage Systems >Add

Step 4.    Click Overview > Getting Started, and then click ADD under Add Storage System.

Step 5.    Specify the vCenter Server instance where the storage will be located.

Step 6.    In the IP Address/Hostname field, enter the storage cluster management IP.

Step 7.    Confirm Port 443 to Connect to this storage system.

Step 8.    Enter admin for the username and the admin password for the cluster.

Step 9.    Click Save to add the storage configuration to ONTAP tools.

Graphical user interface, text, application, email, websiteDescription automatically generated

Step 10.                       Wait for the Storage Systems to update. You might need to click Refresh to complete this update.

Step 11.                       Discover the cluster and SVMs with the cluster admin account:

Graphical user interface, text, application, emailDescription automatically generated

Step 12.                       From the vSphere Client Home page, click Hosts and Clusters.

Step 13.                       Right-click the Epic-DC datacenter, click NetApp ONTAP tools > Update Host and Storage Data.

Graphical user interface, applicationDescription automatically generated

NetApp ONTAP tools displays a Confirm dialog box that informs you that this operation might take a few minutes.

Step 14.                       Click OK.

Procedure 7.     Optimal Storage Settings for ESXi Hosts

Note:   NetApp ONTAP tools enables the automated configuration of storage-related settings for all ESXi hosts that are connected to NetApp storage controllers.

Step 1.    From the VMware vSphere Web Client Home page, click vCenter > Hosts and Clusters.

Step 2.    Select a host and then click Actions > NetApp ONTAP tools > Set Recommended Values.

Step 3.    In the NetApp Recommended Settings dialog box, select all the applicable values for the ESXi host.

Graphical user interface, text, application, emailDescription automatically generated

Note:   This functionality sets values for HBAs and converged network adapters (CNAs), sets appropriate paths and path-selection plug-ins, and verifies appropriate settings for NFS I/O. A vSphere host reboot may be required after applying the settings.

Step 4.    Click OK.

ONTAP Tools 9.10 Provisioning Datastores

Using ONTAP tools, the administrator can provision an NFS, FC, FC-NVMe or iSCSI datastore and attach it to a single host or multiple hosts in the cluster. The following sections describe provisioning a datastore and attaching it to the cluster.

Note:   It is a NetApp best practice to use ONTAP tools to provision datastores for the FlexPod infrastructure. When using VSC to create vSphere datastores, all NetApp storage best practices are implemented during volume creation and no additional configuration is needed to optimize performance of the datastore volumes.

Storage Capabilities

A storage capability is a set of storage system attributes that identifies a specific level of storage performance (storage service level), storage efficiency, and other capabilities such as encryption for the storage object that is associated with the storage capability.

Procedure 1.     Create the Storage Capability Profile

Note:   In order to leverage the automation features of VASA two primary components must first be configured. The Storage Capability Profile (SCP) and the VM Storage Policy. The Storage Capability Profile expresses a specific set of storage characteristics into one or more profiles used to provision a Virtual Machine. The SCP is specified as part of VM Storage Policy. NetApp ONTAP tools comes with several pre-configured SCPs such as Platinum, Bronze, and so on.

Note:   The ONTAP tools for VMware vSphere plug-in also allows customers to set Quality of Service (QoS) rule using a combination of maximum and/or minimum IOPs.

Step 1.    From the vCenter console, click Menu > ONTAP tools.

Step 2.    In the NetApp ONTAP tools click Storage Capability Profiles.

Step 3.    Select the Platinum Storage Capability Profile and select Clone from the toolbar.

NA_Capability_profile

Step 4.    Enter a name for the cloned SCP and add a description if desired. Click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 5.    Select All Flash FAS(AFF) for the storage platform and click NEXT.

Graphical user interface, text, application, chat or text messageDescription automatically generated

Step 6.    Select None to allow unlimited performance or set a the desired minimum and maximum IOPS for the QoS policy group. Click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 7.    On the Storage attributes page, Change the Encryption and Tiering policy to the desired settings and click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 8.    Review the summary page and click FINISH to create the storage capability profile.

Note:   It is recommended to Clone the Storage Capability Profile if you wish to make any changes to the predefined profiles rather than editing the built-in profile.

Procedure 2.     Create a VM Storage Policy

Create a VM storage policy and associate a storage capability profile (SCP) to the datastore that meets the requirements defined in the SCP.

Step 1.    From the vCenter console, click Menu > Policies and Profiles.

Step 2.    Select VM Storage Policies and click CREATE.

A screenshot of a computerDescription automatically generated with medium confidence

Step 3.    Create a name for the VM storage policy and enter a description and click NEXT.

Step 4.    Choose Enable rules for NetApp.clustered.Data.ONTAP.VP.VASA10 storage located under the Datastore specific rules section and click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 5.    On the Placement tab select the SCP created in the previous step and click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 6.    All the datastores with matching capabilities are displayed, click NEXT.

Step 7.    Review the policy summary and click FINISH.

Procedure 3.     Provision NFS Datastore

Step 1.    From the vCenter console, click Menu > ONTAP tools.

Step 2.    From the ONTAP tools Home page, click Overview.

Step 3.    In the Getting Started tab, click Provision.

Step 4.    Click Browse to select the destination to provision the datastore.

Step 5.    Select the type as NFS and Enter the datastore name.

Step 6.    Provide the size of the datastore and the NFS Protocol.

Step 7.    Check the storage capability profile and click NEXT.

Graphical user interface, application, emailDescription automatically generated

Step 8.    Select the desired Storage Capability Profile, cluster name and the desired SVM to create the datastore.

Step 9.    Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 10.                       Select the aggregate name and click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 11.                       Review the Summary and click FINISH.

Graphical user interface, applicationDescription automatically generated

Note:   The datastore is created and mounted on the hosts in the cluster. Click Refresh from the vSphere Web Client to see the newly created datastore or it is also listed in the ONTAP tools home page > Traditional Dashboard > Datastores view. Also, ONTAP tools home page > Reports > Datastore Report should list the newly created datastore.

Procedure 4.     Provision FC Datastore

Step 1.    From the vCenter console, click Menu > ONTAP tools.

Step 2.    From the ONTAP tools Home page, click Overview.

Step 3.    In the Getting Started tab, click Provision.

Step 4.    Click Browse to select the destination to provision the datastore.

Step 5.    Select the type as VMFS and Enter the datastore name.

Step 6.    Provide the size of the datastore and the FC Protocol.

Step 7.    Check the Use storage capability profile and click NEXT.

Graphical user interface, textDescription automatically generated

Step 8.    Select the Storage Capability Profile, Storage System, and the desired Storage VM to create the datastore.

Step 9.    Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 10.                       Select the aggregate name and click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 11.                       Review the Summary and click FINISH.

Graphical user interfaceDescription automatically generated

Step 12.                       Click OK.

Note:   The datastore is created and mounted on all the hosts in the cluster. Click Refresh from the vSphere Web Client to see the newly created datastore.

Procedure 5.     Create Virtual Machine with Assigned VM Storage Policy

Step 1.    Log into vCenter and navigate to the VMs and Templates tab and click to select the datacenter (for example, HC-DC).

Step 2.    Click Actions and click New Virtual Machine.

Step 3.    Click Create a new virtual machine and click NEXT.

Step 4.    Enter a name for the VM and click the HC-DC datacenter.

Step 5.    Choose the HC-Management Data compute Resource.

Step 6.    Select the VM storage policy from the selections and select a compatible datastore. Click NEXT.

Graphical user interface, tableDescription automatically generated

Step 7.    Select Compatibility (for example, ESXi 7.0 U2 or later) and click NEXT.

Step 8.    Select the Guest OS and click NEXT.

Step 9.    Customize the hardware for the VM and click NEXT.

Step 10.                       Review the details and click FINISH.

Virtual Volumes (vVols)

NetApp VASA Provider enables customers to create and manage VMware virtual volumes (vVols). A vVols datastore consists of one or more FlexVol volumes within a storage container (also called "backing storage"). A virtual machine can be spread across one vVols datastore or multiple vVols datastores. All of the FlexVol volumes within the storage container must use the same protocol (NFS, iSCSI, or FCP) and the same SVMs.

Note:   Lab testing has shown that if a virtual machine (VM) has one or more disks in vVol datastores and the VM is migrated to another host, under heavy load the VM can be stunned or frozen for 45 or more seconds.

Procedure 1.     Verify NDMP Vserver Scope Mode

Step 1.    View NDMP scope mode with the following command:

system services ndmp node-scope-mode status

Step 2.    If NDMP node-scope is enabled, disable NDMP node-scoped mode:

system services ndmp node-scope-mode off

Step 3.    Enable NDMP services on the vserver:

vserver add-protocols -protocols ndmp -vserver Infra_svm
vserver services ndmp on -vserver Infra_svm

Procedure 2.     Create the Storage Capability Profile

Note:   Select one or more VASA Provider storage capability profiles for a vVols datastore. A default storage capability profile can also be specified for any vVols datastores that are automatically created in that storage container.

Step 1.    From the vCenter console, click Menu > ONTAP tools.

Step 2.    From the ONTAP tools Home page, click Storage Capability Profiles.

Step 3.    Select the Platinum Storage Capability Profile and select Clone from the toolbar.

Graphical user interface, applicationDescription automatically generated

Step 4.    Enter a name for the cloned SCP and add a description if desired. Click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 5.    Select All Flash FAS(AFF) for the storage platform and click NEXT.

Step 6.    Select None to allow unlimited performance or set a the desired minimum and maximum IOPS for the QoS policy group. Selecting a value for Max IOPS enables customers to use the QoS functionality.

Note:   When applied for a virtual datastore, a QoS policy with "MAX IOPS" value is created for each data vVols.

Note:   When you select ONTAP Service Level, then the existing adaptive QoS policies of ONTAP are applied to a data vVols. You can select one of three service levels: Extreme, Performance, or Value. The ONTAP service level is applicable only to vVols datastores.

Step 7.    On the Storage attributes page, change the Encryption and Tiering policy to the desired settings and click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 8.    Review the summary page and click FINISH to create the storage capability profile.

Procedure 3.     Create a VM Storage Policy

Create a VM storage policy and associate a storage capability profile (SCP) to the datastore that meets the requirements defined in the SCP.

Step 1.    From the vCenter console, click Menu > Policies and Profiles.

Step 2.    Select VM Storage Policies and click Create VM Storage Policy.

Graphical user interface, text, applicationDescription automatically generated

Step 3.    Enter a new name for the VM storage Policy and click NEXT.

Step 4.    Select Enable rules for NetApp.clustered.Data.ONTAP.VP.VASA.10 storage and NetApp.clustered.Data.ONTAP.VP.vvol storage, located under the Datastore specific rules section and click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 5.    Select a pre-defined SCP or SCP created in the previous step under BOTH NetApp.clustered.Data.ONTAP.VP.VASA10 and NetApp.clustered.Data.ONTAP.VP.vvol rules. Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 6.    The datastores with matching capabilities are displayed, click NEXT.

Step 7.    Review the Policy Summary and click Finish.

Procedure 4.     Provision a vVols Datastore

Step 1.    From the vCenter console, click Menu > ONTAP tools.

Step 2.    From the NetApp ONTAP tools Home page, click Overview.

Step 3.    In the Getting Started tab, click Provision.

Step 4.    Click Browse to select the destination cluster to provision the datastore.

Step 5.    Select the type as vVols and enter the datastore name.

Step 6.    Select NFS for protocol and click NEXT.

Graphical user interfaceDescription automatically generated

Step 7.    Select the Storage capability profile previously created for vVols.

Step 8.    Select the Storage system and the Storage SVM. Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 9.    Select Create new volumes.

Step 10.                       Enter the name and size for one or more vVols and Click ADD.

Note:   You can create multiple vVols for a datastore.

Graphical user interface, text, application, emailDescription automatically generated 

Step 11.                       Verify the correct Default storage capability profile is selected. Click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 12.                       Review all the fields on the summary page and click FINISH.

Graphical user interface, applicationDescription automatically generated

Step 13.                       Verify in the vVols Datastore report the vVols is mounted correctly, under ONTAP tools > Reports > vVols Datastore Report.

Graphical user interface, text, application, emailDescription automatically generated

Note:   You might need to rediscover the storage systems (ONTAP tools > Storage Systems > Rediscover All). In some cases, log-out and log-back-in for vCenter might be required.

Note:   vVols can also be configured for FC or ISCSI protocol using the steps outlined above and selecting a different protocol on the first screen.

Procedure 5.     Update a vVols Datastore

Step 1.    Go to the Storage tab in vCenter.

Step 2.    Select the vVol datastore.

Step 3.    Click ACTIONs.

Step 4.    Under NetApp ONTAP tools:

a.     Expand Storage on a vVols Datastore.

b.     Remove Storage on a vVols Datastore.

c.     Edit Properties of vVols Datastore.

d.     Mount vVols Datastore.

e.     Delete vVols Datastore.

Graphical user interface, text, application, email, websiteDescription automatically generated

Procedure 6.     Create a Virtual Machine on a vVols Datastore

Step 1.    Navigate to vSphere Client > VMs and Templates > Actions > New Virtual Machine.

Step 2.    Select Create a new virtual machine and click NEXT.

Step 3.    Enter the name for the VM (for example, VM_VVOL_01) and select the datacenter (for example, Epic-DC). Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 4.    Select the cluster and click NEXT.

Step 5.    Select the VM Storage Policy and vVol datastore. Verify Compatibility checks succeeded.

TableDescription automatically generated

Step 6.    Click NEXT.

Step 7.    Select Compatibility (for example, ESXi 7.0 U3 and later) and click NEXT.

Step 8.    Select Guest OS Family (for example, Linux) and Guest OS version (for example, Red Hat Enterprise Linux 9) and click NEXT.

Step 9.    Select the hardware parameters (CPU, Memory, HDD and so on.) and click NEXT.

Step 10.                       Review the settings and click FINISH.

Graphical user interface, text, application, emailDescription automatically generated

Step 11.                       Verify that the VM is created successfully and install operating system on the VM.

NetApp SnapCenter 4.6

SnapCenter Software is a centralized and scalable platform that provides application-consistent data protection for applications, databases, host file systems, and VMs running on ONTAP systems anywhere in the Hybrid Cloud.

NetApp SnapCenter Architecture

The SnapCenter platform is based on a multitier architecture that includes a centralized management server (SnapCenter Server) and a SnapCenter host agent. The host agent that performs virtual machine and datastore backups for VMware vSphere is the SnapCenter Plug-in for VMware vSphere. It is packaged as a Linux appliance (Debian-based Open Virtual Appliance format) and is no longer part of the SnapCenter Plug-ins Package for Windows. Additional information on deploying SnapCenter server for application backups can be found below.

This CVD focuses on deploying and configuring the SnapCenter plug-in for VMware vSphere to protect virtual machines and VM datastores.

Note:   Customers must install SnapCenter Server and the necessary plug-ins to support application-consistent backups for Microsoft SQL, Microsoft Exchange, Oracle databases and SAP HANA. Application-level protection is beyond the scope of this deployment guide.

Refer to the SnapCenter documentation for more information or the application specific CVD’s and technical reports for detailed information on how to deploy SnapCenter for a specific application configuration.

      SnapCenter Documentation: https://docs.netapp.com/us-en/snapcenter/index.html

      Deploy FlexPod Datacenter for Microsoft SQL Server 2019 with VMware 7.0 on Cisco UCS B200 M6 and NetApp ONTAP 9.8: https://www.cisco.com/c/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/flexpod-sql-2019-vmware-on-ucs-netapp-ontap-wp.html

      SnapCenter Plug-in for VMware vSphere Documentation: https://docs.netapp.com/us-en/sc-plugin-vmware-vsphere-45/index.html

Install SnapCenter Plug-In 4.6 for VMware vSphere

NetApp SnapCenter Plug-in for VMware vSphere is a Linux-based virtual appliance which enables the SnapCenter Plug-in for VMware vSphere to protect virtual machines and VMware datastores.

Host and Privilege Requirements for the SnapCenter Plug-In for VMware vSphere

Review the following requirements before installing the SnapCenter Plug-in for VMware vSphere virtual appliance:

      SnapCenter Plug-in for VMware vSphere is deployed as a Linux based virtual appliance.

      Virtual appliance must be deployed on the vCenter Server.

      Virtual appliance must not be deployed in a folder name with special characters.

      A separate, unique instance of the virtual appliance must be deployed for each vCenter Server.

Table 16.   Port Requirements

Port

Requirement

8080(HTTPS) bidirectional

This port is used to manage the virtual appliance

8144(HTTPs) bidirectional

Communication between SnapCenter Plug-in for VMware vSphere and vCenter

443 (HTTPS)

Communication between SnapCenter Plug-in for VMware vSphere and vCenter

License Requirements for SnapCenter Plug-In for VMware vSphere

The following licenses are required to be installed on the ONTAP storage system to backup and restore VM’s in the virtual infrastructure:

Table 17.   SnapCenter Plug-in for VMware vSphere License Requirements

Product

License Requirements

ONTAP

SnapManager Suite: Used for backup operations

One of these: SnapMirror or SnapVault (for secondary data protection regardless of the type of relationship)

ONTAP Primary Destinations

To perform protection of VMware VMs and datastores the following licenses should be installed:

SnapRestore: used for restoring operations

FlexClone: used for mount and attach operations

ONTAP Secondary Destinations

To perform protection of VMware VMs and datastores only:

FlexClone: used for mount and attach operations

VMware

vSphere Standard, Enterprise, or Enterprise Plus

A vSphere license is required to perform restore operations, which use Storage vMotion. vSphere Essentials or Essentials Plus licenses do not include Storage vMotion.

Note:   It is recommended (but not required) to add SnapCenter Standard licenses to secondary destinations. If SnapCenter Standard licenses are not enabled on secondary systems, SnapCenter cannot be used after a failover operation. A FlexClone license on secondary storage is required to perform mount and attach operations. A SnapRestore license is required to perform restore operations.

Procedure 1.     Download and Deploy the SnapCenter Plug-In 4.6 for VMware vSphere

Step 1.    Download SnapCenter Plug-in for VMware vSphere OVA file from NetApp support site: (https://mysupport.netapp.com).

Procedure 2.     Deploy the SnapCenter Plug-In 4.6 for VMware vSphere using Ansible

Step 1.    Clone the repository from https://github.com/NetApp-Automation/SnapCenter-Plug-in-for-VMware-vSphere

Step 2.    Follow the instructions in the README file in the repository to ensure the Ansible environment is configured properly.

Step 3.    Update the following variable files:

hosts

vars/snapcenter_vmware_plugin_main.yml

group_vars/vcenter

Step 4.    To invoke the ansible scripts use the following command:

ansible-playbook -i hosts Setup_SnapCenter_VMware_Plugin.yml

Procedure 3.     Manually deploy the SnapCenter Plug-In 4.6 for VMware vSphere

Step 1.    From VMware vCenter, navigate to the VMs and Templates tab, right-click the data center (for example, Epic-DC) and select Deploy OVF Template.

Step 2.    Specify the location of the OVF Template and click NEXT.

Step 3.    On the Select a name and folder page, enter a unique name (for example, NX-SNAPCTR) and location (data center for example, Epic-DC) for the VM and click NEXT to continue.

Graphical user interface, text, application, emailDescription automatically generated

Step 4.    On the Select a compute resource page, select the cluster, and click NEXT.

Step 5.    On the Review details page, verify the OVA template details and click NEXT.

Step 6.    On the License agreements page, read and check the box I accept all license agreements. Click NEXT.

Step 7.    On the Select storage page, select a datastore, change the datastore virtual disk format to Thin Provision and click NEXT.

TableDescription automatically generated

Step 8.    On the Select networks page, select s destination network for example, IB-MGMT and then click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 9.    On the Customize template page, under Register to existing vCenter, enter the vCenter credentials.

Step 10.                       In Create SCV credentials, create a username (for example, admin) and password for the SCV maintenance user.

Step 11.                       In Setup Network Properties, enter the network information.

Step 12.                       In Setup Date and Time, provide the NTP server address(es) and select the time zone where the vCenter is located.

Step 13.                       Click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 14.                       On the Ready to complete page, review the page and click FINISH. The VM deployment will start. After the VM is deployed successfully, proceed to the next step.

Graphical user interface, application, tableDescription automatically generated

Step 15.                       Navigate to the SnapCenter VM, right-click, and select Power > Power On to start the virtual appliance.

Step 16.                       While the virtual appliance is powering on, click Install VMware tools.

Graphical user interface, text, applicationDescription automatically generated

Step 17.                       After the SnapCenter VM installation is complete and VM is ready to use, proceed to the next step.

Step 18.                       Log into SnapCenter Plug-in for VMware vSphere using the IP address (https://<ip_address_of_SnapCenter>:8080 ) displayed on the appliance console screen with the credentials that you provided in the deployment wizard.

Step 19.                       Verify on the Dashboard that the virtual appliance has successfully connected to vCenter and the SnapCenter Plug-in for VMware vSphere is successfully enabled and connected.

Graphical user interface, applicationDescription automatically generated

Procedure 4.     SnapCenter Plug-In for VMware vSphere in vCenter Server

Note:   After successfully installing the SnapCenter Plug-in for VMware vSphere, configure SnapCenter and make it ready to backup virtual machines.

Step 1.    Navigate to VMware vSphere Web Client URL https://<vCenter Server>.

Note:   If currently logged into vCenter, logoff, close the open tab and sign-on again to access the newly installed SnapCenter Plug-in for VMware vSphere.

Step 2.    After logging on, a blue banner will be displayed indicating the SnapCenter plug-in was successfully deployed. Click Refresh to activate the plug-in.

Step 3.    On the VMware vSphere Web Client page, select Menu > SnapCenter Plug-in for VMware vSphere to launch the SnapCenter Plug-in for VMware GUI.

Graphical user interface, text, applicationDescription automatically generated

Procedure 5.     Add Storage System

Step 1.    Click Storage Systems.

Graphical user interface, applicationDescription automatically generated

Step 2.    Click Add Storage System to add a cluster or SVM.

Step 3.    Enter vCenter, Storage System, user credentials, and other required information in following dialog box.

Step 4.    Check the box for Log SnapCenter server events to syslog and Send AutoSupport Notification for failed operation to storage system.

Graphical user interfaceDescription automatically generated

Step 5.    Click ADD.

Procedure 6.     Create Backup Policies for Virtual Machines and Datastores

Step 1.    From SnapCenter GUI, Click Policies.

Step 2.    On the Policies page, click +Create to create a new policy.

Step 3.    On the New Backup Policy page, follow these steps:

Step 4.    Enter the policy name and a description.

Step 5.    Enter the backups to keep.

Step 6.    From the Frequency drop-down list, select the backup frequency (hourly, daily, weekly, monthly, and on-demand only).

Step 7.    Expand the Advanced options and select VM Consistency and Include datastore with independent disks.

Step 8.    Click ADD.

Graphical user interfaceDescription automatically generated

Note:   Customers can create multiple policies as required for different sets of VMs or datastores.

Procedure 7.     Create Resource Groups

Resource groups are groups of virtual machines or datastores that are backed up together. A backup policy is associated with the resource group to back up the virtual machines and retain a certain number of backups as defined in the policy.

Step 1.    In SnapCenter Plug-in Navigator, click Resource Groups.

Step 2.    Click +Create to create a new Resource Group.

Graphical user interface, applicationDescription automatically generated

Note:   To create a resource group for one virtual machine, click VMs and Templates, right-click a virtual machine, select NetApp SnapCenter from the drop-down list, and then select Create Resource Group from the secondary drop-down list.

Note:   To create a resource group for one datastore, click Storage, right-click a datastore, select NetApp SnapCenter from the drop-down list, and then select Create Resource Group from the secondary drop-down list.

Step 3.    In the General Info & notification page, enter the resource group name and complete the notification settings.

Step 4.    Select the Custom snapshot format option and select the desired label for example, $ResourceGroup to have the resource group name appended to the snapshot name during snapshot operation.

Step 5.    Click NEXT.

Graphical user interface, text, applicationDescription automatically generated

Step 6.    Select a datastore as the parent entity to create a resource group of virtual machines.

Step 7.    Select the virtual machines from the available list. Click NEXT.

Graphical user interface, applicationDescription automatically generated

Note:   Entire datastores can be backed up by selecting data center (for example, Epic-DC) in the parent entity list box and selecting/adding the datastore.

Step 8.    From the Spanning Disks options, select the Always include all spanning datastores option and click NEXT.

Graphical user interface, text, applicationDescription automatically generated

Step 9.    From the Policies tab, select the policy created in the last step to associate the policy with the resource group and click NEXT.

TableDescription automatically generated

Step 10.                       From Schedules, select the schedule for the selected policy and click NEXT.

Graphical user interfaceDescription automatically generated

Step 11.                       Review the summary and click FINISH to complete the creation of the resource group.

Graphical user interface, applicationDescription automatically generated

Procedure 8.     View Virtual Machine Backups from vCenter by Using SnapCenter Plug-In

Backups of the virtual machines included in the resource group occurs according to the schedule of the policies associated with the resource group.

Step 1.    Log into the VMware vCenter GUI.

Step 2.    Navigate to the VMs and Templates tab.

Step 3.    Select a VM that is the member of the previously created Resource Group.

Step 4.    Select the Configure tab.

Step 5.    Under SnapCenter Plug-in for vSphere, select the Backups tab to view all the backups available for the VM.

Graphical user interface, text, application, emailDescription automatically generated

Step 6.    Navigate to Menu > SnapCenter Plug-in for VMware vSphere and select Dashboard to view recent job activity, backup jobs and configuration details.

Graphical user interface, websiteDescription automatically generated

Step 7.    Click Resource Groups and select a resource group. In the right pane, the completed backups are displayed.

Graphical user interfaceDescription automatically generated

Procedure 9.     Create On-Demand Backup

Step 1.    From the vCenter GUI, select Menu > SnapCenter Plugin for VMware vSphere.

Step 2.    Click Resource Groups.

Step 3.    Select a resource group and click Run Now to run the backup immediately.

Graphical user interface, text, applicationDescription automatically generated

Procedure 10.  Restore from vCenter by Using SnapCenter Plug-In

Note:   The SnapCenter Plug-in for VMware vSphere provides native backup, recovery, and cloning of virtualized applications.

Step 1.    Log into VMware vCenter.

Step 2.    Navigate to VMs and Templates, select a VM and right-click to access the context menu. Select NetApp SnapCenter > Restore.

Graphical user interface, text, applicationDescription automatically generated

Step 3.    Select a backup to restore. Click Next.

TableDescription automatically generated with medium confidence

Step 4.    From the Restore Scope drop-down list, select either Entire virtual machine to restore the virtual machine with all Virtual Machine Disks (VMDKs) or select Particular Virtual Disk to restore the VMDK without affecting the virtual machine configuration and other VMDKs.

Step 5.    Select the ESXi host that the VM should be restored to and check the box to restart the VM (if needed).

Step 6.    Click NEXT.

Graphical user interface, text, applicationDescription automatically generated

Step 7.    Select the destination datastore and click NEXT.

TableDescription automatically generated

Step 8.    Review the Summary and click FINISH to complete the restore process.

Graphical user interface, text, applicationDescription automatically generated

NetApp Active IQ Unified Manager 9.10

NetApp Active IQ Unified Manager enables customers to monitor and manage the health and performance of ONTAP storage systems and virtual infrastructure from a single interface. NetApp Active IQ Unified Manager provides a graphical interface that displays the capacity, availability, protection, and performance status of the monitored storage systems. NetApp Active IQ Unified Manager is required to integrate NetApp storage with Cisco Intersight.

This section describes the steps to deploy NetApp Active IQ Unified Manager 9.10 as a virtual appliance. The following table lists the recommended configuration for the virtual machine to install and run NetApp Active IQ Unified Manager to ensure acceptable performance.

Table 18.   Virtual Machine Configuration

Hardware Configuration

Recommended Settings

RAM

12 GB

Processors

4 CPUs/ vCPUs

CPU Cycle Capacity

9572 MHz total

Free Disk Space/virtual disk size

5 GB - Thin provisioned

 

152 GB - Thick provisioned

 

Note:   There is a limit to the number of nodes that a single instance of NetApp Active IQ Unified Manager can monitor before you need to install a second instance of NetApp Active IQ Unified Manager. See the Unified Manager Best Practices Guide (TR-4621) for more details.

Procedure 1.     Install NetApp Active IQ Unified Manager 9.10 using Ansible

Step 1.    Follow the Pre-Requisites from https://github.com/NetApp-Automation/NetApp-AIQUM

Step 2.    Download ansible git: git clone https://github.com/NetApp-Automation/NetApp-AIQUM.git

Step 3.    To invoke the ansible scripts use the following command:

ansible-playbook aiqum.yml -t aiqum_setup

Procedure 2.     Install NetApp Active IQ Unified Manager 9.10 Manually

Step 1.    Download NetApp Active IQ Unified Manager for VMware vSphere OVA file from:

Step 2.    https://mysupport.netapp.com/site/products/all/details/activeiq-unified-manager/downloads-tab.

Step 3.    In the VMware vCenter GUI, click VMs and Templates and then click Actions> Deploy OVF Template.

Step 4.    Specify the location of the OVF Template and click NEXT.

Step 5.    On the Select a name and folder page, enter a unique name for the VM, and select a deployment location, and then click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 6.    On the Select a compute resource screen, select the cluster where VM will be deployed and click NEXT.

Step 7.    On the Review details page, verify the OVA template details and click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 8.    On the License agreements page, read and check the box for I accept all license agreements. Click NEXT.

Step 9.    On the Select storage page, select following parameters for the VM deployment:

a.     Select the disk format for the VMDKs (for example, Think Provisioning).

b.     Select a VM Storage Policy (for example, Datastore Default).

c.     Select a datastore to store the deployed OVA template (for example, infra_datastore_2).

Step 10.                       Click NEXT.

TableDescription automatically generated

Step 11.                       On the Select networks page, select the destination network (for example, IB-MGMT) and click NEXT.

Step 12.                       On the Customize template page, provide network details such as hostname, IP address, gateway, and DNS.

Step 13.                       Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Note:   Scroll through the customization template to ensure all required values are entered.

Step 14.                       On the Ready to complete page, review the settings and click FINISH. Wait for the VM deployment to complete before proceeding to the next step.

Graphical user interface, application, emailDescription automatically generated

Step 15.                       Select the newly created NetApp Active IQ Unified Manager VM, right-click and select Power > Power On.

Step 16.                       While the virtual machine is powering on, click the prompt in the yellow banner to Install VMware tools.

Graphical user interface, text, application, emailDescription automatically generated

Note:   Because of timing, VMware tools might not install correctly. In that case VMware tools can be manually installed after NetApp Active IQ Unified Manager VM is up and running.

Step 17.                       Open the VM console for the NetApp Active IQ Unified Manager VM and configure the time zone information when displayed.

TextDescription automatically generated

Step 18.                       Create a maintenance user account when prompted by specifying a user account name and password.

Note:   Save the maintenance user account credentials in a secure location. These credentials will be used for the initial GUI login and to make any configuration changes to the appliance settings in future.

Note:   If the systems complaints about Network information not valid, enter the values for hostname, IP address and DNS information when prompted.

A screenshot of a cell phoneDescription automatically generated

Step 19.                       Log into NetApp Active IQ Unified Manager using the IP address or URL displayed on the deployment screen using the maintenance user credentials created in the last step.

TextDescription automatically generated

Procedure 3.     Configure NetApp Active IQ Unified Manager

Step 1.    Launch a web browser and log into NetApp Active IQ Unified Manager.

Graphical user interface, applicationDescription automatically generated

Step 2.    Enter the email address that Unified Manager will use to send alerts, enter the mail server configuration, and the IP address or hostname of the NTP server. Click Continue and complete the AutoSupport configuration.

Graphical user interface, text, application, emailDescription automatically generated

Step 3.    Configure AutoSupport for Unified Manager by clicking Agree and Continue.

Graphical user interface, application, timeline, TeamsDescription automatically generated

Step 4.    Check the box for Enable API Gateway and click Continue to setup the API gateway for NetApp Active IQ Unified Manager.

Graphical user interface, application, timeline, TeamsDescription automatically generated

Step 5.    Enter the ONTAP cluster hostname or IP address and the admin login credentials then click Add.

Graphical user interface, timelineDescription automatically generated with medium confidence

Step 6.    A security prompt will be displayed to authorize the cluster certificate. Click Yes to trust the certificate.

Graphical user interface, text, application, emailDescription automatically generated

Step 7.    When prompted to trust the self-signed certificate from Active IQ Unified Manager, click Yes to finish and add the storage system.

Note:   The initial discovery process can take up to 15 minutes to complete.

Procedure 4.     Add Local Users to NetApp Active IQ Unified Manager

Step 1.    Navigate to the General section and click Users.

Graphical user interface, applicationDescription automatically generated

Step 2.    Click Add and complete the requested information:

a.     Select Local User for the Type.

b.     Enter a username and password.

c.     Add the user’s email address.

d.     Select the appropriate role for the new user.

Step 3.    Click Save to add the new user to NetApp Active IQ Unified Manager.

Graphical user interface, text, application, emailDescription automatically generated

Procedure 5.     Configure Remote Authentication

Simplify user management and authentication for Active IQ Unified Manager by integrating it with Microsoft Active Directory. Connect the Active IQ Unified Manager to Active Directory and perform user authentication with the Active Directory domain.

Note:   You must be logged on as the maintenance user created during the installation or another user with Application Administrator privileges to configure remote authentication.

Step 1.    Navigate to the General section and click Remote Authentication.

Step 2.    Select the option to Enable remote authentication and define a remote user or remote group.

Graphical user interface, applicationDescription automatically generated

Step 3.    Select Active Directory from the authentication service list.

Step 4.    Enter the Active Directory service account name and password. The account name can be in the format of domain\user or user@domain.

Step 5.    Enter the base DN where your Active Directory users reside.

Step 6.    If Active Directory LDAP communications are protected via SSL enable the Use Secure Connection option.

Step 7.    Add one or more Active Directory domain controllers by clicking Add and entering the IP or FQDN of the domain controller.

Step 8.    Click Save to enable the configuration.

Note:   If you don’t know the base DN to your Active Directory user organizational unit, contact the Active Directory administrator at your organization to provide this information.

Related image, diagram or screenshot

Step 9.    Click Test Authentication and enter an Active Directory username and password to test authentication with the Active Directory authentication servers.

Graphical user interface, application, WordDescription automatically generated

A result message displays indicating authentication was successful:

Graphical user interface, text, application, emailDescription automatically generated

Procedure 6.     Add a Remote User to NetApp Active IQ Unified Manager

Step 1.    Navigate to the General section and click Users.

Step 2.    Click Add and click Remote User from the Type drop-down list.

Graphical user interface, applicationDescription automatically generated

Step 3.    Enter the following information into the form:

a.     The username of the Active Directory user.

b.     Email address of the user.

c.     Select the appropriate role for the user

Step 4.    Click Save when finished to add the remote user to NetApp Active IQ Unified Manager.

Graphical user interfaceDescription automatically generated with low confidence

Procedure 7.     Add the vCenter Server to NetApp Active IQ Unified Manager

NetApp Active IQ Unified Manager provides visibility into vCenter and the virtual machines running inside the datastores backed by ONTAP storage. Virtual machines and storage are monitored to enable fast identification of performance issues within the various components of the virtual infrastructure stack.

Note:   Before adding vCenter into NetApp Active IQ Unified Manager, the log level of the vCenter server must be changed.

Step 1.    In the vSphere client navigate to VMs and Templates and select the vCenter instance from the top of the object tree.

Step 2.    Click the Configure tab, expand the settings, and click General.

Graphical user interface, application, TeamsDescription automatically generated

Step 3.    Click EDIT.

Step 4.    In the pop-up window under Statistics, locate the 5 minutes Interval Duration row and change the setting to Level 3 under the Statistics Level column. Click SAVE.

Graphical user interfaceDescription automatically generated

Step 5.    Return to Active IQ Unified Manager and navigate to the VMware section located under Inventory.

Related image, diagram or screenshot

Step 6.    Expand the section and select vCenter and click Add.

Step 7.    Enter the VMware vCenter server details and click Save.

Graphical user interface, applicationDescription automatically generated

Step 8.    A dialog box will appear asking to authorize the certificate. Click Yes to accept the certificate and add the vCenter server.

Graphical user interface, text, application, email, TeamsDescription automatically generated

Note:   It may take up to 15 minutes to discover the vCenter server. Performance data can take up to an hour after discovery to become available.

Procedure 8.     View Virtual Machine Inventory

Note:   The virtual machine inventory is automatically added to NetApp Active IQ Unified Manager during discovery of the vCenter server. Virtual machines can be viewed in a hierarchical display detailing storage capacity, IOPS and latency for each component in the virtual infrastructure to troubleshoot the source of any performance related issues.

Step 1.    Navigate to the VMware section located under Inventory, expand the section, and click Virtual Machines.

Graphical user interface, applicationDescription automatically generated

Step 2.    Select a VM and click the blue caret to expose the topology view. Review the compute, network, and storage components and their associated IOPS and latency statistics.

A screenshot of a computerDescription automatically generated

Step 3.    Click Expand Topology to see the entire hierarchy of the virtual machine and its virtual disks as it is connected through the virtual infrastructure stack. The VM components are mapped from vSphere and compute through the network to the storage.

Graphical user interface, diagramDescription automatically generated

Procedure 9.     Review Security Compliance with NetApp Active IQ Unified Manager

NetApp Active IQ Unified Manager identifies issues and makes recommendations to improve the security posture of ONTAP. NetApp Active IQ Unified Manager evaluates ONTAP storage based on recommendations made in the Security Hardening Guide for ONTAP 9. Items are identified according to their level of compliance with the recommendations. All events identified do not inherently apply to all environments, for example, FIPS compliance. Review the Security Hardening Guide for NetApp ONTAP 9 (TR-4569) for additional information and recommendations for securing ONTAP 9.

A screenshot of a computerDescription automatically generated

Step 1.    Navigate to the URL of the Active IQ Unified Manager installation and login.

Step 2.    Select the Dashboard from the left menu bar in Active IQ Unified Manager.

Step 3.    Locate the Security card and note the compliance level of the cluster and SVM. Click the blue arrow to expand the findings.

Graphical user interface, applicationDescription automatically generated

Step 4.    Locate Individual Cluster section and the Cluster Compliance card. From the drop-down list select View All.

Graphical user interface, applicationDescription automatically generated

Step 5.    Select an event from the list and click the name of the event to view the remediation steps.

Graphical user interfaceDescription automatically generated

Step 6.    Remediate the risk if desired and perform the suggested actions to fix the issue.

Graphical user interface, text, application, emailDescription automatically generated

Remediate Security Compliance Findings

NetApp Active IQ identifies several security compliance risks after installation that can be immediately corrected to improve the security posture of ONTAP.

Procedure 1.     Correct Cluster Risks

Step 1.    To associate an NTP server with the cluster, run the ONTAP command:

cluster time-service ntp server create -server <ntp server host name or ip address>

Step 2.    Enable the login banner on the cluster:

security login banner modify -vserver <clustername> -message "Access restricted to authorized users"

NetApp Active IQ

NetApp Active IQ is a data-driven service that leverages artificial intelligence and machine learning to provide analytics and actionable intelligence for ONTAP storage systems. NetApp Active IQ uses AutoSupport data to deliver proactive guidance and best practices recommendations to optimize storage performance and minimize risk.

Additional NetApp Active IQ documentation is available on the Active IQ Documentation Resources web page.

Procedure 1.     Install NetApp Active IQ

Note:   NetApp Active IQ is automatically enabled when you configure AutoSupport on the ONTAP storage controllers.

Step 1.    Obtain the controller serial numbers from your ONTAP system with the following command:

system node show -fields serialnumber

Step 2.    Navigate to the Active IQ portal at https://activeiq.netapp.com/

Step 3.    Login with you NetApp support account ID.

Step 4.    At the welcome screen enter the cluster name or one of controller serial numbers in the search box. NetApp Active IQ will automatically begin searching for the cluster and display results below.

Graphical user interface, applicationDescription automatically generated

Step 5.    Select the cluster name to launch the main dashboard.

Step 6.    From the drop-down list select Set Page as Default Landing View.

Graphical user interface, applicationDescription automatically generated

Procedure 2.     Add a Watchlist to the Discovery Dashboard

Note:   The system level dashboard is the default view for systems in NetApp Active IQ.

Step 1.    Click Discovery Dashboard in the toolbar at the top of the NetApp Active IQ screen.

TableDescription automatically generated with medium confidence

Step 2.    Click Create Watchlist and enter a name for the watchlist.

Step 3.    Select the radio button to add systems by serial number and enter the cluster serial numbers to the watchlist.

Step 4.    Check the box for Make this my default watchlist if desired and click Create Watchlist.

Graphical user interface, text, application, emailDescription automatically generated

Step 5.    Click Manage watchlists and then click the ellipsis on the cluster watchlist card you created and click View in Discovery Dashboard.

Graphical user interface, applicationDescription automatically generated

Step 6.    View the health and risk overview for the cluster.

Procedure 3.     Create Active IQ Digital Advisor Dashboard

The Active IQ Digital advisor provides a summary dashboard and system wellness score based on the health and risks that Active IQ have identified. The dashboard provides a quick way to identify and get proactive recommendations on how to mitigate risks in the storage environment including links to technical reports and mitigation plans.

Step 1.    At the cluster dashboard, click Active IQ Digital Advisor from the top menu.

Related image, diagram or screenshot

Step 2.    Select the watchlist created in the previous step and click Next.

Graphical user interface, text, applicationDescription automatically generated

Step 3.    Accept the dashboard default name and select all the available widgets.

Step 4.    Check the box Make this the default dashboard and click Create.

Graphical user interface, text, application, emailDescription automatically generated

Step 5.    Review the enhanced dashboard including the Wellness Score and any recommended actions or risks.

Graphical user interface, applicationDescription automatically generated

Step 6.    Switch between the Actions and Risks tabs to view the risks broken down by category or a list of all risks with their impact and links to corrective actions.

Graphical user interface, applicationDescription automatically generated

Graphical user interface, text, application, emailDescription automatically generated

Step 7.    Click the link in the Corrective Action column to read the bug information or knowledge base article about how to remediate the risk.

Note:   Additional tutorials and video walk-throughs of Active IQ features can be viewed on the Active IQ documentation web page.

Deploy Cisco Intersight Assist Appliance

Cisco IntersightTM is a management platform delivered as a service with embedded analytics for your Cisco and third-party IT infrastructure. This platform offers an intelligent level of management that enables IT organizations to analyze, simplify, and automate their environments in more advanced ways than the prior generations of tools. Cisco Intersight provides an integrated and intuitive management experience for resources in the traditional data center and at the edge. With flexible deployment options to address complex security needs, getting started with Intersight is quick and easy.

Cisco Intersight offers flexible deployment either as Software as a Service (SaaS) on Intersight.com or running on your premises as Cisco Intersight Virtual Appliance. The virtual appliance provides the benefits of Cisco Intersight while allowing more flexibility for those with additional data locality and security requirements. The remainder of this section details Intersight deployment as SaaS on Intersight.com. To learn more about the virtual appliance, see the Cisco Intersight Virtual Appliance Getting Started Guide.

Procedure 1.     Configure Cisco Intersight

Step 1.    If you do not already have a Cisco Intersight account, to claim your Cisco UCS system into a new account on Cisco Intersight, connect to https://intersight.com. If you have an existing Intersight account, connect to https://intersight.com and sign in with your Cisco ID, select the appropriate account, and skip to step 6.

Step 2.    Click Create an account.

Step 3.    Sign in with your Cisco ID.

Step 4.    Read, scroll through, and accept the End User License Agreement and click Next.

Step 5.    Enter an Account Name and click Create.

Step 6.    Click ADMIN > Targets. Click Claim Target. Select Cisco UCS Domain (UCSM Managed) and click Start. Fill in the Device ID and Claim Code and click Claim. The Device ID and Claim Code can be obtained by connecting to Cisco UCS Manager and selecting Admin > All > Device Connector. The Device ID and Claim Code are on the right.

Step 7.    To claim your Cisco UCS system into an existing Intersight account, log into the account at https://Intersight.com. Click Administration > Devices. Click Claim a New Device. Under Direct Claim, fill in the Device ID and Claim Code and click Claim. The Device ID and Claim Code can be obtained by connecting to Cisco UCS Manager and selecting Admin > All > Device Connector. The Device ID and Claim Code are on the right.

Step 8.    From the Cisco Intersight window, click Related image, diagram or screenshot and then click Licensing. If this is a new account, all servers connected to the UCS Domain will appear under the Base license Tier. If you have purchased Cisco Intersight licenses and have them in your Cisco Smart Account, click Register and follow the prompts to register this Cisco Intersight account to your Cisco Smart Account. Cisco Intersight also offers a one-time 90-day trial of Premier licensing for new accounts. Click Start Trial and then Start to begin this evaluation. The remainder of this section will assume Premier licensing.

Step 9.    From the Licensing Window, click Set Default Tier. From the drop-down list select Premier for Tier and click Set.

Graphical user interface, application, TeamsDescription automatically generated

Step 10.                       Click Refresh to refresh the Intersight window with Premier, Advantage, and Essentials features added.

Step 11.                       Click Related image, diagram or screenshot in the Intersight window and click Guided Help > Site Tour. Follow the prompts for a tour of Cisco Intersight.

Step 12.                       The Essentials tier of Cisco Intersight includes a Cisco driver check against the Cisco Hardware Compatibility List (HCL). In the Servers list, select one of the servers in your VMware FlexPod-Management cluster by clicking the server name. Review the detailed General and Inventory information for the server. Click the HCL tab. Review the server information, the version of VMware ESXi, and the Cisco VIC driver versions.

A screenshot of a computerDescription automatically generated with medium confidence

Step 13.                       Using the Intersight Assist personality of the Cisco Intersight Virtual Appliance, VMware vCenter and NetApp Storage can be monitored (Advantage Licensing Tier) and configured (Premier Licensing Tier). To install Intersight Assist from an Open Virtual Appliance (OVA) in your VMware FlexPod-Management Cluster, first download the latest release of the Cisco Intersight Virtual Appliance for vSphere OVA from https://software.cisco.com/download/home/286319499/type/286323047/release/1.0.9-342.

Refer to: https://www.cisco.com/c/en/us/td/docs/unified_computing/Intersight/b_Cisco_Intersight_Appliance_Getting_Started_Guide/b_Cisco_Intersight_Appliance_Install_and_Upgrade_Guide_chapter_00.html and set up the DNS entries for the Intersight Assist hostname as specified under Before you begin.

Step 14.                       From Hosts and Clusters in the VMware vCenter HTML5 client, right-click the FlexPod-Management cluster and click Deploy OVF Template.

Step 15.                       Specify a URL or either browse to the intersight-appliance-installer-vsphere-1.0.9-432.ova or latest release file. Click NEXT.

Graphical user interface, text, application, emailDescription automatically generated

Step 16.                       Name the Intersight Assist VM and select the location. Click NEXT.

Step 17.                       Select the FlexPod-Management cluster and click NEXT.

Step 18.                       Review details, click Ignore, and click NEXT.

Step 19.                       Select a deployment configuration (Tiny for just Intersight Assist, Small for Intersight Assist and IWO) and click NEXT.

Graphical user interface, applicationDescription automatically generated

Step 20.                       Select infra_datastore01 for storage and select the Thin Provision virtual disk format. Click NEXT.

Step 21.                       Select IB-MGMT Network for the VM Network. Click NEXT.

Step 22.                       Fill in all values to customize the template. Click NEXT.

Step 23.