Explore Cisco
How to Buy

Have an account?

  •   Personalized content
  •   Your products and support

Need an account?

Create an account

FlexPod Datacenter with Oracle 19c RAC Databases on Cisco UCS and NetApp AFF with NVMe over FibreChannel

Available Languages

Download Options

  • PDF
    (16.8 MB)
    View with Adobe Reader on a variety of devices
  • ePub
    (19.7 MB)
    View in various apps on iPhone, iPad, Android, Sony Reader, or Windows Phone
  • Mobi (Kindle)
    (9.9 MB)
    View on Kindle device or Kindle app on multiple devices

Bias-Free Language

The documentation set for this product strives to use bias-free language. For the purposes of this documentation set, bias-free is defined as language that does not imply discrimination based on age, disability, gender, racial identity, ethnic identity, sexual orientation, socioeconomic status, and intersectionality. Exceptions may be present in the documentation due to language that is hardcoded in the user interfaces of the product software, language used based on RFP documentation, or language that is used by a referenced third-party product. Learn more about how Cisco is using Inclusive Language.

Available Languages

Download Options

  • PDF
    (16.8 MB)
    View with Adobe Reader on a variety of devices
  • ePub
    (19.7 MB)
    View in various apps on iPhone, iPad, Android, Sony Reader, or Windows Phone
  • Mobi (Kindle)
    (9.9 MB)
    View on Kindle device or Kindle app on multiple devices

Table of Contents

 

Related image, diagram or screenshot

FlexPod Datacenter with Oracle 19c RAC Databases on Cisco UCS and NetApp AFF with NVMe over FibreChannel

Deployment Guide for Oracle 19c RAC Databases on Cisco Unified Computing System and NetApp AFF A800 Storage using Modern SANs on NVMe/FC NOTE: Works with document’s Advanced Properties “First Published” property. Click File | Properties | Advanced Properties | Custom.

Published: April 2021

Related image, diagram or screenshot

In partnership with

Related image, diagram or screenshot

About the Cisco Validated Design Program

The Cisco Validated Design (CVD) program consists of systems and solutions designed, tested, and documented to facilitate faster, more reliable, and more predictable customer deployments. For more information, go to:

http://www.cisco.com/go/designzone.

ALL DESIGNS, SPECIFICATIONS, STATEMENTS, INFORMATION, AND RECOMMENDATIONS (COLLECTIVELY, "DESIGNS") IN THIS MANUAL ARE PRESENTED "AS IS," WITH ALL FAULTS. CISCO AND ITS SUPPLIERS DISCLAIM ALL WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OR ARISING FROM A COURSE OF DEALING, USAGE, OR TRADE PRACTICE. IN NO EVENT SHALL CISCO OR ITS SUPPLIERS BE LIABLE FOR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, OR INCIDENTAL DAMAGES, INCLUDING, WITHOUT LIMITATION, LOST PROFITS OR LOSS OR DAMAGE TO DATA ARISING OUT OF THE USE OR INABILITY TO USE THE DESIGNS, EVEN IF CISCO OR ITS SUPPLIERS HAVE BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.

THE DESIGNS ARE SUBJECT TO CHANGE WITHOUT NOTICE. USERS ARE SOLELY RESPONSIBLE FOR THEIR APPLICATION OF THE DESIGNS. THE DESIGNS DO NOT CONSTITUTE THE TECHNICAL OR OTHER PROFESSIONAL ADVICE OF CISCO, ITS SUPPLIERS OR PARTNERS. USERS SHOULD CONSULT THEIR OWN TECHNICAL ADVISORS BEFORE IMPLEMENTING THE DESIGNS. RESULTS MAY VARY DEPENDING ON FACTORS NOT TESTED BY CISCO.

CCDE, CCENT, Cisco Eos, Cisco Lumin, Cisco Nexus, Cisco StadiumVision, Cisco TelePresence, Cisco WebEx, the Cisco logo, DCE, and Welcome to the Human Network are trademarks; Changing the Way We Work, Live, Play, and Learn and Cisco Store are service marks; and Access Registrar, Aironet, AsyncOS, Bringing the Meeting To You, Catalyst, CCDA, CCDP, CCIE, CCIP, CCNA, CCNP, CCSP, CCVP, Cisco, the Cisco Certified Internetwork Expert logo, Cisco IOS, Cisco Press, Cisco Systems, Cisco Systems Capital, the Cisco Systems logo, Cisco Unified Computing System (Cisco UCS), Cisco UCS B-Series Blade Servers, Cisco UCS C-Series Rack Servers, Cisco UCS S-Series Storage Servers, Cisco UCS Manager, Cisco UCS Management Software, Cisco Unified Fabric, Cisco Application Centric Infrastructure, Cisco Nexus 9000 Series, Cisco Nexus 7000 Series. Cisco Prime Data Center Network Manager, Cisco NX-OS Software, Cisco MDS Series, Cisco Unity, Collaboration Without Limitation, EtherFast, EtherSwitch, Event Center, Fast Step, Follow Me Browsing, FormShare, GigaDrive, HomeLink, Internet Quotient, IOS, iPhone, iQuick Study,  LightStream, Linksys, MediaTone, MeetingPlace, MeetingPlace Chime Sound, MGX, Networkers, Networking Academy, Network Registrar, PCNow, PIX, PowerPanels, ProConnect, ScriptShare, SenderBase, SMARTnet, Spectrum Expert, StackWise, The Fastest Way to Increase Your Internet Quotient, TransPath, WebEx, and the WebEx logo are registered trademarks of Cisco Systems, Inc. and/or its affiliates in the United States and certain other countries. (LDW6)

All other trademarks mentioned in this document or website are the property of their respective owners. The use of the word partner does not imply a partnership relationship between Cisco and any other company. (0809R)

© 2021 Cisco Systems, Inc. All rights reserved.

 

Executive Summary

The Cisco Unified Computing System™ (Cisco UCS®) is a next-generation data center platform that unites computing, network, storage access, and virtualization into a single cohesive system. Cisco UCS is an ideal platform for the architecture of mission critical database workloads such as Oracle RAC. The combination of Cisco UCS, NetApp and Oracle Real Application Cluster Database architecture can accelerate your IT transformation by enabling faster deployments, greater flexibility of choice, efficiency, high availability, and lower risk.

Cisco and NetApp have partnered to deliver FlexPod, which serves as the foundation for a variety of workloads and enables efficient architectural designs that are based on customer requirements. The FlexPod Datacenter with Cisco UCS and NetApp All Flash AFF system is a converged infrastructure platform that combines best-of breed technologies from Cisco and NetApp into a powerful converged platform for enterprise applications such as Oracle. Cisco and NetApp work closely with Oracle to support the most demanding transactional and response-time-sensitive databases required by today’s businesses.

Cisco Validated Designs (CVDs) consist of systems and solutions that are designed, tested, and documented to facilitate and improve customer deployments. This CVD document describes the Cisco and NetApp® FlexPod® solution, which is a validated approach for deploying highly available Oracle RAC Database environment. Cisco and NetApp have validated the reference architecture with various Database workloads like OLTP (Online Transactional Processing) and Data Warehouse in Cisco’s UCS Datacenter lab. This document shows the hardware and software configuration of the components involved, results of various tests and offers implementation and a framework for implementing Oracle RAC Databases on NVMe/FC using Cisco UCS and NetApp Storage System.

Solution Overview

Introduction

This Cisco Validated Design (CVD) describes how the Cisco Unified Computing System™ (Cisco UCS®) can be used in conjunction with NetApp® AFF Storage A800 System to implement a mission-critical application such as an Oracle Multitenant Real Application Cluster (RAC) 19c Database solution using modern SANs on NVMe over Fabrics (NVMe over Fibre-Channel or NVMe/FC). 

Digital transformations are driving an increased number of new applications, with more sources of data. Organizations of all kinds rely on their relational databases for both transaction processing (OLTP) and analytics (OLAP), but many still have challenges in meeting their goals of high availability, security, and performance. Applications must be able to move quickly from development to a reliable, scalable platform. An ideal solution integrates best-in-class components that can scale compute and storage independently to meet the needs of dynamic business requirements.

FlexPod Datacenter with NetApp All Flash AFF is comprised of compute (database, application, and management servers from Cisco), network (three-layer network and SAN technologies from Cisco), and storage (NetApp All Flash AFF storage systems). This CVD documents validation of the real-world performance, ease of management, and agility of the FlexPod Datacenter with Cisco UCS and All Flash AFF in high-performance Oracle RAC Databases environments using NVMe over Fibre-Channel (NVMe/FC).

Audience

The intended audience for this document includes, but is not limited to, sales engineers, field consultants, database administrators, IT managers, oracle database architects, and customers who want to deploy Oracle RAC 19c database solution on FlexPod Converged Infrastructure with NetApp clustered Data ONTAP® and the Cisco UCS platform. A working knowledge of Oracle RAC Database, Linux, Storage technology, and Network is assumed but is not a prerequisite to read this document.

Purpose of this Document

This FlexPod solution for Oracle databases delivers industry-leading storage, unprecedented scalability, continuous data access, and automated data management for immediate responses to business opportunities. The goal of this document is to determine the Oracle database server read latency, peak sustained throughput and IOPS of this FlexPod reference architecture system while running the Oracle OLTP and OLAP workloads.

This document provides a step-by-step configuration and implementation guide for the FlexPod Datacenter with Cisco UCS Compute Servers, Cisco Fabric Interconnect Switches, Cisco MDS Switches, Cisco Nexus Switches and NetApp AFF Storage to deploy an Oracle RAC Database solution. The following are the objectives of this reference document:

   Provide reference FlexPod architecture design guidelines for the Oracle RAC Databases solution

   Demonstrate simplicity and agility with the software-driven architecture and high performance of Cisco UCS compute Servers

   Build, validate, and predict performance of Servers, Network and Storage platform on various types of workload

What’s New in the Release?

This version of the FlexPod CVD introduces the NetApp Storage AFF A800 that brings the low latency and high performance of NVMe technology to the storage network along with Cisco UCS B200 M5 Blade Servers 5th Generation to deploy Oracle RAC Database Release 19c using modern SANs on NVMe over Fabrics (NVMe over FibreChannel or NVMe/FC)

It incorporates the following features:

   Support for the NVMe/FC on Cisco UCS and NetApp Storage

   Implementation of FC and NVMe/FC on the same architecture

   Validation of Oracle RAC 19c Container and Non-Container Database deployments

   Support for the Cisco UCS Infrastructure and UCS Manager Software Release 4.1(3b)

   Support for the release of NetApp ONTAP® 9.7

Solution Summary

   Nonvolatile Memory Express (NVMe) is an optimized, high-performance, scalable interface designed to work with current and the next-generation NVM technologies. The NVMe interface is defined to enable host software to communicate with nonvolatile memory over PCI Express (PCIe). It was designed from the ground up for low-latency solid state media, eliminating many of the bottlenecks seen in the legacy protocols for running enterprise applications. NVMe devices are connected to the PCIe bus inside a server. NVMe-oF extends the high-performance and low-latency benefits of NVMe across network fabrics that connect servers and storage. NVMe-oF takes the lightweight and streamlined NVMe command set, and the more efficient queueing model, and replaces the PCIe transport with alternate transports, like Fibre Channel, RDMA over Converged Ethernet (RoCE v2), TCP.

   NVMe over Fibre Channel (NVMe/FC) is implemented through the Fibre Channel NVMe (FC-NVMe) standard which is designed to enable NVMe based message commands to transfer data and status information between a host computer and a target storage subsystem over a Fibre Channel network fabric. FC-NVMe simplifies the NVMe command sets into basic FCP instructions. Because Fibre Channel is designed for storage traffic, functionality such as discovery, management and end-to-end qualification of equipment is built into the system.

   Almost all high-performance latency sensitive applications and workloads are running on FCP today. Because NVMe/FC and Fibre Channel networks use the same underlying transport protocol (FCP), they can use common hardware components. It’s even possible to use the same switches, cables, and ONTAP target port to communicate with both protocols at the same time. The ability to use either protocol by itself or both at the same time on the same hardware makes transitioning from FCP to NVMe/FC both simple and seamless.

   Large-scale block flash-based storage environments that use Fibre Channel are the most likely to adopt NVMe over FC. FC-NVMe offers the same structure, predictability, and reliability characteristics for NVMe-oF that Fibre Channel does for SCSI. Plus, NVMe-oF traffic and traditional SCSI-based traffic can run simultaneously on the same FC fabric.

   In this FlexPod solution, we will showcase Cisco UCS System with NetApp AFF Storage Array running on NVMe over FibreChannel (NVMe/FC) which can provide efficiency and performance of NVMe, and the benefits of all-flash robust scale out storage system that combines low-latency performance with comprehensive data management, built-in efficiencies, integrated data protection, multiprotocol support, and nondisruptive operations.

FlexPod System Overview

Built on groundbreaking technology from NetApp and Cisco, the FlexPod converged infrastructure platform meets and exceeds the challenges of simplifying deployments for best-in-class data center infrastructure. FlexPod is a defined set of hardware and software that serves as an integrated foundation for both virtualized and non-virtualized solutions. Composed of pre-validated storage, networking, and server technologies, FlexPod is designed to increase IT responsiveness to organizational needs and reduce the cost of computing with maximum uptime and minimal risk. Simplifying the delivery of data center platforms gives enterprises an advantage in delivering new services and applications.

FlexPod provides the following differentiators:

   Flexible design with a broad range of reference architectures and validated designs

   Elimination of costly, disruptive downtime through Cisco UCS and NetApp® ONTAP®

   Leverage a pre-validated platform to minimize business disruption and improve IT agility and reduce deployment time from months to weeks

   Cisco Validated Designs (CVDs) and NetApp Validated Architectures (NVAs) covering a variety of use cases

Cisco and NetApp have carefully validated and verified the FlexPod solution architecture and its many use cases while creating a portfolio of detailed documentation, information, and references to assist customers in transforming their data centers to this shared infrastructure model.

Figure 1.       FlexPod System Overview

A picture containing text, computer, screenshotDescription automatically generated

FlexPod datacenter architecture includes three components:

   Cisco Unified Computing System (UCS)

   Cisco MDS and Nexus Switches

   NetApp AFF Storage Systems

A benefit of the FlexPod architecture is the ability to customize or "flex" the environment to suit a customer's requirements. A FlexPod can easily be scaled as requirements and demand change. The unit can be scaled both up (adding resources to a FlexPod unit) and out (adding more FlexPod units). This document highlights the resiliency, cost benefit, and ease of deployment of a Fibre Channel storage solution to deploy Oracle RAC Database environments on FlexPod Infrastructure.

Solution Deployment

This solution provides an end-to-end architecture with Cisco Unified Computing System (UCS), Oracle, and NetApp technologies to demonstrates the benefits for running Oracle Multitenant RAC Databases 19c environment with excellent performance, scalability and high availability using FC and NVMe/FC.

The reference architecture covered in this document is built on the NetApp All Flash AFF A800 for Storage, Cisco B200 M5 Blade Servers for Compute, Cisco Nexus 9336C-FX2 Switches, Cisco MDS 9132T Fibre Channel Switches and Cisco Fabric Interconnects 6454 Fabric Interconnects for System Management in a single package. The design is flexible enough that the networking, computing, and storage can fit in one data center rack or be deployed according to a customer's data center design. The reference architecture reinforces the "wire-once" strategy, because as additional storage is added to the architecture, no re-cabling is required from the hosts to the Cisco UCS fabric interconnect.

The processing capabilities of CPUs have increased much faster than the processing demands of most database workloads. Sometimes databases are limited by CPU work, but it is generally a result of the processing limits of a single core and is not a limitation of the CPU. The result is an increasing number of idle cores on database servers that still must be licensed for the Oracle Database software. This underutilization of CPU resources is a waste of capital expenditure, not only in terms of licensing costs, but also in terms of the cost of the server itself, heat output, and so on.  The Cisco UCS servers comes with different CPU options in terms of higher clock-speed which would help the database workloads and customer will have the option of using Higher clock-speed CPU with lower cores to keep the Oracle licensing costs down.

Physical Topology

This solution consists of the following set of hardware combined into a single stack:

   Compute: Cisco UCS B200 M5 Blade Servers with Cisco Virtual Interface Cards (VICs) 1440

   Network: Cisco Nexus 9336C-FX2, Cisco MDS 9132T Fibre Channel and Cisco UCS Fabric Interconnect 6454 for network and management connectivity

   Storage: NetApp AFF A800 Storage

In this solution design, we have deployed two Cisco UCS 5108 Blade Server Chassis with 8 identical Intel Xeon CPU based Cisco UCS B200 M5 Blade Servers for hosting the 8-Node Oracle RAC Databases. The Cisco UCS B200 M5 Server has Virtual Interface Card (VIC) 1440 with port expander and they were connected to eight ports from each Cisco Fabric extender 2408 of the Cisco UCS Chassis to the Cisco Fabric Interconnects, which were in turn connected to the Cisco MDS Switches for upstream SAN connectivity to access the NetApp AFF storage.

Figure 2 shows the architecture diagram of the FlexPod components to deploy an eight node Oracle RAC 19c Database solution. This reference design is a typical network configuration that can be deployed in a customer's environments. The best practices and setup recommendations are described later in this document.

Figure 2.       FlexPod Architecture Design

DiagramDescription automatically generated

As shown in Figure 2, a pair of Cisco UCS 6454 Fabric Interconnects (FI) carries both storage and network traffic from the Cisco UCS B200 M5 server with the help of Cisco Nexus 9336C-FX2 and Cisco MDS 9132T switches. Both the Fabric Interconnects and the Cisco Nexus switches are clustered with the peer link between them to provide high availability.

As illustrated in Figure 2, 16 (8 x 25G link per chassis) links from the blade server chassis go to Fabric Interconnect – A. Similarly, 16 (8 x 25G link per chassis) links from the blade server chassis go to Fabric Interconnect – B. Fabric Interconnect – A links are used for Oracle Public Network Traffic (VLAN-134) shown as green lines while Fabric Interconnect – B links are used for Oracle Private Interconnect Traffic (VLAN 10) shown as red lines. Two virtual Port-Channels (vPCs) are configured to provide public network and private network traffic paths for the server blades to northbound nexus switches.

FC and NVMe/FC Storage access from both Fabric Interconnects to MDS Switches and NetApp Storage Array are shown as orange lines. Four 32Gb links are connected from FI – A to MDS – A Switch. Similarly, four 32Gb links are connected from FI – B to MDS – B Switch. The NetApp Storage AFF A800 have eight active FC connection goes to the Cisco MDS Switches. Four FC ports are connected to MDS-A, and other four FC ports are connected to MDS-B Switch. The NetApp Controller CT1 and Controller CT2 SAN ports 2a and 2b are connected to MDS – A Switch while the Controller CT1 and Controller CT2 SAN ports 2c and 2d are connected to MDS – B Switch. Also, two FC Port-Channels (PC) are configured to provide storage network paths from the server blades to storage array. Each PC has VSANs created for application and storage network data access.

*       For Oracle RAC configuration on Cisco Unified Computing System, we recommend keeping all private interconnects network traffic local on a single Fabric interconnect. In such a case, the private traffic will stay local to that fabric interconnect and will not be routed via northbound network switch. In that way, all the inter server blade (or RAC node private) communications will be resolved locally at the fabric interconnects and this significantly reduces latency for Oracle Cache Fusion traffic.

Additional 1Gb management connections will be needed for an out-of-band network switch that sits apart from this FlexPod infrastructure. Each UCS FI, MDS and Nexus switch is connected to the out-of-band network switch, and each AFF controller also has two connections to the out-of-band network switch.

Although this is the base design, each of the components can be scaled easily to support specific business requirements. For example, more servers or even blade chassis can be deployed to increase compute capacity, additional disk shelves can be deployed to improve I/O capability and throughput, and special hardware or software features can be added to introduce new features. This document guides you through the detailed steps for deploying the base architecture, as shown in the above figure. These procedures cover everything from physical cabling to network, compute, and storage device configurations.

Design Topology

This section describes the hardware and software components used to deploy an eight node Oracle RAC 19c Databases Solution on this architecture.

The inventory of the components used in this solution architecture is as per the table below.

Table 1.          Hardware Inventory and Bill of Material

Name

Model/Product ID

Description

Quantity

Cisco UCS Blade Server Chassis

UCSB-5108-AC2

Cisco UCS AC Blade Server Chassis, 6U with Eight Blade Server Slots

2

Cisco UCS Fabric Extender

UCS-IOM-2408

Cisco UCS 2408 8x25 Gb Port IO Module

4

Cisco UCS B200 M5 Blade Server

UCSB-B200-M5

Cisco UCS B200 M5 2 Socket Blade Server

8

Cisco UCS VIC 1440

UCSB-MLOM-40G-04

Cisco UCS VIC 1440 Blade MLOM

8

Cisco UCS Port Expander Card

UCSB-MLOM-PT-01

Port Expander Card for Cisco UCS MLOM

8

Cisco UCS 6454 Fabric Interconnect

UCS-FI-6454

Cisco UCS 6454 Fabric Interconnect

2

Cisco Nexus Switch

N9K-9336C-FX2

Cisco Nexus 9336C-FX2 Switch

2

Cisco MDS Switch

DS-C9132T-8PMESK9

Cisco MDS 9132T 32-Gbps 32-Port Fibre Channel Switch

2

NetApp AFF Storage

AFF A800

NetApp AFF A-Series All Flash Arrays

1

In this solution design, we used 8 identical Cisco UCS B200 M5 Blade Servers for hosting an 8 Node Oracle RAC Databases. The Cisco UCS B200 M5 Server configuration is listed in Table 2.

Table 2.          Cisco UCS B200 M5 Blade Server

Cisco UCS B200 M5 Server Configuration

Processor

2 x Intel(R) Xeon(R) Gold 6248 2.50 GHz 150W 20C 27.50MB Cache DDR4 2933MHz 1TB

UCS-CPU-I6248

Memory

8 x Samsung 64GB DDR4-2933-MHz LRDIMM/4Rx4/1.2v

UCS-ML-X64G4RT-H

Cisco UCS VIC 1440

Cisco UCS VIC 1440 Blade MLOM

UCSB-MLOM-40G-04

Cisco UCS Port Expander Card

Port Expander Card for Cisco UCS MLOM

UCSB-MLOM-PT-01

Processor

2 x Intel(R) Xeon(R) Gold 6248 2.50 GHz 150W 20C 27.50MB Cache DDR4 2933MHz 1TB

UCS-CPU-I6248

In this solution, we configured two vNIC and 4 vHBA on each host to carry all the network and storage traffic.

Table 3.          vNIC and vHBA Configured on Each Linux Host

vNIC Details

vNIC (eth0)

Management and Public Network Traffic Interface for Oracle RAC. MTU = 1500

vNIC (eth1)

Private Server-to-Server Network (Cache Fusion) Traffic Interface for Oracle RAC. MTU = 9000

vHBA0

FC Network Traffic & Boot from SAN through MDS-A Switch

vHBA1

FC Network Traffic & Boot from SAN through MDS-B Switch

vHBA2

NVMe/FC Network Traffic (Oracle RAC Storage Traffic) through MDS-A Switch

vHBA3

NVMe/FC Network Traffic (Oracle RAC Storage Traffic) through MDS-B Switch

For this solution, we configured 2 VLANs to carry public and private network traffic as well as two VSANs to carry FC and NVMe/FC storage traffic as listed in Table 4.

Table 4.          VLAN and VSAN Configurations

VLAN and VSAN Configuration

VLAN

Name

ID

Description

Default VLAN

1

Native VLAN

Public VLAN

134

VLAN for Public Network Traffic

Private VLAN

10

VLAN for Private Network Traffic

VSAN

Name

ID

Description

VSAN-A

151

FC and NVMe/FC Network Traffic through for Fabric Interconnect A

VSAN-B

152

FC and NVMe/FC Network Traffic through for Fabric Interconnect B

This FlexPod solution consist of NetApp All Flash AFF Series Storage as described in Table 5.

Table 5.          NetApp AFF A800 Storage Configuration

Storage Components

Description

AFF Flash Array

NetApp All Flash AFF A800 Storage Array (24 x 1.75 TB NVMe SSD Drives)

Capacity

41.82 TB

Connectivity

8 x 32 Gb/s redundant FC, NVMe/FC

1 Gb/s redundant Ethernet (Management port)

Physical

4 Rack Units

For this FlexPod solution, we used the following versions of the software and firmware releases (Table 6).

Table 6.          Software and Firmware Revisions

Software and Firmware

Version

Cisco UCS Manager System

4.1(3b)

Cisco UCS Adapter VIC 1440

Package Version – 4.1 (3b)

Running Version – 5.1 (3a)

Cisco eNIC (Cisco VIC Ethernet NIC Driver)

(modinfo enic)

4.0.0.14-802.74

(kmod-enic-4.0.0.14-802.74.rhel8u2.x86_64.rpm)

Cisco fNIC (Cisco VIC FC HBA Driver)

(modinfo fnic)

2.0.0.69-178.0

(kmod-fnic-2.0.0.69-178.0.rhel8u2.x86_64.rpm)

Oracle Linux

Oracle Linux Release 8 Update 2 for x86 (64 bit)

Linux Kernel

Linux 4.18.0-193.el8.x86_64

Oracle Database 19c Grid Infrastructure for Linux x86-64

19.10.0.0.0

Oracle Database 19c Enterprise Edition for Linux x86-64

19.10.0.0.0

Cisco Nexus 9336C-FX2 NXOS

9.3(3)

Cisco MDS 9132T System

8.4(1)

NetApp Storage AFF A800

ONTAP 9.7P7

FIO

fio-3.7-3.el8.x86_64

Oracle Swingbench

2.5.971

SLOB

2.5.2.4

Solution Configuration

Figure 3 shows the high-level overview and steps for configuring various components to deploy and test the Oracle RAC Database 19c on FlexPod reference architecture.

Figure 3.       High-Level Overview

Graphical user interface, diagramDescription automatically generated

Cisco Nexus Switch Configuration

This section details the high-level steps to configure Cisco Nexus Switches as shown in Figure 4.

Figure 4.       Cisco Nexus Switch Configuration

Graphical user interface, applicationDescription automatically generated

The following procedures describe how to configure the Cisco Nexus switches for use in a base FlexPod environment. This procedure assumes you’re using Cisco Nexus 9336C-FX2 switches deployed with the 100Gb end-to-end topology.

Initial Setup

*       On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.

Cisco Nexus A Switch

To set up the initial configuration for the Cisco Nexus A switch on <nexus-A-hostname>, follow these steps:

Abort Power on Auto Provisioning and continue with normal setup? (yes/no) [n]: yes

Do you want to enforce secure password standard (yes/no) [y]: Enter

Enter the password for admin: <password>

Confirm the password for admin: <password>

Would you like to enter the basic configuration dialog (yes/no): yes

Create another login account (yes/no) [n]: Enter

Configure read-only SNMP community string (yes/no) [n]: Enter

Configure read-write SNMP community string (yes/no) [n]: Enter

Enter the switch name: <nexus-A-hostname>

Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: Enter

Mgmt0 IPv4 address: <nexus-A-mgmt0-ip>

Mgmt0 IPv4 netmask: <nexus-A-mgmt0-netmask>

Configure the default gateway? (yes/no) [y]: Enter

IPv4 address of the default gateway: <nexus-A-mgmt0-gw>

Configure advanced IP options? (yes/no) [n]: Enter

Enable the telnet service? (yes/no) [n]: Enter

Enable the ssh service? (yes/no) [y]: Enter

Type of ssh key you would like to generate (dsa/rsa) [rsa]: Enter

Number of rsa key bits <1024-2048> [1024]: Enter

Configure the ntp server? (yes/no) [n]: y

NTP server IPv4 address: <global-ntp-server-ip>

Configure default interface layer (L3/L2) [L3]: L2

Configure default switchport interface state (shut/noshut) [noshut]: Enter

Configure CoPP system profile (strict/moderate/lenient/dense/skip) [strict]: Enter

Would you like to edit the configuration? (yes/no) [n]: Enter

Cisco Nexus B Switch

Similarly, follow the steps from section Cisco Nexus B Switch to setup the initial configuration for the Cisco Nexus B Switch and change the relevant switch hostname and management IP address.

Configure Global Settings

To set global configuration, follow these steps on both the nexus switches:

1.    Login as admin user into the Nexus Switch A and run the following commands to set global configurations on Switch A:

configure terminal

feature interface-vlan

feature hsrp

feature lacp

feature vpc

feature udld

spanning-tree port type network default

spanning-tree port type edge bpduguard default

port-channel load-balance src-dst l4port

policy-map type network-qos jumbo

  class type network-qos class-default

    mtu 9216

system qos

  service-policy type network-qos jumbo

vrf context management

  ip route 0.0.0.0/0 10.29.135.1

copy run start

2.    Login as admin user into the Nexus Switch B and run the same above commands to set global configurations on Nexus Switch B.

*       Make sure to run copy run start to save the configuration on each switch after the configuration is completed.

VLANs Configuration

To create the necessary virtual local area networks (VLANs), follow these steps on both Nexus switches.

1.    Login as admin user into the Nexus Switch A.

2.    Create VLAN 134 for Public Network Traffic and VLAN 10 for Private Network Traffic:

configure terminal

vlan 134

name Oracle_RAC_Public_Network

no shutdown

vlan 10

name Oracle_RAC_Private_Network

no shutdown

interface Ethernet1/31

  description connect to uplink switch

  switchport access vlan 134

  speed 1000

copy run start

3.    Login as admin user into the Nexus Switch B and create VLAN 134 for Oracle RAC Public Network Traffic and VLAN 10 for Oracle RAC Private Network Traffic.

*       Make sure to run copy run start to save the configuration on each switch after the configuration is completed.

Virtual Port Channel (vPC) Summary for Network Traffic

A port channel bundles individual links into a channel group to create a single logical link that provides the aggregate bandwidth of up to eight physical links. If a member port within a port channel fails, traffic previously carried over the failed link switches to the remaining member ports within the port channel. Port channeling also load balances traffic across these physical interfaces. The port channel stays operational as long as at least one physical interface within the port channel is operational. Using port channels, Cisco NX-OS provides wider bandwidth, redundancy, and load balancing across the channels

In the Cisco Nexus Switch topology, a single vPC feature is enabled to provide HA, faster convergence in the event of a failure, and greater throughput. Cisco Nexus vPC configurations with the vPC domains and corresponding vPC names and IDs for Oracle Database Servers is shown in Table 7.

Table 7.          vPC Summary

vPC Domain

vPC Name

vPC ID

1

Peer-Link

1

1

vPC FI-A

51

1

vPC FI-B

52

As listed in Table 7, a single vPC domain with Domain ID 1 is created across two Nexus switches to define vPC members to carry specific VLAN network traffic. In this topology, we defined a total number of 3 vPCs.

vPC ID 1 is defined as Peer link communication between the two Nexus switches. vPC IDs 51 and 52 are configured for both Cisco UCS fabric interconnects. Please follow these steps to create this configuration.

*       A port channel bundles up to eight individual interfaces into a group to provide increased bandwidth and redundancy.

Create vPC Peer-Link

Related image, diagram or screenshot

For vPC 1 as Peer-link, we used interfaces 1 and 2 for Peer-Link. You may choose an appropriate number of ports based on your needs. To create the necessary port channels between devices, follow these steps on both the Nexus Switches:

4.    Login as admin user into the Nexus Switch A:

configure terminal

vpc domain 1

  peer-keepalive destination 10.29.134.53 source 10.29.134.52

  auto-recovery

interface port-channel 1

  description vPC peer-link

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type network

  vpc peer-link

interface Ethernet1/1

  description Peer link 100g connected to N9K-B-Eth1/1

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  channel-group 1 mode active

interface Ethernet1/2

  description Peer link 100g connected to N9K-B-Eth1/2

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  channel-group 1 mode active

exit

copy run start

5.    Login as admin user into the Nexus Switch B and repeat the above steps to configure second nexus switch. Make sure to changes the description of interfaces and peer-keepalive destination and source IP addresses.

Create vPC Configuration between Nexus Switches and Fabric Interconnects

This section describes how to create and configure port channel 51 and 52 for network traffic between the Nexus and Fabric Interconnect Switches.

DiagramDescription automatically generated

Table 8 lists the vPC IDs, allowed VLAN IDs, and Ethernet uplink ports.

Table 8.          vPC IDs and VLAN IDs

vPC Description

vPC ID

Fabric Interconnects Ports

Nexus Switch Ports

Allowed VLANs

Port Channel FI-A

51

FI-A Port 1/49

N9K-A Port 1/25

134, 10

Note: VLAN 10 Needed for Failover

FI-A Port 1/50

N9K-B Port 1/25

Port Channel FI-B

52

FI-B Port 1/49

N9K-A Port 1/26

10, 134

Note: VLAN 134 Needed for Failover

FI-B Port 1/50

N9K-B Port 1/26

Verify the ports connectivity on both the nexus switches as shown below:

A picture containing text, plaque, blackDescription automatically generated

A picture containing text, plaque, blackDescription automatically generated

To configure port channels on Nexus Switches, follow these steps:

6.    Login as admin user into the Nexus Switch A and perform the following steps:

configure terminal

interface port-channel51

  description Port-Channel FI-A

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 51

  no shutdown

interface port-channel52

  description Port-Channel FI-B

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 52

  no shutdown

interface Ethernet1/25

  description 100g link to Fabric-Interconnect A port 49

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 51 mode active

  no shutdown

interface Ethernet1/26

  description 100g link to Fabric-Interconnect B Port 49

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 52 mode active

  no shutdown

copy run start

7.    Login as admin user into the Nexus Switch B and run the following commands to configure the second Nexus switch:

configure terminal

interface port-channel51

  description Port-Channel FI-A

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 51

  no shutdown

interface port-channel52

  description Port-Channel FI-B

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 52

  no shutdown

 

interface Ethernet1/25

  description 100g link to Fabric-Interconnect A port 50

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 51 mode active

  no shutdown

interface Ethernet1/26

  description 100g link to Fabric-Interconnect B Port 50

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 52 mode active

  no shutdown

copy run start

Verify All vPC Status

To verify all the port-channel status, run the following commands into the Nexus Switches:

1.    Nexus Switch A Port-Channel Summary:

Graphical user interface, text, application, chat or text messageDescription automatically generated

2.    Nexus Switch B Port-Channel Summary:

Graphical user interface, text, application, chat or text messageDescription automatically generated

3.    Nexus Switch A vPC Status:

TextDescription automatically generated

4.    Nexus Switch B vPC Status:

TextDescription automatically generated

Cisco UCS Configuration

This section details the Cisco UCS configuration that was done as part of the infrastructure buildout. The racking, power, and installation of the chassis are described in the installation guide (see https://www.cisco.com/c/en/us/support/servers-unified-computing/ucs-manager/products-installation-guides-list.html).

*       It is beyond the scope of this document to explain the Cisco UCS infrastructure setup and connectivity. The documentation guides and examples are available here: https://www.cisco.com/c/en/us/support/servers-unified-computing/ucs-manager/products-installation-and-configuration-guides-list.html

Graphical user interface, diagramDescription automatically generated

*       This document details all the tasks to configure Cisco UCS but only some screenshots are included.

Using logical servers that are disassociated from the physical hardware removes many limiting constraints around how servers are provisioned. Cisco UCS Service Profiles contain values for a server's property settings, including virtual network interface cards (vNICs), MAC addresses, boot policies, firmware policies, fabric connectivity, external management, and HA information. The service profiles represent all the attributes of a logical server in Cisco UCS model. By abstracting these settings from the physical server into a Cisco Service Profile, the Service Profile can then be deployed to any physical compute hardware within the Cisco UCS domain. Furthermore, Service Profiles can, at any time, be migrated from one physical server to another. Furthermore, Cisco is the only hardware provider to offer a truly unified management platform, with Cisco UCS Service Profiles and hardware abstraction capabilities extending to both blade and rack servers.

High-Level Steps to Configure Base Cisco UCS

The following are the high-level steps involved for a Cisco UCS configuration:

1.    Perform Initial Setup of Fabric Interconnects for a Cluster Setup

2.    Upgrade UCS Manager Software to Version 4.1(3b)

3.    Synchronize Cisco UCS to NTP

4.    Configure Fabric Interconnects for Chassis and Blade Discovery

5.    Configure Global Policies

6.    Configure Server Ports

7.    Configure LAN and SAN

8.    Configure Ethernet LAN Uplink Ports

9.    Create Uplink Port Channels to Nexus Switches

10.  Configure FC SAN Uplink Ports

11.  Configure VLANs

12.  Configure VSANs

13.  Create FC Uplink Port Channels to MDS Switches

14.  Enable FC Uplink VSAN Trunking (FCP)

15.  Configure IP, UUID, Server, MAC, WWNN and WWPN Pools

16.  IP Pool Creation

17.  UUID Suffix Pool Creation

18.  Server Pool Creation

19.  MAC Pool Creation

20.  WWNN and WWPN Pool

21.  Set Jumbo Frames in both the Fabric Interconnect

22.  Configure Server BIOS Policy

23.  Create Adapter Policy

24.  Create Adapter Policy for Public and Private Network Interfaces

25.  Create Adapter Policy for NVMe FC Storage Network Interfaces

26.  Configure Update Default Maintenance Policy

27.  Configure Host Firmware Policy

28.  Configure vNIC and vHBA Template

29.  Create Public vNIC Template

30.  Create Private vNIC Template

31.  Create Storage FC Storage vHBA Template

32.  Create Server Boot Policy for SAN Boot

The details for each of these steps are documented in the following sections.

Perform Initial Setup of Cisco UCS 6454 Fabric Interconnects for a Cluster Setup

This section provides detailed procedures for configuring the Cisco Unified Computing System (Cisco UCS) for use in a FlexPod environment. The steps are necessary to provision the Cisco UCS B-Series and C-Series servers and should be followed precisely to avoid improper configuration.

Configure FI-A and FI-B

To configure the UCS Fabric Interconnects, follow these steps.

1.    Verify the following physical connections on the fabric interconnect:

a.     The management Ethernet port (mgmt0) is connected to an external hub, switch, or router

b.    The L1 ports on both fabric interconnects are directly connected to each other

c.     The L2 ports on both fabric interconnects are directly connected to each other

2.    Connect to the console port on the first Fabric Interconnect and follow these steps:

Enter the configuration method. (console/gui) ? console

Enter the setup mode; setup newly or restore from backup. (setup/restore) ? setup

You have chosen to setup a new Fabric interconnect. Continue? (y/n): y

Enforce strong password? (y/n) [y]: Enter

Enter the password for admin: <password>

Confirm the password for admin: <password>

Is this Fabric interconnect part of a cluster(select 'no' for standalone)? (yes/no) [n]: y

Enter the switch fabric (A/B) []: A

Enter the system name:  <ucs-cluster-name>

Physical Switch Mgmt0 IP address : <ucsa-mgmt-ip>

Physical Switch Mgmt0 IPv4 netmask : <ucsa-mgmt-mask>

IPv4 address of the default gateway : <ucsa-mgmt-gateway>

Cluster IPv4 address : <ucs-cluster-ip>

Configure the DNS Server IP address? (yes/no) [n]: y

DNS IP address : <dns-server-1-ip>

Configure the default domain name? (yes/no) [n]: y

Default domain name : <ad-dns-domain-name>

Join centralized management environment (UCS Central)? (yes/no) [n]: Enter

Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): yes

3.    Review the settings printed to the console. Answer yes to apply and save the configuration.

4.    Wait for the login prompt to make the configuration has been saved to Fabric Interconnect A.

5.    Connect console port on the second Fabric Interconnect B and follow these steps:

  Enter the configuration method. (console/gui) ? console

  Installer has detected the presence of a peer Fabric interconnect. This Fabric interconnect will be added to the cluster. Continue (y/n) ? y

  Enter the admin password of the peer Fabric interconnect: <password>

  Connecting to peer Fabric interconnect... done

  Retrieving config from peer Fabric interconnect... done

  Peer Fabric interconnect Mgmt0 IPv4 Address: <ucsa-mgmt-ip>

  Peer Fabric interconnect Mgmt0 IPv4 Netmask: <ucsa-mgmt-mask>

  Cluster IPv4 address          : <ucs-cluster-ip>

  Peer FI is IPv4 Cluster enabled. Please Provide Local Fabric Interconnect Mgmt0 IPv4 Address

  Physical Switch Mgmt0 IP address : <ucsb-mgmt-ip>

  Local fabric interconnect model(UCS-FI-6454)

  Peer fabric interconnect is compatible with the local fabric interconnect. Continuing with the installer...

  Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): yes

6.    Review the settings printed to the console. Answer yes to apply and save the configuration.

7.    Wait for the login prompt to make the configuration has been saved to Fabric Interconnect B.

Log into Cisco UCS Manager

To log into the Cisco Unified Computing System (UCS) environment, follow these steps:

1.    Open a web browser and navigate to the Cisco UCS fabric interconnect cluster address.

2.    Click the Launch UCS Manager link under HTML to launch Cisco UCS Manager.

3.    If prompted to accept security certificates, accept as necessary.

A screenshot of a computerDescription automatically generated with medium confidence

4.    When prompted, enter admin as the username and enter the administrative password.

5.    Click Login to log into Cisco UCS Manager.

Configure Cisco UCS Call Home

It is highly recommended by Cisco to configure Call Home in Cisco UCS Manager.  Configuring Call Home will accelerate resolution of support cases.  To configure Call Home, follow these steps:

1.    In Cisco UCS Manager, click Admin.

2.    Select All > Communication Management > Call Home.

3.    Change the State to On.

4.    Fill in all the fields according to your Management preferences and click Save Changes and OK to complete configuring Call Home.

Upgrade Cisco UCS Manager Software to Version 4.1 (3b)

This solution was configured on Cisco UCS 4.1(3b) software release. To upgrade the Cisco UCS Manager software and the Cisco UCS Fabric Interconnect software to version 4.1, go to https://software.cisco.com/download/home/283612660/type/283655658/release/4.1(3b)

For more information about Install and Upgrade Guides, go to: https://www.cisco.com/c/en/us/support/servers-unified-computing/ucs-manager/products-installation-guides-list.html

Synchronize Cisco UCS to NTP

To synchronize the Cisco UCS environment to the NTP server, follow these steps:

1.    In Cisco UCS Manager, in the navigation pane, click the Admin tab.

2.    Select All > Time zone Management.

3.    In the Properties pane, select the appropriate time zone in the Time zone menu.

4.    Click Save Changes and then click OK.

5.    Click Add NTP Server.

6.    Enter the NTP server IP address and click OK.

7.    Click OK to finish.

Configure Fabric Interconnect for Chassis and Server Discovery

Cisco UCS 6454 Fabric Interconnects are configured for redundancy. It provides resiliency in case of failures. The first step to establish connectivity between blades and Fabric Interconnects.

Configure Global Policies

The chassis discovery policy determines how the system reacts when you add a new chassis. We recommend using the platform max value as shown. Using platform max helps ensure that Cisco UCS Manager uses the maximum number of IOM uplinks available.

To configure global policies, follow these steps:

1.    Go to Equipment > Policies > Global Policies > Chassis/FEX Discovery Policies. As shown in the screenshot below, select Action as Platform Max from the drop-down list and set Link Grouping to Port Channel.

Graphical user interface, text, application, emailDescription automatically generated

2.    Click Save Changes.

3.    Click OK.

Configure Server Ports

Configure Server Ports to initiate Chassis and Blade discovery. To configure server ports, follow these steps:

1.    Go to Equipment > Fabric Interconnects > Fabric Interconnect A > Fixed Module > Ethernet Ports.

2.    Select the ports (for this solution ports are 17-32) which are connected to the Cisco IO Modules of the two Cisco UCS B-Series 5108 Chassis.

3.    Right-click and select Configure as Server Port.

4.    Click Yes to confirm and click OK.

Graphical user interface, application, table, ExcelDescription automatically generated

5.    Repeat steps 1-4 for Fabric Interconnect B.

6.    After configuring Server Ports, acknowledge both the Chassis. Go to Equipment > Chassis > Chassis 1 >  General > Actions > select Acknowledge Chassis. Similarly, acknowledge the chassis 2.

7.    After acknowledging both the chassis, Re-acknowledge all the servers placed in the chassis. Go to Equipment > Chassis 1 > Servers > Server 1 > General > Actions > select Server Maintenance > select option Re-acknowledge and click OK. Similarly, repeat the process to Re-acknowledge all the eight Servers.

8.    Once the acknowledgement of the Servers completed, verify Port-channel of Internal LAN. Go to tab LAN > Internal LAN > Internal Fabric A > Port Channels as shown below.

Graphical user interface, tableDescription automatically generated

9.    Verify the same for Internal Fabric B.

Configure LAN and SAN on Cisco UCS Manager

Configure Ethernet Uplink Ports and Fibre Channel (FC) Storage ports on Cisco UCS as explained below.

Configure Ethernet LAN Uplink Ports

To configure network ports used to uplink the Fabric Interconnects to the Nexus switches, follow these steps:

1.    In Cisco UCS Manager, in the navigation pane, click the Equipment tab.

2.    Select Equipment > Fabric Interconnects > Fabric Interconnect A > Fixed Module.

3.    Expand Ethernet Ports.

4.    Select ports (for this solution ports are 49-50) that are connected to the Nexus switches, right-click them, and select Configure as Network Port.

5.    Click Yes to confirm ports and click OK.

6.    Verify the Ports connected to Nexus upstream switches are now configured as network ports.

7.    Repeat steps 1-6 for Fabric Interconnect B. The screenshot shows the network uplink ports for Fabric A.

Graphical user interface, application, tableDescription automatically generated

Now two uplink ports have been created on each Fabric Interconnect as shown above. These ports will be used to create Virtual Port Channel in the next section.

Create Uplink Port Channels to Cisco Nexus Switches

In this procedure, two port channels were created: one from Fabric A to both Nexus switch and one from Fabric B to both Nexus switch. To configure the necessary port channels in the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the LAN tab in the navigation pane.

2.    Under LAN > LAN Cloud, expand node Fabric A tree:

a.     Right-click Port Channels.

b.    Select Create Port Channel.

c.     Enter 51 as the unique ID of the port channel.

d.    Enter FI-A as the name of the port channel.

C:\Users\havyas\Desktop\Screenshot 2018-06-14 17.08.46.png

e.    Click Next.

f.      Select Ethernet ports 49-50 for the port channel.

g.    Click >> to add the ports to the port channel

3.    Click Finish to create the port channel and then click OK.

Graphical user interface, text, application, emailDescription automatically generated

4.    Repeat steps 1-3 for Fabric Interconnect B, substituting 52 for the port channel number and FI-B for the name.

Configure FC SAN Uplink Ports

To enable the Fibre channel ports, follow these steps for the FI 6454:

1.    In Cisco UCS Manager, click Equipment.

2.    Select Equipment > Fabric Interconnects > Fabric Interconnect A (primary).

3.    Select Configure Unified Ports.

4.    Click Yes on the pop-up window warning that changes to the fixed module will require a reboot of the fabric interconnect and changes to the expansion module will require a reboot of that module.

5.    Within the Configured Fixed Ports pop-up window move the gray slider bar from the left to the right to select either 4, 8, or 12 ports to be set as FC Uplinks.

Graphical user interface, tableDescription automatically generated

6.    For this solution, we configured the first four ports on the FI as FC Uplink ports. Click OK, then click Yes, then click OK to continue

*       Applying this configuration will cause the immediate reboot of Fabric Interconnect and/or Expansion Module(s).

7.    Select Equipment > Fabric Interconnects > Fabric Interconnect B (primary).

8.    Select Configure Unified Ports.

9.    Click Yes on the pop-up window warning that changes to the fixed module will require a reboot of the fabric interconnect and changes to the expansion module will require a reboot of that module.

10.  Within the Configured Fixed Ports pop-up window move the gray slider bar from the left to the right to select either 4, 8, or 12 ports to be set as FC Uplinks.

11.  Click OK then click Yes then click OK to continue.

12.  Wait for both Fabric Interconnects to reboot.

13.  Log back into Cisco UCS Manager.

Configure VLAN

In this solution, two VLANs were created: one for private network (VLAN 10) traffic, one for public network (VLAN 134) traffic. These two VLANs will be used in the vNIC templates that are discussed later.

To configure the necessary virtual local area networks (VLANs) for the Cisco UCS environment, follow these steps:

*       It is very important to create both VLANs as global across both fabric interconnects. This way, VLAN identity is maintained across the fabric interconnects in case of NIC failover.

1.    In Cisco UCS Manager, click the LAN tab in the navigation pane.

2.    Select LAN >  LAN Cloud.

3.    Right-click VLANs.

4.    Select Create VLANs.

5.    Enter Public_Traffic as the name of the VLAN to be used for Public Network Traffic.

6.    Keep the Common/Global option selected for the scope of the VLAN.

7.    Enter 134 as the ID of the VLAN ID.

8.    Keep the Sharing Type as None.

Graphical user interface, text, applicationDescription automatically generated

9.    Click OK and then click OK again.

10.  Create the second VLAN: for private network (VLAN 10) traffic and remaining two storage VLANs for storage network (VALN 11 and 12) traffic as shown below:

Graphical user interface, text, application, emailDescription automatically generated

These two VLANs will be used in the vNIC templates that are described in this CVD.

Configure VSAN

In this solution, we created two VSANs. VSAN-A 151 and VSAN-B 152 for FC SAN Boot and NVMe/FC Storage Access.

To configure the necessary virtual storage area networks (VSANs) for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the SAN tab in the navigation pane.

2.    Select SAN > SAN Cloud > Fabric A > VSANs

3.    Under VSANs, right-click on VSANs.

4.    Select Create VSAN.

5.    Enter VSAN-A as the name of the VSAN.

6.    Leave FC Zoning set at Disabled.

7.    Select Fabric A for the scope of the VSAN.

8.    Enter VSAN ID as 151.

Graphical user interface, text, application, emailDescription automatically generated

9.    Click OK and then click OK again

10.  Repeat steps 1-9 to create the VSAN 152 on FI-B.

*       Enter a unique VSAN ID and a corresponding FCoE VLAN ID that matches the configuration in the MDS switch for Fabric A.  It is recommended to use the same ID for both parameters and to use something other than 1.

Create FC Uplink Port Channels to MDS Switches

In this solution, we created two FC Port Chanel. The first FC Port Channel is between FI-A to MDS-A and the second FC Port Channel is between FI-B to MDS-B.

To create FC Port channel, follow these steps on the FI 6454:

1.    In Cisco UCS Manager, click the SAN tab.

2.    Select SAN > SAN Cloud > Fabric A > FC Port Channels > and then right-click on the FC Port Channel.

3.    Enter the name of Port Channel as FC-PC-A and unique ID as 251 and click Next

4.    Select the appropriate ports of FI-A which are going to MDS-A and click the button >> to select those ports as a member of the Port Channel. For this solution, we configured all the four ports as a Port Channel ports as shown in the screenshot below:

Graphical user interface, applicationDescription automatically generated

5.    Click Finish to create this FC Port Channel for FI-A.

6.    Repeat steps 1-5 to create the FC Port Channel on FI-B with related FC Ports going to MDS-B.

*       We configured the FI-B Port Channel as FC-PC-B with unique ID 252 as shown below.

Graphical user interface, text, application, emailDescription automatically generated

7.    Select VSAN-A 151 for FC-PC-A and click Save Changes.

8.    Similarly, select VSAN-B 152 for FC-PC-B and click Save Changes.

*       The MDS Switch is configured in the following section and after the appropriate VSAN and FC ports configuration, the FC Ports and Port-Channel will become ACTIVE.

Enable FC Uplink VSAN Trunking (FCP)

To enable VSAN trunking on the FC Uplinks in the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click SAN.

2.    Expand SAN > SAN Cloud.

3.    Choose Fabric A and in the Actions pane choose Enable FC Uplink Trunking.

4.    Click Yes on the Confirmation and Warning and then click OK.

5.    Choose Fabric B and in the Actions pane choose Enable FC Uplink Trunking.

6.    Click Yes on the Confirmation and Warning. Click OK to finish.

*       Enabling VSAN trunking is optional. It is important that the Cisco Nexus VSAN trunking configuration match the configuration set in Cisco UCS Manager.

Configure IP, UUID, Server, MAC, WWNN and WWPN Pools

IP Pool Creation

An IP address pool on the out of band management network must be created to facilitate KVM access to each compute node in the UCS domain. To create a block of IP addresses for server KVM access in the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, in the navigation pane, click the LAN tab.

2.    Select Pools > root > IP Pools >click Create IP Pool.

3.    We have named IP Pool as Ora19C-KVM Pool for this solution.

4.    Select option Sequential to assign IP in sequential order then click Next.

5.    Click Add IPv4 Block.

6.    Enter the starting IP address of the block and the number of IP addresses required, and the subnet and gateway information as shown in the screenshot according to your environment.

Graphical user interface, textDescription automatically generated

7.    Click Next and then click Finish to create the IP block.

UUID Suffix Pool Creation

To configure the necessary universally unique identifier (UUID) suffix pool for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the Servers tab in the navigation pane.

2.    Select Pools > root.

3.    Right-click UUID Suffix Pools and then select Create UUID Suffix Pool.

4.    Enter ORA19C-UUID as the name of the UUID Pool name.

5.    Optional: Enter a description for the UUID pool.

6.    Keep the prefix at the derived option and select Sequential in as Assignment Order then click Next.

7.    Click Add to add a block of UUIDs.

8.    Create a starting point UUID as per your environment.

9.    Specify a size for the UUID block that is sufficient to support the available blade or server resources.

Graphical user interface, text, application, chat or text messageDescription automatically generated

10.  Clink OK then click Finish to complete the UUID Pool congiruration.

Server Pool Creation

To configure the necessary server pool for the Cisco UCS environment, follow these steps:

*       Consider creating unique server pools to achieve the granularity that is required in your environment.

1.    In Cisco UCS Manager, click the Servers tab in the navigation pane.

2.    Select Pools > root > Right-click Server Pools > Select Create Server Pool.

3.    Enter ORA19C-SERVER-POOL as the name of the server pool.

4.    Optional: Enter a description for the server pool then click Next.

5.    Select all 8 servers to be used for the Oracle RAC management and click >> to add them to the server pool.

TableDescription automatically generated

6.    Click Finish and then click OK.

MAC Pool Creation

In this solution, we created two MAC Pool as ORA19C-PUBLIC-FI-A and ORA19C-PRIVATE-FI-B to provide MAC addresses for Public and Private Network Interfaces.

To configure the necessary MAC address pools for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the LAN tab in the navigation pane.

2.    Select Pools > root > right-click MAC Pools under the root organization.

3.    Select Create MAC Pool to create the MAC address pool.

4.    Enter ORA19C-PUBLIC-FI-A as the name for MAC pool.

5.    Enter the seed MAC address and provide the number of MAC addresses to be provisioned.

Graphical user interface, text, application, chat or text messageDescription automatically generated

6.    Click OK and then click Finish.

7.    In the confirmation message, click OK.

8.    Create MAC Pool B and assign unique MAC Addresses as shown below.

Graphical user interface, text, application, emailDescription automatically generated

WWNN and WWPN Pool Creation

In this solution, we configured one WWNN Pool to provide SAN access point for Linux hosts. To configure the necessary WWNN pools for the Cisco UCS environment, follow these steps:

*       These WWNN and WWPN entries will be used to access storage through SAN configuration

1.    In Cisco UCS Manager, click the SAN tab in the navigation pane.

2.    Select Pools > Root > WWNN Pools > right-click WWNN Pools > Select Create WWNN Pool.

3.    Assign name as ORA19C-WWNN-A and Assignment Order as sequential and click Next.

4.    Click Add and create a WWN Block as shown below.

Graphical user interface, text, applicationDescription automatically generated

5.    Click OK and then Finish.

In this solution, we created two WWPNs; ORA19C-WWPN-A and ORA19C-WWPN-B Pool, for the World Wide Port Name. To configure the necessary WWPN pools for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the SAN tab in the navigation pane.

2.    Select Pools > Root > WWPN Pools > right-click WWPN Pools > select Create WWPN Pool.

3.    Assign the name ORA19C-WWPN-A and Assignment Order as sequential.

4.    Click Next and then click Add to add block of Ports.

5.    Enter Block for WWN and size.

Graphical user interface, text, applicationDescription automatically generated

6.    Click OK and then Finish.

7.    Configure the ORA19C-WWPN-B Pool as well and assign the unique block IDs as shown below.

Graphical user interface, text, application, emailDescription automatically generated

*       When there are multiple UCS domains sitting in adjacency, it is important that these blocks; the WWNN, WWPN, and MAC, hold differing values between each set.

Set Jumbo Frames in both the Cisco Fabric Interconnect

To configure jumbo frames and enable quality of service in the Cisco UCS fabric, follow these steps:

1.    In Cisco UCS Manager, click the LAN tab in the navigation pane.

2.    Select LAN > LAN Cloud > QoS System Class.

3.    Click the General tab.

4.    On the Best Effort row, enter 9216 in the box under the MTU column.

5.    Click Save Changes in the bottom of the window.

Graphical user interface, application, tableDescription automatically generated

6.    Click OK.

*       The only the Fibre Channel and Best Effort QoS System Classes are enabled in this FlexPod implementation. The Cisco UCS and Nexus switches are intentionally configured this way so that all IP traffic within the FlexPod will be treated as Best Effort. 

*       Enabling the other QoS System Classes without having a comprehensive, end-to-end QoS setup in place can cause difficult to troubleshoot issues.

Configure Server BIOS Policy

*       All of the Server BIOS policies may have to be required on your setup. Please follow the steps according to your environment and requirement. The following changes were made on the test bed where Oracle RAC installed. Please validate and change as needed.

*       For more detailed on BIOS Settings, go to https://www.cisco.com/c/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/white-paper-c11-744678.html

*       It is recommended to disable C states in the BIOS and in addition, Oracle recommends disabling it from OS level as well by modifying grub entries. We will cover the OS level settings in the Operating system configuration section.

To create a server BIOS policy for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click Servers.

2.    Select Policies > root.

3.    Right-click BIOS Policies.

4.    Select Create BIOS Policy.

5.    Enter ORA19C_BIOS as the BIOS policy name

6.    Select and click the newly created BIOS Policy.

7.    Click the Advanced tab, leaving the Processor tab selected within the Advanced tab.

8.    Set the following within the Processor tab:

a.     CPU Hardware Power Management: HWPM Native Mode

b.    CPU Performance: Enterprise

c.     Energy Efficient Turbo: Disabled

d.    IMC Inteleave: Auto

e.    Sub NUMA Clustering: Disabled

f.      Package C State Limit: C0 C1 State

g.    Processor C State: Disabled

h.     Processor C1E: Disabled

i.      Processor C3 Report: Disabled

j.      Processor C6 Report: Disabled

k.     Processor C7 Report: Disabled

l.      LLC Prefetch: Disabled

m.   Demand Scrub: Disabled

n.     Patrol Scrub: Disabled

o.    Workload Configuration: IO Sensitive

Graphical user interface, applicationDescription automatically generated

9.    Set the following within the RAS Memory tab:

a.     Memory RAS configuration: ADDDC Sparing

10.  Click Save Changes and then click OK.

Create Adapter Policy

*       In this solution, we created two adapter policies; “Ethernet Adapter Policy” and “Fibre Channel Adapter Policy.” We also configured “Ethernet Adapter Policy” for the Public and Private Network Interface Traffic and the second one “Fibre Channel Adapter Policy” for NVMe FC Storage Network Interface Traffic as explained in the following sections.

Create Adapter Policy for Ethernet Traffic (Public and Private Network Interfaces)

To create an Adapter Policy for the UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the Servers tab in the navigation pane.

2.    Select Policies > root > right-click Adapter Policies.

3.    Select Create Ethernet Adapter Policy.

4.    Provide a name for the Ethernet adapter policy as ORA19C-Linux. Change the following fields and click Save Changes:

a.     Resources:

     Transmit Queues: 8

     Ring Size: 4096

     Receive Queues: 8

     Completion Queues: 16

     Interrupts: 32

b.    Options:

     Receive Side Scaling (RSS): Enabled

     Configure the adapter policy as shown below.

Graphical user interfaceDescription automatically generated

*       RSS distributes network receive processing across multiple CPUs in multiprocessor systems. This can be one of the following.

*       Disabled—Network receive processing is always handled by a single processor even if additional processors are available.

*       Enabled—Network receive processing is shared across processors whenever possible.

Create Adapter Policy for Fibre Chanel (NVMe/FC Storage Network Interfaces)

To create an Adapter Policy for the UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the Servers tab in the navigation pane.

2.    Select Policies > root > right-click Adapter Policies.

3.    Select Create Fibre Channel Adapter Policy.

4.    Provide a name for the FC Adapter Policy and change the following fields and clink Save changes when you are finished:

a.     Resources:

     I/O Queues: 16

b.    Options:

     vHBA Type: FC NVMe Initiator

     Max LUNs Per Target: 1024

The NVMe/FC Adapter is configured as shown below:

Graphical user interface, text, application, emailDescription automatically generated

Configure Update Default Maintenance Policy

To update the default Maintenance Policy, follow these steps:

1.    In Cisco UCS Manager, click the Servers tab in the navigation pane.

2.    Select Policies > root > Maintenance Policies > Default.

3.    Change the Reboot Policy to User Ack.

4.    Click Save Changes.

5.    Click OK to accept the changes.

Configure Host Firmware Policy

Firmware management policies allow the administrator to choose the corresponding packages for a given server configuration. These policies often include packages for adapter, BIOS, board controller, FC adapters, host bus adapter (HBA) option ROM, and storage controller properties.

To create the default firmware management policy in the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click Servers.

2.    Expand Policies > root.

3.    Expand Host Firmware Packages and right-click to “Create Host Firmware Policy.”

4.    Give the policy name as 4.1-3b and select the Blade and Rack Packages as shown below.

Graphical user interface, applicationDescription automatically generated

5.    Click OK, to create the host firmware package for this UCSM version.

Configure vNIC and vHBA Template

*       For this solution, we created two vNIC template for Public Network and Private Network Traffic. We will use these vNIC templates during creation of Service Profile later in this section.

Create Public and Private vNIC Template

To create vNIC (virtual network interface card) template for the UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the LAN tab in the navigation pane.

2.    Select Policies > root > vNIC Templates > Right-click to vNIC Template and Select "Create vNIC Template."

3.    Enter ORA19C-vNIC-A as the vNIC template name and keep Fabric A selected.

4.    Select the Enable Failover checkbox for high availability of the vNIC.

*       Selecting Failover is a critical step to improve link failover time by handling it at the hardware level, and to guard against NIC any potential for NIC failure not being detected by the virtual switch.

5.    Select Template Type as Updating Template

6.    Under VLANs, select the checkboxes default and Public_Traffic and set Native-VLAN as the Public_Traffic.

Graphical user interface, applicationDescription automatically generated

7.    Keep MTU value 1500 for Public Network Traffic

8.    In the MAC Pool list, select ORA19C-PUBLIC-FI-A.

9.    Click OK to create the vNIC template as shown below.

Graphical user interface, text, applicationDescription automatically generated

10.  Click OK to finish.

11.  Similarly, create another vNIC template for Private Network Traffic with few changes.

12.  Enter ORA19C-vNIC-B as the vNIC template name for Private Network Traffic.

13.  Select the Fabric B and Enable Failover for Fabric ID options.

Graphical user interface, applicationDescription automatically generated

14.  Select Template Type as Updating Template.

15.  Under VLANs, select the checkboxes default and Private_Traffic and set Native-VLAN as the Private_Traffic.

16.  Set MTU value to 9000 and MAC Pool as ORA19C-PRIVATE-FI-B.

17.  Click OK to create the vNIC template.

Create Storage vHBA Template

For this solution, we created two vHBA as ORA19C-vHBA-A and ORA19C-vHBA-B.

To create virtual host bus adapter (vHBA) templates for the Cisco UCS environment, follow these steps:

1.    In Cisco UCS Manager, click the SAN tab in the navigation pane.

2.    Select Policies > root > right-click vHBA Templates > select “Create vHBA Template” to create vHBA.

3.    Enter the name ORA19C-vHBA-A and keep Fabric A selected.

4.    Select VSAN as VSAN-A and template type to Updating Template.

5.    Select WWPN Pool as ORA19C-WWPN-A from the drop-down list as shown below.

Graphical user interface, applicationDescription automatically generated

6.    Click OK to create first vHBA.

7.    Create the second vHBA and change the name as ORA-vHBA-B.

8.    Select the Fabric ID as B, template type to Updating Template and WWPN as ORA19C-WWPN-B as shown below.

Graphical user interface, text, application, emailDescription automatically generated

Create Server Boot Policy for SAN Boot

All Oracle nodes were set to boot from SAN for the Cisco Validated Design as part of the Service Profile template. The benefits of booting from SAN are numerous; disaster recovery, lower cooling, and power requirements for each server since a local drive is not required, and better performance, name just a few. We strongly recommend using “Boot from SAN” to realize full benefits of Cisco UCS stateless computing feature such as service profile mobility.

This process applies to a Cisco UCS environment in which the storage SAN ports are configured in the following sections.

Create Local Disk Configuration Policy

A Local disk configuration for the Cisco UCS is necessary if the servers in the environments have a local disk. To configure Local disk policy, follow these steps:

1.    Go to Cisco UCS Manager and then go to Servers > Policies > root > Boot Policies.

2.    Right-click and select Create Boot Policy. Enter SAN-Boot for the name of the boot policy.

3.    Change the mode to “No Local Storage.”

4.    Click OK to create the policy as shown below.

Graphical user interface, text, applicationDescription automatically generated

Create SAN Boot Policy

To create SAN Boot Policy, you need to enter the WWPN of NetApp Storage as explained below.

The screenshot below shows both the NetApp AFF A800 Controller FC Ports and related WWPN.

Graphical user interface, applicationDescription automatically generated

For this solution, we created four FC Logical Interfaces (LIFs) on Storage Controller Node 1 (node1_lif02a, node1_lif02b, node1_lif02c and node1_lif02d) and four FC LIFs on Storage Controller Node 2 (node2_lif02a, node2_lif02b, node2_lif02c and node2_lif02d) as shown below.

A picture containing tableDescription automatically generated

The SAN boot policy configures the SAN Primary's primary-target to be network interface node1_lif02a on NetApp storage cluster and SAN Primary's secondary-target to be network interface node2_lif02a on NetApp storage cluster. Similarly, the SAN Secondary’s primary-target to be network interface node1_lif02c on NetApp storage cluster and SAN Secondary's secondary-target to be network interface node2_lif02c on NetApp storage cluster. This multiple FC ports selection will allow the server node to be highly availability in case of any storage controller failure. Login into NetApp storage controller and verify all the port information is correct. This information can be found in the NetApp Storage GUI under Network > Network Interfaces.

You have to create SAN Primary (hba0) and SAN Secondary (hba1) in SAN Boot Policy by entering the WWPN of NetApp Storage LIFs as explained below.

To create Boot Policy for the Cisco UCS environment, follow these steps:

1.    Go to Cisco UCS Manager and then go to Servers > Policies > root > Boot Policies and then right-click to Create Boot Policy as shown below.

C:\Users\havyas\Desktop\Snap 2018-11-08 at 17.32.50.jpg

2.    Expand the Local Devices drop-down menu and Choose Add CD/DVD.

3.    Expand the vHBAs drop-down menu and select Add SAN Boot.

4.    In the Add SAN Boot dialog box, select Type as “Primary” and name vHBA as “hba0.” Click OK to add SAN Boot.

C:\Users\havyas\Desktop\Snap 2018-11-08 at 17.36.28.jpg

*       The SAN boot paths and targets will include primary and secondary options in order to maximize resiliency and number of paths

5.    Select Add SAN Boot Target to enter WWPN address of storage LIF. Keep 0 as the value for Boot Target LUN. Enter the WWPN of NetApp Storage cluster interface node1_lif02a and add SAN Boot Primary Target.

Graphical user interface, text, application, chat or text messageDescription automatically generated

6.    Add secondary SAN Boot target into same hba0, enter boot target LUN as 0 and WWPN of NetApp Storage cluster interface node2_lif02a and add SAN Boot Secondary Target.

Graphical user interface, text, application, chat or text messageDescription automatically generated

7.    From the vHBA drop-down list and Choose Add SAN Boot.

8.    In the Add SAN Boot dialog box, enter "hba1" in the vHBA field.

C:\Users\havyas\Desktop\CISCO\Project_WORK\2017_07_FlashStack_X_CVD\Screenshots\Screenshot_3\Screenshot 2017-11-04 01.19.56 (2).png

9.    Click OK to SAN Boot. Then choose add SAN Boot Target

10.  Enter 0 as the value for Boot Target LUN. Enter the WWPN of NetApp Storage cluster interface node1_lif02c and add SAN Boot Primary Target.

Graphical user interface, text, application, chat or text messageDescription automatically generated

11.  Add secondary SAN Boot target into same hba1 and enter boot target LUN as 0 and WWPN of NetApp Storage cluster interface node2_lif02c and add SAN Boot Secondary Target.

Graphical user interface, text, application, chat or text messageDescription automatically generated

12.  After creating the FC boot policies, you can review the boot order in the UCS Manager GUI as shown below.

Graphical user interface, text, application, emailDescription automatically generated

13.  Click OK to finish creating SAN_Boot Policy.

*       For this solution, we created one Boot Policy as “SAN_Boot.” For all eight Oracle Database RAC Nodes (flex1, flex2, flex3, flex4, flex5, flex6, flex7 and flex8), we will assign this boot policy to the Service Profiles as explained in the following section.

Create and Configure Service Profile Template

Service profile templates enable policy-based server management that helps ensure consistent server resource provisioning suitable to meet predefined workload needs.

The Cisco UCS service profiles with SAN boot policy provides the following benefits:

   Scalability - Rapid deployment of new servers to the environment in a very few steps.

   Manageability - Enables seamless hardware maintenance and upgrades without any restrictions.

   Flexibility - Easy to repurpose physical servers for different applications and services as needed.

   Availability - Hardware failures are not impactful and critical.  In rare case of a server failure, it is easier to associate the logical service profile to another healthy physical server to reduce the impact.

For this solution, we will create one Service Profile Template “ORA19C_FLEXPOD” using boot policy created earlier to utilize four LIF ports from NetApp Storage for high-availability in case of any FC links go down.

Create Service Profile Template

To create a service profile template, follow these steps:

1.    In the Cisco UCS go to Servers > Service Profile Templates > root and right-click to Create Service Profile Template as shown below:

Graphical user interface, text, applicationDescription automatically generated

2.    Enter the Service Profile Template name, select the UUID Pool that was created earlier, and click Next.

3.    Select Local Disk Configuration Policy to SAN-Boot as no Local Storage.

Graphical user interface, text, applicationDescription automatically generated

4.    In the networking window, select Expert and click Add to create vNICs that the server should use to connect to the LAN

*       In this solution, we created two vNIC. We have given name to first vNIC as “eth0” and second vNIC as “eth1”.

5.    As shown below in the Create vNIC menu, enter the name “eth0” and check the box for “Use vNIC Template” option. Select vNIC Template “ORA19C-vNIC-A” with Ethernet Adapter Policy as “ORA19C-Linux” which was created earlier.

Graphical user interface, text, application, chat or text messageDescription automatically generated

6.    Add the second vNIC “eth1” and check the box for “Use vNIC Template” option. Select vNIC Template ORA19C-vNIC-B with Ethernet Adapter Policy as ORA19C-Linux.

Graphical user interface, text, application, chat or text messageDescription automatically generated

As shown below, we configured two vNICs as eth0 and eth1 so that Servers should use to connect to the LAN.

Graphical user interface, applicationDescription automatically generated

7.    When vNICs are created, click Next.

8.    In the SAN Connectivity menu, select Expert to configure the SAN connectivity. Select WWNN (World Wide Node Name) pool, which was created earlier.

Graphical user interface, applicationDescription automatically generated

9.    Click Add to add vHBAs as explained below.

*       For this solution, we configured 4 vHBA as 2 vHBA for SAN Boot and 2 vHBA for NVMe/FC. vHBA0 and vHBA1 are configured for carrying FC Network Traffic & Boot from SAN through MDS-A and MDS-B Switch while vHBA2 and vHBA3 are configured for NVMe/FC Network Traffic (Oracle RAC Storage Traffic) through MDS-A and MDS-B Switch.

The following four vHBA are created as follows:

   hba0 using vHBA Template ORA19C-vHBA-A and FC Adapter Policy as “Linux”

   hba1 using vHBA Template ORA19C-vHBA-B and FC Adapter Policy as “Linux”

   hba2 using vHBA Template ORA19C-vHBA-A and FC Adapter Policy as “ORA19C-FCNVMe”

   Hba3 using vHBA Template ORA19C-vHBA-B and FC Adapter Policy as “ORA19C-FCNVMe”

Graphical user interface, applicationDescription automatically generated

Graphical user interface, text, applicationDescription automatically generated

Graphical user interface, text, applicationDescription automatically generated

Graphical user interface, text, application, emailDescription automatically generated

Four vHBAs are configured as shown below.

Graphical user interface, applicationDescription automatically generated

10.  Click Next to proceed into zoning menu. 

11.  For this Oracle RAC Configuration, the Cisco MDS 9132T is used for zoning. So, skip zoning and click Next.

12.  In vNIC/vHBA Placement menu, keep the option as Let System Perform Placement.

13.  For this solution, we didn’t configure any vMedia Policy, click Next.

14.  In the Server Boot Order menu, select SAN_Boot for the Boot Policy which was created earlier and click Next.

Graphical user interface, text, application, emailDescription automatically generated

15.  The maintenance policy was not selected in this configuration, click Next.

16.  In Server Assignment menu, select “ORA19C-SERVER-POOL” as Pool Assignment which we created earlier and “all-chassis” option into Server Pool Qualification selection. In the Firmware Management option, click “4.1-3b” as Host Firmware Package which we created earlier. Click Next.

Graphical user interface, text, application, emailDescription automatically generated

17.  Select the BIOS Policy ORA19C_BIOS in the BIOS.

Graphical user interface, text, applicationDescription automatically generated

18.  Select Management IP Address and then click the Management IP Address Policy “Ora19c-KVM” which was created earlier.

19.  Click Finish to create Service Profile Template “ORA19C-FlexPod.”

Now you’ve created one Service profile template “ORA19C_FlexPod” having four vHBAs and two vNICs. This service profile template will be used to create eight service profiles for eight oracle RAC nodes as FLEX1, FLEX2, FLEX3, FLEX4, FLEX5, FLEX6, FLEX7 and FLEX8 as explained in the next section.

Create Service Profiles from Template and Associate to Servers

Create Service Profiles from Template

*       We created eight Service profiles for all eight Oracle RAC nodes as explained below.

For all eight Linux Oracle RAC Nodes (flex1, flex2, flex3, flex4, flex5, flex6, flex7 and flex8), you will create eight Service Profiles as FLEX1, FLEX2, FLEX3, FLEX4, FLEX5, FLEX6, FLEX7 and FLEX8 from the template “ORA19C_FlexPod.”

To create Service Profiles from Template, follow these steps:

1.    Go to Servers > Service Profiles > root > and right-click Create Service Profiles from Template.

2.    Select the Service profile template ORA19C_FlexPod, previously created and name the service profile “FLEX.”

3.    To create eight service profiles, enter Number of Instances as 8. This process will create service profiles as FLEX1, FLEX2, FLEX3, FLEX4, FLEX5, FLEX6, FLEX7 and FLEX8

 

Graphical user interface, applicationDescription automatically generated

4.    When the service profiles are created, associate them to the servers as described in the following section.

Associate Service Profiles to the Servers

To associate service profiles to the servers, follow these steps:

1.    Under the server tab, right-click the name of service profile you want to associate with the server and select the option "Change Service Profile Association.”

2.    In the Change Service Profile Association page, from the Server Assignment drop-down list, select the existing server that you would like to assign, and click OK.

C:\Users\havyas\Desktop\Screenshot 2018-06-18 15.24.54.png

You have assigned, service profiles FLEX1 to Chassis 1 Server 1, service profile FLEX2 to Chassis 1 Server 2, service profiles FLEX3 to Chassis 1 Server 3 and service profiles FLEX4 to Chassis 1 Server 4.

Similarly, you have assigned, service profiles FLEX5 to Chassis 2 Server 1, service profile FLEX6 to Chassis 2 Server 2, service profiles FLEX7 to Chassis 2 Server 3 and service profiles FLEX8 to Chassis 2 Server 4.

1.    Make sure all the service profiles are associated.

2.    As shown above, make sure all the server nodes have no major or critical fault and all are in operable state as well.

This completes the configuration required for Cisco UCS Manager Setup.

*       Additional server pools, service profile templates, and service profiles can be created in the respective organizations to add more servers to the FlexPod unit. All other pools and policies are at the root level and can be shared among the organizations.

Configure Cisco MDS Switch

This section provides a detailed procedure for configuring the Cisco MDS 9132T Switches.

*       Follow these steps precisely because failure to do so could result in an improper configuration.

Graphical user interface, diagramDescription automatically generated

We connected the MDS Switches to Fabric Interconnects and NetApp AFF A800 Storage System as shown below.

DiagramDescription automatically generated

For this solution, we connected four ports (ports 1 to 4) of MDS Switch A to Fabric Interconnect A (ports 1-4). Similarly, we connected four ports (ports 1 to 4) of MDS Switch B to Fabric Interconnect B (ports 1-4). All ports carry 32 Gb/s FC Traffic. Table 9 lists the port connectivity of the Cisco MDS Switches to the Fabric Interconnects.

Table 9.          MDS Switch Connectivity to the Fabric Interconnects

MDS Switch

MDS Switch Port

FI Ports

Fabric Interconnect

MDS Switch A

FC Port 1/1

FI-A Port 1/1

Fabric Interconnect A (FI-A)

FC Port 1/2

FI-A Port 1/2

FC Port 1/3

FI-A Port 1/3

FC Port 1/4

FI-A Port 1/4

MDS Switch B

FC Port 1/1

FI-B Port 1/1

Fabric Interconnect B (FI-B)

FC Port 1/2

FI-B Port 1/2

FC Port 1/3

FI-B Port 1/3

FC Port 1/4

FI-B Port 1/4

For this solution, we connected four ports (ports 5 to 8) of MDS Switch A to the NetApp AFF A800 Storage controller. Similarly, we connected four ports (ports 5 to 8) of MDS Switch B to the NetApp AFF A800 Storage controller. All ports carry 32 Gb/s FC Traffic. Table 10 lists the port connectivity of the Cisco MDS Switches to NetApp AFF A800 Controller.

Table 10.       MDS Switch Connectivity to the NetApp AFF A700s Storage

MDS Switch

MDS Switch Port

NetApp Storage Controller

NetApp Controller Ports

NetApp Port Description

MDS Switch A

FC Port 1/5

NetApp AFF A800 Controller 1

FC Port – 2a

FlexPod-A800-CT1 – Port 2a

FC Port 1/6

NetApp AFF A800 Controller 1

FC Port – 2b

FlexPod-A800-CT1 – Port 2b

FC Port 1/7

NetApp AFF A800 Controller 2

FC Port – 2a

FlexPod-A800-CT2 – Port 2a

FC Port 1/8

NetApp AFF A800 Controller 2

FC Port – 2b

FlexPod-A800-CT2 – Port 2b

MDS Switch B

FC Port 1/5

NetApp AFF A800 Controller 1

FC Port – 2c

FlexPod-A800-CT1 – Port 2c

FC Port 1/6

NetApp AFF A800 Controller 1

FC Port – 2d

FlexPod-A800-CT1 – Port 2d

FC Port 1/7

NetApp AFF A800 Controller 2

FC Port – 2c

FlexPod-A800-CT2 – Port 2c

FC Port 1/8

NetApp AFF A800 Controller 2

FC Port – 2d

FlexPod-A800-CT2 – Port 2d

Configure features

To configure feature on MDS Switches, follow this step:

1.    Login as admin user into the MDS Switch A and MDS Switch B and run the following commands:

config terminal

feature npiv

feature fport-channel-trunk

copy running-config startup-config

Configure VSANs and Ports

To create VSANs, follow these steps:

1.    Login as admin user into MDS Switch A.

2.    Create VSAN 151 for Storage Traffic and configure ports by running the following commands:

config terminal

vsan database

vsan 151

vsan 151 name "VSAN-FI-A"

vsan 151 interface fc 1/1-12

zone smart-zoning enable vsan 151

exit

 

interface port-channel251

  switchport trunk allowed vsan 151

  switchport description ORA19C-FlexPod-FI-A

  switchport rate-mode dedicated

  switchport trunk mode off

  no shutdown

 

interface fc1/1

  switchport description ORA19C-FlexPod-FI-A-1/1

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

 

interface fc1/2

  switchport description ORA19C-FlexPod-FI-A-1/2

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

 

interface fc1/3

  switchport description ORA19C-FlexPod-FI-A-1/3

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

 

interface fc1/4

  switchport description ORA19C-FlexPod-FI-A-1/4

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

 

interface fc1/5

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-01-2a

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/6

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-01-2b

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/7

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-02-2a

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/8

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-02-2b

  switchport trunk mode off

  port-license acquire

  no shutdown

copy running-config startup-config

3.    Login as admin user into MDS Switch B.

4.    Create VSAN 152 for Storage Traffic and configure ports by running the following commands:

config terminal

vsan database

vsan 152

vsan 152 name "VSAN-FI-B"

vsan 152 interface fc 1/1-12

zone smart-zoning enable vsan 152

exit

 

interface port-channel252

  switchport trunk allowed vsan 152

  switchport description ORA19C-FlexPod-FI-B

  switchport rate-mode dedicated

  switchport trunk mode off

 

interface fc1/1

  switchport description ORA19C-FlexPod-FI-B-1/1

  switchport trunk mode off

  port-license acquire

  channel-group 252 force

  no shutdown

 

interface fc1/2

  switchport description ORA19C-FlexPod-FI-B-1/2

  switchport trunk mode off

  port-license acquire

  channel-group 252 force

  no shutdown

 

interface fc1/3

  switchport description ORA19C-FlexPod-FI-B-1/3

  switchport trunk mode off

  port-license acquire

  channel-group 252 force

  no shutdown

 

interface fc1/4

  switchport description ORA19C-FlexPod-FI-B-1/4

  switchport trunk mode off

  port-license acquire

  channel-group 252 force

  no shutdown

 

interface fc1/5

  switchport trunk allowed vsan 152

  switchport description FlexPod-A800-01-2c

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/6

  switchport trunk allowed vsan 152

  switchport description FlexPod-A800-01-2d

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/7

  switchport trunk allowed vsan 152

  switchport description FlexPod-A800-02-2c

  switchport trunk mode off

  port-license acquire

  no shutdown

 

interface fc1/8

  switchport trunk allowed vsan 152

  switchport description FlexPod-A800-02-2d

  switchport trunk mode off

  port-license acquire

  no shutdown

Configure Zoning

This procedure sets up the Fibre Channel connections between the Cisco MDS 9132T switches, the Cisco UCS Fabric Interconnects, and the NetApp AFF Storage systems. Before you configure the zoning details, decide how many paths are needed for each LUN and extract the WWPN numbers for each of the HBAs from each server.

For this solution, we created 4 vHBAs on each server node. As listed in Table 3 and into UCS Service Profile Template configuration section, we have configured vHBA0 and vHBA1 for carrying FC Network Traffic & Boot from SAN through MDS-A and MDS-B Switch. We have configured vHBA2 and vHBA3 for NVMe/FC Network Traffic (Oracle RAC Storage Traffic) through MDS-A and MDS-B Switch.

To create and configure the Fibre channel zoning, follow these steps:

1.    Log into the Cisco UCS Manager > Equipment > Chassis > Servers and select the desired server. On the right-hand menu, click the Inventory tab and HBA's tab to get the WWPN of HBA's as shown below:

A screenshot of a computerDescription automatically generated with medium confidence

2.    Login into the NetApp storage controller and extract the WWPN of FC LIFs configured and verify all the port information is correct. This information can be found in the NetApp Storage GUI under Network > Network Interfaces.

Now you have configured 2 SVMs. One SVM named “Infra-SVM” was configured to carry FC network traffic for SAN Boot while the second SVM, named “ORA19C-SVM” was configured to run NVMe/FC Network Traffic for Oracle RAC Databases. The screenshot below shows the allowed protocols configured for both SVMs.

A screenshot of a computerDescription automatically generated with medium confidence

The screenshot below shows the network interface, WWPN and ports connectivity configured for NetApp AFF A800 Storage Controller.

For SVM “Infra-SVM”, four FC Logical Interfaces (LIFs) are created on storage controller cluster node 1 (node1_lif02a, node1_lif02b, node1_lif02c and node1_lif02d) and four Fibre Channel LIFs are created on storage controller cluster node 2 (node2_lif02a, node2_lif02b, node2_lif02c and node2_lif02d).

For SVM “ORA19C-SVM”, two NVMe Logical Interfaces (LIFs) are created on storage controller cluster node 1 (node1_lif02a and node1_lif02c) and two NVMe LIFs are created on storage controller cluster node 2 (node2_lif02a and node2_lif02c).

Graphical user interface, application, table, ExcelDescription automatically generated

*       You can also obtain this information by login to the storage cluster and run the network interface show command.

Create Device Aliases for Zoning on MDS Switch A

To configure device aliases and zones for FC and NVMe/FC Network data paths on MDS switch A, follow these steps:

1.    Login as admin user and run the following commands MDS switch A:

configure terminal

device-alias database

  device-alias name Flex1-hba0 pwwn 20:00:00:25:b5:99:aa:00

  device-alias name Flex1-hba2 pwwn 20:00:00:25:b5:99:aa:01

  device-alias name Flex2-hba0 pwwn 20:00:00:25:b5:99:aa:02

  device-alias name Flex2-hba2 pwwn 20:00:00:25:b5:99:aa:03

  device-alias name Flex3-hba0 pwwn 20:00:00:25:b5:99:aa:04

  device-alias name Flex3-hba2 pwwn 20:00:00:25:b5:99:aa:05

  device-alias name Flex4-hba0 pwwn 20:00:00:25:b5:99:aa:06

  device-alias name Flex4-hba2 pwwn 20:00:00:25:b5:99:aa:07

  device-alias name Flex5-hba0 pwwn 20:00:00:25:b5:99:aa:08

  device-alias name Flex5-hba2 pwwn 20:00:00:25:b5:99:aa:09

  device-alias name Flex6-hba0 pwwn 20:00:00:25:b5:99:aa:0a

  device-alias name Flex6-hba2 pwwn 20:00:00:25:b5:99:aa:0b

  device-alias name Flex7-hba0 pwwn 20:00:00:25:b5:99:aa:0c

  device-alias name Flex7-hba2 pwwn 20:00:00:25:b5:99:aa:0d

  device-alias name Flex8-hba0 pwwn 20:00:00:25:b5:99:aa:0e

  device-alias name Flex8-hba2 pwwn 20:00:00:25:b5:99:aa:0f

  device-alias name FlexPod-A800-01-2a pwwn 20:01:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-01-2b pwwn 20:02:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2a pwwn 20:05:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2b pwwn 20:06:00:a0:98:b9:25:08

  device-alias name A800-NVMe-01-2a pwwn 20:14:00:a0:98:b9:25:08

  device-alias name A800-NVMe-02-2a pwwn 20:16:00:a0:98:b9:25:08

device-alias commit

copy run start

Create Device Aliases for Zoning on MDS Switch B

To configure device aliases and zones for the FC and NVMe/FC Network data paths on MDS switch B, follow these steps:

1.    Login as admin user and run the following commands on MDS switch B:

configure terminal

device-alias database

  device-alias name Flex1-hba1 pwwn 20:00:00:25:b5:99:bb:00

  device-alias name Flex1-hba3 pwwn 20:00:00:25:b5:99:bb:01

  device-alias name Flex2-hba1 pwwn 20:00:00:25:b5:99:bb:02

  device-alias name Flex2-hba3 pwwn 20:00:00:25:b5:99:bb:03

  device-alias name Flex3-hba1 pwwn 20:00:00:25:b5:99:bb:04

  device-alias name Flex3-hba3 pwwn 20:00:00:25:b5:99:bb:05

  device-alias name Flex4-hba1 pwwn 20:00:00:25:b5:99:bb:06

  device-alias name Flex4-hba3 pwwn 20:00:00:25:b5:99:bb:07

  device-alias name Flex5-hba1 pwwn 20:00:00:25:b5:99:bb:08

  device-alias name Flex5-hba3 pwwn 20:00:00:25:b5:99:bb:09

  device-alias name Flex6-hba1 pwwn 20:00:00:25:b5:99:bb:0a

  device-alias name Flex6-hba3 pwwn 20:00:00:25:b5:99:bb:0b

  device-alias name Flex7-hba1 pwwn 20:00:00:25:b5:99:bb:0c

  device-alias name Flex7-hba3 pwwn 20:00:00:25:b5:99:bb:0d

  device-alias name Flex8-hba1 pwwn 20:00:00:25:b5:99:bb:0e

  device-alias name Flex8-hba3 pwwn 20:00:00:25:b5:99:bb:0f

  device-alias name FlexPod-A800-01-2c pwwn 20:03:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-01-2d pwwn 20:04:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2c pwwn 20:07:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2d pwwn 20:08:00:a0:98:b9:25:08

  device-alias name A800-NVMe-01-2c pwwn 20:15:00:a0:98:b9:25:08

  device-alias name A800-NVMe-02-2c pwwn 20:21:00:a0:98:b9:25:08

device-alias commit

copy run start

For each of the SVM (Infra-SVM and ORA19C-SVM) and its corresponding WWPN, you will create its individual zoning (FC Zoning for Boot and NVMe/FC Zoning for NVMe/FC network traffic).

Create Zoning for Boot

To configure the SAN Boot on each node, configure the zoning on both MDS switches as detailed in the following sections.

Cisco MDS Switch A

To configure the zones on MDS Switch A, follow these steps:

1.    Login as admin user.

2.    Create the zones for each server:

configure terminal

zone name Flex1A-Boot vsan 151

    member device-alias Flex1-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex2A-Boot vsan 151

    member device-alias Flex2-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex3A-Boot vsan 151

    member device-alias Flex3-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex4A-Boot vsan 151

    member device-alias Flex4-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex5A-Boot vsan 151

    member device-alias Flex5-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex6A-Boot vsan 151

    member device-alias Flex6-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex7A-Boot vsan 151

    member device-alias Flex7-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex8A-Boot vsan 151

    member device-alias Flex8-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

3.    Create Zoneset and add all the members:

zoneset name Flex-A vsan 151

    member Flex1A-Boot

    member Flex2A-Boot

    member Flex3A-Boot

    member Flex4A-Boot

    member Flex5A-Boot

    member Flex6A-Boot

    member Flex7A-Boot

    member Flex8A-Boot

4.    Activate the Zoneset and save the configuration:

zoneset activate name Flex-A vsan 151

copy run start

Cisco MDS Switch B

To configure zones on MDS Switch B, follow these steps:

1.    Login as admin user.

2.    Create the zones for each server:

configure terminal

zone name Flex1B-Boot vsan 152

    member device-alias Flex1-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex2B-Boot vsan 152

    member device-alias Flex2-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex3B-Boot vsan 152

    member device-alias Flex3-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex4B-Boot vsan 152

    member device-alias Flex4-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex5B-Boot vsan 152

    member device-alias Flex5-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex6B-Boot vsan 152

    member device-alias Flex6-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex7B-Boot vsan 152

    member device-alias Flex7-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

zone name Flex8B-Boot vsan 152

    member device-alias Flex8-hba1 init

    member device-alias FlexPod-A800-01-2c target

    member device-alias FlexPod-A800-02-2c target

3.    Create Zoneset and add all the members:

zoneset name Flex-B vsan 152

    member Flex1B-Boot

    member Flex2B-Boot

    member Flex3B-Boot

    member Flex4B-Boot

    member Flex5B-Boot

    member Flex6B-Boot

    member Flex7B-Boot

    member Flex8B-Boot

4.    Activate the Zoneset and save the configuration.

zoneset activate name Flex-B vsan 152

copy run start

Create Zoning for NVMe/FC

To configure the NVMe/FC on each node, configure the zoning on both MDS switches as detailed in the following sections.

Cisco MDS Switch A

To configure NVMe/FsC zones on MDS Switch A, follow these steps:

1.    Login as admin user.

2.    Create the zones for each server:

configure terminal

zone name Flex1A-NVMe vsan 151

    member device-alias Flex1-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex2A-NVMe vsan 151

    member device-alias Flex2-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex3A-NVMe vsan 151

    member device-alias Flex3-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex4A-NVMe vsan 151

    member device-alias Flex4-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex5A-NVMe vsan 151

    member device-alias Flex5-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex6A-NVMe vsan 151

    member device-alias Flex6-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex7A-NVMe vsan 151

    member device-alias Flex7-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex8A-NVMe vsan 151

    member device-alias Flex8-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

3.    Add all the members into Zoneset:

zoneset name Flex-A vsan 151

    member Flex1A-NVMe

    member Flex2A-NVMe

    member Flex3A-NVMe

    member Flex4A-NVMe

    member Flex5A-NVMe

    member Flex6A-NVMe

    member Flex7A-NVMe

    member Flex8A-NVMe

4.    Activate the Zoneset and save the configuration.

zoneset activate name Flex-A vsan 151

copy run start

Cisco MDS Switch B

To configure NVMe/FC zones on MDS Switch B, follow these steps:

1.    Login as admin user.

2.    Create the zones for each server:

configure terminal

zone name Flex1B-NVMe vsan 152

    member device-alias Flex1-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex2B-NVMe vsan 152

    member device-alias Flex2-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex3B-NVMe vsan 152

    member device-alias Flex3-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex4B-NVMe vsan 152

    member device-alias Flex4-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex5B-NVMe vsan 152

    member device-alias Flex5-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex6B-NVMe vsan 152

    member device-alias Flex6-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex7B-NVMe vsan 152

    member device-alias Flex7-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

zone name Flex8B-NVMe vsan 152

    member device-alias Flex8-hba3 init

    member device-alias A800-NVMe-01-2c target

    member device-alias A800-NVMe-02-2c target

3.    Create Zoneset and add all the members.

zoneset name Flex-B vsan 152

    member Flex1B-NVMe

    member Flex2B-NVMe

    member Flex3B-NVMe

    member Flex4B-NVMe

    member Flex5B-NVMe

    member Flex6B-NVMe

    member Flex7B-NVMe

    member Flex8B-NVMe

4.    Activate the Zoneset and save the configuration.

zoneset activate name Flex-B vsan 152

copy run start

Verify FC Ports on MDS Switch

To verify the FC ports on the MDS switch, follow these steps:

1.    Login as admin user into MDS Switch A and run the “show flogi database vsan 151” to verify all FC ports.

TextDescription automatically generated

2.    Login as admin user into MDS Switch B and run the “show flogi database vsan 152” to verify all FC ports.

TextDescription automatically generated

This concludes both MDS Switch configurations.

Configure NetApp AFF A800 Storage

The high-level steps for configuring the NetApp Storage for this solution is shown below:

Graphical user interface, applicationDescription automatically generated

NetApp Storage Connectivity

*       It is beyond the scope of this document to explain the detailed information about the NetApp storage connectivity and infrastructure configuration. For installation and setup instruction for the NetApp AFF A800 System, go to: https://docs.netapp.com/platstor/index.jsp?topic=%2Fcom.netapp.doc.hw-a800-install-setup%2FGUID-91FA78D3-A39E-451D-BB17-6476972A0716.html

For more information, go to the Cisco site: https://www.cisco.com/c/en/us/solutions/design-zone/data-center-design-guides/flexpod-design-guides.html

This section describes the storage layout and design considerations for the database deployment. The screenshot below shows the SVM (formally known as Vserver) and FC Interfaces configuration. For this solution, we configured two SVM. The first SVM named as “Infra-SVM” was configured to carry FC network traffic for SAN Boot while the second SVM named as “ORA19C-SVM” was configured to run NVMe/FC Network Traffic for Oracle RAC Databases. The screenshot below shows the allowed protocols configured for both the SVMs.

A screenshot of a computerDescription automatically generated with medium confidence

For the FC SVM (Infra-SVM), four FC Logical Interfaces (LIFs) are created on storage controller cluster node 1 (node1_lif02a, node1_lif02b, node1_lif02c and node1_lif02d) and four Fibre Channel LIFs are created on storage controller cluster node 2 (node2_lif02a, node2_lif02b, node2_lif02c and node2_lif02d) as shown below.

Graphical user interface, text, applicationDescription automatically generated

For the NVMe SVM (ORA19C-SVM), two NVMe Logical Interfaces (LIFs) are created on storage controller cluster node 1 (node1_lif02a and node1_lif02c) and two NVMe LIFs are created on storage controller cluster node 2 (node2_lif02a and node2_lif02c).

Graphical user interface, text, applicationDescription automatically generated

As shown above, for both the storage controller nodes (FlexPod-A800-CT1 and FlexPod-A800-CT2), we have used ports 2a, 2b and 2c, 2d to configure LIFs. WWPN of these LIFs are used for zoning into the MDS switches for storage to MDS connectivity as explained earlier.

For all the database deployment, we have configured two aggregates (one aggregate on each storage node) and each aggregate contains 12 SSD (1.75 TB Each) drives that were subdivided into RAID DP groups as shown below.

Graphical user interface, text, emailDescription automatically generated

We created 8 LUNs for SAN boot and mapped those with the corresponding Linux host initiators to boot from SAN. For database deployment, we created multiple subsystem and namespaces. We also distributed equal number of subsystems on the storage controller by placing those into the aggregate equally. We will explain the subsystem configuration in the database creation section.

Operating System and Database Deployment

The design goal of the reference architecture was to best represent a real-world environment as closely as possible. As explained in the earlier section, service profile was created within Cisco UCS Manager to rapidly deploy all the stateless servers to deploy an eight node Oracle RAC. SAN boot LUNs for these servers were hosted on the NetApp Storage Cluster to provision the OS on top it. Zoning was performed on the Cisco MDS switches to enable the initiators discover the targets during boot process.

Each Server node has a dedicated single LUN to install operating system. For this solution, we have installed Oracle Linux Server 8.2 (RHCK 4.18.0-193.el8.x86_64) on these LUNs and configured NVMe/FC connectivity with all the prerequisite packages for installing Oracle Software to create an eight node Oracle Multitenant RAC 19c database solution.

The high-level steps to configure Linux Hosts and deploy the Oracle RAC Database solution is shown below:

DiagramDescription automatically generated

Configure OS

*       The detailed installation process is not contained in this document, but the following section describes the key steps for OS installation.

To Configure OS, follow these steps:

1.    Download the Oracle Linux 8.2 OS image from https://edelivery.oracle.com/linux

2.    Launch KVM console on desired server by going to tab Equipment > Chassis > Chassis 1 > Servers > Server 1 > from right side windows General > and select KVM Console to open KVM.

Graphical user interface, applicationDescription automatically generated

3.    Click Accept security and open KVM. Activate Virtual Devices and map the Oracle Linux ISO image from the top right corner menu options, and then reboot the server.

4.    When the Server starts booting, it will detect the NetApp Storage active FC paths as shown below. If you see the following message in the KVM console while the server is rebooting along with the target WWPNs, it confirms the setup and zoning is done correctly and boot from SAN will be successful.

TextDescription automatically generated

5.    During server boot order, it will detect the virtual media connected as Oracle Linux ISO DVD media and it should launch the Oracle Linux OS installer. Select language and assign the Installation destination as NetApp Storage LUN. Apply hostname and click “Configure Network” to configure all network interfaces. Alternatively, you can only configure “Public Network” in this step. You can configure additional interfaces as part of post OS install steps.

6.    As a part of additional RPM package, we recommend selecting “Customize Now” option and relevant packages according to your environment.

7.    After the OS install finishes, reboot the server, and complete the appropriate registration steps. You can choose to synchronize the time with ntp server. Alternatively, you can choose to use Oracle RAC cluster synchronization daemon (OCSSD). Both ntp and OCSSD are mutually exclusive and OCSSD will be setup during GRID install if ntp is not configured.

Set Default Kernel to RHCK

Oracle Linux 8 Update 2 ships with two kernels:

   UEK R6 (kernel-uek-5.4.17-2011.1.2.el8uek) for x86_64 (Intel & AMD) and aarch64 (Arm) platform

   RHCK (kernel-4.18.0-193.el8) for x86_64 (Intel & AMD) platform

After installing Oracle Linux 8.2 on all the server nodes (flex1, flex2, flex3, flex4, fle5, flex6, flex7 and flex8), to configure the default kernel to RHCK, follow these steps:

1.    Check the list of installed kernel:

[root@flex1 ~]# ls -al /boot/vmlinuz*

lrwxrwxrwx  1 root root      32 Feb 21 23:43 /boot/vmlinuz -> /boot/vmlinuz-4.18.0-193.el8.gv+

-rwxr-xr-x. 1 root root 9226480 Oct 15 17:07 /boot/vmlinuz-0-rescue-8ca6c4cb17ac4fa5aad3028bfb737418

-rwxr-xr-x. 1 root root 9226480 Apr 29  2020 /boot/vmlinuz-4.18.0-193.el8.x86_64

-rwxr-xr-x. 1 root root 8923392 Apr 20  2020 /boot/vmlinuz-5.4.17-2011.1.2.el8uek.x86_64

2.    Set the default kernel and reboot the node:

[root@flex1 ~]# grubby --set-default=/boot/vmlinuz-4.18.0-193.el8.x86_64

[root@flex1 ~]# systemctl reboot

3.    After the node reboots, verify the default kernel boot:

[root@flex1 ~]# grubby --default-kernel

/boot/vmlinuz-4.18.0-193.el8.x86_64

4.    Repeat steps 1-3 and configure the RHCL as the default kernel boot on all the nodes.

Configure Public and Private Network Interfaces

If you have not configured network interfaces during OS installation, then configure it now. Each node must have at least two network interface or network adapters. One network interface is for the public network traffic and the second interface is for the private network traffic (the node interconnects). The server nodes will access FC and NVMe/FC network traffic through vHBA.

To configure public and private network interfaces, follow these steps:

1.    Login as a root user into each node and go to /etc/sysconfig/network-scripts/

2.    Configure Public network and Private network IP addresses according to your environments.

*       Configure the Private and Public network with the appropriate IP addresses on all the Oracle RAC nodes

Install ENIC and FNIC Linux OS Driver

For this solution, we configured the following ENIC and FNIC drivers:

   ENIC: version:        4.0.0.14-802.74

   FNIC: version:        2.0.0.69-178.0

To install the ENIC and FNIC linux OS driver, follow these steps:

1.    Download the supported UCS Linux Drivers “ucs-bxxx-drivers-linux.4.1.3b.iso” from this link: https://software.cisco.com/download/home/283853163/type/283853158/release/4.1(3b)

2.    Mount the Driver ISO to Linux Host KVM and Install the relevant supported ENIC and FNIC driver for the Linux OS.

3.    Check the current ENIC and FNIC version:

[root@flex1 ~]# cat /sys/module/enic/version

[root@flex1 ~]# cat /sys/module/fnic/version

[root@flex1 ~]# rpm -qa | grep enic

[root@flex1 ~]# rpm -qa | grep fnic

4.    Install the supported ENIC and FNIC driver from RPM:

[root@flex1 software]# rpm -ivh kmod-enic-4.0.0.14-802.74.rhel8u2.x86_64.rpm

[root@flex1 software]# rpm -ivh kmod-fnic-2.0.0.69-178.0.rhel8u2.x86_64.rpm

5.    Reboot the server and verify that the new driver is running:

[root@flex1 ~]# rpm -qa | grep enic

kmod-enic-4.0.0.14-802.74.rhel8u2.x86_64

 

[root@flex1 ~]# rpm -qa | grep fnic

kmod-fnic-2.0.0.69-178.0.rhel8u2.x86_64

 

[root@flex1 ~]# modinfo enic | grep version

version:        4.0.0.14-802.74

rhelversion:    8.2

srcversion:     7C8E065228B97E368868B1A

vermagic:       4.18.0-193.el8.x86_64 SMP mod_unload modversions

 

[root@flex1 ~]# modinfo fnic | grep version

version:        2.0.0.69-178.0

rhelversion:    8.2

srcversion:     2133F5E0E629F689B9695F8

vermagic:       4.18.0-193.el8.x86_64 SMP mod_unload modversions

6.    Repeat steps 1-5 and configure the linux drivers on all the nodes.

*       You should use a matching ENIC and FNIC pair. Check the Cisco UCS supported driver release for more information about the supported kernel version; https://www.cisco.com/c/en/us/support/docs/servers-unified-computing/ucs-manager/116349-technote-product-00.html

Configure NVMe/FC for Linux OS

The 4 vHBA are configured on each of the Linux host. HBA-0 and HBA-1 are for FC SAN Boot while HBA-2 and HBA-3 are for carrying NVMe/FC Oracle RAC Storage Traffic. This section describes the high-level steps to configure and enable NVMe/FC on one of the Linux Host.

*       Repeat the following steps and configure the NVMe/FC for OS on all the nodes.

Install NVME CLI

NVME host discovery of NVME targets is triggered by NVME CLI that makes system call into NVME Linux core kernel module to discover NVME namespaces behind an NVME target. This CLI is part of “nvme-cli-1.*.x86_64.rpm” that can be installed from OS Installer DVD. In order to enable namespace discovery, the nvme CLI rpm has to be installed, following the installation of FNIC driver.

To install the NVME CLI, follow this step:

1.    Install “nvme-cli” tool rpm on each node:

[root@flex1 ~]# rpm -ivh nvme-cli-1.9-5.el8.x86_64

[root@flex1 ~]# rpm -qa nvme-cli

nvme-cli-1.9-5.el8.x86_64

Set Up Host NQN

NVME hosts and targets are distinguished through their NQN. Fnic NVME host reads its host nqn from the file /etc/nvme/hostnqn. Upon successful installation of the nvme-cli package, the hostnqn file will be created automatically for some OS versions, such as RHEL.

*       If the /etc/nvme/hostnqn file is not present after name-cli installed, then create the file manually

To set up the host NQN, follow these steps:

1.    Generate hostnqn through uuidgen script:

[root@flex1 ~]# uuidgen

c9efe809-d3d0-4028-9018-c1a648cd40c4

2.    Append the unique UUID to NVME standard format of hostnqn:

[root@flex1 ~]# echo “nqn.2014-08.org.nvmexpress:uuid:c9efe809-d3d0-4028-9018-c1a648cd40c4” > /etc/nvme/hostnqn

Setup Native Multipathing

For this solution, we enabled and configured “Native Multipathing” for NVMe/FC which is provided by nvme-core. To enable native multipathing provided by nvme-core, follow these steps:

1.    Enable multipathing on host by running the below command:

/sbin/mpathconf --enable

2.    Setup GRUB to boot the kernel with NVME-multipath support:

grubby --args=nvme_core.multipath=Y --update-kernel /boot/vmlinuz-4.18.0-193.el8.x86_64

3.    Save the file and reboot the host:

Set Up NVMe Retry Scheme

The existing NVME core may sometimes fail to discover targets/namespaces due to timing issues. In order to make this operation fail-safe fnic automatic connect utility implements retry mechanism to retry target and namespace discovery through NVME core. To configure the retry logic, follow these steps:

1.    Configure Number of retry attempts:

echo "5" > /etc/nvme/connect_retry_attempts

2.    Configure Delay between each retry attempt (in seconds):

echo "10" > /etc/nvme/connect_retry_delay

Load FNIC Driver

To load the FNIC driver module and verify it,  run the following commands on the Linux host:

modprobe fnic

/sbin/lsmod | grep fnic

Related image, diagram or screenshot

*       When the modules are installed, they should come up automatically on system reboot.

After installing and configuring fnic driver, configure the storage array for NVMe/FC as explained in next section.

Configure Storage NVMe Subsystem

To configure subsystem on storage, follow these steps:

1.    Login as admin user into NetApp Storage Array and go to > Storage > NVMe > Subsystems > and click +Create.

Graphical user interface, text, application, emailDescription automatically generated

2.    For this solution, configure four Subsystem on NVMe SVM “ORA19C-SVM” as “orasub1”, “orasub2”, “orasub3” and “orasub4”. On each of the subsystem, select Host OS as “Linux” and add all the eight-host unique “hostnqn” as shown below.

Graphical user interfaceDescription automatically generated

The overview of all the subsystems is shown below:

Graphical user interface, textDescription automatically generated

You will configure the NVMe Namespaces and associate them to these four subsystems later in the database creation section.

Discover Fabric Ports and Targets from OS

After configuring the storage NVMe Subsystem, connect the Linux hosts to NVMe Storage target as explained below.

*       “fcc” is a tool that is packaged with the driver and can be used to list the FC HBAs and discovered remote ports and luns. Running the fcc without any arguments lists all FC (scsi) hosts, remote ports, and LUNs as sown below:

A black screen with white textDescription automatically generated with low confidence

1.    To list NVME hosts and NVME targets discovered by each host (FC-NVME only), run the following commands:

A picture containing text, black, screenshotDescription automatically generated

2.    The FNIC driver automatically discovers and connects to NVME controllers. NVME storage controllers are automatically connected, and namespaces are discovered. User may also trigger manually to connect to NVME namespaces using “fcc connect all” command. If NVME storage connectivity changes during run-time, NVME controller discovery can also be manually triggered using “fcc discover all” command to discover any target changes. The following command can be used to check the available paths for an NVME sub system:

Graphical user interface, textDescription automatically generated

This concludes the NVMe/FC setup for the first Linux host. Repeat steps 1 and 2 on each remaining Linux hosts to enable and configure NVMe/FC storage connectivity.

Next, configure the operating system prerequisites to install Oracle Grid and Oracle Database Software as explained in the following sections.

Configure BIOS

This section describes how to optimize the BIOS settings to meet requirements for the best performance and energy efficiency for the Cisco UCS M5 generation of blade servers.

For more information about BIOS settings, refer to: https://www.cisco.com/c/dam/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/whitepaper_c11-740098.pdf

*       For this solution, we configured the BIOS settings as described in section Configure Server BIOS Policy. Apply these settings according to your environments.

Disable C-States for OLTP Workloads

OLTP systems are often decentralized to avoid single points of failure. Spreading the work over multiple servers can also support greater transaction processing volume and reduce response time. Make sure to disable Intel IDLE driver in the OS configuration section. When Intel idle driver is disabled, the OS uses acpi_idle driver to control the C-States.

*       For latency sensitive workloads, it is recommended to always disable c-states in both OS and BIOS to ensure c-states are disabled

If the CPU gets into a deeper C-state and not able to get out to deliver full performance quickly, then result is unwanted latency spikes for workloads. To address this, it is recommended to disable C states in the BIOS and in addition, Oracle recommends disabling it from OS level as well by modifying grub entries. To configure the BIOS options by modifying in the “/etc/default/grub” file, run the following commands:

[root@flex1 ~]$ cat /etc/default/grub

GRUB_TIMEOUT=5

GRUB_DISTRIBUTOR="$(sed 's, release .*$,,g' /etc/system-release)"

GRUB_DEFAULT=saved

GRUB_DISABLE_SUBMENU=true

GRUB_TERMINAL_OUTPUT="console"

GRUB_CMDLINE_LINUX="crashkernel=auto resume=/dev/mapper/ol-swap rd.lvm.lv=ol/root rd.lvm.lv=ol/swap rhgb quiet biosdevname=0 net.ifnames=0 nvme-core.multipath=Y numa=off transparent_hugepage=never intel_idle.max_cstate=0 processor.max_cstate=0"

GRUB_DISABLE_RECOVERY="true"

GRUB_ENABLE_BLSCFG=true

Configure OS Prerequisites for Oracle Software

To successfully install Oracle RAC Database 19c software, configure the operating system prerequisites on all eight nodes as explained in this section

*       Follow the steps according to your environment and requirements. For more information, see the Install and Upgrade Guide for Linux for Oracle Database 19C: https://docs.oracle.com/en/database/oracle/oracle-database/19/cwlin/configuring-operating-systems-for-oracle-grid-infrastructure-on-linux.html#GUID-B8649E42-4918-49EA-A608-446F864EB7A0

Prerequisites RPM Installation

To configure the operating system prerequisites using RPM for Oracle 19c software on all nodes, install the “oracle-database-preinstall-19c" rpm package. You can also download the required packages from http://public-yum.oracle.com/oracle-linux-7.html.

If you plan to use the “oracle-database-preinstall-19c" rpm package to perform all your prerequisite setup automatically, then login as root user and issue the following command on all the RAC nodes.

[root@orarac1 ~]# yum install oracle-database-preinstall-19c

*       If you have not used the " oracle-database-preinstall-19c " package, then you will have to manually perform the prerequisites tasks on all the nodes.

Additional Prerequisites Configuration

After configuring the automatic or manual prerequisites steps, you have to configure a few additional steps to complete the prerequisites for the Oracle database software installations on all eight nodes as described below.

Disable SELinux

As most of the Organizations might already be running hardware-based firewalls to protect their corporate networks, we disabled Security Enhanced Linux (SELinux) and the firewalls at the server level for this reference architecture.

You can set secure Linux to permissive by editing the "/etc/selinux/config" file, making sure the SELINUX flag is set as follows:

SELINUX=permissive

Disable Firewall

Check the status of the firewall by running following commands. (The status displays as active (running) or inactive (dead)). If the firewall is active / running, enter this command below to stop it:

systemctl status firewalld.service

systemctl stop firewalld.service

Also, to completely disable the firewalld service, so it does not reload when you restart the host machine, run the following command:

systemctl disable firewalld.service

Create the Grid User

Run this command to create a grid user:

useradd –u 54322 –g oinstall –G dba grid

Set the User Passwords

Run these commands to change the password for Oracle and Grid Users:

passwd oracle

passwd grid

Configure Multipath

For this solution, configure the DM-Multipath only for the FC Boot LUNs. For the NVMe/FC Storage path, you need to configure the native multipathing as explained previously.

*       You must configure “enable_foreign” in “/etc/multipath.conf” for dm-multipath to prevent dm-multipath from claiming NVMe/FC namespace devices. We recommend you use in-kernel NVMe multipath for ONTAP namespaces and dm-multipath for ONTAP LUNs.

*       For DM-Multipath Configuration and best practice, refer to NetApp Support: https://library.netapp.com/ecmdocs/ECMP1217221/html/GUID-34FA2578-0A83-4ED3-B4B3-8401703D65A6.html

To configure multipath, follow these steps:

1.    Add or modify the /etc/multipath.conf file accordingly to provide the alias name of each LUN id presented from NetApp Storage as shown below into all 8 nodes:

TextDescription automatically generated

2.    Run multipath –ll command to view all the LUN IDs and enter that wwid information accordingly on each node:

TextDescription automatically generated

*       Make sure the LUNs wwid address reflects the correct value for all eight nodes in “/etc/multipath.conf”. We made sure the multipathing packages were installed and enabled for automatic restart across reboots.

Configure UDEV rules for IO Policy

You need to configure the UDEV rules to assign permission and IO Policy in all Oracle RAC nodes to access the NetApp Storage LUNs and Namespaces. This includes the device details along with required permissions to enable grid and oracle user to have read/write privileges on these devices.

To configure the UDEV rules on all Oracle RAC Nodes, follow this step:

1.    Assign IO Policy by creating a new file named “71-nvme-iopolicy-netapp-ONTAP.rules” with the following entries on all the nodes:

[root@flex1 ~]# cat /etc/udev/rules.d/71-nvme-iopolicy-netapp-ONTAP.rules

### Enable round-robin for NetApp ONTAP

ACTION=="add", SUBSYSTEM=="nvme-subsystem", ATTR{model}=="NetApp ONTAP Controller", ATTR{iopolicy}="round-robin"

Configure /etc/hosts

To configure /etc/hosts, follow these steps:

1.    Login as a root user into the node and edit the “/etc/hosts” file.

2.    Provide the details for Public IP Address, Private IP Address, SCAN IP Address, and Virtual IP Address for all the nodes. Configure these settings in each Oracle RAC Nodes as shown below:

[root@flex1 ~]# cat /etc/hosts

127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4

###     Public IP       ###

10.29.134.71    flex1   flex1.ciscoucs.com

10.29.134.72    flex2   flex2.ciscoucs.com

10.29.134.73    flex3   flex3.ciscoucs.com

10.29.134.74    flex4   flex4.ciscoucs.com

10.29.134.75    flex5   flex5.ciscoucs.com

10.29.134.76    flex6   flex6.ciscoucs.com

10.29.134.77    flex7   flex7.ciscoucs.com

10.29.134.78    flex8   flex8.ciscoucs.com

 

###       Virtual IP           ###

10.29.134.79    flex1-vip       flex1-vip.ciscoucs.com

10.29.134.80    flex2-vip       flex2-vip.ciscoucs.com

10.29.134.81    flex3-vip       flex3-vip.ciscoucs.com

10.29.134.82    flex4-vip       flex4-vip.ciscoucs.com

10.29.134.83    flex5-vip       flex5-vip.ciscoucs.com

10.29.134.84    flex6-vip       flex6-vip.ciscoucs.com

10.29.134.85    flex7-vip       flex7-vip.ciscoucs.com

10.29.134.86    flex8-vip       flex8-vip.ciscoucs.com

 

###       Private IP           ###

10.10.10.71     flex1-priv      flex1-priv.ciscoucs.com

10.10.10.72     flex2-priv      flex2-priv.ciscoucs.com

10.10.10.73     flex3-priv      flex3-priv.ciscoucs.com

10.10.10.74     flex4-priv      flex4-priv.ciscoucs.com

10.10.10.75     flex5-priv      flex5-priv.ciscoucs.com

10.10.10.76     flex6-priv      flex6-priv.ciscoucs.com

10.10.10.77     flex7-priv      flex7-priv.ciscoucs.com

10.10.10.78     flex8-priv      flex8-priv.ciscoucs.com

 

###       SCAN IP              ###

10.29.134.87    flex-scan       flex-scan.ciscoucs.com

10.29.134.88    flex-scan       flex-scan.ciscoucs.com

10.29.134.89    flex-scan       flex-scan.ciscoucs.com

You must configure the following addresses manually in your corporate setup:

   A Public IP Address for each node

   A Virtual IP address for each node

   Three single client access name (SCAN) address for the oracle database cluster

All the steps above were performed on all of the eight nodes. These steps complete the prerequisite for Oracle Database 19c Installation at OS level on Oracle RAC Nodes.

Configure NetApp Storage Host Group and LUNs for OCR and Voting Disk

To create and configure the NVMe Namespaces for storing OCR and Cluster Files, follow these steps:

1.    Login as Admin user into the NetApp array and go to Storage > NVMe > NVMe Namespaces > and click +Create.

Graphical user interface, text, application, emailDescription automatically generated

*       For this solution, we created two namespaces. Namespace “asm1” was configured on “orasub1” and namespace “asm2” was configured on “orasub2” with each namespace was 100 GB in size as shown in above screenshot for storing OCR and Voting Disk files for all the RAC databases. Also, each namespace was spread across both the aggregate.

2.    When all the above O.S level prerequisites and namespaces are configured, install the Oracle Grid Infrastructure as a grid user. Download Oracle Database 19c Release (19.3) for Linux x86-64 and Oracle Database 19c Release Grid Infrastructure (19.3) for Linux x86-64 software from Oracle Software site. Copy these software binaries to Oracle RAC Node 1 and Unzip all files into appropriate directories.

These steps complete the prerequisite for Oracle Database 19c Installation at OS level on Oracle RAC Nodes.

Oracle Database 19c GRID Infrastructure Deployment

This section describes the high-level steps for the Oracle Database 19c RAC installation. This document provides a partial summary of details that might be relevant.

*       It is not within the scope of this document to include the specifics of an Oracle RAC installation; you should refer to the Oracle installation documentation for specific installation instructions for your environment. For more information, use this link for Oracle Database 19c install and upgrade guide: https://docs.oracle.com/en/database/oracle/oracle-database/19/cwlin/index.html

For this solution, you will install the Oracle Grid and Database software on all the eight nodes (flex1 to flex8).

Oracle 19c Release 19.3 Grid Infrastructure (GI) was installed on the first node as a grid user. The installation also configured and added the remaining 7 nodes as a part of the GI setup. The Oracle Automatic Storage Management (ASM) in Flex mode was configured. Complete this procedure to install Oracle Grid Infrastructure software for Oracle Standalone Cluster.

Create Directory Structure

*       Download and copy the Oracle Grid Infrastructure image files to the local node only. During installation, the software is copied and installed on all other nodes in the cluster.

To create the directory structure appropriately according to your environment, run the following commands:

For example:

mkdir -p /u01/app/grid

mkdir -p /u01/app/19.3.0/grid

mkdir -p /u01/app/oraInventory

mkdir -p /u01/app/oracle/product/19.3.0/dbhome_1

 

chown -R grid:oinstall /u01/app/grid

chown -R grid:oinstall /u01/app/19.3.0/grid

chown -R grid:oinstall /u01/app/oraInventory

chown -R oracle:oinstall /u01/app/oracle

As the grid user, download the Oracle Grid Infrastructure image files and extract the files into the Grid home:

cd /u01/app/19.0.0/grid

unzip -q download_location/grid.zip

Configure UDEV rules for ASM Disks

You need to configure UDEV rules to have read/write privileges on these storage namespaces for grid user. This includes the device details and corresponding uuid of the storage namespaces.

Assign Owner and Permission on NVMe Targets by creating a new file named “80-nvme.rules” with the following entries on all the nodes:

[root@flex1 ~]# cat /etc/udev/rules.d/80-nvme.rules

# All ASM Volumes For GRID Users

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.7636791b-944f-4adb-941e-0d962639f718", SYMLINK+="asm1", GROUP:="oinstall", OWNER:="grid", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.1251e1c3-3735-4546-8b39-6654c83620f7", SYMLINK+="asm2", GROUP:="oinstall", OWNER:="grid", MODE:="660"

Run Cluster Verification Utility

This step verifies that all the prerequisites are met to install the Oracle Grid Infrastructure Software. Oracle Grid Infrastructure ships with the Cluster Verification Utility (CVU) that can run to validate the pre and post installation configurations.

To run this utility, login as Grid User in Oracle RAC Node 1 and go to the directory where oracle grid software binaries are located. Run script named as “runcluvfy.sh” as follows:

./runcluvfy.sh stage -pre crsinst -n flex1,flex2,flex3,flex4,flex5,flex6,flex7,flex8 –verbose

Configure HugePages

HugePages is a method to have larger page size that is useful for working with a very large memory. For Oracle Databases, using HugePages reduces the operating system maintenance of page states, and increases Translation Lookaside Buffer (TLB) hit ratio.

Advantage of HugePages:

   HugePages are not swappable so there is no page-in/page-out mechanism overhead.

   HugePages uses fewer pages to cover the physical address space, so the size of "bookkeeping"(mapping from the virtual to the physical address) decreases, so it is requiring fewer entries in the TLB and so TLB hit ratio improves.

   HugePages reduces page table overhead. Also, HugePages eliminated page table lookup overhead: Since the pages are not subject to replacement, page table lookups are not required.

   Faster overall memory performance: On virtual memory systems each memory operation is two abstract memory operations. Since there are fewer pages to work on, the possible bottleneck on page table access is clearly avoided.

For our configuration, we used HugePages for all the OLTP and DSS workloads. Please refer to the oracle guidelines to configure appropriately as below:

https://docs.oracle.com/en/database/oracle/oracle-database/19/unxar/administering-oracle-database-on-linux.html#GUID-CC72CEDC-58AA-4065-AC7D-FD4735E14416

After configuration, you are ready to install the Oracle Grid Infrastructure and Oracle Database 19c software. For this solution, we installed Oracle home binaries on the boot LUN of the nodes. The OCR, Data, and Redo Log files reside in the namespace configured on netapp storage array.

Install and Configure Oracle Database Grid Infrastructure Software

*       It is not within the scope of this document to include the specifics of an Oracle RAC installation. However, a partial summary of details is provided that might be relevant. Please refer to the Oracle installation documentation for specific installation instructions for your environment.

To install Oracle Database Grid Infrastructure Software, follow these steps:

1.    Go to grid home where the Oracle 19c Grid Infrastructure software binaries are located and launch the installer as the "grid" user.

2.    Start the Oracle Grid Infrastructure installer by running the following command:

./gridSetup.sh

3.    Select option “Configure Oracle Grid Infrastructure for a New Cluster”, then click Next.

Graphical user interface, text, application, emailDescription automatically generated

4.    Select cluster configuration options “Configure an Oracle Standalone Cluster”, then click Next

5.    In next window, enter the Cluster Name and SCAN Name fields. Enter the names for your cluster and cluster scan that are unique throughout your entire enterprise network. You can also select to Configure GNS if you have configured your domain name server (DNS) to send to the GNS virtual IP address name resolution requests.

6.    In Cluster node information window, click the "Add" button to add all the eight nodes Public Hostname and Virtual Hostname as shown below:

Graphical user interface, text, application, emailDescription automatically generated

7.    As shown above, you will see all nodes listed in the table of cluster nodes. Click the SSH Connectivity button at the bottom of the window. Enter the operating system username and password for the Oracle software owner (grid). Click Setup.

8.    A message window appears, indicating that it might take several minutes to configure SSH connectivity between the nodes. After some time, another message window appears indicating that password-less SSH connectivity has been established between the cluster nodes. Click OK to continue

9.    In Network Interface Usage screen, select the usage type for each network interface displayed as shown below:

Graphical user interface, text, applicationDescription automatically generated

10.  In storage option, Select the option as “Use Oracle Flex ASM for storage” then click on “Next”. For this solution we have choose “No” option into separate ASM disk group for the Grid Infrastructure Management Repository data.

11.  In the Create ASM Disk Group window, select the “ASM1” & “ASM2” namespaces which are configured into NetApp Storage to store OCR and Voting disk files. Enter the name of disk group as “OCRVOTE” and select appropriate external redundancy options as shown below:

Graphical user interface, text, application, emailDescription automatically generated

*       For this solution, we did not configure Oracle ASM Filter Driver.

12.  Choose the password for the Oracle ASM SYS and ASMSNMP account, then click Next.

13.  Select the option “Do not use Intelligent Platform Management Interface (IPMI)”. Click Next.

14.  You can configure to have this instance of Oracle Grid Infrastructure and Oracle Automatic Storage Management to be managed by Enterprise Manager Cloud Control. For this solution we did not select this option. Click Next.

*       You can choose to set it up according to your requirements.

15.  Select the appropriate operating system group names for Oracle ASM according to your environments.

16.  Specify the oracle base and inventory directory to use for the Oracle Grid Infrastructure installation and then click Next. The Oracle base directory must be different from the Oracle home directory. Click Next and choose Inventory Directory according to your setup.

17.  Click Automatically run configuration scripts to run scripts automatically and enter the relevant root user credentials. Click Next.

18.  Wait while the prerequisite checks complete. If you have any issues, click the "Fix & Check Again" button. If any of the checks have a status of Failed and are not fixable, then you must manually correct these issues. After you have fixed the issue, you can click the Check Again button to have the installer re-check the requirement and update the status. Repeat as needed until all the checks have a status of Succeeded. Click Next.

19.  Review the contents of the Summary window and then click Install. The installer displays a progress indicator enabling you to monitor the installation process.

Graphical user interface, applicationDescription automatically generated

20.  Wait for the grid installer configuration assistants to complete.

21.  When the configuration complete successfully, click Close to finish and exit the grid installer.

22.  When GRID install is successful, login to each of the nodes and perform minimum health checks to make sure that Cluster state is healthy. After your Oracle Grid Infrastructure installation is complete, you can install Oracle Database on a cluster node for high availability or install Oracle RAC.

TextDescription automatically generated

Oracle Database Installation

After successful Oracle GRID software install, we recommend installing Oracle Database 19c software only. You can create databases using DBCA or database creation scripts at later stage.

*       It is not within the scope of this document to include the specifics of an Oracle RAC database installation. However, we will provide partial summary of details that might be relevant. Please refer to the Oracle database installation documentation for specific installation instructions for your environment.  https://docs.oracle.com/en/database/oracle/oracle-database/19/ladbi/index.html

To install Oracle Database software, follow these steps:

1.    Start the “./runInstaller” command from the Oracle Database 19c installation media where Oracle database software is located.

2.    Select option “Set Up Software Only” into configuration Option.

Related image, diagram or screenshot

3.    Select option "Oracle Real Application Clusters database installation" and click Next.

4.    Select nodes in the cluster where installer should install Oracle RAC. For this setup, install the software on all eight nodes as shown below:

Graphical user interface, text, applicationDescription automatically generated

5.    Click "SSH Connectivity..." and enter the password for the "oracle" user. Click Setup to configure passwordless SSH connectivity and click Test to test it when it is complete. When the test is complete, click Next.

6.    Select Database Edition Options according to your environments and then click Next.

7.    Enter appropriate Oracle Base, then click Next.

8.    Select the desired operating system groups and then click Next.

9.    Select option Automatically run configuration script from the option Root script execution menu and click Next.

10.  Wait for the prerequisite check to complete. If there are any problems click "Fix & Check Again" or try to fix those by checking and manually installing required packages. Click Next.

11.  Verify the Oracle Database summary information and then click Install.

Graphical user interface, application, tableDescription automatically generated

12.  Wait for the installation of Oracle Database finish successfully, then click Close to exit of the installer.

Overview of Oracle Flex ASM

Oracle ASM is Oracle's recommended storage management solution that provides an alternative to conventional volume managers, file systems, and raw devices. Oracle ASM is a volume manager and a file system for Oracle Database files that reduces the administrative overhead for managing database storage by consolidating data storage into a small number of disk groups. The smaller number of disk groups consolidates the storage for multiple databases and provides for improved I/O performance.

Oracle Flex ASM enables an Oracle ASM instance to run on a separate physical server from the database servers. With this deployment, larger clusters of Oracle ASM instances can support more database clients while reducing the Oracle ASM footprint for the overall system.

DiagramDescription automatically generated

When using Oracle Flex ASM, Oracle ASM clients are configured with direct access to storage. With Oracle Flex ASM, you can consolidate all the storage requirements into a single set of disk groups. All these disk groups are mounted by and managed by a small set of Oracle ASM instances running in a single cluster. You can specify the number of Oracle ASM instances with a cardinality setting. The default is three instances.

Prior to Oracle 12c, if ASM instance on one of the RAC nodes crashes, all the instances running on that node will crash too. This issue has been addressed in Flex ASM; Flex ASM can be used even if all the nodes are hub nodes. However, GNS configuration is mandatory for enabling Flex ASM. You can check what instances relate to a simple query as shown below:

Graphical user interface, textDescription automatically generated

As you can see from the query (above), instance1 (FLEX1), instance2 (FLEX2) and instance8 (FLEX8) are connected to +ASM. There are a few more commands you can run to check cluster and FLEX ASM details as shown below:

TextDescription automatically generated

Refer to the Oracle documentation for more information: https://docs.oracle.com/en/database/oracle/oracle-database/19/ostmg/manage-flex-asm.html#GUID-DE759521-9CF3-45D9-9123-7159C9ED4D30

Oracle Database Multitenant Architecture

The multitenant architecture enables an Oracle database to function as a multitenant container database (CDB). A CDB includes zero, one, or many customer-created pluggable databases (PDBs). A PDB is a portable collection of schemas, schema objects, and non-schema objects that appears to an Oracle Net client as a non-CDB. All Oracle databases before Oracle Database 12c were non-CDBs.

A container is logical collection of data or metadata within the multitenant architecture. The following figure represents possible containers in a CDB.

DiagramDescription automatically generated

The multitenant architecture solves several problems posed by the traditional non-CDB architecture. Large enterprises may use hundreds or thousands of databases. Often these databases run on different platforms on multiple physical servers. Because of improvements in hardware technology, especially the increase in the number of CPUs, servers can handle heavier workloads than before. A database may use only a fraction of the server hardware capacity. This approach wastes both hardware and human resources. Database consolidation is the process of consolidating data from multiple databases into one database on one computer. The Oracle Multitenant option enables you to consolidate data and code without altering existing schemas or applications.

For more information on Oracle Database Multitenant Architecture, please refer to: https://docs.oracle.com/en/database/oracle/oracle-database/19/multi/introduction-to-the-multitenant-architecture.html#GUID-267F7D12-D33F-4AC9-AA45-E9CD671B6F22

*       In this solution, we configured both type of databases to check performance of Non-Container Databases and Container Databases as explained in the next scalability test section.

Scalability Test and Results

Before configuring a database for workload tests, it is extremely important to validate that this is indeed a balanced configuration that is capable of delivering expected performance. In this solution, we will test and validate node and user scalability on all 8 node Oracle RAC Databases with various database benchmarking tools as explained below.

Hardware Calibration Test using FIO

FIO is short for Flexible IO, a versatile IO workload generator. FIO is a tool that will spawn a number of threads or processes doing a particular type of I/O action as specified by the user. For our solution, we use FIO to measure the performance of a NetApp storage device over a given period of time. For the FIO Tests, we created 4 Subsystems with total 32 Namespaces (each subsystem having 8 Namespaces) and each of the subsystem was 500 GB in size equally distributed across both the aggregates. These 32 Namespace were shared across all the eight nodes for read/write IO operations.

We run various FIO tests for measuring IOPS, Latency and Throughput performance of this solution by changing block size parameter into the FIO test. For each FIO test, we also changed read/write ratio as 0/100% read/write, 50/50% read/write, 70/30% read/write, 90/10% read/write and 100/0% read/write to scale the performance of the system. We also ran the tests for at least 4 hours to help ensure that this configuration is capable of sustaining this type of load for longer period of time.

IOPS Tests

The chart below shows results for the random read/write FIO test for the 8k block size representing OLTP type of workloads.

Related image, diagram or screenshot

For the 100/0% read/write test, we achieved around 1591k IOPS with the read latency around 0.95 millisecond. Similarly, for the 90/10% read/write test, we achieved around 1723k IOPS with the read latency around 0.92 millisecond and the write latency around 0.78 millisecond. For the 70/30% read/write test, we achieved around 1326k IOPS with the read latency around 0.44 millisecond and the write latency around 1.5 millisecond. For the 50/50% read/write test, we achieved around 864k IOPS with the read latency around 0.38 millisecond and the write latency around 2.05 millisecond. For the 0/100% read/write test, we achieved around 410k IOPS with the write latency around 2.57 millisecond. Reads and writes consume system resources differently. The system under test benefited from slightly better resource distribution in the 90/10 R/W test, resulting in slightly improved peak IOPS in this test compared with the 100/0 R/W test

Bandwidth Tests

The bandwidth tests are carried out with 512k IO Size and represents the DSS database type workloads. The chart below shows results for the sequential read/write FIO test for the 512k block size.

Related image, diagram or screenshot

For the 100/0% read/write test, we achieved around 13.26 GB/s throughput with the read latency around 2.4 millisecond. Similarly, for the 90/10% read/write test, we achieved around 14.34 GB/s throughput with the read latency around 2.3 millisecond and the write latency around 1.4 millisecond. For the 70/30% read/write test, we achieved around 14.98 GB/s throughput with the read latency around 2.1 millisecond and the write latency around 2.2 millisecond. For the 50/50% read/write test, we achieved around 14 GB/s throughput with the read latency around 1.5 millisecond and the write latency around 3.25 millisecond. For the 0/100% read/write test, we achieved around 7.2 GB/s throughput with the write latency around 4.6 millisecond.

We did not see any performance dips or degradation over the period of run time. It is also important to note that this is not a benchmarking exercise, and these are practical and out of box test numbers that can be easily reproduced by anyone. At this time, we are ready to create OLTP database(s) and continue with database tests.

Database Creation with DBCA

We used Oracle Database Configuration Assistant (DBCA) to create three OLTP (SLOB, SOEPDB and SOEFIN) and one DSS (PDBSH) databases for SLOB and SwingBench test calibration. Alternatively, you can use Database creation scripts to create the databases as well.

For all the database deployment, we have configured two aggregates (one aggregate on each storage node) into a single SVM, and each aggregate contains 11 SSD (1.75 TB Each) drives that were subdivided into RAID DP groups, plus one spare drive as explained earlier in the storage configuration section.

For each RAC database, we created total number of 32 Namespaces. We distributed equal number of namespaces on the storage nodes by placing those namespaces into both the aggregates. All the database  files were also spread evenly across the two nodes of the storage system so that each storage node served data for the databases. Table 11 lists the storage layout of all the namespaces configuration for all the databases.

Table 11.       Namespace Storage Layout

Database Name

Subsystem

Namespace

Size (GB)

Aggregate

Notes

ASM

orasub1

asm1

100

aggr1_node1

OCR & Voting Disk

orasub2

asm2

100

aggr1_node2

SLOB (Non CDB Database)

orasub1

dataslob01

300

aggr1_node1

SLOB Database Data Files

orasub2

dataslob02

300

aggr1_node1

orasub3

dataslob03

300

aggr1_node1

orasub4

dataslob04

300

aggr1_node1

orasub1

dataslob05

300

aggr1_node2

orasub2

dataslob06

300

aggr1_node2

orasub3

dataslob07

300

aggr1_node2

orasub4

dataslob08

300

aggr1_node2

orasub1

dataslob09

300

aggr1_node1

orasub2

dataslob10

300

aggr1_node1

orasub3

dataslob11

300

aggr1_node1

orasub4

dataslob12

300

aggr1_node1

orasub1

dataslob13

300

aggr1_node2

orasub2

dataslob14

300

aggr1_node2

orasub3

dataslob15

300

aggr1_node2

orasub4

dataslob16

300

aggr1_node2

orasub1

dataslob17

300

aggr1_node1

orasub2

dataslob18

300

aggr1_node1

orasub3

dataslob19

300

aggr1_node1

orasub4

dataslob20

300

aggr1_node1

orasub1

dataslob21

300

aggr1_node2

orasub2

dataslob22

300

aggr1_node2

orasub3

dataslob23

300

aggr1_node2

orasub4

dataslob24

300

aggr1_node2

orasub1

redoslob01

50

aggr1_node1

SLOB Database Redo Log Files

orasub2

redoslob02

50

aggr1_node1

orasub3

redoslob03

50

aggr1_node1

orasub4

redoslob04

50

aggr1_node1

orasub1

redoslob05

50

aggr1_node2

orasub2

redoslob06

50

aggr1_node2

orasub3

redoslob07

50

aggr1_node2

orasub4

redoslob08

50

aggr1_node2

CDBDB (Container Database)

orasub1

datacdb01

200

aggr1_node1

CDBDB Database Data Files

orasub2

datacdb02

200

aggr1_node2

orasub3

datacdb03

200

aggr1_node1

orasub4

datacdb04

200

aggr1_node2

orasub1

redocdb01

50

aggr1_node1

CDBDB Database Redo Log Files

orasub2

redocdb02

50

aggr1_node1

orasub3

redocdb03

50

aggr1_node1

orasub4

redocdb04

50

aggr1_node1

orasub1

redocdb05

50

aggr1_node2

orasub2

redocdb06

50

aggr1_node2

orasub3

redocdb07

50

aggr1_node2

orasub4

redocdb08

50

aggr1_node2

PDBSOE (Pluggable Database on Container CDBDB Database)

orasub1

pdbsoe01

400

aggr1_node1

PDBSOE Database Data Files

orasub2

pdbsoe02

400

aggr1_node1

orasub3

pdbsoe03

400

aggr1_node1

orasub4

pdbsoe04

400

aggr1_node1

orasub1

pdbsoe05

400

aggr1_node2

orasub2

pdbsoe06

400

aggr1_node2

orasub3

pdbsoe07

400

aggr1_node2

orasub4

pdbsoe08

400

aggr1_node2

orasub1

pdbsoe09

400

aggr1_node1

orasub2

pdbsoe10

400

aggr1_node1

orasub3

pdbsoe11

400

aggr1_node1

orasub4

pdbsoe12

400

aggr1_node1

orasub1

pdbsoe13

400

aggr1_node2

orasub2

pdbsoe14

400

aggr1_node2

orasub3

pdbsoe15

400

aggr1_node2

orasub4

pdbsoe16

400

aggr1_node2

orasub1

pdbsoe17

400

aggr1_node1

orasub2

pdbsoe18

400

aggr1_node1

orasub3

pdbsoe19

400

aggr1_node1

orasub4

pdbsoe20

400

aggr1_node1

orasub1

pdbsoe21

400

aggr1_node2

orasub2

pdbsoe22

400

aggr1_node2

orasub3

pdbsoe23

400

aggr1_node2

orasub4

pdbsoe24

400

aggr1_node2

PDBFIN (Pluggable Database on Container CDBDB Database)

orasub1

pdbfin01

300

aggr1_node1

PDBFIN Database Data Files

orasub2

pdbfin02

300

aggr1_node1

orasub3

pdbfin03

300

aggr1_node1

orasub4

pdbfin04

300

aggr1_node1

orasub1

pdbfin05

300

aggr1_node2

orasub2

pdbfin06

300

aggr1_node2

orasub3

pdbfin07

300

aggr1_node2

orasub4

pdbfin08

300

aggr1_node2

orasub1

pdbfin09

300

aggr1_node1

orasub2

pdbfin10

300

aggr1_node1

orasub3

pdbfin11

300

aggr1_node1

orasub4

pdbfin12

300

aggr1_node1

orasub1

pdbfin13

300

aggr1_node2

orasub2

pdbfin14

300

aggr1_node2

orasub3

pdbfin15

300

aggr1_node2

orasub4

pdbfin16

300

aggr1_node2

orasub1

pdbfin17

300

aggr1_node1

orasub2

pdbfin18

300

aggr1_node1

orasub3

pdbfin19

300

aggr1_node1

orasub4

pdbfin20

300

aggr1_node1

orasub1

pdbfin21

300

aggr1_node2

orasub2

pdbfin22

300

aggr1_node2

orasub3

pdbfin23

300

aggr1_node2

orasub4

pdbfin24

300

aggr1_node2

DSSDB (Container Database)

orasub1

datadss01

200

aggr1_node1

DSSDB Database Data Files

orasub2

datadss02

200

aggr1_node2

orasub3

datadss03

200

aggr1_node1

orasub4

datadss04

200

aggr1_node2

orasub1

redodss01

50

aggr1_node1

DSSDB Database Redo Log Files

orasub2

redodss02

50

aggr1_node1

orasub3

redodss03

50

aggr1_node1

orasub4

redodss04

50

aggr1_node1

orasub1

redodss05

50

aggr1_node2

orasub2

redodss06

50

aggr1_node2

orasub3

redodss07

50

aggr1_node2

orasub4

redodss08

50

aggr1_node2

PDBSH (Pluggable Database on Container DSSDB Database)

orasub1

pdbsh01

400

aggr1_node1

PDBSH Database Data Files

orasub2

pdbsh02

400

aggr1_node1

orasub3

pdbsh03

400

aggr1_node1

orasub4

pdbsh04

400

aggr1_node1

orasub1

pdbsh05

400

aggr1_node2

orasub2

pdbsh06

400

aggr1_node2

orasub3

pdbsh07

400

aggr1_node2

orasub4

pdbsh08

400

aggr1_node2

orasub1

pdbsh09

400

aggr1_node1

orasub2

pdbsh10

400

aggr1_node1

orasub3

pdbsh11

400

aggr1_node1

orasub4

pdbsh12

400

aggr1_node1

orasub1

pdbsh13

400

aggr1_node2

orasub2

pdbsh14

400

aggr1_node2

orasub3

pdbsh15

400

aggr1_node2

orasub4

pdbsh16

400

aggr1_node2

orasub1

pdbsh17

400

aggr1_node1

orasub2

pdbsh18

400

aggr1_node1

orasub3

pdbsh19

400

aggr1_node1

orasub4

pdbsh20

400

aggr1_node1

orasub1

pdbsh21

400

aggr1_node2

orasub2

pdbsh22

400

aggr1_node2

orasub3

pdbsh23

400

aggr1_node2

orasub4

pdbsh24

400

aggr1_node2

As shown in Table 11, each database has a total 32 namespaces. On these 32 namespaces, two disk groups were created to store the “data” and “redolog” files for the database. 24 namespaces were used to create Oracle ASM “Data” disk group and 8 namespaces to create Oracle ASM “redolog” disk group for each database. We have used widely adopted SLOB and Swingbench database performance test tools to test and validate throughput, IOPS, and latency for various test scenarios as explained below.

SLOB Test

The Silly Little Oracle Benchmark (SLOB) is a toolkit for generating and testing I/O through an Oracle database. SLOB is very effective in testing the I/O subsystem with genuine Oracle SGA-buffered physical I/O. SLOB supports testing physical random single-block reads (db file sequential read) and random single block writes (DBWR flushing capability). SLOB issues single block reads for the read workload that are generally 8K (as the database block size was 8K).

For testing the SLOB workload, we created one non-container database as SLOB. For SLOB database, we created total 32 namespace. On these 32 namespaces, we created two disk groups to store the “data” and “redolog” files for the SLOB database. First disk-group “DATASLOB” was created with 24 namespaces (300 GB each) while second disk-group “REDOSLOB” was created with 8 namespaces (50 GB each).

Those ASM disk groups provided the storage required to create the tablespaces for the SLOB Database. We loaded SLOB schema on “DATASLOB” disk-group of up to 3.5 TB in size.

We used SLOB2 to generate our OLTP workload. Each database server applied the workload to Oracle database, log, and temp files. The following tests were performed and various metrics like IOPS and latency were captured along with Oracle AWR reports for each test scenario.

User Scalability Test

SLOB2 was configured to run against all the eight Oracle RAC nodes and the concurrent users were equally spread across all the nodes. We tested the environment by increasing the number of Oracle users in database from a minimum of 64 users up to a maximum of 512 users across all the nodes. At each load point, we verified that the storage system and the server nodes could maintain steady-state behavior without any issues. We also made sure that there were no bottlenecks across servers or networking systems.

The User Scalability test was performed with 64, 128, 192, 256, 384 and 512 users on 8 Oracle RAC nodes by varying read/write ratio as explained below:

   Varying workloads

     100% read (0% update)

     90% read (10% update)

     70% read (30% update)

     50% read (50% update)

Table 12 lists the total number of IOPS (both read and write) available for user scalability test when run with 64, 128, 192, 256, 384 and 512 Users on the SLOB database.

Table 12.       Total IOPS for SLOB User Scalability Tests

Users

Read/Write % (100-0)

Read/Write % (90-10)

Read/Write % (70-30)

Read/Write % (50-50)

64

388,706

405,485

457,021

488,826

128

736,713

763,770

746,430

756,207

192

989,616

1,047,593

1,018,685

898,215

256

1,227,841

1,264,029

1,090,525

938,379

384

1,457,965

1,493,350

1,128,892

965,426

512

1,544,274

1,579,193

1,156,161

959,938

The following graphs demonstrate the total number of IOPS while running SLOB workload for various concurrent users for each test scenario.

Related image, diagram or screenshot

The graph above shows the linear scalability with increased users and similar IOPS from 64 users to 512 users with 100% Read/Write, 90% Read/Write, 70% Read/Write and 50% Read/Write.

The AWR screenshot shown below was captured from a 100% Read (0% update) Test scenario while running SLOB test for 512 users. The screenshot shows a section from the Oracle AWR report from the run that highlights Physical Reads/Sec and Physical Writes/Sec for each instance.

Graphical user interfaceDescription automatically generated

The screenshot above highlights that IO load is distributed across all the cluster nodes performing workload operations. Due to variations in workload randomness, we conducted multiple runs to ensure consistency in behavior and test results.

The screenshot shown below was captured from a 70% Read (30% update) Test scenario while running SLOB test for 512 users. The snapshot shows a section from AWR report from the run that highlights Physical Reads/Sec and Physical Writes/Sec for each instance.

Graphical user interfaceDescription automatically generated

The following graph illustrates the latency exhibited by the NetApp AFF A800 Storage across different workloads. All the workloads experienced less than 1 millisecond latency and it varies based on the workloads. As expected, the 50% read (50% update) test exhibited higher latencies as the user counts increases.

Related image, diagram or screenshot

The following screenshot was captured from 100 % Read (0% Update) Test scenario while running SLOB test for 512 users. The snapshot shows a section of AWR report from the run that highlights top timed Events.

Graphical user interface, applicationDescription automatically generated

Swingbench Test

Swingbench is a simple to use, free, Java-based tool to generate various type of database workloads and perform stress testing using different benchmarks in Oracle database environments. Swingbench can be used to demonstrate and test technologies such as Real Application Clusters, Online table rebuilds, Standby databases, online backup, and recovery, etc. In this solution, we have used Swingbench tool for running various type of workload and check the overall performance of this reference architecture.

Swingbench provides four separate benchmarks, namely, Order Entry, Sales History, Calling Circle, and Stress Test. For the tests described in this solution, Swingbench Order Entry (SOE) benchmark was used for representing OLTP type of workload and the Sales History (SH) benchmark was used for representing DSS type of workload.

The Order Entry benchmark is based on SOE schema and is TPC-C like by types of transactions. The workload uses a very balanced read/write ratio around 60/40 and can be designed to run continuously and test the performance of a typical Order Entry workload against a small set of tables, producing contention for database resources.

The Sales History benchmark is based on the SH schema and is like TPC-H. The workload is query (read) centric and is designed to test the performance of queries against large tables.

Typically encountered in the real-world deployments, we tested a combination of scalability and stress related scenarios that ran across all of the 8-node Oracle RAC cluster, as follows:

   OLTP database user scalability workload representing small and random transactions

   DSS database workload representing larger transactions

   Mixed databases (OLTP and DSS) workloads running simultaneously

For the Swingbench workload tests, we created two Container Database as “CDBDB” and “DSSDB”. We configured the “CDBDB” container database and created two Pluggable Databases as “PDBSOE” and “PDBFIN” to run the Swingbench SOE workload representing OLTP type of workload characteristics. We configured the “DSSDB” container databases and created one Pluggable Databases as “PDBSH” to run the Swingbench SH workload representing DSS type of workload characteristics.

For this solution, we deployed multiple pluggable databases (PDBSOE and PDBFIN) plugged into one container (CDBDB) database and one pluggable database (PDBSH) plugged into one container (DSSDB) database to demonstrate the multitenancy capability, performance, and sustainability for this reference architecture.

In “CDBDB” container database, we created two pluggable databases as both the databases have similar workload characteristics. By consolidating multiple pluggable databases under the same container database allows easier management, efficiently sharing computational and memory resources, separation of administrative tasks, easier database upgrades as well as fewer patches and upgrades.

For the OLTP databases, we created and configured SOE schema of 4.1 TB for the PDBSOE Database and 2.8 TB for the PDBFIN Database. And for the DSS database, we created and configured SH schema of 5.1 TB for the PDBSH Database.

The first step after the databases creation is calibration; about the number of concurrent users, nodes, throughput, IOPS and latency for database optimization. For this FlexPod solution, we ran the swingbench workloads on various combination of databases and captured the system performance as follows:

   One OLTP Database Performance

   Multiple (Two) OLTP Databases Performance

   One DSS Database Performance

   Multiple OLTP & DSS Databases Performance

One OLTP Database Performance

For one OLTP database workload featuring Order Entry schema, we created one container database CDBDB and one pluggable database PDBSOE as explained earlier. We used 64 GB size of SGA for this database and also, we ensured that the HugePages were in use. We ran the swingbench SOE workload with varying the total number of users on this database from 100 Users to 800 Users. Each user scale iteration test was run for at least 3 hours and for each test scenario, we captured the Oracle AWR reports to check the overall system performance below:

User Scalability

Table 13 lists the Transaction Per Minutes (TPM), IOPS, Latency and System Utilization for the CDBDB Database while running the workload from 100 users to 800 users across all the eight RAC nodes.

Table 13.       User Scale Test on One OLTP Database

Number of Users

Transactions

Storage IOPS

Latency (milliseconds)

CPU Utilization (%)

Per Seconds (TPS)

Per Minutes (TPM)

Reads/Sec

Writes/Sec

Total IOPS

100

17,767

1,066,020

107,896

50,652

158,549

0.35

12.3

200

24,265

1,455,900

132,854

64,238

197,092

0.38

18.2

300

29,018

1,741,104

166,685

81,738

248,422

0.44

20.1

400

37,924

2,275,428

217,275

106,337

323,611

0.48

26.6

500

41,698

2,501,898

238,711

116,856

355,566

0.49

33.8

600

42,922

2,575,320

241,854

119,253

361,107

0.50

34.2

700

44,108

2,646,456

251,694

123,392

375,086

0.51

36.2

800

45,728

2,743,668

260,438

127,882

388,320

0.54

42.7

The following chart shows the IOPS and Latency for the CDBDB Database while running the workload from 100 users to 800 users across all eight RAC nodes.

Related image, diagram or screenshot

The chart below shows the TPM and System Utilization for the same above tests on CDBDB Database for running the workload from 100 users to 800 users.

Related image, diagram or screenshot

We also ran the maximum number of users (800) test for 24-hour period to check the system performance. The screenshot below highlights the database summary while running the swingbench SOE workload for 24-hour test duration. The container database “CDBDB” was running with one pluggable database as “PDBSOE.”

A computer screen captureDescription automatically generated with low confidence

The screenshot below captured from Oracle AWR report shows the “Top Timed Events” for the CDBDB database for the entire 24-hour duration of the test.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report highlights the Physical Reads/Sec, Physical Writes/Sec and Transactions per Seconds for the Container CDBDB Database. We captured about 365k IOPS (243k Reads/s and 123k Writes/s) with the 44k TPS while running multiple databases workloads.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the CDBDB database “IO Profile” for the “Reads/s” and “Writes/s” requests for the entire 24-hour duration of the test. The Total Requests (Read and Write Per Second) were around “377k” with Total (MB) Read+Write Per Second was around “3212” MB/s for the CDBDB database while running the workload test on one database.

Graphical user interface, applicationDescription automatically generated

The screenshot below shows the NetApp Storage array “Q S P S (qos statistics performance show)” when one OLTP database was running the workload. The screenshot shows the average IOPS “380k” with the average throughput of “3.3 GB/s” with the average latency around “0.6 millisecond”. The storage cluster utilization during the above test was around 55% which was an indication that storage hasn’t reached the threshold and could take more load.

Graphical user interfaceDescription automatically generated

For the entire 24-hour test, we observed the system performance (IOPS and Throughput) was consistent throughout and we did not observe any dips in performance while running one OLTP database stress test.

Multiple (Two) OLTP Databases Performance

For running multiple OLTP database workload, we created one container database CDBDB and two pluggable database PDBSOE and PDBFIN as explained earlier. We ran the swingbench SOE workload on both the databases at the same time with varying the total number of users on both the databases from 400 Users to 1200 Users. Each user scale iteration test was run for at least 3 hours and for each test scenario, we captured the Oracle AWR reports to check the overall system performance below.

Table 14 lists the IOPS and System Utilization for each of the pluggable databases while running the workload from total of 400 users to 1200 users across all the eight RAC nodes.

Table 14.       IOPS and System Utilization for Pluggable Databases

Users

IOPS for SOE

IOSP for OLTP

Total IOPS

System Utilization (%)

400

195,832

144,473

340,305

40.1

600

238,797

165,021

403,818

45.2

800

242,408

167,956

410,364

46.8

1000

250,463

171,547

422,010

48.7

1200

250,549

170,831

421,380

48.9

The chart below shows the IOPS and System Utilization for the overall CDBDB Database while running the database workload on both the databases at the same time.

Related image, diagram or screenshot

As shown in the chart above, we observed both the databases were linearly scaling the IOPS after increasing and scaling more and more users. We observed average 422k IOPS with overall system utilization around 49% when scaling maximum number of users on multiple database workload test. After increasing users beyond certain level, we observed more GC cluster events and overall similar IOPS around 420k.

Table 15 lists the Transactions per Seconds (TPS) and Transactions per Minutes (TPM) for each of the pluggable databases while running the workload from total of 400 users to 1200 users across all 8 RAC nodes.

Table 15.       Transactions per Seconds and Transactions per Minutes

Users

TPS for SOE

TPS for OLTP

Total TPS

Total TPM

400

24,602

16,406

41,008

2,460,480

600

28,182

18,758

46,940

2,816,400

800

29,023

19,349

48,372

2,902,320

1000

29,126

20,431

49,557

2,973,420

1200

29,143

20,302

49,445

2,966,700

The chart below shows the Transactions per Seconds (TPS) for the same tests (above) on CDBDB Database for running the workload on both pluggable databases.

Related image, diagram or screenshot

We also ran the maximum number of users test for 24-hour period to check the system overall performance. The screenshot below highlights the database summary while running the swingbench SOE workload for 24-hour test duration on both the databases. The container database “CDBDB” was running with two pluggable databases as “PDBSOE” and “PDBFIN.”

A picture containing text, black, screenshotDescription automatically generated

The screenshot shown below was captured from Oracle AWR report while running the Swingbench SOE workload tests on both OLTP databases for 24-hours. The screenshot shows the “OS Statistics by Instance” while the system was running mixed workload. As shown below, the workload was equally spread across all the databases clusters while the average CPU utilization was around 48% overall.

Graphical user interfaceDescription automatically generated

The screenshot shown below captured from Oracle AWR report shows the “Top Timed Events” for the CDBDB database for the entire 24-hour duration of the test.

Graphical user interfaceDescription automatically generated

The screenshot shown below captured from Oracle AWR report highlights the Physical Reads/Sec, Physical Writes/Sec and Transactions per Seconds for the Container CDBDB Database. We captured about 432k IOPS (286k Reads/s and 146k Writes/s) with the 51k TPS while running multiple databases workloads.

Graphical user interfaceDescription automatically generated

The screenshot below shows the NetApp Storage array “Q S P S (qos statistics performance show)” when two OLTP database was running the workload at the same time. The screenshot shows the average IOPS “430k” with the average throughput of “3.7 GB/s” with the average latency around “0.75 millisecond”.  In the multi-database use-case the same behavior of storage cluster utilization (~55%) was observed.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the CDBDB database “IO Profile” for the “Reads/s” and “Writes/s” requests for the entire duration of the test. As the screenshots shows, the Total Requests (Read and Write Per Second) were around “438k” with Total (MB) Read+Write Per Second was around “3766” MB/s for the CDBDB database while running the workload test on two databases.

Graphical user interface, textDescription automatically generated

The screenshot below captured from Oracle AWR report shows the CDBDB database “Interconnect Client Statistics Per Second” for the entire duration of the test. As the screenshots shows, Interconnect Sent and Received Statistics were average around “2185 MB/s” while running both the OLTP database workload test.

Graphical user interfaceDescription automatically generated

For the entire 24-hour test, we observed the system performance (IOPS, Latency and Throughput) was consistent throughout and we did not observe any dips in performance while running multiple OLTP database stress test.

One DSS Database Performance

DSS database workloads are generally sequential in nature, read intensive and exercise large IO size. DSS database workload runs a small number of users that typically exercise extremely complex queries that run for hours. For running oracle database multitenancy architecture, we configured one container database as DSSDB and into that container, we created one pluggable database as PDBSH as explained earlier.

We configured 5.1 TB of PDBSH pluggable database by loading Swingbench “SH” schema into Datafile Tablespace. The screenshot below shows the database summary for the “DSSDB” database running for 24-hour duration. The container database “DSSDB” was also running with one pluggable databases “PDBSH” and the pluggable database was running the Swingbench SH workload for the entire 24-hour duration of the test.

A computer screen captureDescription automatically generated with low confidence

The screenshot below captured from Oracle AWR report shows the DSSDB database “IO Profile” for the “Reads/s” and “Writes/s” requests for the entire 24-hour duration of the test. As the screenshots shows, the Total MB (Read and Write Per Second) were around “10,527 MB/s” for the DSSDB database while running this test.

A picture containing text, blackDescription automatically generated

The screenshot shown below shows the NetApp storage array performance (Q S P S (qos statistics performance show)) captured while running Swingbench SH workload on one DSS database. The screenshot shows the average throughput of “11.8 GB/s” with the average latency around “9 millisecond.”

Graphical user interfaceDescription automatically generated

As shown above, the database performance was consistent throughout the test and we did not observe any dips in performance for entire period of 24-hour test. The storage cluster utilization in this case was around 45% during the DSS only workload.

Multiple OLTP and DSS Databases Performance

In this test, we ran Swingbench SOE workloads on both the OLTP (PDBSOE + PDBFIN) databases and Swingbench SH workload on one DSS (PDBSH) Database at the same time and captured the overall system performance. We captured the system performance on small random queries presented via OLTP databases as well as large and sequential transactions submitted via DSS database workload as documented below.

The screenshot below shows the database summary for the “CDBDB” database running for a 12-hour duration. The container database “CDBDB” was running with both the pluggable databases “PDBSOE” and “PDBFIN” and both the pluggable databases were running the Swingbench SOE workload for the entire 12-hour duration of the test.

A picture containing text, screenshot, blackDescription automatically generated

The screenshot below shows the database summary for the “DSSDB” database running for a 12-hour duration. The container database “DSSDB” was also running with one pluggable databases “PDBSH” and the pluggable database was running the Swingbench SH workload for the entire 12-hour duration of the test.

A computer screen captureDescription automatically generated with low confidence

The screenshot shown below was captured from Oracle AWR report while running the Swingbench SOE and SH workload tests on all the three databases for 12-hours. The screenshot shows the “OS Statistics by Instance” while the system was running mixed workload. As shown below, the workload was equally spread across all the databases clusters while the average CPU utilization was around 32% overall.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the “Top Timed Events” for the CDBDB database while running Swingbench SOE workloads on both the pluggable (PDBSOE and PDBFIN) databases for the entire 12-hour duration of the test.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report highlights the Physical Reads/Sec, Physical Writes/Sec and Transactions per Seconds for the Container CDBDB Database. We captured around 326k IOPS (217k Reads/s and 109k Writes/s) with the 38k TPS while running multiple databases workloads.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the CDBDB database “IO Profile” for the “Reads/s” and “Writes/s” requests for the entire 12-hour duration of the test. As the screenshots shows, the Total Requests (Read and Write Per Second) were around “325k” with Total (MB) Read+Write Per Second was around “2813” MB/s for the CDBDB database while running the mixed workload test.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the CDBDB database “Interconnect Client Statistics Per Second” for the entire 12-hour duration of the test. As the screenshots shows, Interconnect Sent and Received Statistics were average around “1690 MB/s” while running the mixed workload test.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the “Top Timed Events” for the DSSDB database while running Swingbench SH workloads on the pluggable (PDBFIN) database for the entire 12-hour duration of the test.

Graphical user interfaceDescription automatically generated

The screenshot below captured from Oracle AWR report shows the DSSDB database “IO Profile” for the “Reads/s” and “Writes/s” requests for the entire 12-hour duration of the test. As the screenshots shows, the Total MB (Read and Write Per Second) were around “3700 MB/s” for the DSSDB database while running this test.

Graphical user interfaceDescription automatically generated

The screenshot below shows the NetApp Storage array “Q S P S (qos statistics performance show)” when all the databases were running the workloads at the same time. The screenshot shows the average IOPS “325k” with the average throughput of “7 GB/s” with the average latency around “1 millisecond.”

Graphical user interfaceDescription automatically generated with medium confidence

The screenshot below also shows the NetApp Storage array “statistics”. The screenshot shows the average CPU busy around “63%” with “7.5 GB/s” disk read and “1.4 GB/s” disk write when all the databases were running the workloads at the same time. The storage cluster utilization was the highest with both OLTP and DSS running together generating around ~9GB/sec throughput.

Graphical user interfaceDescription automatically generated with low confidence

The screenshot below shows the NetApp Array GUI when all the databases were running the workloads at the same time.

Chart, line chartDescription automatically generated

When we ran multiple (OLTP and DSS) databases workloads together, we achieved average around “335k” IOPS, “8.9 GB/s” Throughput with the average latency around “2 milliseconds.” For the entire 12-hour tests, we observed the system performance (IOPS and Throughput) was consistent throughout and we did not observe any dips in performance while running these tests.

Resiliency and Failure Tests

The goal of these tests was to ensure that reference architecture withstands commonly occurring failures due to either unexpected crashes, hardware failures or human errors. We conduct many hardware (disconnect power), software (process kills) and OS specific failures that simulate the real world scenarios under stress condition. In the destructive testing, we will also demonstrate unique failover capabilities of Cisco UCS components. Table 16 highlights the test cases.

Table 16.       Hardware Failover Tests

Test Scenario

Tests Performed

Test 1: Cisco UCS Chassis IOM Links Failure

Run the system on full Database workload.

Disconnect one or two links from each Chassis 1 IOM and Chassis 2 IOM by pulling it out and reconnect it after 10-15 minutes. Capture the impact on overall database performance.

Test 2: FI – A Failure

Run the system on full Database workload.

Power Off “FI – A” and check the network traffic on “FI – B” and capture the impact on overall database performance.

Test 3: FI – B Failure

Run the system on full Database workload.

Power Off “FI – B” and check the network traffic on “FI – A” and capture the impact on overall system performance.

Test 4: MDS – A Switch Failure

Run the system on full Database workload.

Power Off “MDS – A” switch and check the network and storage traffic on “MDS – B” switch. Capture the impact on overall database performance.

Test 5: MDS – B Switch Failure

Run the system on full Database workload.

Power Off “MDS – B” switch and check the network and storage traffic on “MDS – A” switch. Capture the impact on overall system performance

Test 6: Storage Controller Links Failure

Run the system on full Database workload.

Disconnect one or two FC links from each of the NetApp Storage Controller by pulling it out and reconnect it after 10-15 minutes. Capture the impact on overall database performance.

Test 7: Server Node Failure

Run the system on full Database workload.

Power Off one Linux Host and check the impact on database performance.

The architecture below illustrates various failure scenario which can be occurred due to either unexpected crashes or hardware failures.

DiagramDescription automatically generated

As shown above, Scenario 1 and/or scenario 2 represents the Chassis IOM link failures. Also, scenario 3 and/or 4 represents the Chassis all IOM links failure. Scenario 5 represents the UCS FI – A failure and similarly, scenario 6 represents the MDS Switch – A failure. Scenario 7 represents the NetApp Storage Controllers link failures and Scenario 8 represents one of the Server Node Failure.

As previously explained, we configured to carry Oracle Public Network traffic on “VLAN 134” through FI – A and Oracle Private Interconnect Network traffic on “VLAN 10” through FI – B under normal operating conditions before the failover tests.

The snapshots below show a complete infrastructure details of MAC address and VLAN information for UCS FI – A and FI – B Switches before failover test. Log into FI – A and type “connect nxos a” then type “show mac address-table” to see all the VLAN connection on the switch as:

A picture containing text, green, screenshotDescription automatically generated

Similarly, log into FI – B and type “connect nxos b” then type “show mac address-table” to see all the VLAN connection on the switch as follows:

A picture containing text, green, screenshotDescription automatically generated

*       All the Hardware failover tests were conducted during all the three databases (PDBSOE, PDBFIN and PDBSH) running Swingbench mixed workloads.

Test 1 – Cisco UCS Chassis IOM Links Failure

We conducted IOM Links failure test on Cisco UCS Chassis 1 and Chassis by disconnecting one of the server port link cable from the Chassis as shown below.

DiagramDescription automatically generated

Unplug two server port cables from Chassis 1 and Chassis 2 each and check the MAC address and VLAN traffic information on both UCS FI. The screenshot below shows network traffic on FI – A when one link from Chassis 1 and one link from Chassis 2 IOM Failed.

A picture containing text, electronics, screenshotDescription automatically generated

Also, we logged into the storage array and checked the database workload performance as shown below.

Chart, line chartDescription automatically generated

As shown in the screenshot above, we noticed no disruption in public, private and storage network traffic even after multiple IOM links failure from both the Chassis because of the Cisco UCS Port-Channel Feature.

Test 2: FI – A Failure

We conducted a hardware failure test on FI – A by disconnecting power cable to the Fabric Interconnect Switch.

The figure below illustrates how during FI – A switch failure, the respective nodes (flex1, flex2, flex3 and flex4) on chassis 1 and nodes (flex5, flex6, flex7 and flex8) on chassis 2 will fail over the public network interface MAC addresses and its VLAN network traffic 134 to FI – B. However, storage traffic VSANs from FI – A switch were not able to failover to FI – B because of those storage interfaces traffic is not capable of failing over to another switch.

DiagramDescription automatically generated

Log into FI – B and type “connect nxos” then type “show mac address-table” to see all VLAN connection on FI – B.

A picture containing text, electronicsDescription automatically generated

In the screenshot above, we noticed when the FI – A failed, it would route all the Public Network traffic of VLAN 134 to FI – B. So, FI – A Failover did not cause any disruption to Private and Public Network Traffic. However, Storage Network Traffic for VSAN 151 were not able to fail-over to another FI Switch and thus we lost half of the storage traffic connectivity from the Oracle RAC Databases to Storage Array. The below screenshot shows the NetApp Storage Array performance of the mixed workloads on all the databases while one of the FI failed.

Chart, line chartDescription automatically generated

We also recorded performance of the databases from the storage array “Q S P S (qos statistics performance show)” when all the databases were running the workloads and FI – A failure occurred as shown below.

Graphical user interfaceDescription automatically generated

When we disconnect power from FI – A, it causes momentary impact on performance on overall total IOPS, latency on OLTP as well as throughput on DSS database for few seconds but notice that we did not see any interruption in any Private, Public and Storage Network Traffic on IO Service Requests to the storage.

We noticed this behavior because each server node can failover vNICs from one fabric interconnect switch to another fabric interconnect switch but there is no vHBA storage traffic failover from one fabric interconnect switch to another fabric interconnect switch. Therefore, in case of any one fabric interconnect failure, we would lose half of the number of vHBAs or storage paths and consequently we observed momentary databases performance impact for few seconds on the overall system as shown above.

After plugging back power cable to FI – A Switch, the respective nodes (flex1, flex2, flex3 and flex4) on chassis 1 and nodes (flex5, flex6, flex7 and flex8) on chassis 2 will route back the MAC addresses and its VLAN public network traffic 134 to FI – A. After FI – A arrives in normal operating state, all the nodes to storage connectivity, the operating system level multipath configuration will bring back all the path back to active and database performance will resume to peak performance.

Test 3: FI – B Failure

Similarly, we conducted a hardware failure test on FI – B by disconnecting power cable to the Fabric Interconnect B Switch.

The figure below illustrates how during FI – B switch failure, the respective nodes (flex1, flex2, flex3 and flex4) on chassis 1 and nodes (flex5, flex6, flex7 and flex8) on chassis 2 will fail over the private network interface MAC addresses and its VLAN network traffic 10 to FI – A. However, storage traffic VSANs from FI – B switch were not able to failover to FI – A because of those storage interfaces traffic is not capable of failing over to another switch.

DiagramDescription automatically generated

Log into FI – A and type “connect nxos” then type “show mac address-table” to see all VLAN connection on FI – A.

Graphical user interfaceDescription automatically generated

We noticed in the screenshot above, when the FI – B failed, it would route all the Private Network traffic of VLAN 10 to FI – A. So, FI – B Failover did not cause any disruption to Private and Public Network Traffic. However, Storage Network Traffic for VSAN 152 were not able to fail-over to another FI Switch and thus we lost half of the storage traffic connectivity from the Oracle RAC Databases to Storage Array. The screenshot below shows the NetApp Storage Array performance of the mixed workloads on all the databases while one of the FI failed.

Line chartDescription automatically generated with medium confidence

We also recorded performance of the databases from the storage array “Q S P S (qos statistics performance show)” when all the databases were running the workloads and FI – B failure occurred. Similar to previous FI – A failure test, in this FI – B failure test, we captured momentary impact on performance on overall total IOPS, latency and throughput for few seconds but we did not see any interruption in any Private, Public and Storage Network Traffic on IO Service Requests to the storage.

We noticed this behavior because each server node can failover vNICs from one fabric interconnect switch to another fabric interconnect switch but there is no vHBA storage traffic failover from one fabric interconnect switch to another fabric interconnect switch. Therefore, in case of any one fabric interconnect failure, we would lose half of the number of vHBAs or storage path and consequently we observe momentary databases performance impact for few seconds on the overall system as shown above.

After plugging back power cable to FI – B Switch, the respective nodes (flex1, flex2, flex3 and flex4) on chassis 1 and nodes (flex5, flex6, flex7 and flex8) on chassis 2 will route back the MAC addresses and its VLAN private network traffic 10 to FI – B. After FI – B arrives in normal operating state, all the nodes to storage connectivity, the operating system level multipath configuration will bring back all the path back to active and database performance will resume to peak performance.

Test 4 and 5: MDS Switch Failure

We conducted hardware failure test on MDS Switch – A by disconnecting power cable to the Switch and checking the storage network traffic on MDS Switch – B and overall system as shown below.

DiagramDescription automatically generated

Similar to FI failure tests, we observed some impact on all three databases performance as we lost half of the VSAN (VSAN-A 151) traffic. While VSAN-A (151) stays locally into the switch and only carry storage traffic through the MDS switch A, VSAN-A doesn’t failover to MDS Switch B therefore we reduced server to storage connectivity into half during MDS Switch A failure. However, failure in MDS Switch did not cause any disruption to Private and Public Network Traffic.

We also recorded performance of the databases from the storage array “Q S P S “where we observed momentary impact on performance on overall IOPS, latency on OLTP as well as throughput on DSS database for few seconds.

After plugging back power cable to MDS Switch A, the operating system level multipath configuration will bring back all the path back to active and database performance will resume to peak performance.

Test 6: Storage Controller Links Failure

We performed storage controller link failure test by disconnecting two of the FC links from the NetApp Array from each of the storage controller as shown below:

DiagramDescription automatically generated

Similar to FI and MDS failure tests, storage link failure did not cause any disruption to Private, Public and Storage Network Traffic. After plugging back FC links to storage controller, MDS Switch and Storage array links comes back online, and the operating system level multipath configuration will bring back all the path back to active and database performance will resume to peak performance.

Test 7: Server Node Failure

In this test, we power down one node from the RAC cluster and run the swingbench workload on all of the databases to check the overall system performance. We didn’t observe any performance impact on overall database IOPS, latency and throughput after losing one node from the system.

We completed an additional failure scenario and validated that there is no single point of failure in this reference design.

Summary

The Cisco Unified Computing System™ (Cisco UCS®) is a next-generation data center platform that unites computing, network, storage access, and virtualization into a single cohesive system. Cisco UCS is an ideal platform for the architecture of mission critical database workloads such as Oracle RAC. The FlexPod Datacenter with NetApp All Flash AFF system is a converged infrastructure platform that combines best-of breed technologies from Cisco and NetApp into a powerful converged platform for enterprise applications. The pre-validated FlexPod architecture delivers proven value, agility, and performance that drive higher productivity, faster decision making, and greater opportunities for growth.

An essential feature for Oracle databases deployed on shared enterprise system is the ability to deliver consistent and dependable high performance. High performance must be coupled with non-disruptive operations, high availability, scalability, and storage efficiency. Customers can depend on Cisco UCS and NetApp Clustered Data ONTAP Storage to provide these essential elements. Built on clustered Data ONTAP unified scale-out architecture, AFF consistently meets or exceeds the high-performance demands of Oracle databases. It also provides rich data management capabilities, such as integrated data protection and non-disruptive upgrades and data migration. These features allow customers to eliminate performance silos and seamlessly integrate AFF into a shared infrastructure.

Clustered Data ONTAP 9.7 delivers an enhanced inline compression capability that significantly reduces the amount of flash storage required and carries near-zero effects on system performance and doesn’t impact the application performance. The inline de-duplication won’t help databases much but doesn’t cause any harm in terms of CPU cycles when it is turned on and can help other non-database objects sharing the same storage cluster. The combination of Cisco UCS, NetApp and Oracle Real Application Cluster Database architecture can provide the following benefits to accelerate your IT transformation

Cisco UCS stateless computing architecture provided by the Service Profile capability of Cisco UCS allows fast, non-disruptive workload changes to be executed simply and seamlessly across the integrated UCS infrastructure and Cisco x86 servers.

A single platform built from unified compute, fabric, and storage technologies, allowing you to scale to large-scale data centers without architectural changes

Enabling faster deployments, greater flexibility of choice, efficiency, high availability, and lower risk.

Appendix

MDS 9132T Switch Configuration

ORA19C-FLEXPOD-MDS-A# show running-config

version 8.4(1)

power redundancy-mode redundant

feature npiv

feature fport-channel-trunk

feature telnet

role name default-role

  description This is a system defined role and applies to all users.

  rule 5 permit show feature environment

  rule 4 permit show feature hardware

  rule 3 permit show feature module

  rule 2 permit show feature snmp

  rule 1 permit show feature system

ip domain-lookup

ip host ORA19C-FLEXPOD-MDS-A  10.29.134.50

aaa group server radius radius

snmp-server user admin network-admin auth md5 0xeafc179c4eb2875cdbca434a323e13b2 priv 0xeafc179c4eb2875cdbca434a323e13b2 localizedkey

vsan database

  vsan 151 name "VSAN-FI-A"

device-alias mode enhanced

device-alias database

  device-alias name Flex1-hba0 pwwn 20:00:00:25:b5:99:aa:00

  device-alias name Flex1-hba2 pwwn 20:00:00:25:b5:99:aa:01

  device-alias name Flex2-hba0 pwwn 20:00:00:25:b5:99:aa:02

  device-alias name Flex2-hba2 pwwn 20:00:00:25:b5:99:aa:03

  device-alias name Flex3-hba0 pwwn 20:00:00:25:b5:99:aa:04

  device-alias name Flex3-hba2 pwwn 20:00:00:25:b5:99:aa:05

  device-alias name Flex4-hba0 pwwn 20:00:00:25:b5:99:aa:06

  device-alias name Flex4-hba2 pwwn 20:00:00:25:b5:99:aa:07

  device-alias name Flex5-hba0 pwwn 20:00:00:25:b5:99:aa:08

  device-alias name Flex5-hba2 pwwn 20:00:00:25:b5:99:aa:09

  device-alias name Flex6-hba0 pwwn 20:00:00:25:b5:99:aa:0a

  device-alias name Flex6-hba2 pwwn 20:00:00:25:b5:99:aa:0b

  device-alias name Flex7-hba0 pwwn 20:00:00:25:b5:99:aa:0c

  device-alias name Flex7-hba2 pwwn 20:00:00:25:b5:99:aa:0d

  device-alias name Flex8-hba0 pwwn 20:00:00:25:b5:99:aa:0e

  device-alias name Flex8-hba2 pwwn 20:00:00:25:b5:99:aa:0f

  device-alias name A800-NVMe-01-2a pwwn 20:14:00:a0:98:b9:25:08

  device-alias name A800-NVMe-02-2a pwwn 20:16:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-01-2a pwwn 20:01:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-01-2b pwwn 20:02:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2a pwwn 20:05:00:a0:98:b9:25:08

  device-alias name FlexPod-A800-02-2b pwwn 20:06:00:a0:98:b9:25:08

device-alias commit

system default zone distribute full

zone smart-zoning enable vsan 151

zoneset distribute full vsan 151

!Active Zone Database Section for vsan 151

zone name Flex1A-Boot vsan 151

    member device-alias Flex1-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex2A-Boot vsan 151

    member device-alias Flex2-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex3A-Boot vsan 151

    member device-alias Flex3-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex4A-Boot vsan 151

    member device-alias Flex4-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex5A-Boot vsan 151

    member device-alias Flex5-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex6A-Boot vsan 151

    member device-alias Flex6-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex7A-Boot vsan 151

    member device-alias Flex7-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex8A-Boot vsan 151

    member device-alias Flex8-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex1A-NVMe vsan 151

    member device-alias Flex1-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex2A-NVMe vsan 151

    member device-alias Flex2-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex3A-NVMe vsan 151

    member device-alias Flex3-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex4A-NVMe vsan 151

    member device-alias Flex4-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex5A-NVMe vsan 151

    member device-alias Flex5-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex6A-NVMe vsan 151

    member device-alias Flex6-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex7A-NVMe vsan 151

    member device-alias Flex7-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex8A-NVMe vsan 151

    member device-alias Flex8-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zoneset name Flex-A vsan 151

    member Flex1A-Boot

    member Flex2A-Boot

    member Flex3A-Boot

    member Flex4A-Boot

    member Flex5A-Boot

    member Flex6A-Boot

    member Flex7A-Boot

    member Flex8A-Boot

    member Flex1A-NVMe

    member Flex2A-NVMe

    member Flex3A-NVMe

    member Flex4A-NVMe

    member Flex5A-NVMe

    member Flex6A-NVMe

    member Flex7A-NVMe

    member Flex8A-NVMe

zoneset activate name Flex-A vsan 151

do clear zone database vsan 151

!Full Zone Database Section for vsan 151

zone name Flex1A-Boot vsan 151

    member device-alias Flex1-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex2A-Boot vsan 151

    member device-alias Flex2-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex3A-Boot vsan 151

    member device-alias Flex3-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex4A-Boot vsan 151

    member device-alias Flex4-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex5A-Boot vsan 151

    member device-alias Flex5-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex6A-Boot vsan 151

    member device-alias Flex6-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex7A-Boot vsan 151

    member device-alias Flex7-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex8A-Boot vsan 151

    member device-alias Flex8-hba0 init

    member device-alias FlexPod-A800-01-2a target

    member device-alias FlexPod-A800-02-2a target

zone name Flex1A-NVMe vsan 151

    member device-alias Flex1-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex2A-NVMe vsan 151

    member device-alias Flex2-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex3A-NVMe vsan 151

    member device-alias Flex3-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex4A-NVMe vsan 151

    member device-alias Flex4-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex5A-NVMe vsan 151

    member device-alias Flex5-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex6A-NVMe vsan 151

    member device-alias Flex6-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex7A-NVMe vsan 151

    member device-alias Flex7-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zone name Flex8A-NVMe vsan 151

    member device-alias Flex8-hba2 init

    member device-alias A800-NVMe-01-2a target

    member device-alias A800-NVMe-02-2a target

zoneset name Flex-A vsan 151

    member Flex1A-Boot

    member Flex2A-Boot

    member Flex3A-Boot

    member Flex4A-Boot

    member Flex5A-Boot

    member Flex6A-Boot

    member Flex7A-Boot

    member Flex8A-Boot

    member Flex1A-NVMe

    member Flex2A-NVMe

    member Flex3A-NVMe

    member Flex4A-NVMe

    member Flex5A-NVMe

    member Flex6A-NVMe

    member Flex7A-NVMe

    member Flex8A-NVMe

interface mgmt0

  ip address 10.29.134.50 255.255.255.0

interface port-channel251

  switchport trunk allowed vsan 151

  switchport description ORA19C-FlexPod-FI-A

  switchport rate-mode dedicated

  switchport trunk mode off

vsan database

  vsan 151 interface port-channel251

  vsan 151 interface fc1/5

  vsan 151 interface fc1/6

  vsan 151 interface fc1/7

  vsan 151 interface fc1/8

  vsan 151 interface fc1/9

  vsan 151 interface fc1/10

  vsan 151 interface fc1/11

  vsan 151 interface fc1/12

switchname ORA19C-FLEXPOD-MDS-A

cli alias name autozone source sys/autozone.py

line console

line vty

boot kickstart bootflash:/m9100-s6ek9-kickstart-mz.8.4.1.bin

boot system bootflash:/m9100-s6ek9-mz.8.4.1.bin

interface fc1/1

  switchport description ORA19C-FlexPod-FI-A-1/1

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

interface fc1/2

  switchport description ORA19C-FlexPod-FI-A-1/2

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

interface fc1/3

  switchport description ORA19C-FlexPod-FI-A-1/3

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

interface fc1/4

  switchport description ORA19C-FlexPod-FI-A-1/4

  switchport trunk mode off

  port-license acquire

  channel-group 251 force

  no shutdown

interface fc1/5

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-01-2a

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/6

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-01-2b

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/7

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-02-2a

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/8

  switchport trunk allowed vsan 151

  switchport description FlexPod-A800-02-2b

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/9

  switchport trunk allowed vsan 151

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/10

  switchport trunk allowed vsan 151

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/11

  switchport trunk allowed vsan 151

  switchport trunk mode off

  port-license acquire

  no shutdown

interface fc1/12

  switchport trunk allowed vsan 151

  switchport trunk mode off

  port-license acquire

  no shutdown

ip default-gateway 10.29.134.1

Cisco Nexus 9336C-FX2 Switch Configuration

ORA19C-FLEXPOD-N9K-A# show running-config

version 9.3(3) Bios:version 05.40

hostname ORA19C-FLEXPOD-N9K-A

policy-map type network-qos jumbo

  class type network-qos class-default

    mtu 9216

cfs eth distribute

feature udld

feature interface-vlan

feature hsrp

feature lacp

feature vpc

feature lldp

ip domain-lookup

system default switchport

system qos

  service-policy type network-qos jumbo

vlan 1,10,134

vlan 10

  name Oracle_RAC_Private_Network

vlan 134

  name Oracle_RAC_Public_Network

spanning-tree port type edge bpduguard default

spanning-tree port type network default

vrf context management

  ip route 0.0.0.0/0 10.29.134.1

port-channel load-balance src-dst l4port

vpc domain 1

  peer-keepalive destination 10.29.134.53 source 10.29.134.52

  auto-recovery

interface Vlan1

interface Vlan134

  no shutdown

  ip address 10.29.134.253/24

interface port-channel1

  description vPC peer-link

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type network

  vpc peer-link

interface port-channel51

  description Port-Channel FI-A

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 51

interface port-channel52

  description Port-Channel FI-B

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  vpc 52

interface Ethernet1/1

  description Peer link 100g connected to N9K-B-Eth1/1

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  channel-group 1 mode active

interface Ethernet1/2

  description Peer link 100g connected to N9K-B-Eth1/2

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  channel-group 1 mode active

interface Ethernet1/25

  description 100g link to Fabric-Interconnect A port 49

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 51 mode active

interface Ethernet1/26

  description 100g link to Fabric-Interconnect B Port 49

  switchport mode trunk

  switchport trunk allowed vlan 1,10,134

  spanning-tree port type edge trunk

  mtu 9216

  channel-group 52 mode active

interface Ethernet1/31

  description connect to uplink switch

  switchport access vlan 134

  speed 1000

interface mgmt0

  vrf member management

  ip address 10.29.134.52/24

line console

line vty

boot nxos bootflash:/nxos.9.3.3.bin

no system default switchport shutdown

Multipath Configuration “/etc/multipath.conf”

[root@flex1 ~]# cat /etc/multipath.conf

defaults {

        find_multipaths yes

        user_friendly_names yes

        enable_foreign NONE

}

blacklist {

}

multipaths {

        multipath {

                wwid    3600a098038304437575d507956514145

                alias   Flex1-OS

        }

}

Configuration of “/etc/sysctl.conf”

[root@flex1 ~]# cat /etc/sysctl.conf

vm.nr_hugepages=78000

# oracle-database-preinstall-19c setting for fs.file-max is 6815744

fs.file-max = 6815744

# oracle-database-preinstall-19c setting for kernel.sem is '250 32000 100 128'

kernel.sem = 250 32000 100 128

# oracle-database-preinstall-19c setting for kernel.shmmni is 4096

kernel.shmmni = 4096

# oracle-database-preinstall-19c setting for kernel.shmall is 1073741824 on x86_64

kernel.shmall = 1073741824

# oracle-database-preinstall-19c setting for kernel.shmmax is 4398046511104 on x86_64

kernel.shmmax = 4398046511104

# oracle-database-preinstall-19c setting for kernel.panic_on_oops is 1 per Orabug 19212317

kernel.panic_on_oops = 1

# oracle-database-preinstall-19c setting for net.core.rmem_default is 262144

net.core.rmem_default = 262144

# oracle-database-preinstall-19c setting for net.core.rmem_max is 4194304

net.core.rmem_max = 4194304

# oracle-database-preinstall-19c setting for net.core.wmem_default is 262144

net.core.wmem_default = 262144

# oracle-database-preinstall-19c setting for net.core.wmem_max is 1048576

net.core.wmem_max = 1048576

# oracle-database-preinstall-19c setting for net.ipv4.conf.all.rp_filter is 2

net.ipv4.conf.all.rp_filter = 2

# oracle-database-preinstall-19c setting for net.ipv4.conf.default.rp_filter is 2

net.ipv4.conf.default.rp_filter = 2

# oracle-database-preinstall-19c setting for fs.aio-max-nr is 1048576

fs.aio-max-nr = 1048576

# oracle-database-preinstall-19c setting for net.ipv4.ip_local_port_range is 9000 65500

net.ipv4.ip_local_port_range = 9000 65500

Configuration of “/etc/security/limits.d/oracle-database-preinstall-19c.conf”

[root@flex1 ~]# cat /etc/security/limits.d/oracle-database-preinstall-19c.conf

# oracle-database-preinstall-19c setting for nofile soft limit is 1024

oracle   soft   nofile    1024

# oracle-database-preinstall-19c setting for nofile hard limit is 65536

oracle   hard   nofile    65536

# oracle-database-preinstall-19c setting for nproc soft limit is 16384

# refer orabug15971421 for more info.

oracle   soft   nproc    16384

# oracle-database-preinstall-19c setting for nproc hard limit is 16384

oracle   hard   nproc    16384

# oracle-database-preinstall-19c setting for stack soft limit is 10240KB

oracle   soft   stack    10240

# oracle-database-preinstall-19c setting for stack hard limit is 32768KB

oracle   hard   stack    32768

# oracle-database-preinstall-19c setting for memlock hard limit is maximum of 128GB on x86_64 or 3GB on x86 OR 90 % of RAM

### oracle   hard   memlock    237010305

### oracle   hard   memlock    237010882

oracle   hard   memlock    474832847

# oracle-database-preinstall-19c setting for memlock soft limit is maximum of 128GB on x86_64 or 3GB on x86 OR 90% of RAM

### oracle   soft   memlock    237010305

### oracle   soft   memlock    237010882

oracle   soft   memlock    474832847

# oracle-database-preinstall-19c setting for data soft limit is 'unlimited'

oracle   soft   data    unlimited

# oracle-database-preinstall-19c setting for data hard limit is 'unlimited'

oracle   hard   data    unlimited

Configuration of “/etc/udev/rules.d/71-nvme-iopolicy-netapp-ONTAP.rules”

[root@flex1 ~]# cat /etc/udev/rules.d/71-nvme-iopolicy-netapp-ONTAP.rules

### Enable round-robin for NetApp ONTAP

ACTION=="add", SUBSYSTEM=="nvme-subsystem", ATTR{model}=="NetApp ONTAP Controller", ATTR{iopolicy}="round-robin"

Configuration of “/etc/udev/rules.d/80-nvme.rules”

[root@flex1 ~]# cat /etc/udev/rules.d/80-nvme.rules

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.7636791b-944f-4adb-941e-0d962639f718", SYMLINK+="asm1", GROUP:="oinstall", OWNER:="grid", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.1251e1c3-3735-4546-8b39-6654c83620f7", SYMLINK+="asm2", GROUP:="oinstall", OWNER:="grid", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.e2c15660-bc0c-4bfb-a360-02011bb55b06", SYMLINK+="datacdb01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.374102da-60ab-436f-96cd-0a8b4d6d75c8", SYMLINK+="datacdb02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.4c1b87f3-f3e2-432e-b13d-435adc5e75d5", SYMLINK+="datacdb03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.b9232ee5-afe2-4237-913a-f6b213d382f6", SYMLINK+="datacdb04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.6fc2542d-ba50-42a2-907b-e7375709ee0f", SYMLINK+="datadss01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.710065e5-c4b1-43af-ae1b-d28662d921fb", SYMLINK+="datadss02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.90336e21-bec1-4da5-b64e-d988e8f689c3", SYMLINK+="datadss03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.d51f7c79-f7aa-4155-9c92-7e4532e11560", SYMLINK+="datadss04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.4a7533cc-f86a-4787-be48-05263a929c42", SYMLINK+="dataslob01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.ffc6a134-0f79-4748-8dfe-73bf7a778729", SYMLINK+="dataslob02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.231aa3d6-6a14-4ea1-9e4f-32145a97817f", SYMLINK+="dataslob03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.0b53d533-6af1-4e1b-ad64-424c0ceaa569", SYMLINK+="dataslob04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.938e9f50-0030-471c-b4f7-9f2cd0f45aec", SYMLINK+="dataslob05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.faf7a66b-2b20-4a28-8f20-8a2ddd06856d", SYMLINK+="dataslob06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.76a65ede-9e0b-4976-8e3d-26c0e6ba859d", SYMLINK+="dataslob07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.ad3cc536-4fc1-4844-880a-b5ea6dd74a8c", SYMLINK+="dataslob08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f373f271-7a63-418f-ac52-171b0d8fc302", SYMLINK+="dataslob09", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.a2df6615-c3ab-4233-93fb-49c0f12eece3", SYMLINK+="dataslob10", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.af831613-fe2c-4fe2-a495-df3de7b92bb1", SYMLINK+="dataslob11", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.45cc5b21-f780-4ec2-8c3a-a35baa28b1f4", SYMLINK+="dataslob12", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c4e18aaa-2b0a-44a9-92c4-6c5aaa9e819d", SYMLINK+="dataslob13", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.90cb2720-f977-4a00-8af7-6ba61c95cb56", SYMLINK+="dataslob14", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5307ccf1-9b73-430f-812a-a3effa547943", SYMLINK+="dataslob15", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.59618b5f-921f-4a21-b7b2-3e6b7c6e30e7", SYMLINK+="dataslob16", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.b579894d-db9d-400b-9f15-c6317533af0f", SYMLINK+="dataslob17", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5199ba10-848c-42c7-bd7d-c1a91b9a91d2", SYMLINK+="dataslob18", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.04af9123-01c3-41cb-a315-65736f0684e8", SYMLINK+="dataslob19", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3afda788-8222-47ae-a153-400c322c1408", SYMLINK+="dataslob20", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f75037d6-cc82-4e26-9b2f-47621a21ead0", SYMLINK+="dataslob21", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.04ee767c-77e3-4e63-921d-f62efdd2c20d", SYMLINK+="dataslob22", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.6814b2f4-0733-413d-b8f5-1e37b3c6846d", SYMLINK+="dataslob23", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.65b82f55-b9d6-41d6-a4f2-55dd8f88a174", SYMLINK+="dataslob24", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.6b229c62-ae32-42f7-8fd3-abfb6cb35408", SYMLINK+="pdbfin01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f0d78685-078e-4d64-bd18-0a00a9376581", SYMLINK+="pdbfin02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5fba1916-9f1e-44e5-88dd-cf318988a8e4", SYMLINK+="pdbfin03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3b2d3903-5381-49b5-90b2-76fb08744fbe", SYMLINK+="pdbfin04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.a139aabf-1a98-4b36-aae9-6d18456b68b1", SYMLINK+="pdbfin05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5f65d480-14ea-40e6-81df-8c6d76dd76a3", SYMLINK+="pdbfin06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.48c7945c-a9ac-4e08-938f-af923e949aa2", SYMLINK+="pdbfin07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.e8cf19ef-1842-4f36-bdf6-d92b0009ad13", SYMLINK+="pdbfin08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.e66f783e-5769-4a39-92c8-64d641bdfe4f", SYMLINK+="pdbfin09", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.367ad79d-631d-408f-adb5-d3fac028fe9b", SYMLINK+="pdbfin10", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.d7cd43e9-a2e5-43f9-b186-a98e2e630c57", SYMLINK+="pdbfin11", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.b0f03462-806f-4adc-83d1-75137e450efe", SYMLINK+="pdbfin12", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.49a82540-18a0-4331-9830-ffe7bb670a32", SYMLINK+="pdbfin13", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.7ac01cc2-8fe6-4a84-bb10-b18ce3e7ae15", SYMLINK+="pdbfin14", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.b9cf3604-5b01-4542-8400-a1574554fc59", SYMLINK+="pdbfin15", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.26545951-0b48-475f-9d6e-8a74f0374d40", SYMLINK+="pdbfin16", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c26457e9-eb27-48e4-aeaa-02360902f0f8", SYMLINK+="pdbfin17", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.edbe0357-ee10-423b-ba8a-79a5f661367a", SYMLINK+="pdbfin18", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.97b8825c-d622-4e43-93c9-c9f26d57898b", SYMLINK+="pdbfin19", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.30c52011-eaf0-4fed-971a-f136b82882c3", SYMLINK+="pdbfin20", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f317569a-0e7f-40b1-b565-be11fa0e311e", SYMLINK+="pdbfin21", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.bae4f114-86e8-4f0d-a39f-b9bf017bb211", SYMLINK+="pdbfin22", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.171eff22-a071-4502-9c95-ef8bea0d6d51", SYMLINK+="pdbfin23", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3674baf6-f803-446e-9e96-2b5bd858ef1c", SYMLINK+="pdbfin24", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.56e99801-e527-46b7-993c-a047707dd937", SYMLINK+="pdbsh01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.8207dcc9-5d7a-4a21-aa6d-bcdab22a1ac8", SYMLINK+="pdbsh02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.cc32591e-47fb-4ae0-8ffe-434501787335", SYMLINK+="pdbsh03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.402b14c8-7e79-4973-afca-f09d061681a7", SYMLINK+="pdbsh04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.67858129-a63d-4780-9a45-5cbfc271648d", SYMLINK+="pdbsh05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c5bfa12e-6c38-4c4a-a283-8f4b7f936452", SYMLINK+="pdbsh06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.9611a6eb-0867-4bae-9387-10270a5226a5", SYMLINK+="pdbsh07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c13cd79f-a445-45b6-a24b-9c88857ef96e", SYMLINK+="pdbsh08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.43d072c9-d76a-4c5e-a1d4-1b109e54a273", SYMLINK+="pdbsh09", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.a3916b8b-5fde-453f-8900-466a7b507b37", SYMLINK+="pdbsh10", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.01840de9-4ad6-4665-a938-f5a6ef1466cc", SYMLINK+="pdbsh11", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.ab58e2f5-52b8-4944-9bff-035117445749", SYMLINK+="pdbsh12", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.965a7b77-d0ff-4618-b7a3-854bc33d00c2", SYMLINK+="pdbsh13", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.8193ccea-fa8b-471d-a5f9-567c16f02896", SYMLINK+="pdbsh14", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.d7a78b72-dd7b-499b-8106-45593a3530b0", SYMLINK+="pdbsh15", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.4e953209-16f9-4a5b-a0a1-cc67148fd73f", SYMLINK+="pdbsh16", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.fa2bf4df-4112-4cb7-bf05-d53eaa3d0ee8", SYMLINK+="pdbsh17", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.2ae34a39-33c4-459f-bfc1-04a31a29eea8", SYMLINK+="pdbsh18", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.12efdae5-d8f0-449f-aef7-a5e6c3df316d", SYMLINK+="pdbsh19", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f32fe1c9-88da-4091-ac4b-d0038c4ea006", SYMLINK+="pdbsh20", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.69e16e6d-6a44-46c9-84ef-02a7ab26f3d1", SYMLINK+="pdbsh21", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f02733b8-1d5f-4ec4-add7-9961068c8eea", SYMLINK+="pdbsh22", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c4f9a590-92db-48c3-b8e7-1f0646a50c66", SYMLINK+="pdbsh23", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.cc7b6a92-0e7d-40e1-bb9a-c4575945591c", SYMLINK+="pdbsh24", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.311f295a-6232-4d03-a58d-dfcfb9ef2e49", SYMLINK+="pdbsoe01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.a93a0661-1872-489e-b2ef-14c726d914bb", SYMLINK+="pdbsoe02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.06ce3ec7-d163-4971-a864-033034630d16", SYMLINK+="pdbsoe03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c25759df-e940-4910-9068-27a853ab73b8", SYMLINK+="pdbsoe04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.dafe4860-42a9-4b1b-96e1-26149efebec8", SYMLINK+="pdbsoe05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f2faffe2-3dda-4cc6-a691-2cc4430a34d4", SYMLINK+="pdbsoe06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c2afc715-6688-410e-8d9c-fe8eac7c621f", SYMLINK+="pdbsoe07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.18d67b41-3c97-4fd2-b082-c9827fa55727", SYMLINK+="pdbsoe08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.0c1be47a-683b-4a0e-80b5-9e5e9a7426af", SYMLINK+="pdbsoe09", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.e7a04d6e-d783-4379-82b1-fedea30fae6b", SYMLINK+="pdbsoe10", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.741306b6-ec1d-4059-a594-65575383bd16", SYMLINK+="pdbsoe11", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.ea6297c5-0e81-412b-97a3-9960b06bfa0c", SYMLINK+="pdbsoe12", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.29645ebd-6a00-4e57-abeb-20de0551a2e2", SYMLINK+="pdbsoe13", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.07a78a88-2469-4849-9681-31fe884d8ea8", SYMLINK+="pdbsoe14", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.7501b410-8424-4e19-b240-a4b19a822574", SYMLINK+="pdbsoe15", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.6c985ea5-0fea-4ac8-a845-dfa3ea6f4493", SYMLINK+="pdbsoe16", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.78f59c23-e674-4ff6-95af-3d7fd2486831", SYMLINK+="pdbsoe17", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.d1ff7a09-1ef7-4371-b5dd-5957c8ca0417", SYMLINK+="pdbsoe18", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.885f272e-9ab1-4f3b-b6ce-2e417c6cb679", SYMLINK+="pdbsoe19", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.29f1816f-9f19-42e4-b7b8-8f339b05b633", SYMLINK+="pdbsoe20", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.cfdd83c7-479e-473e-8f4b-1c90f23405fc", SYMLINK+="pdbsoe21", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.9f29dd56-b6ed-45a1-9a00-cac28d6ddb3f", SYMLINK+="pdbsoe22", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.4374abc9-e1b7-42e0-bb54-d06c2b94216e", SYMLINK+="pdbsoe23", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.1d0694ca-f96f-444e-b5df-d9e4fba06e7d", SYMLINK+="pdbsoe24", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.0b1f4c21-4267-494f-a80e-a8606faa6ad1", SYMLINK+="redocdb01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.a1d3752b-5187-4d42-9ec5-aefeeb6ea741", SYMLINK+="redocdb02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.fae25270-8c19-45b3-86af-6073d30a76dc", SYMLINK+="redocdb03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.81cf333d-6a1e-4585-92db-b4f8833bd0f6", SYMLINK+="redocdb04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.d4ef48cd-5492-4071-84a2-48711be2293c", SYMLINK+="redocdb05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.aa4b4ce8-e38c-47ef-8b43-7dc62f8370f7", SYMLINK+="redocdb06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5e3d67e1-c312-4b35-ab6b-ec6dfc19e2c4", SYMLINK+="redocdb07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.b5d5d31c-4783-49ad-9684-91c10cf146ed", SYMLINK+="redocdb08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.82b7679a-a533-40ce-a95c-3994582b955a", SYMLINK+="redodss01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.09717858-7bd6-42df-9785-61f2052dedc5", SYMLINK+="redodss02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.347ab1a7-ac68-4981-9305-f0bb8cfdd1db", SYMLINK+="redodss03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3dd650cc-a441-4e47-8bec-6dfa526b0a61", SYMLINK+="redodss04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.48de5a34-9e5d-4084-a674-dfa354d9eddd", SYMLINK+="redodss05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.cb3c7194-5eb5-47a4-a037-f88fad986cbe", SYMLINK+="redodss06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.db7a16b6-839d-41e2-8fa6-fc592dc21953", SYMLINK+="redodss07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.761d202d-0d89-4352-b2ad-bb1e72961a6b", SYMLINK+="redodss08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.be8fd420-abbd-48fe-b104-666f0df4319a", SYMLINK+="redoslob01", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.c9fa0299-e631-4ec6-8526-1c34431f6410", SYMLINK+="redoslob02", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.e45cd643-3e39-47a9-b39f-20ebeecd73c9", SYMLINK+="redoslob03", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.f655315e-cbec-4f21-9f88-cbeaf58646f2", SYMLINK+="redoslob04", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.86cdec61-9a51-4756-8c6d-e354ddf5310a", SYMLINK+="redoslob05", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3063ae18-4ac7-4b6f-b36d-91bf38fcfb1b", SYMLINK+="redoslob06", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.3a783797-eff0-471a-9580-04b4bd284c6a", SYMLINK+="redoslob07", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

KERNEL=="nvme[0-9]*n[0-9]*", ENV{DEVTYPE}=="disk", ENV{ID_WWN}=="uuid.5253642b-0dbd-4ec3-a97a-0312ed38074f", SYMLINK+="redoslob08", GROUP:="oinstall", OWNER:="oracle", MODE:="660"

References

The following references were used in preparing this documents.

Cisco Unified Computing System

https://www.cisco.com/c/en/us/products/servers-unified-computing/index.html

Cisco UCS B200 M5 Servers

https://www.cisco.com/c/en/us/products/collateral/servers-unified-computing/ucs-b-series-blade-servers/datasheet-c78-739296.html

Oracle Database 19c

https://docs.oracle.com/en/database/oracle/oracle-database/index.html

NetApp AFF A-Series All Flash Storage

https://www.netapp.com/us/products/storage-systems/all-flash-array/aff-a-series.aspx

Cisco UCS Data Center Design Guides

https://www.cisco.com/c/en/us/solutions/design-zone/data-center-design-guides/data-center-design-guides-all.html#Hyperconverged

FlexPod Converged Infrastructure

https://www.cisco.com/c/en/us/solutions/data-center-virtualization/flexpod/index.html#~tab-resources

https://www.netapp.com/us/products/converged-systems/flexpod-converged-infrastructure.aspx

NetApp Support

https://mysupport.netapp.com/

About the Authors

Tushar Patel, Principal Engineer, CSPG UCS Product Management and Data Center Solutions Engineering Group, Cisco Systems, Inc.

Tushar Patel is a Principal Engineer in Cisco Systems CSPG UCS Product Management and Data Center Solutions Engineering Group and a specialist in Flash Storage technologies and Oracle RAC RDBMS. Tushar has over 25 years of experience in Flash Storage architecture, Database architecture, design, and performance. Tushar also has strong background in Intel X86 architecture, hyper converged systems, Storage technologies and Virtualization. He has worked with large number of enterprise customers, evaluate, and deploy mission critical database solutions. Tushar has presented to both internal and external audiences at various conferences and customer events.

Hardikkumar Vyas, Technical Marketing Engineer, CSPG UCS Product Management and Data Center Solutions Engineering Group, Cisco Systems, Inc.

Hardikkumar Vyas is a Solution Engineer in Cisco Systems CSPG UCS Product Management and Data Center Solutions Engineering Group for configuring, implementing, and validating infrastructure best practices for highly available Oracle RAC databases solutions on Cisco UCS Servers, Cisco Nexus Products, and various Storage Technologies. Hardikkumar Vyas holds a master’s degree in Electrical Engineering and has over 8 years of experience working with Oracle RAC Databases and associated applications. Hardikkumar Vyas’s focus is developing database solutions on different platforms, perform benchmarks, prepare reference architectures, and write technical documents for Oracle RAC Databases on Cisco UCS Platforms.

Acknowledgements

For their support and contribution to the design, validation, and creation of this Cisco Validated Design, the authors would like to thank:

   Bobby Oommen, Technical Lead, FlexPod Solutions, NetApp

Feedback

For comments and suggestions about this guide and related guides, join the discussion on Cisco Community at https://cs.co/en-cvds.

NOTE: Available paragraph styles are listed in the Quick Styles Gallery in the Styles group on the Home tab. Alternatively, they can be accessed via the Styles window (press Alt + Ctrl + Shift + S).

Learn more