Installation Overview
The HA solution for Cisco Prime Optical is available in local redundancy and geographical redundancy configurations for Linux. For information about the different HA configurations available for Prime Optical, including hardware configuration and network diagrams, see the following sections:
Note
The servers in the following figures contain mirrored internal root disks. The servers must be the same make, model, and storage capacity. The 100BASE-T heartbeat cables between the primary and secondary servers in the dual-node configuration are cross-over (reversing) Ethernet cables. We recommend that you use a fault-tolerant, switched/routed network for communication with the HA servers.
Local Redundancy Configuration
The local redundancy configuration provides an automatic failover solution to specific software and single hardware failures without the need to reconfigure IP addresses on your switched/routed network.
The following figures illustrate a local redundancy HA configuration on Prime Optical for Linux environments.
- Figure 1-1 is a network diagram of Prime Optical in a local redundancy HA configuration.
- Figure 1-2 illustrates the hardware configuration of Prime Optical in a local redundancy HA configuration.
Figure 1-1 Prime Optical in a Local Redundancy HA Configuration
In a local redundancy configuration (Figure 1-1), there is a single cluster consisting of a primary and a secondary server.
Figure 1-2 Local Redundancy
In a local redundancy configuration, the primary and secondary servers share a common disk array (that is, the servers are both connected to a single disk array). This single disk array is where the Prime Optical application, Oracle application, and Prime Optical database are installed.
The heartbeat link connects the primary and secondary servers to monitor the health of both servers. The heartbeat link is achieved between the primary and secondary servers using two cross-over Ethernet cables. During normal operation, the Prime Optical and Oracle applications run on the primary server (normally the active server). If a fault occurs on the primary server, the Prime Optical and Oracle applications run on the secondary server; at this stage, the secondary server becomes the active server.
The primary server can be switched back to be the active server after the fault on the primary server is fixed.
Single-Node Cluster Geographical Redundancy Configuration
The geographical redundancy configuration allows you to locate two Prime Optical instances at geographically remote sites. One server instance is active; the other server instance is standby. The HA agent switches to the standby Element Management System (EMS) instance if an unrecoverable failure occurs on the active EMS instance.
In a single-node cluster geographical redundancy configuration, there are two clusters with different names (one on each node), each containing a server.
Both of the nodes status must be as:
For RHEL-6.x
CLUSTER STATUS SHOULD BE LIKE BELOW:
Primary server: HA1-107.cisco.com
Cluster Status for HA1_CLUSTER @ Wed Feb 11 17:40:18 2015
ha1-107.cisco.com 1 Online, Local, rgmanager
Service Name Owner (Last) State
------- ---- ----- ------ -----
service:ctm_heartbeat ha1-107.cisco.com started
service:ctm_monitoring ha1-107.cisco.com started
service:ctm_service ha1-107.cisco.com started
Standby Server: HA2-107.cisco.com
Cluster Status for HA2_CLUSTER @ Wed Feb 11 17:42:26 2015
ha2-107.cisco.com 1 Online, Local, rgmanager
Service Name Owner (Last) State
------- ---- ----- ------ -----
service:ctm_heartbeat ha2-107.cisco.com started
service:ctm_monitoring ha2-107.cisco.com started
service:ctm_service (ha2-107.cisco.com) disabled
##################################################
For RHEL-7.x
CLUSTER STATUS SHOULD BE LIKE BELOW:
Primary server: HA1-107.cisco.com
Last updated: Mon Dec 26 18:41:14 2016 Last change: Sat Dec 24 22:12:35 2016 by root via crm_resource on ha1-107.cisco.com
Current DC: ha1-107.cisco.com (version 1.1.13-10.el7-44eb2dd) - partition with quorum
1 node and 10 resources configured
Online: [ ha1-107.cisco.com ]
Resource Group: ctm_service
FSCheck (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/FsCheckAgent.py): Started ha1-107.cisco.com
NTW_IF (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/NtwIFAgent.py): Started ha1-107.cisco.com
CTM_RSYNC (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/RsyncAgent.py): Started ha1-107.cisco.com
REPL_IF (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ODG_IFAgent.py): Started ha1-107.cisco.com
ORACLE_REPLICATOR (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ODG_ReplicatorAgent.py): Started ha1-107.cisco.com
CTM_SID (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/OracleAgent.py): Started ha1-107.cisco.com
CTM_SRV (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/CtmAgent.py): Started ha1-107.cisco.com
CTM_APACHE (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ApacheAgent.py): Started ha1-107.cisco.com
Resource Group: ctm_heartbeat
CTM_HEARTBEAT (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/HeartBeat.py): Started ha1-107.cisco.com
Resource Group: ctm_monitoring
FLASHBACK (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/FlashBackMonitor.py): Started ha1-107.cisco.com
ha1-107.cisco.com: Online
pacemaker: active/enabled
Stand By server: Ha2-107.cisco.com
Last updated: Mon Dec 26 18:46:54 2016 Last change: Sat Dec 24 22:10:44 2016 by root via crm_resource on ha2-107.cisco.com
Current DC: ha1-107.cisco.com (version 1.1.13-10.el7-44eb2dd) - partition with quorum
1 node and 10 resources configured
Online: [ ha2-107.cisco.com ]
Resource Group: ctm_service
FSCheck (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/FsCheckAgent.py): (target-role:Stopped) Stopped
NTW_IF (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/NtwIFAgent.py): (target-role:Stopped) Stopped
CTM_RSYNC (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/RsyncAgent.py): (target-role:Stopped) Stopped
REPL_IF (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ODG_IFAgent.py): (target-role:Stopped) Stopped
ORACLE_REPLICATOR (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ODG_ReplicatorAgent.py): (target-role:Stopped) Stopped
CTM_SID (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/OracleAgent.py): (target-role:Stopped) Stopped
CTM_SRV (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/CtmAgent.py): (target-role:Stopped) Stopped
CTM_APACHE (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/ApacheAgent.py): (target-role:Stopped) Stopped
Resource Group: ctm_heartbeat
CTM_HEARTBEAT (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/HeartBeat.py): Started ha2-107.cisco.com
Resource Group: ctm_monitoring
FLASHBACK (lsb:../../..//cisco/PrimeOpticalServer/HA/bin/FlashBackMonitor.py): Started ha2-107.cisco.com
ha2-107.cisco.com: Online
pacemaker: active/enabled
The following figures illustrate a geographical redundancy HA configuration on Prime Optical:
- Figure 1-3 is a network diagram of Prime Optical in a geographical redundancy HA configuration for Linux.
- Figure 1-4 illustrates the hardware configuration for Prime Optical in a single-node geographical redundancy HA configuration.
Figure 1-3 Prime Optical in a Geographical Redundancy HA Environment for Linux
Figure 1-4 Geographical Redundancy (Single-Node Cluster)
In a single-node cluster geographical redundancy configuration, the clusters do not share a disk array (that is, the servers are connected to different disk arrays and do not share data). The Prime Optical application, Oracle application, and Prime Optical database are installed on each cluster. The Prime Optical database contains identical data on both clusters due to the volume replication (Oracle Active Data Guard on Linux).
The heartbeat link connects the primary and secondary clusters to monitor the health of both clusters. The heartbeat link is realized between the primary and secondary clusters through the WAN. The protocol used for the heartbeat is TCP/IP. During normal operation, the Prime Optical and Oracle applications run on the primary cluster (normally the active cluster). If a fault occurs on the primary cluster, the Prime Optical and Oracle applications can run on the secondary cluster; at this stage, the secondary cluster becomes the active cluster.
The primary cluster can be switched back manually to be the active cluster after the fault on the primary cluster is fixed.
The Replication link manages the Oracle Data Guard (ODG) communication between the active database (read/write) and standby database (read only).
Prime Optical Integration with Cisco Prime Central
After Prime Optical has been installed in standalone mode, you still have the option to integrate it with Prime Central. When Prime Optical is installed as part of the Prime Central suite, you can launch Prime Optical from the Prime Central portal. For more information about the Prime Central portal, see http://www.cisco.com/c/en/us/support/cloud-systems-management/prime-central/tsd-products-support-series-home.html.
Note
If you are installing Prime Optical with Prime Central, DNS must be enabled on the machine where Prime Optical is being installed. If the machine cannot resolve hostnames into IP addresses through DNS, Prime Optical registration will fail.
To integrate Prime Optical with Prime Central after Prime Optical has been installed in standalone mode, as the root user, run the./DMIntegrator.sh script. Follow the procedure described in "Configuring Domain Managers as Suite Components” in the Cisco Prime Central 1.1 Quick Start Guide.
What’s New in Prime Optical High Availability Guide 10.7
Table 1-1 Whats’s New in Prime Optical High Availability Guide 10.7
|
|
For More information, See...
|
Troubleshooting and upgrade pcs package |
By default, RHEL-7.x pcs package upgrade leads to Resource-*.sh that does not work. |
Step 3 |