Install Cisco Crosswork Planning

This chapter explains the processes to install Cisco Crosswork Planning on a single VM or node. In a single VM-based solution, the Cisco Crosswork Platform Infrastructure, Cisco Crosswork Planning infrastructure, and Cisco Crosswork Planning Design and Cisco Crosswork Planning Collector applications are bundled together in a single OVA image. Once you complete the installation procedure, all these components are installed on your machine by default.

  • The Cisco Crosswork Platform Infrastructure is a microservices-based platform and is the foundation required for running Crosswork applications.

  • Cisco Crosswork Planning Collector provides a set of microservices to create, maintain, and archive a model of the current network.

  • Cisco Crosswork Planning Design provides a set of microservices for network planning, designing, and visualization.

  • Cisco Crosswork Planning Infrastructure provides a set of microservices to manage Smart Licensing and Archive services.

This section contains the following topics:

Installation Parameters

This section explains the important parameters that you must specify while installing Cisco Crosswork Planning. Ensure that you have relevant information to provide for each of the parameters mentioned in the table.


Attention


Use the latest template file that comes with the Cisco Crosswork Planning build file.


Table 1. General parameters

Parameter

Description

Cw_VM_Image

The name of Crosswork VM image in vCenter.

This value is set as an option when running the installer tool and does not need to be set in the template file.

ClusterIPStack

The IP stack protocol: IPv4 or IPv6

vm_sizes

Create a custom profile as per your requirement. The following two profiles are supported in Cisco Crosswork Planning:

vm_sizes = {
  "Small" = {
    vcpus = 16,
    cpu_reservation = 24000,
    memory = 128000
  },
  "Large" = {
    vcpus = 24,
    cpu_reservation = 32000,
    memory = 256000
  }
}
vcpus

The number of virtual CPU instances allocated for virtual machine.

cpu_reservation

The guaranteed minimum CPU resource allocation for virtual machine.

memory

The amount of memory allocation for virtual machine.

ManagementVIP

The Management Virtual IP for the Crosswork VM.

ManagementVIPName

Name of the Management Virtual IP for the Crosswork VM. This is an optional parameter used to reach Crosswork Management VIP via DNS name. If this parameter is used, the corresponding DNS record must exist in the DNS server.

ManagementIPNetmask

The Management IP subnet in dotted decimal format (IPv4 or IPv6).

ManagementIPGateway

The Gateway IP on the Management Network (IPv4 or IPv6). The address must be reachable, otherwise the installation will fail.

DataVIP

The Data Virtual IP for the Crosswork VM.

DataVIPName

Name of the Data Virtual IP for the Crosswork VM. This is an optional parameter used to reach Crosswork Data VIP via DNS name. If this parameter is used, the corresponding DNS record must exist in the DNS server.

DataIPNetmask

The Data IP subnet in dotted decimal format (IPv4 or IPv6).

DataIPGateway

The Gateway IP on the Data Network (IPv4 or IPv6). The address must be reachable, otherwise the installation will fail.

DNS

The IP address of the DNS server (IPv4 or IPv6). The address must be reachable, otherwise the installation will fail.

DomainName

The domain name used for the VM.

K8sServiceNetwork

The network address for the kubernetes service network. By default, the CIDR range is fixed to '/16'.

K8sPodNetwork

The network address for the kubernetes pod network. By default, the CIDR range is fixed to '/16'.

CWPassword

Password to log into Cisco Crosswork.

Use a strong VM Password (8 characters long, including upper & lower case letters, numbers, and at least one special character). Avoid using passwords similar to dictionary words (for example, "Pa55w0rd!") or relatable words. While they satisfy the criteria, such passwords are weak and will be rejected resulting in failure to setup the VM.

VMSize

VM size. You can choose small or large in accordance with the desired profile. This parameter is required only for deploying via the docker installer tool.

IsSeed

Set to "True".

InitNodeCount

Set value to 1.

InitLeaderCount

Set value to 1.

NTP

NTP server address or name. The address must be reachable, otherwise the installation will fail.

SchemaVersion

The configuration Manifest schema version. This indicates the version of the installer to use with this template.

Schema version should map to the version packaged with the sample template in the installer tool on cisco.com. You should always build a new template from the default template provided with the release you are deploying, as template requirements may change from one release to the next.

ManagerDataFsSize

Refers to the data disk size for the Crosswork node (in GB). This is an optional parameter and the default value is 450 (valid range is from 450 to 8000), if not explicitly specified.

Please use the default value unless recommended otherwise.

ThinProvisioned

Set to "false" for production deployments.

LogFsSize

Log partition size (in Giga Bytes). Minimum value is 20 GB and Maximum value is 1000 GB.

If left blank, the default value (20 GB) is selected.

BackupMinPercent

Minimum percentage of the data disk space to be used for the size of the backup partition. The default value is 35 (valid range is from 1 to 80).

Please use the default value unless recommended otherwise.

Note

 

The final backup partition size will be calculated dynamically. This parameter defines the minimum.

ManagerDataFsSize

Refers to the data disk size for the Crosswork node (in Giga Bytes). This is an optional parameter and the default value is 485 (valid range is from 485 to 8000), if not explicitly specified.

Please use the default value unless recommended otherwise.

EnableHardReservations

Determines the enforcement of VM CPU and Memory profile reservations. This is an optional parameter and the default value is true, if not explicitly specified.

If set as true, the VM's resources are provided exclusively. In this state, the installation will fail if there are insufficient CPU cores, memory or CPU cycles.

If set as false (only set for lab installations), the VM's resources are provided on best efforts. In this state, insufficient CPU cores can impact performance or cause installation failure.

RamDiskSize

Size of the RAM disk.

This parameter is only used for lab installations (value must be at least 2). When a non-zero value is provided for RamDiskSize, the HSDatastore value is not used.

SchemaVersion

The configuration Manifest schema version. This indicates the version of the installer to use with this template.

Schema version should map to the version packaged with the sample template in the installer tool on cisco.com. You should always build a new template from the default template provided with the release you are deploying, as template requirements may change from one release to the next.

LogFsSize

Log partition size (in Giga Bytes). Minimum value is 20 GB and Maximum value is 1000 GB.

If left blank, the default value (20 GB) is selected.

EnableSkipAutoInstallFeature

Any pods marked as skip auto install will not be brought up until a dependent application/pod explicitly asks for it.

Set to "True".

EnforcePodReservations

Enforces minimum resource reservations for the pod. If left blank, the default value ("True") is selected.

Timezone

Enter the timezone. Input is a standard IANA time zone (for example, "America/Chicago").

If left blank, the default value (UTC) is selected.

This is an optional parameter.

DefaultApplicationResourceProfile

Resource profile for application pods. If left blank, resource profile defaults to the deployment's VM profile (recommended option).

DefaultInfraResourceProfile

Resource profile for infra pods. If left blank, resource profile defaults to the deployment's VM profile (recommended option).

IsRunDiagnoticsScriptForCheck

Used to enable/disable execution of the diagnostic script. The values are "true" (default value) and "false".

You are recommended to select the default value.

IgnoreDiagnoticsCheckFailure

Used to set the system response in case of a diagnostic check failure.

If set to "true" (default value), the diagnostic check is ignored and installation will continue. If set to "false", the installation is terminated.

You are recommended to select the default value.

Table 2. VM parameters

Parameter

Description

VMName

Name of the VM

NodeType

Type of VM. Choose "Hybrid".

ManagementIPAddress

The Management IP address of the VM (IPv4 or IPv6).

DataIPAddress

The Data IP address of the VM (IPv4 or IPv6).

Table 3. VMware template parameters

Parameter

Description

VCenterAddress

The vCenter IP or host name.

VCenterUser

The username needed to log into vCenter.

VCenterPassword

The password needed to log into vCenter.

DCname

The name of the Data Center resource to use.

Example: DCname = "cp"

MgmtNetworkName

The name of the vCenter network to attach to the VM's Management interface.

This network must already exist in VMware or the installation will fail.

DataNetworkName

The name of the vCenter network to attach to the VM's Data interface.

This network must already exist in VMware or the installation will fail. However, if this is a single NIC installation using only the Management interface, then you need to specify DataNetworkName = "".

HostedCwVMs

The ID of the VM to be hosted by the ESXi host or resource.

Host

The ESXi host, or ONLY the vCenter VM/resource group name where the VM is to be deployed.

The primary option is to use the host IP or name (all the hosts should be under the data center). If the hosts are under a VM in the data center, only provide the VM name (all hosts within the VM will be picked up).

The subsequent option is to use a resource group. In this case, a full path should be provided.

Example: Host = "Main infrastructure/Resources/00_trial"

Datastore

The datastore name available to be used by this host or resource group.

The primary option is to use host IP or name. The subsequent option is to use a resource group.

Example: Datastore = "SDRS-DCNSOL-prodexsi/bru-netapp-01_FC_Prodesx_ds_15"

HSDatastore

The high speed datastore available for this host or resource group.

If the vCenter server has a separate high-speed datastore, then this should be set to the name of that high-speed data store. If the vCenter server uses only a high-speed datastore, then both Datatore and HSDatastore can be set to the same name.

Note: The use of high-speed datastore is mandatory for Cisco Crosswork Planning.

Install Cisco Crosswork Planning Using the vCenter vSphere UI

This topic explains how to deploy Cisco Crosswork Planning on a single VM using the vCenter user interface.

This is the recommended method for installing Cisco Crosswork Planning on a single VM.

Procedure


Step 1

Download the latest available Cisco Crosswork Planning image file (*.ova) to your system.

Step 2

With VMware ESXi running, log in to the VMware vSphere Web Client. On the left navigation pane, choose the ESXi host where you want to deploy the VM.

Step 3

In the vSphere UI, go to Host > Configure > Networking > Virtual Switches and select the virtual switch for the Management Network that will be used to access the UI of the VM. In the virtual switch, select Edit > Security, and configure the following DVS port group properties:

  • Set Promiscuous mode as Reject

  • Set MAC address changes as Reject

Confirm the settings and repeat the process for the virtual switch that will be used for the Data Network.

Step 4

Review and confirm that your network settings meet the requirements.

Ensure that the networks that you plan to use for Management network and Data network are connected to the host. Contact your Cisco Experience team for assistance.

Step 5

Choose Actions > Deploy OVF Template.

Caution

 

The default VMware vCenter deployment timeout is 15 minutes. If vCenter times out during deployment, the resulting VM will not be bootable. To prevent this, we recommend that you document the choices (such as IP address, gateway, DNS server, etc.) so that you can enter the information quickly and avoid any issues with the VMware configuration.

Step 6

The VMware Deploy OVF Template window appears, with the first step, 1 - Select an OVF template, highlighted. Click Choose Files to navigate to the location where you downloaded the OVA image file and select it. Once selected, the file name is displayed in the window.

Step 7

Click Next. The Deploy OVF Template window is refreshed, with 2 - Select a name and folder now highlighted. Enter a name and select the respective data center for the Cisco Crosswork Planning VM you are creating.

We recommend that you include the Cisco Crosswork Planning version and build number in the name, for example: Cisco Crosswork Planning 7.0 Build 152.

Step 8

Click Next. The Deploy OVF Template window is refreshed, with 3 - Select a compute resource highlighted. Select the host for your Cisco Crosswork Planning VM.

Step 9

Click Next. The VMware vCenter Server validates the OVA. Network speed will determine how long validation takes. After the validation is complete, the Deploy OVF Template window is refreshed, with 4 - Review details highlighted.

Step 10

Review the OVF template that you are deploying. Note that this information is gathered from the OVF, and cannot be modified.

Note

 

Even though the Product field displays as "Cisco Crosswork Network Automation", once the installation procedure completes, the product name is displayed as "Cisco Crosswork Planning" in the application UI. The term "Cisco Crosswork Network Automation" is used universally for all the products under Crosswork portfolio.

Note

 

You may see alerts regarding the OVF package containing advanced configuration options and/or about trusted certificates. These are common and you can safely select the "Ignore" option.

Step 11

Click Next. The Deploy OVF Template window is refreshed, with 5 - License agreements highlighted. Review the End User License Agreement and if you agree, click the I accept all license agreements checkbox. Otherwise, contact your Cisco Experience team for assistance.

Step 12

Click Next The Deploy OVF Template window is refreshed, with 6 - Configuration highlighted. Choose the desired deployment configuration.

Important

 

For single VM deployment, the supported configurations are IPv4 Network and IPv6 Network using two NICs.

Figure 1. Select a deployment configuration

Step 13

Click Next. The Deploy OVF Template window is refreshed, with 7 - Select Storage highlighted. Choose the relevant option from the Select virtual disk format drop-down list. From the table, choose the datastore you want to use, and review its properties to ensure there is enough available storage.

Figure 2. Select Storage

Note

 

For production deployment, choose the Thick Provision Eager Zeroed option because this will preallocate disk space and provide the best performance. For lab purposes, we recommend the Thin Provision option because it saves disk space.

Step 14

Click Next. The Deploy OVF Template window is refreshed, with 8 - Select networks highlighted. From the Destination Network drop-down list, select the proper networks for the Management Network and the Data Network.

Figure 3. Select a deployment configuration

Important

 

Keep Admin Network and NBI Network at default values.

Step 15

Click Next. The Deploy OVF Template window is refreshed, with 9 - Customize template highlighted.

  1. Expand the Management Network settings. Provide information for the IPv4 or IPv6 deployment (as per your selection).

  2. Expand the Data Network settings. Provide information for the IPv4 or IPv6 deployment (as per your selection).

    Figure 4. Customize template settings
  3. Expand the Deployment Credentials settings. Enter relevant values for the VM Username and Password.

    Note

     

    Avoid using passwords that resemble dictionary words (for example, 'Pa55w0rd!') or easily guessable patterns. While such passwords might meet the initial criteria, they are considered weak and could cause the VM setup to fail without a clear explanation. To ensure a successful installation, use a complex password with a minimum of 8 characters that combines uppercase and lowercase letters, numbers, and special characters in a non-predictable sequence.

  4. Expand the DNS and NTP Servers settings. According to your deployment configuration (IPv4 or IPv6), the fields that are displayed are different. Provide information in the following three fields:

    • DNS IP Address: The IP addresses of the DNS servers you want the Cisco Crosswork server to use. Separate multiple IP addresses with spaces.

    • DNS Search Domain: The name of the DNS search domain.

    • NTP Servers: The IP addresses or host names of the NTP servers you want to use. Separate multiple IPs or host names with spaces.

    Note

     

    The DNS and NTP servers must be reachable using the network interfaces you have mapped on the host. Otherwise, the configuration of the VM will fail.

  5. The default Disk Configuration settings should work for most environments. Change the settings only if you are instructed to by the Cisco Customer Experience team.

  6. Expand Crosswork Configuration and enter your legal disclaimer text (users will see this text if they log into the CLI).

  7. Expand Crosswork Cluster Configuration. Provide relevant values for the following fields:

    • VM Type: Choose Hybrid.

    • Cluster Seed node: Choose True.

    • Crosswork Management Cluster Virtual IP: Enter the Management Virtual IP address and Management Virtual IP DNS name.

    • Crosswork Data Cluster Virtual IP: Enter the Data Virtual IP address. and the Data Virtual IP DNS name.

    • Initial node count: Set to 1.

    • Initial leader node count: Set to 1.

    • Location of VM: Enter the location of VM.

    • Installation type: Not applicable to single VM deployment. Do not select any checkbox.

    • Enable Skip Auto Install Feature: Set to True.

    • Ignore Diagnose Failure?: Use the default value (True).

    • Enable Diagnostics Script Check Run?: Use the default value (True).

    • Default Application Resource Profile: Use the default value (Empty).

    • Default Infra Resource Profile: Use the default value (Empty).

    • Auto Action Manifest Definition: Use the default value (Empty).

Step 16

Click Next. The Deploy OVF Template window is refreshed, with 10 - Ready to Complete highlighted.

Step 17

Review your settings and then click Finish if you are ready to begin deployment. Wait for the deployment to finish before continuing. To check the deployment status:

  1. Open a VMware vCenter client.

  2. In the Recent Tasks tab of the host VM, view the status of the Deploy OVF template and Import OVF package jobs.

Step 18

Once the deployment is completed, right-click on the VM and select Edit Settings. The Edit Settings dialog box is displayed. Under the Virtual Hardware tab, update these attributes based on your profile requirement.

  • CPU: change to 16 (for small profile) or 24 (for large profile)

  • Memory: change to 128 GB (for small profile) or 256 GB (for large profile)

Click OK to save the changes.

Step 19

Power on the Cisco Crosswork Planning VM. To power on, expand the host’s entry, click the Cisco Crosswork Planning VM, and then choose Actions > Power > Power On.

The time taken to create the VM can vary based on the size of your deployment profile and the performance characteristics of your hardware.


Install Cisco Crosswork Planning via the OVF Tool

This topic explains how to deploy Cisco Crosswork Planning on a single VM using the OVF tool. You must modify the list of mandatory and optional parameters in the script as per your requirements and run the OVF tool.


Note


The file names mentioned in this topic are sample names and may differ from the actual file names on cisco.com.


Follow these steps to log in to the Cisco Crosswork Planning VM from SSH:

Before you begin

  • In your vCenter data center, go to Host > Configure > Networking > Virtual Switches and select the virtual switch. In the virtual switch, select Edit > Security, and ensure that the following DVS port group properties are as shown:

    • Set Promiscuous mode as Reject

    • Set MAC address changes as Reject

    Confirm the settings and repeat the process for each virtual switch used by Cisco Crosswork Planning.

  • Ensure you are using the OVF tool version 4.4 or higher.

Procedure


Step 1

On the machine where you have the OVF tool installed, use the following command to confirm that you have OVF tool version 4.4:

ovftool --version

Step 2

Create the script file (see example below) and provide relevant information as per your target environment (such as IP addresses, gateway, netmask, password, and VCENTER_PATH, etc.).

Note

 

The file names mentioned in this topic are sample names and may differ from the actual file names on cisco.com.

cat cp_install.sh
#!/usr/bin/env bash
Host="X.X.X.X"
DM="thick"
DS="DS36"
Deployment="cw_ipv4"
DNSv4="10.10.0.99"
NTP="ntp.cisco.com"
Timezone="US/Pacific"
EnforcePodReservations="True"
EnableSkipAutoInstallFeature="True"
Domain="cisco.com"
Disclaimer="ACCESS IS MONITORED"
VM_NAME="cp"
DataNetwork="DataNet"
ManagementNetwork="MgmtNet"
DataIPv4Address="x.x.x.x"
DataIPv4Gateway="x.x.x.x"
DataIPv4Netmask="x.x.x.x"
ManagementIPv4Address="x.x.x.x"
ManagementIPv4Gateway="x.x.x.x"
ManagementIPv4Netmask="x.x.x.x"
K8sServiceNetworkV4="10.75.0.0"
K8sPodNetworkV4="10.225.0.0"
Password="CLI Password"
Username="cw-admin"
ManagementVIP="x.x.x.x"
DataVIP="x.x.x.x"
VMType="Hybrid"
IsSeed="True"
InitNodeCount="1"
InitMasterCount="1"
 
CP_OVA_PATH=$1
 
VCENTER_LOGIN="Administrator%40vsphere%2Elocal:Password%40123%21@x.x.x.x"
VCENTER_PATH="DC1/host"
 
ovftool --version
ovftool --acceptAllEulas --skipManifestCheck --X:injectOvfEnv -ds=$DS \
--numberOfCpus:"*"=24 --viCpuResource=:32000: \
--memorySize:"*"=256000 --viMemoryResource=:256000: \
--diskMode=$DM --overwrite --powerOffTarget --powerOn --noSSLVerify \
--allowExtraConfig \
--deploymentOption=$Deployment \
--prop:"DNSv4=${DNSv4}" \
--prop:"NTP=${NTP}" \
--prop:"Timezone=${Timezone}" \
--prop:"EnforcePodReservations=${EnforcePodReservations}" \
--prop:"EnableSkipAutoInstallFeature=${EnableSkipAutoInstallFeature}" \
--prop:"Domain=${Domain}" \
--prop:"Disclaimer=${Disclaimer}" \
--name=$VM_NAME \
--net:"Data Network=${DataNetwork}" \
--net:"Management Network=${ManagementNetwork}" \
--prop:"DataIPv4Address=${DataIPv4Address}" \
--prop:"DataIPv4Gateway=${DataIPv4Gateway}" \
--prop:"DataIPv4Netmask=${DataIPv4Netmask}" \
--prop:"ManagementIPv4Address=${ManagementIPv4Address}" \
--prop:"ManagementIPv4Gateway=${ManagementIPv4Gateway}" \
--prop:"ManagementIPv4Netmask=${ManagementIPv4Netmask}" \
--prop:"K8sServiceNetworkV4=${K8sServiceNetworkV4}" \
--prop:"K8sPodNetworkV4=${K8sPodNetworkV4}" \
--prop:"CWPassword=${Password}" \
--prop:"CWUsername=${Username}" \
--prop:"ManagementVIP=${ManagementVIP}" \
--prop:"DataVIP=${DataVIP}" \
--prop:"VMType=${VMType}" \
--prop:"IsSeed=${IsSeed}" \
--prop:"InitNodeCount=${InitNodeCount}" \
--prop:"InitMasterCount=${InitMasterCount}" \
$CP_OVA_PATH \
vi://$VCENTER_LOGIN/$VCENTER_PATH/$Host

Note

 

The recommended CPU parameters for Small and Large profiles are as follows:

  • Small profile:
    --numberOfCpus:"*"=16 --viCpuResource=:24000: \
    --memorySize:"*"=128000 --viMemoryResource=:128000: \
  • Large profile:

    --numberOfCpus:"*"=24 --viCpuResource=:32000: \
    --memorySize:"*"=256000 --viMemoryResource=:256000: \

Step 3

Download the OVA and install scripts from cisco.com. For the purpose of these instructions, we use the file name as signed-cw-na-planning-7.0.0-904-release-240215.

Use the following command to extract the files from the tar bundle:

tar -xvzf signed-cw-na-planning-7.0.0-904-release-240215.ova

The OVA is extracted:

cp]# ls -al
-rw-r--r--   1 root root 15416145920 Mar 28 11:12 signed-cw-na-planning-7.0.0-904-release-
240215.ova
-rwxr-xr-x   1 root root        2324 Apr  2 14:06 cp_install.sh

Step 4

Use the following command to make the scripts executable:

chmod +x {filename}

For example:

chmod +x cp_install.sh

Step 5

Execute the script with the OVA file name as parameter:

cp]# ./cp_install.sh signed-cw-na-planning-7.0.0-904-release-240215.ova
VMware ovftool 4.4.0 (build-16360108)
Opening OVA source: signed-cw-na-planning-7.0.0-904-release-240215.ova
<Removed some output >
Completed successfully

The time taken to create the VM can vary based on the size of your deployment profile and the performance characteristics of your hardware.


Install Cisco Crosswork Planning Using the Docker Installer Tool

This section explains the procedure to install Cisco Crosswork Planning on a single VM using the docker installer tool. This method is less recommended compared to using the vCenter UI or the OVF tool for installation.

Before you begin

  • Make sure that your environment meets all the vCenter requirements specified in VMware Settings.

  • The edited template in the /data directory contains sensitive information (VM passwords and the vCenter password). The operator needs to manage access to this content. Store the templates used for your install in a secure environment or edit them to remove the passwords.

  • The install.log, install_tf.log, and .tfstate files will be created during the install and stored in the /data directory. If you encounter any trouble with the installation, provide these files to the Cisco Customer Experience team when opening a case.

  • The install script is safe to run multiple times. Upon error, input parameters can be corrected and re-run. You must remove the install.log, install_tf.log, and tfstate files before each re-run. Running the installer tool multiple times may result in the deletion and re-creation of VMs.

  • In case you are using the same installer tool for multiple Cisco Crosswork Planning installations, it is important to run the tool from different local directories, allowing for the deployment state files to be independent. The simplest way for doing so is to create a local directory for each deployment on the host machine and map each one to the container accordingly.

  • Docker version 19 or higher is required while using the installer tool. For more information on Docker, see https://docs.docker.com/get-docker/.

  • To change the install parameters or to correct the parameters following the installation errors, it is important to distinguish whether the installation has managed to deploy the VM or not. Deployed VM is evidenced by the output of the installer similar to:

    vsphere_virtual_machine.crosswork-IPv4-vm["1"]: Creation complete after 2m50s [id=4214a520-c53f-f29c-80b3-25916e6c297f]

Known limitations:

  • The vCenter host VMs defined must use the same network names (vSwitch) across all hosts in the data center.

  • The vCenter storage folders or datastores organized under a virtual folder structure, are not supported currently. Ensure that the datastores referenced are not grouped under a folder.

Procedure


Step 1

In your Docker capable machine, create a directory where you will store everything you will use during this installation.

Note

 

If you are using a Mac, ensure that the directory name is in lower case.

Step 2

Download the installer bundle (.tar.gz file) and the OVA file from cisco.com to the directory you created previously. For the purpose of these instructions, we will use the file name as signed-cw-na-planning-installer-7.0.0-85-release700-240823.tar.gz and signed-cw-na-planning-7.0.0-85-release700-240823.ova.

Attention

 

The file names mentioned in this topic are sample names and may differ from the actual file names on cisco.com.

Step 3

Use the following command to unzip the installer bundle:

tar -xvf signed-cw-na-planning-installer-7.0.0-85-release700-240823.tar.gz

The contents of the installer bundle is unzipped to a new directory (for example, signed-cw-na-planning-installer-7.0.0-85-release700). This new directory will contain the installer image (cw-na-planning-installer-7.0.0-85-release700-240823.tar.gz) and files necessary to validate the image.

Step 4

Review the contents of the README file to understand everything that is in the package and how it will be validated in the following steps.

Step 5

Use the following command to verify the signature of the installer image:

Note

 

Use python --version to find out the version of python on your machine.

If you are using Python 2.x, use the following command to validate the file:

python cisco_x509_verify_release.py -e <.cer file> -i <.tar.gz file> -s <.tar.gz.signature
 file> -v dgst -sha512

If you are using Python 3.x, use the following command to validate the file:

python3 cisco_x509_verify_release.py3 -e <.cer file> -i <.tar.gz file> -s <.tar.gz.signature 
file> -v dgst -sha512

Note

 

If you do not have Python installed, go to python.org and download the version of Python that is appropriate for your work station.

Step 6

Use the following command to load the installer image file into your Docker environment.

docker load -i <.tar.gz file>

For example:

docker load -i cw-na-planning-installer-7.0.0-85-release700-240823.tar.gz

Step 7

Run the docker image list or docker images command to get the "image ID" (which is needed in the next step).

For example:

docker images

The result will be similar to the following: (section we will need is underlined for clarity)

My Machine% docker images
REPOSITORY                        TAG                                           IMAGE ID             CREATED        SIZE
dockerhub.cisco.com/cw-installer  cw-na-planning-7.0.0-85-release700-240823  316c53341bde      7 days ago     280MB

Note

 

Pay attention to the "CREATED" time stamp in the table presented when you run docker images, as you might have other images present from the installation of prior releases. If you wish to remove these, the docker image rm {image id} command can be used.

Step 8

Launch the Docker container using the following command:

docker run --rm -it -v `pwd`:/data {image id of the installer container}

To run the image loaded in our example, use the following command:

docker run --rm -it -v `pwd`:/data 316c53341bde

Note

 
  • You do not have to enter that full value. In this case, docker run --rm -it -v `pwd`:/data 316 was adequate. Docker requires enough of the image ID to uniquely identify the image you want to use for the installation.

  • In the above command, we are using the backtick (`). Do not use the single quote or apostrophe (') as the meaning to the shell is very different. By using the backtick (recommended), the template file and OVA will be stored in the directory where you are on your local disk when you run the commands, instead of inside the container.

  • When deploying a IPv6 setup, the installer needs to run on an IPv6 enabled container/VM. This requires additionally configuring the Docker daemon before running the installer, using the following method:

    • Linux hosts (ONLY): Run the Docker container in host networking mode by adding the "–network host" flag to the Docker run command line.

      docker run --network host <remainder of docker run options>
  • Centos/RHEL hosts, by default, enforce a strict SELinux policy which does not allow the installer container to read from or write to the mounted data volume. On such hosts, run the Docker volume command with the Z option as shown below:

    docker run --rm -it -v `pwd`:/data:Z <remainder of docker options>

Note

 

The Docker command provided will use the current directory to read the template and the ova files, and to write the log files used during the install. If you encounter either of the following errors you should move the files to a directory where the path is in lowercase (all lowercase, no spaces or other special characters). Then navigate to that directory and rerun the installer.

Error 1:

% docker run --rm -it -v `pwd`:/data 316
docker: invalid reference format: repository name must be lowercase.
See 'docker run --help'

Error 2:

docker: Error response from daemon: Mounts denied: approving /Users/Desktop: file does not exist
ERRO[0000] error waiting for container: context canceled

Step 9

Navigate to the directory with the VMware template.

cd /opt/installer/deployments/7.0.0/vcentre

Step 10

Copy the template file found under /opt/installer/deployments/7.0.0/vcentre/deployment_template_tfvars to the /data folder using a different name.

For example, cp deployment_template_tfvars /data/deployment.tfvars

For the rest of this procedure, we will use deployment.tfvars in all the examples.

Step 11

Edit the template file located in the /data directory in a text editor, to match your planned deployment (for reference, see Sample Manifest Template). The <sample manifest template> includes an example that you can reference for proper formatting. The example is more compact due to the removal of descriptive comments.

Step 12

From the /opt/installer directory, run the installer.

./cw-installer.sh install -m /data/<template file name> -o /data/<.ova file>

For example:

./cw-installer.sh install -m /data/deployment.tfvars -o /data/cw-na-planning-7.0.0-85-release700-240823.ova

Step 13

Read, and then enter "yes" if you accept the End User License Agreement (EULA). Otherwise, exit the installer and contact your Cisco representative.

Step 14

Enter "yes" when prompted to confirm the operation.

Note

 

It is not uncommon to see some warnings like the following during the install:

Warning: Line 119: No space left for device '8' on parent controller '3'.
Warning: Line 114: Unable to parse 'enableMPTSupport' for attribute 'key' on element 'Config'.

If the install process proceeds to a successful conclusion (see sample output below), these warnings can be ignored.

Sample output:

cw_vms = <sensitive>
INFO: Copying day 0 state inventory to CW
INFO: Waiting for deployment status server to startup on 10.90.147.66. 
Elapsed time 0s, retrying in 30s
Crosswork deployment status available at http://{VIP}:30602/d/NK1bwVxGk/crosswork-
deployment-readiness?orgId=1&refresh=10s&theme=dark 
Once deployment is complete login to Crosswork via: https://{VIP}:30603/#/logincontroller 
INFO: Cw Installer operation complete.

Note

 

If the installation fails, open a case with Cisco and provide the .log files that were created in the /data directory (and the local directory where you launched the installer Docker container), to Cisco for review. The two most common reasons for the install to fail are: (a) password that is not adequately complex, and (b) errors in the template file. If the installer fails for any errors in the template (for example, mistyped IP address), correct the error and rerun the install script.


Sample Manifest Template

This section presents the manifest template example for deploying Cisco Crosswork Planning on a single VM.


Note


Ensure that the vm_sizes parameter is set in accordance with your deployment profile. The following two profiles are supported in Cisco Crosswork Planning:
vm_sizes = {
  "Small" = {
    vcpus = 16,
    cpu_reservation = 24000,
    memory = 128000
  },
  "Large" = {
    vcpus = 24,
    cpu_reservation = 32000,
    memory = 256000
  }
}


Cw_VM_Image = "cw-na-planning-7.0.0-39-release700-240806"  # Line added automatically by installer.
ClusterIPStack = "IPv4"
ManagementVIP = "172.20.118.86"
ManagementIPNetmask = "255.255.255.0"
ManagementIPGateway = "172.20.118.1"
DataVIP = "172.20.118.87"
DataIPNetmask = "255.255.255.0"
DataIPGateway = "172.20.118.1"
DNS = "172.24.97.250"
DomainName = "cisco.com"
CWPassword = "*****!"
VMSize = "Small"
NTP = "ntp.esl.cisco.com"
Timezone = "Asia/Calcutta"
DefaultApplicationResourceProfile = "small"
EnableSkipAutoInstallFeature = "True"
# VM size definitions
vm_sizes = {
    "Small" = {
        vcpus = 16,
        cpu_reservation = 32000,
        memory = 128000
    }
}
EnforcePodReservations = "False"
ThinProvisioned = "true"
ManagerDataFsSize = "2000"
BackupMinPercent = "5"
SchemaVersion = "7.0.0"
 
CwVMs = {
    "0" = {
        VMName = "vm1",
        ManagementIPAddress = "172.20.118.88",
        DataIPAddress = "172.20.118.89",
        NodeType = "Hybrid"
    }
}
 
VCentreDC = {
    VCentreAddress = "172.20.118.84",
    VCentreUser = "<your-username>",
    VCentrePassword = "******",
    DCname = "cp",
    MgmtNetworkName = "VM Network",
    DataNetworkName = "cwdata-group",
    VMs = [
        {
            HostedCwVMs = ["0"],
            Host = "172.20.118.81",
            Datastore = "datastore1 (2)",
            HSDatastore = "datastore1 (2)"    
        }
    ]
}
SchemaVersion = "7.0.0"

Migrate Collector Configurations

In Cisco Crosswork Planning, you can migrate the collector configurations from Cisco WAE 7.5.x/7.6.x, as well as from one Cisco Crosswork Planning instance to the other.


Note


When using collectors that have file upload options, ensure to upload the correct files after importing the collector configuration. This is necessary because, after importing the configuration, the server restores only the file name and not the actual file. If the correct file is not used, then the collection fails.


Migrate Collector Configuration from Cisco WAE

Follow these steps to migrate the collector configurations from Cisco WAE 7.5.x/7.6.x to Cisco Crosswork Planning.

Before you begin

Procedure


Step 1

If you have not backed up the configuration, use the following steps to back up and migrate it to a configuration compatible with Cisco Crosswork Planning:

  1. Log in to the machine where Cisco WAE 7.x is installed.

  2. Enter the following command:

    # ./wae_upgrade --export --install-dir <WAE_7.x_INSTALL_DIR> --cfg-dir 
    <dir_to_save_exported_config>
    
    Where:
          --install-dir   indicates the directory where 7.x WAE is installed
          --cfg-dir       indicates the folder where the backup of 7.x configuration 
                          must reside

Step 2

If you already have the backed-up configuration, use the following steps to convert the file into a format compatible with Cisco Crosswork Planning:

  1. Log in to the machine where the Cisco WAE 7.x configuration is backed up.

  2. Enter the following command:

    # ./wae_upgrade --migrate --cfg-dir <dir_containing_7.x_config>
    
    Where:
          --cfg-dir       indicates the folder where the 7.x configuation is backed up.  
                          This configuration will be migrated to Cisco Crosswork Planning 
                          compatible configuration.

Step 3

Import the Cisco Crosswork Planning compatible configuration file to Cisco Crosswork Planning using the following steps:

Note

 

Before migration, ensure that configurations are backed up using the upgrade scripts. Otherwise, the migration will fail.

  1. Log in to the Cisco Crosswork Planning UI.

  2. From the main menu, choose Collector > Migration.

  3. Click Actions and choose Configuration migration.

    The Import configuration file window appears.

    Figure 5. Import Configuration File Window
  4. From the Import type drop-down, choose WAN Automation Engine.

  5. Click Browse and select the Cisco WAE collector configuration file which is compatible with Cisco Crosswork Planning compatible.

  6. (Optional) If you want to overwrite the existing collector configuration, check the Overwrite the existing data check box.

  7. Click Import to import the collector configuration file.


You can monitor status of the import in the Migration page (Collector > Migration). Once the import is successful, the Import status column displays the status of the task as Success.


Note


After migrating from Cisco WAE to Cisco Crosswork Planning, the Telnet and SSH settings are not preserved. You need to manually verify and update these settings, if required.


Configurations Excluded During Migration

There are several configurations which are NOT migrated while moving from Cisco WAE to Cisco Crosswork Planning. For details, see the "Configurations Excluded During Migration" section in the Cisco Crosswork Planning 7.0 Collection Setup and Administration.

Migrate Collector Configuration between Cisco Crosswork Planning Instances


Note


If you are using the SR-PCE collector in your configurations, ensure to update the SR-PCE host and Backup SR-PCE host fields manually after migration. This is necessary because, these fields are not updated while migrating the collector configurations between Cisco Crosswork Planning instances.


Follow these steps to migrate the collector configuration from one Cisco Crosswork Planning instance (source) to the other (target).

Procedure


Step 1

Download the collector configuration file from the source machine you want to migrate the configuration from:

  1. Log into the Cisco Crosswork Planning instance from which you want to migrate the configuration.

  2. From the main menu, choose Collector > Migration.

  3. Click Actions and choose Configuration backup.

    The collector configuration file is downloaded to your local machine.

Step 2

Import the collector configuration file to the target machine where you want to migrate it to:

  1. Log into the Cisco Crosswork Planning instance to which you want to migrate the configuration.

  2. From the main menu, choose Collector > Migration.

  3. Click Actions and choose Configuration migration.

    The Import configuration file window appears.

    Figure 6. Import Configuration File Window
  4. From the Import type drop-down, choose Crosswork planning.

  5. Click Browse and select the collector configuration file that you downloaded in the Step 1 (c).

  6. (Optional) If you want to overwrite the existing collector configuration, check the Overwrite the existing data check box.

  7. Click Import to import the collector configuration file.


You can monitor status of the import in the Migration page (Collector > Migration). Once the import is successful, the Import status column displays the status of the task as Success.


Note


In case of traffic collection, if the traffic poller agent status is displayed as down on the Agent page after migration, even though traffic collection has run successfully, follow these steps on the Collections (Collector > Collections) page:

  1. Select Edit collection for the collection corresponding to the agent.

  2. In the Traffic collection configuration page, uncheck the Traffic collection check box and save the configuration.

  3. Re-enable the Traffic collection checkbox and save the configuration again.