Virtualization for Unified CVP
(→Creating Virtual Machines by Deploying the OVA Templates)
m (1 revision)
Revision as of 17:30, 27 January 2011
Updates to this Page
The following is a list of significant updates to this page:
|September 24, 2010||Distribution of Cisco UCS Unified CVP DocWiki for Tech review.|
|September 25, 2010||Cleaned up IOPS tables, product naming conventions, section consistency with Unified CCE UCS wiki.|
|September 26, 2010||Validated links, performed spell check, noted areas still requiring additional technical review.|
|September 27, 2010||Updated feedback to the following areas: Information for Partners, VMWare and Application Software Requirements, UCS Network Configuration, Configuring ESXi datastore.|
|September 28, 2010||Updated feedback to the following sections: Modified link to Information for Partners; Unified CVP 8.0(1) Support for Virtualization on the ESXi/UCS Platform (Modified bullet that referred to 'Distributed'; Coresidency (Deleted references to realtime; Modified last bullet regarding support on Blade.)|
|September 30, 2010||Unified CVP Virtualization information posted to Unified Communications Virtualization Wiki.|
|October 6, 2010||Updated links to Unified CVP main page and CVP Reporting Server Configuration sub-page.|
|October 8, 2010||Updated Unified CVP Component Co-Residency section.|
|December 8, 2010||Updated page to note that CVP 8.5(1) features the same level of support for virtualization that is available in CVP 8.0(1)|
|January 12, 2011||Added information on using CVP with C-Series servers.|
Information for Partners about Cisco Unified CVP on UCS Deployment Certification and Ordering
It is important that partners who are planning to sell UCS products on Unified Cisco Customer Voice Portal, read the Unified CCE DocWiki page.
This page contains essential information for partners about the following:
- Partner Certification Requirements
- UCS Server Ordering Information
- Important Notes on Cisco UCS Service and Support
Unified CVP Support for Virtualization on the ESXi/UCS Platform
Starting with Release 8.0(1), virtualization of the following deployments and Unified CVP components on Cisco Unified Communications Systems (UCS) B-Series hardware is supported:
- Unified CVP Call Server
- Unified CVP VXML Server
- Unified CVP Reporting Server
- Unified CVP Ops Console
The following deployments and Unified CVP components have not been qualified and are not supported in virtualization:
- H323 Call flow Deployment
- Distributed VXML Server and Call Server deployment where each server runs in a separate VM
- All in one lab deployment with small reporting server
The following VMware features are not supported with Unified CVP:
- VMware Physical to Virtual migration
- VMware Snapshots
- VMware Consolidated Backup
- VMware High Availability (HA)
- VMware Site Recovery Manager
- VMware vCenter Update Manager
- VMware vCenter Converter
Note: Unified CVP components are supported on the UCS platform only in a virtualized environment with ESXi server.
UCS Hardware Requirements
Requirements for Cisco Unified Customer Voice Portal using UCS B200 M1 hardware are located here.
Note: Unified CVP does not support UCS C200 and UCS-C210-M1.
VMware and Application Software Requirements
The following software requirements apply specifically to Unified Customer Voice Portal:
- VMWare ESXi 4.0 only, minimum version 4.0U1. Versions of ESXi prior to 4.0U1 and all versions of ESX are unsupported.
- The Windows, Informix, and other third party software requirements for the Unified CVP applications in the ESXi/UCS platform are the same as in the physical server. For such information, refer to the Hardware and System Software Specification for Cisco Unified Customer Voice Portal, Release 8.5(1).
UCS Network Configuration
IMPORTANT: For instructions on performing the network configuration needed to deploy Cisco Unified Customer Voice Portal (Unified CVP) on a virtualized platform, please see below.
1) TCP Setting modification on Reporting Server It is highly recommended that customers modify the default "Number of Receive Buffers" setting on the Reporting server to maximize full reporting and call load on the virtualized platform. If this setting is not modified after installing the OS, messages will get backed up on the CallServer and Message Queues will fill up resulting in a sharp drop in cps rate.
- Set the "Number of Receive buffers" on the Reporting Server TCP settings to 4096 (max).
- On Reporting server, click on Control Panel->Network Connection.
- Right click on Network Connection.
- Click on advanced tab.
- Under property tab, select "Number of Receive Buffers".
- In the Value pulldown, enter 4096.
- Restart the Reporting Server.
2) Use TCP protocol over UDP for SIP Comprehensive call flow as a SIP Best Practice Use TCP for SIP Comprehensive call flows. Using UDP may cause call setup failures or may result in abnormally long calls under heavy load. Typically TCP should be set on all call legs, but it is a requirement on the ingress leg. This can be set on the POTS dialpeer by setting the session transport to tcp.
3) Cisco VLAN trunking to VMWare For information on best practices for Cisco VLAN trunking to VMware, refer to the VMware website .
Support for Unified Communications Manager Clustering Over the WAN with Unified CVP on UCS Hardware
You can deploy the Unified Communications Manager (Unified CM) Clustering Over the WAN deployment model with Unified CVP on UCS hardware.
When you implement this deployment model, be sure to follow the best practices outlined in the section "Multiple Cisco Unified CM Clusters" in the Cisco Unified CVP Solution Reference Network Design (SRND) .
In addition, note the following expectation for UCS hardware points of failure:
- Multiple points of failure on the Unified CVP UCS HA deployment can cause catastrophic failure, such as ICM software failovers and interruption of service. If multiple points of failure occur, replace the failed redundant components and links immediately.
When deploying Clustering Over the WAN with B-Series hardware, use of the Cisco UCS M81KR Virtual Interface Card is mandatory.
New B-Series deployments using Clustering Over the WAN must use a Nexus 7000 Series / Nexus 5000 Series vPC infrastructure, or a Cisco Catalyst 6500 Series Virtual Switching Supervisor Engine 720-10G.
Notes for Deploying Unified CVP Applications on UCS B-Series Hardware with SAN
In Storage Area Network (SAN) architecture, storage consists of a series of arrays of Redundant Array of Independent Disks (RAIDs). A Logical Unit Number (LUN) that represents a device identifier can be created on a RAID array. A LUN can occupy all or part of a single RAID array, or span multiple RAID arrays.
In a virtualized environment, datastores are created on LUNs. Virtual Machines (VMs) are installed on the SAN datastore.
Keep the following considerations in mind when deploying Unified CVP applications on UCS B-series hardware with SAN:
- A SAN RAID group must be dedicated to Unified CVP database components; these SAN RAID groups must not be shared by other applications. The SAN disk array must be configured as RAID 10, for added performance and fault tolerance. SAN solutions are typically deployed in a shared environment where multiple applications are contending for storage access. If the SAN storage deployment is identified as affecting the functions of the contact center solution, the customer will be required to deploy a direct attached storage solution instead. Moreover, if in the process of troubleshooting, the SAN itself is identified as the problem, the customer must contact the system integrator or the SAN vendor for resolution.
- The SAN must be designed in accordance with the UCS High Availability guidelines documented at http://www.cisco.com/en/US/solutions/collateral/ns340/ns517/ns224/ns944/white_paper_c11-553711.html.
- The SAN design and configuration must meet the Unified CVP IOPS and Performance numbers described in the Performance Requirements.
- The SAN design and configuration must meet the following VMware ESXi disk performance guidelines:
- Disk Command Latency – It should be 15 mSec or less. 15mSec latencies or greater indicates a possible over-utilized, mis-behaving, or mis-configured disk array.
- For UCS B-series, Virtual Machines must be stored on and booted from the SAN.
- Kernel Disk Command Latency – It should be very small in comparison to the Physical Device Command Latency, and it should be close to zero. A high Kernel Command Latency indicates there is a lot of queuing in the ESXi kernel.
- The SAN design and configuration must meet the following Windows performance counters on Unified CVP VMs:
- AverageDiskQueueLength must remain less than (1.5 * (the total number of disks in the array)).
- %Disktime must remain less than 60%.
- Any given SAN array must be designed to have an IOPS capacity exceeding the sum of the IOPS required for all resident UC applications. Unified CVP applications should be designed for the 95th percentile IOPS values published in this wiki. Refer to the Performance Requirements section below for additional information. For other UC applications, please follow their respective IOPS requirements and guidelines.
- Unified CVP requires application storage to be on VMFS. Raw Device Mapping (RDM) is not supported.
- The total size of all Virtual Machines on a disk (total size = VM disk + RAM copy) must not exceed 90% of the capacity of a datastore.
Input/Output Operations per Second (IOPS) Considerations
- IOPS utilization should be monitored for each application to ensure that the aggregate IOPS is not exceeding the capacity of the array. Prolonged buffering of IOPS against an array may result in degraded system performance and delayed reporting data availability. Refer to the Performance Requirements section below for additional information.
- To help keep your system running most efficiently, schedule automatic database purging and database backups to run when your system is least busy.
- The ESXi boot from SAN feature is not supported; it must be installed to internal storage.
Notes for Deploying Unified CVP Applications on UCS C-Series Hardware
This section is applicable to storing Virtual Machines on C-210 local storage. The C-210 Server comes with a default local storage configured with two sets of RAID groups. Disk 1-2 is RAID 1, while the remaining disks are either RAID 5 or RAID10.
- Install ESXi 4.0 on the first disk array group (RAID 1 with disk 1 and disk 2)
- The creation of the virtual machine for the Reporting Server requires a large virtual disk size. Max disk size for a single VM to be 438GB for the large reporting server.
You must follow the steps described below to configure the ESXi data store block size to 2MB.
Steps to configure the ESXi data store block size to 2MB:
- Boot ESXi 4.0 and use VMware vSphere Client to connect to the ESXi host.
- On the Configuration tab for the host, select Storage in the box labeled Hardware. Select the second disk array group with RAID-5 (or RAID-10 ) configuration, and you will see in the formatting of “Datastore Details” that the block size is by default 1MB.
- Right-click this data store and delete it. We will add the data store back in the following steps.
- Click on the “Add Storage…” and select the Disk/LUN.
- The data store that was just deleted will now be available to add, select it.
- In the configuration for this data store you will now be able to select the block size, select 2MB, and finish the adding of the storage to the ESXi host. This storage is now available for deployment of the virtual machines that requires large disk size, such as the Administration and Data Servers.
Steps for Installing Unified CVP Components on Virtual Machines
Follow the steps and references below to install or migrate the Unified CVP components on Virtual Machines.
- Install, setup, and configure the UCS Hardware.
- Configure the UCS Network. See reference at UCS Network Configuration.
- Install and Boot VMWare ESXi. Refer to the appropriate installation guide: Cisco UCS B-Series Blade Servers VMware Installation Guide.
- Create the Unified CVP Virtual Machines from the OVA template. OVA templates are available here
- Install Windows OS and Websphere Application Server (if using VXML Server with WAS) on the created Virtual Machines. (Note: Microsoft Windows Server 2003 Standard Edition should be used for virtual machine guests. See related information in the links below.)
- Install Unified CVP Software components on the configured Virtual Machines. See install reference for installing Unified CVP Components in the Unified CVP Install and Upgrade guide.
Unified CVP Component Capacities and VM Configuration Requirements
This section provides details associated with VM requirements.
This table shows the supported Unified CVP components, their capacities, and the VM computing resource requirements. You must use the OVA virtual machine templates to create the Unified CVP component VMs.
Unified CVP Component Co-Residency and Sample Deployments
You can have one or more Unified CVP VMs co-resident on the same ESXi server. However, you must follow the rules described below:
- You can have any number of Unified CVP virtual machines and combination of co-residency of Unified CVP virtual machines on an ESXi server as long as the sum of all the virtual machine CPU and memory resource allocation is not over committed on the available ESXi server computing resources.
- You must not have CPU overcommitted on the ESXi server that is running Unified CVP application components . The total number of vCPUs among all the virtual machines on an ESXi host must not greater than the total number of CPUs available on the ESXi server. In the case of the Cisco UCS B-200 M1, the total number of CPUs available is 8.
- You must not have memory overcommitted on the ESXi host that is running UC realtime applications. You must allocate minimum 2GB of memory for the ESXi kernel. For example, if an ESXi server on B-200 M1 hardware has 36GB of memory, after you allocate 2GB for the ESXi kernel, you have 34GB available for the virtual machines. The total memory allocated for all the virtual machines on an ESXi server must not be greater than 34GB in this case.
- VM co-residency with third party applications is not supported unless it is specifically described in this page. Currently, the Media Server for CVP is not recommended to be co-loaded on the CVP call server+vxml server VM. The Media Server VM can use a 2 vCPU VM profile like the CVP Operations Console VM profile. However, this is a third party VM and can be tailored to have a different VM profile to satisfy a particular CVP deployment need.
Creating Virtual Machines from OVA VM Templates
Open Virtualization Format (OVF) is an open standard for packaging and distributing virtual appliances. Files in this format have an extension of .ova. The naming convention for the template is PRODUCT_COMPONENT_USER COUNT_VERSION_VMVER.ova
Follow the instructions in the Downloading OVA Templates section below to download the OVA templates from cisco.com to a local datastore that vSphere Client can access.
Downloading OVA Templates
- Proceed to the Cisco Download Page for CVP.
- To download a single OVA file, click the Download File button next to that file. To download multiple OVA files, click the Add to Cart button next to each file that you want to download, then click on the Download Cart link. A Download Cart page appears.
- Click the Proceed with Download button on this page. A Software License Agreement page appears.
- Read the Software License Agreement, then click the Agree button
- On the next page, click on either the Download Manager link (requires Java) or the Non Java Download Option link. A new browser window appears.
- If you selected Download Manager, a Select Location dialog box appears. Specify the location where you want to save the file, and click Open to save the file to your local machine.
- If you selected Non Java Download Option, click the Download link on the new browser window. Specify the location and save the file to your local machine.
Creating Virtual Machines by Deploying the OVA Templates
In the vSphere client, perform the following steps to deploy the Virtual machines.
- Highlight the host or cluster to which you wish the VM to be deployed.
- Select File > Deploy OVF Template.
- Click the Deploy from File radio button and specify the name and location of the file you downloaded in the previous section or click the Deploy from URL radio button and specify the complete URL in the field, then click Next.
- Verify the details of the template, and click Next.
- Give the VM you are about to create a name, and choose an inventory location on your host, then click Next.
- Choose the datastore on which you would like the VM to reside - be sure there is sufficient free space to accommodate the new VM, then click Next.
- Choose a virtual network for the VM, then click Next.
- Verify the deployment settings, then click Finish.
- VM CPU affinity is not supported. You do not need to set CPU affinity for the VMs that are running Unified CVP applications on the VMware ESXi on UCS platform.
- VM resource Reservation - VM resource reservation is not supported for the VMs that are running Unified CVP applications on the VMware ESXi on UCS platform. The VM computing resources should have a default reservation setting, which is no resource reservations.
- You cannot change the computing resource configuration of your VM at any time.
- You can never go below the minimum VM computing resource requirements as defined in the OVA templates.
- It is required that hyperthreading be enabled by default when running CVP on ESXi. ESXi Server hyperthread is enabled by default and this setting should not be modified. Please ensure all VM's with CVP servers has hyperthreading enabled.
Remote Control of the Virtual Machines
For administrative tasks, you can use either Windows Remote Desktop or the VMware Infrastructure Client for remote control.
Installing VMware Tools
The VMware Tools must be installed on each of the VMs and all of the VMware Tools default settings should be used. Please refer to the VMware documentation for instructions on installing or upgrading VMware Tools on the VM with Windows operating system.
Installing Unified CVP Components on Virtual Machines
You can install the Unified CVP components after the configuration of the VMs. Installation of these Unified CVP components on a VM is the same as the installation of these components on physical hardware.
Refer to the Unified CVP documentation for the steps to install Unified CVP components. You can install the supported Virus Scan software, or any other software in the same way as on physical hardware. Refer to the CVP Hardware and System Software Specification for more information.
Configuring the ESXi Data Store Block Size for Administration and Data Server
The creation of the virtual machine for the Unified CVP Reporting Server requires a large virtual disk size. You must follow the steps described below to configure the ESXi data store size to 438 GB for it to handle the Unified CVP Reporting Server data before you deploy the OVAs
Steps to configure the ESXi for CVP Reporting server to RAID10 found at CVP Reporting Server Configuration
- CPU usage (average) should not exceed 60% for the ESXi Server and for each of the individual processors, and for each VM.
- Memory usage (average) should not exceed 80% for the ESXi Server and for each of the VMs.
- VM snapshots are not supported in production since they have significant impact on system performance.
- The SAN must be able to handle the following Unified CVP application disk I/O characteristics.
- Enable hyperthreading on all ESXi servers.
Performance Numbers on UCS Platform
HTTP Performance Metrics
- The following metrics are measured with full reporting
|Call flow||Simultaneous Calls Supported||Calls Per Second|
|VXML Standalone with Req ICM Label (VXML Server on Tomcat)||900||10|
|VXML Standalone on WAS||900||10|
|VXML Standalone with Req ICM Label on WAS||900||10|
HTTPS Performance Metrics
- The following metrics are measured with full reporting
|Call flow||Simultaneous Calls Supported||Calls Per Second|
|VXML Standalone on WAS||275||3|
VM IOPS Measurement
IOPS Summary: (Unit in Number, Data are Blade-based)
|CVP Server||Average||MAX||95th Percentile|
Timekeeping Best Practices for Windows
You should follow the best practices outlined in the VMware Knowledge Base article VMware KB: Timekeeping best practices for Windows.
- ESXi hosts and domain controllers should synchronize the time from the same NTP source.
- When Unified Unified CVP virtual machines join the domain, they synchronize the time with the domain controller automatically using w32time.
- Be sure that Time synchronization between the virtual machine and the host operating system in the VMware Tools tool box GUI of the Windows Server 2003 guest operating system remains deselected; this checkbox is deselected by default.
System Performance Monitoring Using Windows Perfmon Counters
You must comply with the best practices described in the SRND section System Performance Monitoring.