Professional Documents
Culture Documents
Deploy Hitachi Ucp Select For Citrix Xendesktop With Vmware Vsphere
Deploy Hitachi Ucp Select For Citrix Xendesktop With Vmware Vsphere
Feedback
Hitachi Data Systems welcomes your feedback. Please share your thoughts by sending an email message to SolutionLab@hds.com. To assist the routing of this message, use the paper number in the subject and the title of this white paper in the text.
Table of Contents
Solution Overview.......... ..................................................................................... 2 Determining User Workload...................................................................... 2 Logical Design.......... ................................................................................ 4 Solution Components.......... ............................................................................... 7 Hitachi Compute Blade 500.......... ............................................................ 7 Hitachi Unified Storage.......... ...................................................................7 Hitachi Dynamic Provisioning.......... ......................................................... 7 Citrix XenDesktop 5.5.......... ..................................................................... 8 Citrix Provisioning Services 5.6.2.......... ................................................... 8 VMware vSphere 5.0.......... ...................................................................... 8 Brocade Switches.......... ........................................................................... 9 Solution Design.......... ....................................................................................... 10 Infrastructure Cell for Compute Resources.......... .................................. 12 Infrastructure Cell for Storage Resources.......... .................................... 18 Application Cell for Hitachi Unified Compute Platform Select Management........................................................................................... 20 Application Cell for Citrix Provisioning Services.......... ........................... 26 Application Cell for Citrix XenDesktop.......... .......................................... 29 Expansion Cell for Compute Resources................................................. 34 Engineering Validation...................................................................................... 36 Test Methodology.......... ......................................................................... 37 Test Results............................................................................................ 38 Conclusion.......... ............................................................................................... 51
1
1
Deploy Hitachi Unified Compute Platform Select for Citrix XenDesktop with VMware vSphere
Reference Architecture Guide
This reference architecture guide describes a solution using Hitachi Unified Compute Platform Select for Citrix XenDesktop with VMware vSphere 5.0. This solution scales in a repeatable, building block design to deploy desktops using Citrix XenDesktop and Citrix Provisioning Services. This solution provides information to plan and deploy a Citrix XenDesktop 5.5 environment using the following:
Hitachi Unified Storage 150 Hitachi Compute Blade 500 Brocade Ethernet and Fibre Channel networking components VMware vSphere 5.0 as the hypervisor
Hitachi Unified Compute Platform is a family of integrated and flexible reference solutions. Each Unified Compute Platform solution, configured for immediate deployment, runs top tier infrastructure applications without over-purchasing or provisioning unnecessary equipment. The entire solution is stack certified and compatible. You need a working familiarity with techniques and practices used for the products listed in this guide.
NoteTesting of this configuration was in a lab environment. Many things affect production environments beyond prediction or duplication in a lab environment. Follow the recommended practice of conducting proof-of-concept testing for acceptable results in a non-production, isolated test environment that otherwise matches your production environment before your production implementation of this solution.
2
2
Solution Overview
This reference architecture uses the following:
Hitachi Unified Storage 150 Hitachi Compute Blade 500 Hitachi Dynamic Provisioning Citrix XenDesktop 5.5 Citrix Provisioning Services 5.6.2 VMware vSphere 5.0 Brocade 6510 enterprise fabric switches Brocade VDX-6720 Ethernet switches
User Workload
Light
Medium
Heavy
Workload
Used in this solution Microsoft Internet Explorer Microsoft Word Microsoft Excel Microsoft Outlook
Not used in this solution Microsoft Internet Explorer Microsoft Word Microsoft Outlook Microsoft PowerPoint Adobe Acrobat Reader and PDF printer 7-Zip
Not used in this solution Microsoft Internet Explorer Microsoft Excel Microsoft PowerPoint Adobe Acrobat Reader and PDF printer 7-Zip Media Player
3
3 The desktop configuration used for this solution followed the recommendations in Table 2 for a light user workload. See Table 2 for a description of the different workloads.
Table 2. User Desktop Configurations
User Workload Operating System vCPU Allocation Memory Allocation per User Average Steday State IOPS* Estimated Users per Core
Light
Medium
Heavy
Microsoft Windows 7, 64- Microsoft Windows 7, 64- Microsoft Windows 7, 64bit bit bit 1 1 GB to 2 GB 4 to 7 8 to 10 1 1.5 GB to 2 GB 8 to 12 6 to 8 2 4 GB 12 to 25 4 to 6
*The IOPS estimates are an average of when the user is logged on to the virtual desktop and working. It does not take into account start-ups, users logging on, and users logging off.
4
4
Logical Design
Figure 1 on page 5 illustrates the high-level logical design of this solution with the following:
One infrastructure cell for Hitachi Unified Compute Platform Select management One application cell for Citrix provisioning services One application cell for Citrix XenDesktop
5
5
Figure 1
6
6 This solution supports up to 500-streamed desktops designed with the following:
Using the Hitachi Data Systems cell design, this solution can be scaled out to support thousands of desktops.
7
7
Solution Components
These are the major components used in this solution.
8
8 Deploying Dynamic Provisioning avoids the routine issue of hot spots that occur on logical devices (LDEVs). These occur within individual RAID groups when the host workload exceeds the IOPS or throughput capacity of that RAID group. Dynamic provisioning distributes the host workload across many RAID groups, which provides a smoothing effect that dramatically reduces hot spots. When used with Hitachi Unified Storage, Hitachi Dynamic Provisioning has the benefit of thin provisioning. Physical space assignment from the pool to the dynamic provisioning volume happens as needed using 1 GB chunks, up to the logical size specified for each dynamic provisioning volume. There can be a dynamic expansion or reduction of pool capacity without disruption or downtime. You can rebalance an expanded pool across the current and newly added RAID groups for an even striping of the data and the workload.
ESXi 5.0this is a hypervisor that loads directly on a physical server. It partitions one physical machine into many virtual machines that share hardware resources. vCenter Serverthis allows management of the vSphere environment through a single user interface. With vCenter, there are features available such as vMotion, Storage vMotion, Storage Distributed Resource Scheduler, High Availability, and Fault Tolerance.
9
9
Brocade Switches
Brocade and Hitachi Data Systems partner to deliver storage networking and data center solutions. These solutions reduce complexity and cost, as well as enable virtualization and cloud computing to increase business agility. This solution uses the following Brocade products:
Brocade 5460 8 Gb SAN switch for Hitachi Compute Blade 500 Brocade 6510 Brocade VDX 6720 Brocade VDX 6746
10
10
Solution Design
This solution uses a cell architecture. A cell provides the type of a specific type of component necessary to build a solution. The cell architecture defines the compute and storage resources needed to support a defined workload. The cell types in this solution are the following:
Infrastructure Cells
Compute resourcesFoundation for compute components Storage resourcesFoundation for storage components
Application Cells Hitachi Unified Compute Platform Select managementRequired for resources for hosting management services for VMware vSphere, Citrix XenDesktop, or other vendor management software if no management environment currently exists or if specific application management services need isolation Citrix Provisioning ServicesResources for hosting the Citrix Provisioning Services environments Citrix XenDesktopResources for hosting Citrix XenDesktop virtual desktops
Expansion Cell
Compute ResourcesResources for scaling out an application cell for two Citrix XenDesktop application cells
The starting design of this Citrix XenDesktop solution supports 500 light workload users. When fully scaled out with the maximum number of cells, the solution supports 3,500 light workload user desktops.
11
11 Figure 2 shows the cell architecture scale-out configuration that supports 3,500 desktops.
Figure 2
12
12
6 cooling fan modules 4 power supply modules Server blades housed in this chassis are for other cells 2 Brocade VDX 6746 DCB switch modules 2 Brocade 5460 Fibre Channel switch modules
6 ports 8 Gb/sec
13
13
Figure 3
Use the infrastructure cell for compute resources for the following cells:
Infrastructure cell for storage resources Application cell for Hitachi Unified Compute Platform Select management Application cell for Citrix Provisioning Services Application cell for Citrix XenDesktop Expansion cell for compute resources
14
14 The infrastructure cell for compute resources and the infrastructure cell for storage resources are the core infrastructure cells required to build a scalable Citrix XenDesktop solution. The Hitachi Compute Blade 500 chassis in the infrastructure cell for compute resources supports the two configurations in Table 3.
Table 3. Infrastructure Cell for Compute Resources Configurations for the Hitachi Compute Blade 500 Chassis
Configuration Options XenDesktop infrastructure part of the Unified Compute Platform solution
Server Blades 0-1 Application Cell for Hitachi Unified Compute Platform Select management
Application Cell for Citrix XenDesktop XenDesktop infrastructure outside of the Unified Compute Platform solution
The infrastructure cell for compute resources and the infrastructure cell for storage resources support up to three expansion cells for Hitachi Compute Blade 500 before requiring new infrastructure cells. Every infrastructure cell for compute resources requires one infrastructure cell for storage resources.
Chassis Components
The Hitachi Compute Blade 500 chassis is equipped with redundant management modules to provide high availability access to manage and monitor the chassis, switch modules, and server blades. The chassis contains redundant switch modules for high availability and maximum throughput. Hot swappable power and fan modules allow for non-disruptive maintenance.
Network Infrastructure
The network design used in this solution provides ample bandwidth and redundancy for the following:
One fully populated infrastructure cell for compute resources One infrastructure cell for storage resources Up to three expansion cells for Hitachi Compute Blade 500
15
15 Figure 4 illustrates the physical network configuration of the infrastructure cell for compute.
Figure 4
The network design allows for the utilization of advanced features inherent in the Brocade VDX switch family, such as VCS Fabric Technology. This provides:
16
16
SAN Infrastructure
The Hitachi Unified Storage 150 controller used for this solution has 16 ports for connections to the Brocade 6510 enterprise fabric switches. For this solution, zone the infrastructure cell for compute resources to four ports, with two ports per processor. When adding expansion cells for Hitachi Compute Blade 500 to the solution, zone four new open ports on the controller to each new cell. Dedicating four ports on Hitachi Unified Storage 150 to each Hitachi Compute Blade 500 chassis ensures bandwidth between each chassis and Hitachi Unified Storage 150. Figure 5 on page 17 illustrates the physical SAN configuration of the infrastructure cell for compute.
17
17
Figure 5
18
18
1 SFF disk expansion tray Disk drives housed in this tray are for other cells
32 GB cache
Figure 6
19
19 Use an infrastructure cell for storage resources in conjunction with the following cells:
Infrastructure cell for compute resources Application cell for Hitachi Unified Compute Platform Select management Application cell for Citrix Provisioning Services Application cell for Citrix XenDesktop
The infrastructure cell for storage provides the storage infrastructure for all the cells in the solution. After fully populating an infrastructure cell for storage resources, add infrastructure cells for storage resources to scale out the solution. The Hitachi Unified Storage 150 controller in each infrastructure cell for storage resources can support up to eight application cells for Citrix XenDesktop. The Hitachi Unified Storage 1500 SFF disk expansion tray in the infrastructure cell for storage resources can house the following:
An application cell for Hitachi Unified Compute Platform Select management Application cells for Citrix Provisioning Services Hot spare drives
20
20
2 6-core Intel Xeon E5-2640 processors, 2.5 GHz 128 GB RAM 1 Emulex 2-port 10 GbE on-board CNA card 1 Emulex 2-port 8 Gb Fibre Channel mezzanine card
8 600 GB 10k RPM SAS drives RAID-6 (6D+2P) configuration Installed in the disk tray of the infrastructure cell for storage resources
1 600 GB 10k RPM SAS drive Hot spare drive Installed in the disk tray of the infrastructure cell for storage resources
Figure 7
Use an application cell for Hitachi Unified Compute Platform Select management in conjunction with the following cells:
Infrastructure cell for compute resources Infrastructure cell for storage resources
21
21 Use an application cell for Hitachi Unified Compute Platform Select management if either of the following exists:
A VMware vSphere environment does not exist already The Citrix XenDesktop infrastructure and management components need to be isolated
The application cell for Hitachi Unified Compute Platform Select management can be added to either of the following:
The infrastructure cell for compute resources and infrastructure cell for storage resources A separate environment outside of the Citrix XenDesktop solution
If installed outside of the Citrix XenDesktop solution, follow the recommended virtual machine requirements outlined in this solution to size the compute hardware used to host the infrastructure and management virtual machines.
Table 4 lists the software components in the application cell for Hitachi Unified Compute Platform Select management.
Table 4. Software Components for the Application Cell for Hitachi Unified Compute Platform Select Management
Software Citrix XenDesktop VMware vCenter VMware ESXi Microsoft Windows Server Microsoft SQL Server
22
22
Compute Infrastructure
The application cell for Hitachi Unified Compute Platform Select management provides enough capacity to support an emergency high availability event if a single server blade fails. Use VMware High Availability and Distributed Resource Scheduler to configure a cluster dedicated to the application cell for Hitachi Unified Compute Platform Select management to ensure virtual machine failover in the event of a hardware failure. The compute infrastructure of the application cell for Hitachi Unified Compute Platform Select management supports all associated requirements for the following:
Citrix desktop delivery controller Microsoft Active Directory DHCP VMware vCenter
Network Infrastructure
Each 520HB1 server blade is configured with a single onboard two-channel 10 GbE CNA card for network traffic. Split each CNA card into four logical NICs per channel, for a total of eight NICs per blade. This solution only uses three NICs per channel. Set bandwidth and NIC allocation as follows:
VMKernel management network vSwitch 1 GbE per NIC, for a total of 2 GbE
VMKernel vMotion network vSwitch 2 GbE per NIC, for a total of 4 GbE
Virtual machine network vSwitch 7 GbE per NIC, for a total of 14 GbE
23
23 Figure 8 illustrates the CNA and Fibre Channel to switch module mapping for the Hitachi Compute Blade 500.
Figure 8
The following VLANs separate network traffic in the application cell for Hitachi Unified Compute Platform Select management:
Management-VLANChassis management connections and primary management of the ESXi hypervisors vMotion-VLANConfigured for vMotion Desktop-VLANConfigured for the desktop network
Storage Infrastructure
The storage infrastructure of the application cell for Hitachi Unified Compute Platform Select management consists of nine 600 GB 10k RPM SAS drives housed in the disk expansion tray contained in the infrastructure cell for storage resources. Using Hitachi Dynamic Provisioning, configure eight of the disks as follows:
Each deployed application cell for Hitachi Unified Compute Platform Select management uses a separate, dynamic provisioning pool configured with one RAID-6 (6D+2P) group dedicated to a Citrix XenDesktop infrastructure and management virtual machines.
24
24 Zone each 520HB1 server blade in the application cell for Hitachi Unified Compute Platform Select management to Hitachi Unified Storage 150 through the Brocade 5460 Fibre Chanel switch modules. Use a single initiator to multi target zoning for each HBA port on the 520HB1 server blades. Following best practice, configure the SAN environment in a dual fabric topology for redundancy and high availability. This results in four paths available to each ESXi host, providing the following:
Table 5 has the zoning configuration for the application cell for Unified Compute Platform Select management.
Table 5. Application Cell for Hitachi Unified Compute Platform Select Management Zone Configuration
Storage Port 0A 1A 0B 1B 0A 1A 0B 1B
Blade 1-ESX 1
HBA1_1 HBA1_2
Always create a new, dedicated application cell for Unified Compute Platform Select management for every eight application cells for Citrix XenDesktop. (See Application Cell for Citrix XenDesktop on page 29.) Design for high availability compliance and rolling host maintenance if end users require uninterrupted service. Set the multipathing policy for each target to round robin in the ESXi configuration.
25
25
Configuration
Count 1
vCPU1 vRAM8GB Operating system drive20 GB vCPU2 vRAM 8GB Operating system drive20 GB vCPU2 vRAM8GB Operating system drive20 GB Data drive20 GB vCPU4 vRAM 8GB Operating system drive20GB vCPU2 vRAM 4GB Operating System Drive20GB
VMware vCenter
26
26
Virtual machines for Citrix Provisioning Services supporting the Citrix XenDesktop environment Desktop master images
It consists of these components: 2 520HB1 server blades Each server blade has the following:
2 6-core Intel Xeon E5-2640 processors, 2.5 GHz 128 GB RAM 1 Emulex 2-port 10 GbE on-board CNA card 1 Emulex 2-port 8 Gb Fibre Channel mezzanine card
8 600 GB 10k RPM SAS drives RAID-6 (6D+2P) configuration Installed in the disk tray of the infrastructure cell for storage resources
Figure 9
Use the application cell for Citrix Provisioning Services in conjunction with the following cells:
Infrastructure cell for compute resources Infrastructure cell for storage resources
27
27 Table 7 has the software components in the application cell for Citrix Provisioning Services for this architecture.
Table 7. Application Cell for Citrix Provisioning Services Software Components
Compute Infrastructure
Based on the Citrix recommendation of 500 desktop connections per provisioning server, the compute components in the cell are sized to host nine provisioning servers. This is to support either of the following:
Eight application cells for Citrix XenDesktop (Application Cell for Citrix XenDesktop on page 29) 4000 desktops in an N+1 configuration
The application cell for Citrix Provisioning Service provides enough capacity to support an emergency high availability event if a single server blade fails. Use a VMware High Availability and Distributed Resource Scheduler cluster dedicated to the application cell for Citrix Provisioning Services to ensure virtual machine failover in the event of a hardware failure.
Network Infrastructure
The application cell for Citrix Provisioning Services uses the same networking configuration described in the application cell for Hitachi Unified Compute Platform Select management. See Application Cell for Hitachi Unified Compute Platform Select Management on page 20.
Storage Infrastructure
The storage in the application cell for Citrix Provisioning Services provides 3.6 TB of space to host the virtual disk files for up to nine Citrix Provisioning Services servers and up to twenty 20 GB desktop images per Citrix Provisioning Services server. Using Hitachi Dynamic Provisioning, configure the storage into a single RAID6 (6D+2P) dynamic provisioning pool dedicated to Citrix Provisioning Services servers.
Each application cell for Citrix Provisioning Services deployed uses a separate dynamic provisioning pool with one RAID-6 (6D+2P) group dedicated to Citrix Provisioning Services servers. The application cell for Citrix Provisioning Services uses the hot spare drive contained in the application cell for Hitachi Unified Compute Platform Select.
28
28 Zone each 520HB1 server blade in the application cell for Citrix Provisioning Services to Hitachi Unified Storage 150 through the Brocade 5460 Fiber Chanel switch modules. Use single initiator to multi-target zoning for each port on the server blades. Following best practice, configure the SAN environment in a dual fabric topology for redundancy and high availability. This results in four paths available to each ESXi host, providing the following:
Table 8 has the zoning configuration for the application cell for Citrix Provisioning Services.
Table 8. Application cell for Citrix Provisioning Service Zone Configuration
Host
Fibre
Zone Name
Storage Port 0A 1A 0B 1B 0A 1A 0B 1B
Eight application cells for Citrix XenDesktop Seven additional Citrix Provisioning Services virtual machines
Add a new Citrix Provisioning Services virtual machine to the environment for either of the following:
Each application cell for Citrix XenDesktop added Every 500 desktops
Use the virtual machine sizing recommendations in this reference architecture guide when adding Citrix Provisioning Services virtual machines to the environment.
29
29 Table 9 has the recommended virtual machine configuration for a Citrix Provisioning Services server supporting 500 desktops.
Table 9. Virtual Machine Sizing Recommendations for 500 Desktops
Configuration
2 6-core Intel Xeon E5-2640 processors, 2.5 GHz 144 GB RAM 1 Emulex 2-port 10 GbE onboard CNA card 1 Emulex 2-port 8 Gb Fibre Channel mezzanine card
24 600 GB 10k RPM SAS drives RAID-10 (2D+2P) configuration Installed in the disk tray of the infrastructure cell for storage resources
1 600 GB 10k RPM SAS drive Hot spare drive Installed in the disk tray of the infrastructure cell for storage resources
1 SFF disk expansion tray Added to the infrastructure cell for storage resources
30
30
Figure 10
Use an application cell for Citrix XenDesktop in conjunction with the following cells:
Infrastructure cell for compute resources Infrastructure cell for storage resources Expansion cell for compute resources (used for scale-out)
To start building a scalable Citrix XenDesktop environment: Add the compute components of the application cell for Citrix XenDesktop to the Hitachi Compute Blade 500 chassis included with the infrastructure cell for compute resources Add the storage components to the infrastructure cell for storage resources
Each application cell for Citrix XenDesktop supports up to 500 light workload users. To scale out the solution, add one of the following to your environment:
Additional application cells for Citrix XenDesktop to the infrastructure cell for compute resources An expansion cell for compute resources
31
31 Up to eight application cells for Citrix XenDesktop can be supported by a single infrastructure cell for compute resources and infrastructure cell for storage resources before you require new infrastructure cells.
Compute Infrastructure
The application cell for Citrix XenDesktop supports a maximum density of 500 light user desktops per cell. In a maximum density configuration, a cell cannot support the failover of desktops in the case of a server blade failure. To provide high availability, reduce the total number of desktops in units of 125 users. For every 125 users you reduce from the total, the environment can sustain one additional server blade failure. For high availability, create a dedicated VMware High Availability and Distributed Resource Scheduler cluster. Then, place the hosts from each application cell for Citrix XenDesktop into the cluster. Based on VMware maximums each cluster can support up to eight application cells for Citrix XenDesktop (32 hosts). Table 10 has the configuration of the Microsoft Windows 7 virtual machine for the streamed desktops.
Table 10. Microsoft Windows 7 Virtual Machine Configuration
Virtual Machine Configuration Option Operating system vCPU Memory Operating system disk size Write cache disk size
Table 11 lists the software components in the application cell for Citrix XenDesktop.
Table 11. Application Cell for Citrix XenDesktop Software Components
Software Citrix Virtual Desktop Agent Citrix Provisioning Services Agent VMware Tools VMware ESXi Microsoft Windows 7
Network Infrastructure
The application cell for Citrix XenDesktop uses the same networking configuration described in the application cell for Hitachi Unified Compute Platform Select management. See Application Cell for Hitachi Unified Compute Platform Select Management on page 20.
32
32
Storage Infrastructure
The storage infrastructure of the application cell for Citrix XenDesktop consists of twenty-four 600 GB 10k RPM SAS drives. Following best practice, the SAN environment was configured in a dual fabric topology for redundancy and high availability. These drives are in two dynamic provisioning pools, each consisting of three RAID-10 (2D+2D) groups created with Hitachi Dynamic Provisioning. Each pool is dedicated for the following:
This supports sustained light workloads and situations such as boot storms and log on storms. The 520HB1 server blades used in this reference architecture use dual-port 8 Gb/ sec Fibre Channel mezzanine cards. There are redundant connections to the Brocade 5460 fabric switch modules. The environment uses single initiator to multi target zoning for each port on the 520HB1 server blades. This results in four paths available to each ESXi host, providing the following:
Set the multipathing policy for each target to round robin in the ESXi configuration. Following best practice, configure the SAN environment in a dual fabric topology for redundancy and high availability. This results in optimal load distribution during an all paths available situation. Table 12 on page 33 shows the zone configuration for the application cell for Citrix XenDesktop.
33
33
Table 12. Application cell for Citrix XenDesktop Zone Configuration
Host HBA Number HBA1_1 HBA1_2 HBA1_1 HBA1_2 HBA1_1 HBA1_2 HBA1_1 HBA1_2
Zone Name esx4_hba1_1_hus_0A_1A esx4_hba1_2_hus_0B_1B esx5_hba1_1_hus_0A_1A esx5_hba1_2_hus_0B_1B esx6_hba1_1_hus_0A_1A esx6_hba1_2_hus_0B_1B esx7_hba1_1_hus_0A_1A esx7_hba1_2_hus_0B_1B
Storage Port* 0A 1A 0B 1B 0A 1A 0B 1B 0A 1A 0B 1B 0A 1A 0B 1B
Blade5-ESX 5
Blade6-ESX 6
Blade7-ESX 7
*The storage target ports for each cell are dependent on the Hitachi Compute Blade 500 chassis in which each is hosted. See Expansion Cell for Compute Resources on page 34 for details.
Use dedicated dynamic provisioning pools for each defined user workload in the environment.
34
34
Housed in the 42U rack enclosure used by the infrastructure cell for compute resources Hitachi Compute Blade 500 Chassis
6 cooling fan modules 4 power supply modules Holds server blades for application cells 2 Brocade VDX 6746 DCB switch modules 2 Brocade 5460 Fibre Channel switch modules
6 ports 8 Gb/sec
Figure 11
Use an expansion cell for compute resources in conjunction with the following cells:
Infrastructure cell for compute resources Application cell for Citrix XenDesktop
Once the chassis in the infrastructure cell for compute resources is full, use an expansion cell for compute resources to provide additional desktop capacity. The expansion cell for compute resources uses the storage and networking infrastructure provided in the infrastructure cells for compute resources and storage resources. House this cell in the infrastructure cell for compute resources rack enclosure.
35
35 You can add up to three expansion cells for Hitachi Compute Blade 500 to infrastructure cells for compute resources and storage resources before requiring new infrastructure cells.
Chassis Components
The expansion cell for compute resources uses the same chassis components contained in the infrastructure cell for compute resources. See Infrastructure Cell for Compute Resources on page 12.
Network Infrastructure
The expansion cell for compute resources uses the same networking configurations as the infrastructure cell for compute resources chassis. See Infrastructure Cell for Compute Resources on page 12.
Storage Infrastructure
Utilizing four of the open storage target ports on Hitachi Unified Storage 150 in the infrastructure cell for storage resources, follow the same storage configuration described for the infrastructure cell for compute resources. Utilize the newly provisioned storage target ports in the zoning configuration. Figure 12 shows the storage target ports of a fully scaled out solution.
Figure 12
36
36
Engineering Validation
This describes the test methodology used to validate the reference architecture with the results of the validation testing. This reference architecture tested the core components of the Hitachi Unified Compute Platform Select for Citrix XenDesktop solution to determine maximum loads per application cell that the solution could support maintaining an acceptable end-user experience. The tested components were validated to support streaming up to 500 desktops with a light user workload per application cell. The actual number of desktops in an environment will vary, depending on workload and high availability requirements. One of the following of each cell was tested in this 500-user reference architecture (Figure 13 on page 37):
Infrastructure cell for compute resources Infrastructure cell for storage resources Application cell for Citrix VDI (500 desktops for light workload users) Application cell for Hitachi Unified Compute Platform Select Application cell for Citrix Provisioning Services Application cell for Citrix XenDesktop
37
37
Figure 13
Test Methodology
Testing used a single chassis to test a 500 seat streamed desktop configuration. Login Consultants Login VSI generated the workload for the lab validation testing. Testing was with a light user workload profile that generated an average of seven IOPS per desktop during steady state.
38
38
Login VSI Configuration Option Application launch delay Total number of virtual machines under test Test mode Log off user(s) after test completion Workload
Value 10 seconds 500 Light workload Disabled Microsoft Outlook, random Microsoft Word document, Microsoft Internet Explorer, Microsoft Excel
Boot Storm
Citrix XenDesktop Power Management controlled the desktop power state during the boot storm tests. The default settings were used:
Maximum active actions40 Max New actions per minute10 Max power actions20%
Cache write pending and core utilization of the storage processor on Hitachi Unified Storage 150 was monitored throughout the entirety of the boot storm test.
Test Results
These are the test results for the environment operating in a logon and steady state condition.
Compute Infrastructure
During the logon and steady state test, the CPU and memory statistics were captured using esxtop to determine the maximum density that the compute hardware could support.
39
39
With the core utilization averaging at 96% during steady state, the server blades ran at maximum CPU utilization. The test determined maximum density per host.
Figure 14
40
40
Swap used on the hypervisor throughout the workload did not rise above 274 MB. The swap in and swap out rates were very low. The overall used memory throughout the 500-user test was close to the available memory. The 520HB1 server blade has adequate memory for light workloads.
Figure 15
With the memory savings provided by the transparent page sharing, the 520HB1 server blades had adequate memory to support 500 desktops running a light workload.
41
41
Storage Infrastructure
To understand the I/O profile of the typical light workload user, multiple performance metrics were collected from the Hitachi Unified Storage 150 controllers to test the following:
Physical disks were saturated Cache was overwhelmed RAID groups performed well
Figure 16
42
42 The low tag count statistics indicate the physical disks queue was not saturated.
Figure 17
Write pending percentage averaged below 4%. Core utilization peaked around 18%.
This data indicates the Hitachi Unified Storage 150 controllers are adequate for 500 light workload streamed desktops, with plenty of headroom for future growth. Figure 18 on page 43 illustrates the processor and cache performance metrics.
43
43
Figure 18
Datastore Performance
Figure 19 and Figure 20 on page 44 illustrate the aggregated performance for the datastore dedicated to the XenDesktop virtual machine disk files and desktop write cache. The following data points are well within acceptable ranges:
The read IOPS peaked at approximately 440 during the logon storm. The write IOPS peaked at approximately 5217 during the logon storm. The read steady state IOPS peaked at approximately 61. The write steady state IOPS peaked at approximately 4579. Latency peaked at approximately 0.9 milliseconds.
44
44
Figure 19
Figure 20
45
45 Metrics were collected for every virtual desktop during the 500-user run. Figure 21 illustrates a representative sample of five guests, showing the combined virtual machine disk files and write cache IOPS statistics for those specific guests.
The IOPS peaked during logon storm for each desktop between one minute and sixty minutes (the first peak for each user). During the remainder of the steady state workload, the desktops averaged seven IOPS.
Figure 21
46
46
Desktop IO Profile
Figure 22 illustrates the percentage of read and write I/O operations that were random versus sequential on the virtual machine disk files and write cache dedicated datastores. These measurements were taken at the back end storage level to show the I/O profile of the workload as seen by the Hitachi Unified Storage 150 controllers.
Figure 22
Application Experience
The Login VSI testing tool reported the time required to open and close the simulated applications during the test.
Load times for all tested applications were less than 1.5 seconds. File open times for all tested applications were less than 0.5 seconds.
These performance metrics are very close to physical desktop performance. This proves that this reference architecture provides adequate performance for 500 users with a light workload profile. Figure 23 on page 47 illustrates the application experience metrics reported by Login VSI.
47
47
Figure 23
Boot Storm
Using Citrix XenDesktop power management to power on the XenDesktop environment, it took approximately 60 minutes until all desktops were registered as ready in Desktop Studio. Figure 24 on page 48 illustrates the aggregated performance of the RAID groups dedicated to the streamed desktops virtual machine disk files and write cache datastores.
Read IOPS peaked at approximately 434. Write IOPs peaked at approximately 4016.
48
48
Figure 24
The data in Figure 24 proves the underlying storage supports a managed poweron of the 500 desktops to have them ready for use in an hour. Figure 25 on page 49 illustrates the percent busy statistics for a representative drive from a dynamic provisioning pool for a virtual machine disk files and write cache during the boot storm test. This physical drive showed the following:
49
49
Figure 25
The percent busy rate seen in Figure 25 remained low during the boot storm test, showing that there is room to support bursts.
Write pending percentage averaged below four percent. Core utilization peaked around eight percent.
50
50
Figure 26
The default Citrix XenDesktop Power Management settings should be sufficient in most environments. If not sufficient, configure the settings under advanced settings of the host details in Desktop Studio. If you decide to increase these settings, make small incremental changes and monitor the affects on the hypervisor and storage to make sure you do not overwhelm the systems.
51
51
Conclusion
This reference architecture guide shows how to design a Hitachi Unified Compute Platform Select for Citrix XenDesktop solution using Hitachi Compute Blade 500 and Hitachi Unified Storage 150. The cell design validated in the Hitachi Data Systems laboratory allows a build-as-you-go model with performance-proven sets of hardware resources. By using this modular approach, you can scale an environment from 500 light workload users to the VMware vSphere configuration maximums. Create a rightsized design that allows purchasing flexibility to meet changing business or project needs. Understand the I/O workload of the desktops in your Citrix XenDesktop environment so you properly design the architecture. This reduces costs and increases ROI. Using Hitachi Dynamic Provisioning to create dynamic provisioning pools in this reference architecture allows increasing or decreasing I/O requirements dynamically, as necessary. Having the capability to provision additional spindles to an already-provisioned datastore within VMware vSphere allows for nondisruptive upgrades to the underlying storage infrastructure. This provides immediate benefits to the environment with no confusing shuffling of virtual machines, datastores, or LUs. Further use of Hitachi Dynamic Provisioning to create dedicated dynamic provisioning pools for different types of XenDesktop users could drive utilization and ROI even higher.
Corporate Headquarters 2845 Lafayette Street, Santa Clara, California 95050-2627 USA www.HDS.com Regional Contact Information Americas: +1 408 970 1000 or info@HDS.com Europe, Middle East and Africa: +44 (0) 1753 618000 or info.emea@HDS.com Asia-Pacific: +852 3189 7900 or hds.marketing.apac@HDS.com
Hitachi Data Systems Corporation 2013. All rights reserved. HITACHI is a trademark or registered trademark of Hitachi, Ltd. Innovate with Information is a trademark or registered trademark of Hitachi Data Systems Corporation. Microsoft, Internet Explorer, Excel, Outlook, PowerPoint, Windows, Windows Server, SQL Server, and Active Directory are trademarks or registered trademarks of Microsoft Corporation. All other trademarks, service marks, and company names are properties of their respective owners. Notice: This document is for informational purposes only, and does not set forth any warranty, expressed or implied, concerning any equipment or service offered or to be offered by Hitachi Data Systems Corporation. AS-185-01, January 2013