Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage Create robust, highly available vSphere 5.1 environments with best- of-breed Fibre Channel HP 3PAR Storage and HP ProLiant DL servers
Table of contents Emulex Solution Implementers Series .............................................................................................................3 Executive summary ..........................................................................................................................................3 Introduction ......................................................................................................................................................5 About this guide ...........................................................................................................................................5 Solution components ........................................................................................................................................6 ESXi 5.1 .......................................................................................................................................................7 HP ProLiant Gen8 servers ...........................................................................................................................7 Deploying the solution components .............................................................................................................8 Pre-installation .................................................................................................................................................9 Updating firmware ........................................................................................................................................9 Configuring network connectivity .................................................................................................................... 10 Planning the network environment for the host .......................................................................................... 11 Configuring storage ........................................................................................................................................ 12 Using the VMFS-5 filesystem for additional functionality ............................................................................ 12 Configuring Fibre Channel connectivity ...................................................................................................... 13 Deploying ESXi 5.1 ........................................................................................................................................ 18 Post-installation .............................................................................................................................................. 19 Configuring the HBA .................................................................................................................................. 19 Using NPIV to identify HBA ports ............................................................................................................... 21 Configuring the storage array ..................................................................................................................... 21 Provisioning virtual LUNs ........................................................................................................................... 22 Performance comparison ............................................................................................................................... 23 Test method ............................................................................................................................................... 24 Results ........................................................................................................................................................... 25 Summary ........................................................................................................................................................ 30 Appendix A Configuring BfS ........................................................................................................................ 31 For more information ...................................................................................................................................... 32
Solution Implementers Series
3 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Emulex Solution Implementers Series This document is part of the Emulex Solution Implementers Series, providing Implementers (IT administrators and system architects) with solution and deployment information on popular server and software platforms. As a leader in I/O adapters Fibre Channel (FC), Ethernet, iSCSI and Fibre Channel over Ethernet (FCoE) the Emulex technology team is taking a lead in providing guidelines for implementing I/O for these solutions. Executive summary With vSphere 5.1, VMware continues to raise the bar for hypervisor products, introducing many new features along with support for more and larger virtual machines (VMs) that can now utilize up to 64 virtual CPUs (vCPUs). vSphere 5.1, like 5.0, does not include a service console operating system (OS); VMware agents and Common Information Model (CIM) providers run directly on the hypervisor layer (VMkernel). There are three options for communicating with VMkernel: VMware vSpheres enhanced command-line interface (vCLI), vSphere PowerCLI or the vSphere Management Assistant (vMA) virtual appliance. vSphere 5.1 provides many new features and enhancements in areas such as storage and networking. Indeed, VMware has several new features, 1 claiming, for example, that vSphere 5.1 can run VMs that are two-times as powerful as those supported by earlier versions and support for new VM hardware formats, such as VM Virtual Hardware Version 9. In storage, there is added support for 16Gb Fibre Channel (16GFC), however, as shown in Table 1, the new, larger VM will place heavy demands on data center infrastructure.
1 http://www.vmware.com/files/pdf/products/vsphere/vmware-what-is-new-vsphere51.pdf Solution Implementers Series
4 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Table 1. Resources supported by various VMware hypervisors Component E S X
1
E S X 2
V M w a r e
I n f r a s t r u c t u r e
3
V M w a r e
v S p h e r e
4
V M w a r e
v S p h e r e
5
V M w a r e
v S p h e r e
5 . 1
S c a l e
f a c t o r
vCPUs 1 2 4 8 32 64 x 64 Memory (GB per VM) 2 3.6 64 256 1,000 1,000 x 500 Network (Gb) < 0.5 0.9 9 30 >36 >36 x 72 Fibre Channel SAN (MB/s per HBA port) 1GFC: 100 2GFC: 200 4GFC: 400 8GFC: 800 8GFC: 800 16GFC: 1,600 x 16
To help you transition to an infrastructure that is capable of supporting the storage needed by new-generation VMs, Emulex has validated the functionality and performance of vSphere 5.1 in conjunction with 16GFC connectivity. The proof-of-concept (POC) environment included the following components: Best-of-breed HP ProLiant DL380 Gen 8 server Dual-port 16GFC adapter produced for HP by Emulex HP 3PAR P10000 V400 Storage SANBlaze 16GFC storage emulator (16GFC connectivity end-to-end) In addition, since 16GFC Storage Area Network (SAN) storage is better suited for the new release of vSphere, this implementers guide outlines the process for deploying vSphere 5.1 with 16GFC connectivity and outlines the results of performance tests carried out in such an environment. The performance test section was executed by the Emulex Technical Marketing team in their labs and it was to validate and test 16GFC connectivity end-to-end. The performance testing demonstrated that, with a 16GFC Emulex Fibre Channel Host Bus Adapter (HBA), I/O performance was significantly higher (as much as 99%) than with 8GFC technology, without requiring additional CPU cycles. Thus, as VM density increases, you are placing more burden on the storage array by migrating from 8GFC to 16GFC since I/O bandwidth has doubled on the array. Intended audience: This document is intended for engineers, system administrators and VMware administrators interested in deploying vSphere 5.1 on an HP ProLiant Gen8 server Solution Implementers Series
5 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
featuring an HP Ethernet 10Gb 2-port 554FLR-SFP+ Adapter (684213-B21) and HP SN1000E 16Gb 2-port PCIe Fibre Channel Host Bus Adapter (QR559A). Testing performed in August September, 2012, is described. Introduction vSphere 5.1 supports an unprecedented number of VMs on a single host; moreover, these VMs can reach unprecedented size depending on the application and workload for each VM. Thus, Emulex expects to see more and more workloads being virtualized, as well as additional resources being assigned to existing VMs in order to meet the needs of particular workloads. As noted by VMware, 2 the VM will only get bigger and faster.
VMware sees more VMs being deployed than ever before, 3 with vSphere 5.1 allowing these VMs to grow as much as two times larger. With this increased density, virtualized environments must be able to provide additional network and storage resources in order to support the increased workload. About this guide This implementers guide describes how to configure a 16GFC SAN with a DL380 Gen8 server in a vSphere environment. Guidelines and instructions are provided for configuring servers, adapters and storage using technical documentation provided by VMware and HP. Information is provided on the following topics: Solution components Networking configuration Storage configuration, including boot from SAN (BfS) Performance testing to compare 8GFC and 16GFC
2 http://www.vmware.com/files/pdf/products/vsphere/vmware-what-is-new-vsphere51.pdf 3 Based on interim results of VMware customer surveys performed in January 2010 and April 2011 Solution Implementers Series
6 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Solution components Emulex built the POC environment shown in Figure 1 in order to validate 16GFC connectivity with vSphere 5.1.
Figure 1. 16GFC POC
Solution Implementers Series
7 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Table 2 outlines the key components deployed in this environment.
Table 2. Test environment Component Device Comments Tested server HP ProLiant DL380 Gen8 Virtualization host running ESXi 5.1; 10 identically-configured VMs Management server vCenter Server 5.1 VM running on ESXi 5.1 host vCenter Server 5.1 is a VM running Windows Server 2008 R2 AD & DNS server Generic Windows Server 2008 system Support for Microsoft Active Directory (AD) and Domain Name System (DNS) Storage HP 3PAR Storage P10000 V400 Storage Fibre Channel connectivity 16GFC HBA HP SN1000E 16Gb 2-port PCIe Fibre Channel Host Bus Adapter (HBA) 16GFC fabric switch 2 x Brocade 6510 16GFC SAN switch
If this is your first time installing VMware products on a ProLiant server, it is important for you to have a basic understanding of each of the solution components so that the terminology used in this guide is clear. vSphere 5.1 VMwares latest hypervisor, vSphere 5.1, extends the core capabilities of vSphere 4.1 and 5.0 to provide the foundation for a cloud infrastructure, whether public or private. Areas where you can expect to see improvements after deploying vSphere 5.1 include server consolidation, performance, management, provisioning and troubleshooting. For more information on vSphere 5.1, refer to Whats New in VMware vSphere 5.1, available at http://www.vmware.com/files/pdf/products/vsphere/vmware-what-is-new-vsphere51.pdf. HP ProLiant Gen8 servers The latest DL380 Gen8 servers are based on Intel Romley processors. These systems continue to be the servers of choice for many HP shops in the VMware space and are widely used from small and medium-sized businesses (SMBs) to large data centers for their high availability, scalability, CPU horsepower and memory capacity. In addition, these 2U rack-mount servers save space and power, making them ideal for large data centers moving to a cloud infrastructure. Solution Implementers Series
8 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Using Fibre Channel for shared storage When deploying vSphere on ProLiant servers, you should always consider using a SAN so that you can take full advantage of the many hypervisor features and enhancements that require shared storage. While VMware supports most of the more common protocols, Fibre Channel is the predominant choice for shared SAN storage. Thus, while storage protocols continue to evolve, introducing options such as iSCSI and NFS storage arrays, this guide focuses on Fibre Channel connectivity. Server sizing HP has developed an automated tool HP Sizing Tool for VMware vSphere that can help you size and scope a server for a particular vSphere deployment. Based on your responses to a questionnaire, the tool provides a quick, consistent method for identifying the best server for your environment. It also creates a bill of material. For more information on this sizer and other HP solutions for VMware, visit www.hp.com/go/vmware. HP 3PAR Storage Highly virtualized from the ground up, HP 3PAR Storage can enhance the benefits of a vSphere environment by taking care of the demands that server virtualization places on the storage infrastructure. HP 3PAR Storage combines highly virtualized, autonomically managed, dynamically tiered storage arrays with advanced internal virtualization capabilities to increase administrative efficiency, system utilization and storage performance. As a result, HP 3PAR Storage can boost the return on a vSphere investment by allowing you to optimize your data center infrastructure, simplify storage administration and maximize virtualization savings. A key area for HP 3PAR P10000 V400 storage arrays which makes a significant impact on performance when migrating from 8GFC to 16GFC is the wide striping architecture. As noted in HPs documentation, wide striping distributes each vSphere storage volume across all array resources. When you increase bandwidth to the array, you are not disk bound on the LUN like you are with traditional arrays that narrow stripe data. Deploying the solution components Having introduced the key components of the POC, this implementers guide now describes how to configure 16GFC connectivity. Guidelines are provided for the following areas: Pre-installation Configuring network connectivity Solution Implementers Series
9 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Configuring storage Deploying vSphere 5.1 Post-installation Pre-installation There are several steps to consider before applying power or installing an operating system on a system. First, you need to ensure that sufficient rack space and appropriate power and cooling are available; you should also verify that firmware is at the latest levels and download any necessary patches and drivers using the HP and VMware links provided in this section. The following pre-installation process offers guidelines for pre-configuring a network to support an ESXi host, as well as suggestions for configuring storage systems and storage area networking. As a best practice, Emulex recommends verifying with HP technical support that you are running the very latest HP firmware and drivers on components, such as HP FlexLOM 4 and PCI adapters. Updating firmware Before deploying vSphere 5.1 on a ProLiant server, you should determine the latest versions of the following firmware and, if necessary, update: ESXi host: o System BIOS o Flexible-LOMs and PCI adapters Storage array and controllers SAN switches You can review the latest firmware levels recommended by HP and VMware at the following sites: HP: Visit www.hp.com/go/vmware and refer to the Certified ProLiants and Certified HP Storage links under Tools/Resources. VMware: Refer to the VMware Compatibility Guides at http://www.vmware.com/resources/guides.html. Note Always contact HP support to identify the latest firmware updates and drivers.
4 Where LOM refers to LAN on motherboard Solution Implementers Series
10 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
As always, plan your deployment or upgrade before installing software; read all the documentation provided by VMware and HP before starting. Planning will speed up the process particularly if you intend to deploy multiple servers. With pre-installation activities complete, you can now start configuring your network. Configuring network connectivity Before installing vSphere 5.1, you need to understand the network requirements for the particular ESXi host and the connectivity supported by the physical server. For example, while many physical servers feature LOM or integrated network interfaces, ports are typically 1Gb Ethernet (1GbE), though newer models such as the DL380 Gen8 server offer 10GbE ports. In order to best use 10GbE, you should understand the requirements of the traffic being carried on the network, as outlined in Table 3.
Table 3. Typical network requirements Traffic type Bandwidth usage Other requirements Management Low Highly reliable, secure channel VMware vMotion High Isolated VMware Fault Tolerance (FT) Medium to -high Highly reliable, low-latency channel iSCSI High Reliable, high-speed channel VM Application-dependent Application-dependent
In general, combining management port traffic, which is relatively light, with VMware vMotion traffic is acceptable in many deployments that utilize four network interface cards (NICs). Since vMotion traffic is heavier, it is not a good practice to combine this with VM traffic; thus, you should consider separating such traffic on different subnets. For simplicity, Emulex placed management and vMotion traffic on the same virtual switch in the POC. In such an implementation, it is a good practice to use virtual LANs (VLANs) to enhance security and isolate traffic. Following VMwares best practices for performance 5 provides a good starting point. Remember that, as time passes, you should regularly revisit your environment to ensure the network configuration is still effective.
5 http://www.vmware.com/pdf/Perf_Best_Practices_vSphere5.1.pdf Solution Implementers Series
11 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Planning the network environment for the host You should plan the network environment for the host in conjunction with your network administrator. The DL380 Gen8 server utilized in the POC was equipped with an HP Ethernet 10Gb 2-port 554FLR-SFP+ Adapter, which was used to configure the single network interface card (NIC) that was needed. In vSphere deployments featuring 1GbE network adapters, it is not uncommon for the use of six or even eight NICs in order to meet VMwares networking requirements for performance, isolation and redundancy. However, for this 10GbE environment, Emulex was able to utilize just two 10GbE ports; VLANs were used in conjunction with VMwares networking I/O control to manage bandwidth. Table 4 shows VMwares best practices for two 10GbE ports. Table 4. VMwares best practices, which were used in the POC Traffic type Port group Teaming option Active uplink Standby uplink NIOC shares VM PG-A LBT dvuplink 1, 2 None 20 iSCSI PG-B LBT dvuplink 1, 2 None 20 FT PG-C LBT dvuplink 1, 2 None 10 Management PG-D LBT dvuplink 1, 2 None 5 vMotion PG-E LBT dvuplink 1, 2 None 20
After setting up network connectivity, you can now configure storage.
Solution Implementers Series
12 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Configuring storage This section provides information on the following topics: Using the new VMFS-5 filesystem 6 to support additional storage functionality Configuring Fibre Channel connectivity Implementing fabric zoning Configuring the storage array Configuring BfS Using the VMFS-5 filesystem for additional functionality Introduced with ESXi 5.0, the VMFS-5 filesystem provides support for VMFS volumes up to 64TB in a single extent. With VMFS-3, 32 extents would have been required to achieve 64TB. Note The volume creation wizard for VMFS-5 uses GUID Partition Table (GPT) format rather than Master Boot Record (MBR), which allows you to create VMFS volumes that are larger than 2TB. GUID refers to a globally-unique identifier.
With the ability to create large VMFS volumes, you must now manage storage array queue depth as well as LUN queue depth. For example, queue depth for an HP SN1000E 16Gb 2-port PCIe Fibre Channel HBA is set by default to 30 and may be adjusted via Emulex OneCommand Manager, the OneCommand Manager plug-in for vCenter Server, or vMA. Other benefits delivered by the VMFS-5 filesystem include: Support for up to 130,000 files rather than 30,000 as before Support for 64TB physical-mode RDM LUNs Virtual mode allows you to create snapshots, which is beneficial when a file exceeds 2TB For space efficiency, there can now be up to 30,000 8KB sub-blocks There is small-file support for files of 1KB or less; in the past, such files would have occupied entire sub-blocks As you plan the deployment of shared storage, take into consideration the benefits of VMFS-5. For example, if you are migrating from a hypervisor that is pre-vSphere 5.1, you may also wish to migrate to VMFS-5 to take advantage of the new features and enhancements.
6 VMFS is a VMware Virtual Machine File System format. Solution Implementers Series
13 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Configuring Fibre Channel connectivity The Emulex Fibre Channel driver is in-box with vSphere, making it is easy to transition to 16GFC connectivity from an earlier platform there is no need to install the driver during the deployment of vSphere 5.1. As a best practice, it is recommended to update the fibre channel driver since the in-box driver will be out of date. Thus, the configuration of 16GFC connectivity via HP SN1000E 16Gb 2-port PCIe Fibre Channel HBAs is a simple process in a vSphere 5.1 environment, with just the following stages: Implement fabric zoning Configure the storage array Configure BfS Before you begin the configuration, it is a best practice to review the appropriate VMware Compatibility Guide to ensure firmware for the storage array is at the latest levels and the array has been certified, as shown in Figure 2.
Figure 2. Showing firmware levels specified in the VMware Compatibility Guide for HP 3PAR P10000 Storage
Solution Implementers Series
14 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
IMPORTANT Whenever you need to update code or firmware, Emulex recommends backing up your data and configuration.
Implementing fabric zoning Zoning has become a standard component of VMware deployments; indeed, most if not all storage array vendors recommend zoning LUNs that are presented to ESXi hosts. Fabric zones can enhance manageability while providing support for advanced features such as vMotion and Fault Tolerance that require multiple hosts to access the same LUN. Zones can also enhance security. For example, consider what might happen if you were to connect a new Microsoft Windows server to the same SAN switch as an existing ESXi host. Without zoning or some other security measure, the new server would be able to access the same storage as the existing host and could potentially overwrite the filesystem, obliterating VM data and files. Thus, since the POC features two 16Gb HBA ports, there should ideally be two or more fabric switches, each configured with a zone that includes one of the ports.
Solution Implementers Series
15 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Figure 3 shows the zoning used in the POC.
Figure 3. Zoning in the POC, with four paths to the LUN
Solution Implementers Series
16 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
The POC utilizes two Brocade 16Gb SAN switches and a total of four zones, as shown in Table 5. The zones are added to an alias zone configuration, which is then activated. Table 4. Zone configuration for the POC HBA Storage controller Alias Zone Zone configuration Switch Port 0
Port A1 Zone 1 ZoneSet 1 ZoneConfig_1 1 Port B2 Zone 2 ZoneSet 1 ZoneConfig_1 2 Port 1
Port A2 Zone 3 ZoneSet 2 ZoneConfig_1 2 Port B1 Zone 4 ZoneSet 2 ZoneConfig_1 1
This zone configuration gives the ESXi host four paths to a single LUN. At this stage of the deployment, no LUNs have been created; thus, LUNs cannot yet be bound to WWN ports on the HP SN1000E2P 16Gb HBA. Setting the storage array HP 3PAR Storage is very popular in VMware environments due to its extensive virtualization capabilities and ease of management. HP has documented deployment and tuning best practices for this simple yet robust and highly available array; for more information, refer to 3PAR Utility Storage with VMware vSphere. Emulex followed HPs best practices when configuring HP 3PAR Storage for the POC. The process included the following stages: Update storage array controller firmware and management software as needed Configure virtual domains Configure virtual LUNs Assign host mode Create hosts Present LUNs to the host If the correct zoning and host mode have been applied, LUNs will be visible to the assigned hosts. There should be four paths to each LUN for optimal performance and redundancy.
Solution Implementers Series
17 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Configuring Boot from SAN (BfS) Enterprise server manufacturers such as HP continue to offer local disk storage; however, with the growth of virtualization and the increased use of BfS, 7 server configurations are evolving. For example, HP offers a diskless server that would allow you to deploy vSphere via a USB flash drive or SD card. BfS capability is often configured in a vSphere environment, where its benefits include: Enhanced manageability Faster deployment Easier backup 8 of the hypervisor Enhanced disaster recovery capabilities The process for configuring BfS via an HP SN1000E2P 16Gb HBA is simple and can be achieved in the following stages: Load the latest boot code to HBA Provision the boot LUNs Configure the ESXi host Specify the desired boot volume Place the HBA first in the boot order This vendor-specific process is described in more detail in Appendix A Configuring BfS.
Note If you plan to install or upgrade ESXi 5.1 with local storage, Emulex recommends disconnecting the Fibre Channel cables from the SAN to prevent the OS from being accidentally installed on the SAN.
Once storage has been configured and you have verified that the hardware has been certified by VMware 9 you can deploy ESXi 5.1.
7 ESXi is installed directly on a LUN instead of local storage 8 Since the array owns the LUN, array-based copies can be made without server intervention. 9 Refer to the VMware Compatibility Guides at http://www.vmware.com/resources/guides.html. Solution Implementers Series
18 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Deploying vSphere 5.1 The main points to remember before beginning a vSphere deployment are as follows: Check all firmware on host and adapters and make updates as needed Check HBA drivers (vSphere 5.1 has inbox drivers for Emulex 16GFC) Plan the networking configuration Plan the Fibre Channel storage configuration and LUNs Decide on local or BfS storage Select the most suitable deployment option for your environment Since vSphere 5.1 has been designed for flexibility, you have a range of options for deploying this hypervisor on a ProLiant server. These options include the following: Interactive installation: Suggested for fewer than five hosts Scripted installation: Unattended deployment for multiple hosts vSphere Auto Deploy installation: Suggested for a large number of ESXi hosts; uses VMware vCenter Server Custom installation: vSphere 5 Image Builder command-line interfaces (CLIs) provide custom updates, patches and drivers For the POC, Emulex elected to use the interactive method, downloading an vSphere 5.1 image from the VMware website to local storage. The deployment process is fairly straightforward and, in many ways, identical to the deployment of earlier versions of ESXi. Since this process is detailed in VMwares technical documentation, it is not described in this guide. It should not take more than a few minutes to install vSphere 5.1. Following the installation of vSphere 5.1 you can configure the management network and, if appropriate, enable lockdown mode via vCenter Server or the ESXi direct console user interface (DCUI). You can then proceed with the post-installation process, which includes configuring the vSphere 5.1 host for 16GFC SAN connectivity.
Solution Implementers Series
19 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Post-installation After vSphere 5.1 has been deployed on the host, you should review the installation and perform any updates that are necessary for example, a vendor may have recently released new firmware or drivers. Next, you should configure NIC teaming, which is not configured automatically unless you are using a scripted installation. If you are using local storage, remember to reconnect the Fibre Channel cables to the SAN and then verify that the host can login to the fabric and view any LUNs that have been assigned. You can now configure the vSphere 5.1 host and storage array for 16GFC SAN connectivity, which may involve the following activities: Planning the network environment for the host Configuring the HBA Configuring the storage array with features such as multipathing Configuring the HBA If you are migrating to vSphere 5.1 or are installing vSphere 5.1 for the first time and have installed the HP SN1000E 16Gb 2-port PCIe HBA in a full-length PCIe 3.0 slot, with the correct small form-factor pluggable (SFP) transceivers, configuration is simple. Since vSphere 5.1 already provides an in-box VMware driver for Fibre Channel, there is no need to install a driver during the initial setup. After the vSphere 5.1 installation, however, you should verify with VMware or Emulex that the in-box driver is the latest and, if necessary, update it.
Solution Implementers Series
20 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
After configuring the HBA, review the Configuration tab of vCenter Server; you should see your device listed under Storage Adapters, as shown in Figure 4.
Figure 4. In this example, ESXi has automatically recognized a 16GFC HBA
Alternatively, you can use vMA to remotely send commands to the ESXi host to verify that the driver has been configured correctly.
Solution Implementers Series
21 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Using NPIV to identify HBA ports N-Port ID Virtualization (NPIV) 10 is supported in vSphere 5.1. This capability allows a single Fibre Channel HBA port to register with the fabric using multiple worldwide port names (WWNs), each having a unique entity. To learn more, refer to the VMware technical note, Configuring and Troubleshooting N-Port ID Virtualization, which provides specific information on Emulex adapters. For additional information, refer to the VMware document, vSphere Storage Guide, which is available at http://www.vmware.com/support/pubs/vsphere-esxi-vcenter-server-pubs.html. Configuring the storage array vSphere 5.1 introduced several new storage features and enhanced others. You should be aware of the following: VMFS-5 o Support for 32 hosts for single read-only file on a VMFS volume vStorage API for array integration (VAAI) Storage Distributed Resource Scheduler (SDRS) o Datastore correlation detector o New I/O metric VMobservedLatency Storage I/O Control (SIOC) o Automatic setting for threshold latency Storage vMotion o Four parallel migrations concurrently Storage protocol enhancements o 16GFC support While these valuable new features are beyond the scope of this guide, be aware that you may be facing additional steps after you install vSphere 5.1. For example, after mounting a LUN and formatting it with VMFS-5, you may need to determine if additional, array-specific agents are required to support features such as VAAI or vSphere APIs for Storage Awareness (VASA). For this POC, Emulex configured the HP 3PAR Storage based on information provided in the HP document, HP 3PAR Storage and VMware vSphere 5 best practices.
10 Maintained by American National Standards Institute (ANSI), Technical Committee T11 Solution Implementers Series
22 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Note For more information on ESXi 5.1 features that are supported by a particular HP array, you should see VMwares HCL to see which specific array features are supported for each model and also consult HP technical support.
Provisioning virtual LUNs You can use the HP 3PAR InForm Management Console (shown in Figure 5) to configure and manage HP 3PAR Storage. Figure 5. View of the HP 3PAR InForm Management Console
For this POC, Emulex started off by configuring a single ESXi Common Provisioning Group (CPG) named VMware and used all drives in the array for provisioning virtual LUNs (VLUNs). Using this CPG group, Emulex had to validate FC connectivity provisioned to the DL380 Gen8 server with a single 600GB LUN to provide support for VMs. The next concern, after deploying vSphere 5.1, is HPs suggested configuration for multipathing. Solution Implementers Series
23 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Configuring multipathing By default with vSphere 5.1, HP 3PAR Storage uses Fixed path policy for active/active storage arrays. This policy maximizes bandwidth utilization by designating the preferred path to each LUN through the appropriate storage controller. According to HP documentation, HP 3PAR Storage also supports Round Robin path policy, which can improve storage performance by load-balancing I/O requests between active paths, sending a fixed number of requests through each in turn. Note The fixed number of I/Os is user-configurable. You should consult HP technical support for their recommendations.
You might consider enabling the arrays Asymmetric Logical Unit Access (ALUA) feature, which can improve storage performance in some environments. Performance comparison As VM density increases, the burden placed on storage by applications running on these VMs will also increase. As a result, both Emulex and VMware have carried out performance tests designed to demonstrate the ability of 16GFC storage to sustain a significantly higher workload than 8GFC storage without increasing CPU utilization on the host.
Solution Implementers Series
24 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Figure 6 shows how storage was configured in the VMware test environment.
Figure 6. Test environment
Test method The Iometer tool was used to compare sequential read and write throughput and CPU effectiveness with 16GFC and 8GFC HBAs. Various block sizes were used. A VM was configured with a single IOmeter worker thread. The VM was hosted on a DL380 Gen8 server that was configured with the following: Two six-core Intel Xeon E5-2640 processors HP SN1000E dual-port 16GFC HBA Emulex LPe12002 dual-port 8GFC HBA A Brocade 6510 16Gb FC switch was also in the POC.
Solution Implementers Series
25 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
The VM was configured as follows: One virtual CPU Guest memory: 4,096MB Virtual disk: 10GB OS and 256GB RDM SCSI controllers: One LSI Logic and one PVSCSI virtual controller VM Virtual Hardware Version 9 Guest OS: Microsoft Windows Server 2008 R2, 64-bit The target RDM LUN for the testing was placed on a SANBlaze VirtualLUN 6.3 16GFC appliance, which is used to emulate Fibre Channel drives in order to characterize read/write performance. The SANBlaze device was configured as follows: HP DL380 G7 256GB RAM 16GFC HBA Results Sequential read throughput Figure 7 shows sequential read throughput for a range of block sizes.
Solution Implementers Series
26 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Figure 7. Sequential read throughput in MB/s in the test environment
With larger block sizes the result is what was expected from 16GFC as we are able to utilize near-line rate. 11 As we double the pipeline in size, we are also able to prove the 16GFC adapter can double the throughput. 12
16GFC out-performed 8GFC by almost 100% with larger blocks. CPU effectiveness reads Figure 8 shows CPU effectiveness (defined as total IOPS divided by CPU utilization) for sequential reads. This metric characterizes the number of CPU cycles taken to complete a particular IOPS total; thus, a lower number is more desirable because it indicates the processor is less stressed.
11 1,560 MB/s for sequential reads 12 750 MB/s for sequential reads 0 200 400 600 800 1000 1200 1400 1600 1800 1K 4K 8K 16K 32K 64K 128K 256K M B s
28 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Sequential write throughput Figure 9 shows sequential write throughput for a range of block sizes.
Figure 9. Sequential write throughput in MB/s in the test environment
With larger block sizes the result is what was expected from 16GFC as we are able to utilize near-line rate. Again, as we double the pipeline in size, we are able to also prove the 16GFC adapter can double the throughput.16GFC out-performed 8GFC by almost 100% with larger blocks.
0 200 400 600 800 1000 1200 1400 1600 1800 1K 4K 8K 16K 32K 64K 128K 256K M B s
30 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Summary vSphere 5.1 clearly adds a broad range of features to the hypervisor; however, taking full advantage of these features requires newly-supported 16GFC SAN connectivity in your virtualization environment. Testing carried out by Emulex indicated that I/O performance with a 16GFC implementation with HP DL380 Gen 8 Servers and HP 3PAR P10000 V400 was significantly higher than with 8GFC, without the need to stress the CPUs. This gives HP DL380 Gen 8 servers the ability to utilize the CPU power to complete others higher priority tasks. Planning the deployment of a vSphere 5.1 HP host or the migration of an existing host to vSphere 5.1 is a relatively simple process; guidelines are provided in this guide. Using HP DL380 Gen8, HP 3PAR P10000 with 16GFC connectivity, rather than 8GFC, results in an infrastructure that can meet the demands of the larger VMs you are now able to create with vSphere 5.1, thus allowing you to introduce additional business-critical applications.
Solution Implementers Series
31 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
Appendix A Configuring BfS This appendix outlines the process for configuring BfS: 1. Collaborate with your SAN administrator on provisioning a boot LUN and presenting it to the vSphere 5.1 host. 2. Download the Universal Boot Code firmware for your adapter, available at http://h20000.www2.hp.com/bizsupport/TechSupport/SoftwareIndex.jsp?lang=en&cc=us&pr odNameId=5219801&prodTypeId=12169&prodSeriesId=5219798&swLang=8&taskId=135& swEnvOID=54. For example, select HP SN1000E 16Gb 2-Port PCIe Fibre Channel Host Bus AdapterCross operating system (BIOS, Firmware, Diagnostics, etc.). 3. Power on the server and, at the prompt, select Ctrl-E. 4. Install the boot code firmware. 5. Reboot the host and, at the prompt, select Ctrl-E. 6. Specify the adapter port from which the system will be booting. 7. Scan the array and select the boot LUN. 8. Save the settings and reboot the host. 9. Insert the vSphere 5.1 CD into the CD drive. 10. Initiate the install. 11. Select the appropriate disk on the SAN when asked where to install the media.
Solution Implementers Series
32 Deploying 16Gb/s Fibre Channel SANs with HP ProLiant DL380 Gen8 servers and HP 3PAR Storage
For more information Storage I/O Performance on VMware vSphere 5.1 over 16 Gigabit Fibre Channel http://www.vmware.com/files/pdf/techpaper/V Mware-vSphere-16Gb-StorageIO-Perf.pdf Whats New in VMware vSphere 5.1 Performance http://www.vmware.com/resources/techresour ces/10309 Whats New in VMware vSphere 5.1 Storage http://www.vmware.com/resources/techresour ces/10308 HP virtualization with VMware, including a section on ProLiant servers www.hp.com/go/vmware HP storage solutions for VMware http://h71028.www7.hp.com/enterprise/w1/en/ solutions/storage-vmware.html 3PAR Utility Storage with VMware vSphere http://www.vmware.com/files/pdf/techpaper/vm w-vsphere-3par-utility-storage.pdf
To help us improve our documents, please provide feedback at implementerslab@emulex.com.
Copyright 2012 Emulex Corporation. The information contained herein is subject to change without notice. The only warranties for Emulex products and services are set forth in the express warranty statements accompanying such products and services. Emulex shall not be liable for technical or editorial errors or omissions contained herein. OneCommand is a registered trademarks of Emulex Corporation. HP is a registered trademark in the U.S. and other countries. VMware is a registered trademark of VMware Corporation.
World Headquarters 3333 Susan Street, Costa Mesa, California 92626 +1 714 662 5600 Bangalore, India +91 80 40156789 | Beijing, China +86 10 68499547 Dublin, Ireland+35 3 (0)1 652 1700 | Munich, Germany +49 (0) 89 97007 177 Paris, France +33 (0) 158 580 022 | Tokyo, Japan +81 3 5322 1348 Wokingham, United Kingdom +44 (0) 118 977 2929