You are on page 1of 35

VBLOCK™SOLUTION FOR GREENPLUM

August 2011

© 2011 VCE Company, LLC. All rights reserved.

1

Table of Contents
Introduction .................................................................................................................................................4 Goal .......................................................................................................................................................4 Audience ................................................................................................................................................4 Scope.....................................................................................................................................................4 Objectives ..............................................................................................................................................4 Greenplum Architecture Overview ............................................................................................................5 Setup .............................................................................................................................................................6 Installation..............................................................................................................................................6 Vblock Series 700 model MX Building Block Configuration Specifications for Greenplum ..................6 Greenplum Design and Configuration Details ..........................................................................................7 Compute – Unified Computing System (UCS) ....................................................................................... 10 UCS Server Blade Provisioning ......................................................................................................... 10 Service Profile Template .................................................................................................................... 10 UCS Firmware .................................................................................................................................... 14 UCS Network Configuration ............................................................................................................... 14 Greenplum Segment Server/ESX Server Provisioning ......................................................................... 16 Symmetrix VMAX Storage Array ............................................................................................................. 19 The Symmetrix VMAX Architecture Overview .................................................................................... 19 Hardware List ..................................................................................................................................... 20 Disk Layout ......................................................................................................................................... 20 Front-end Storage Port Layout ........................................................................................................... 21 RecoverPoint............................................................................................................................................. 22 RecoverPoint VSAN Zoning ............................................................................................................... 23 RecoverPoint Consistency Groups .................................................................................................... 26 TimeFinder/Snap....................................................................................................................................... 28 Replication Manager ........................................................................................................................... 28

© 2011 VCE Company, LLC. All rights reserved.

2

Test Results .............................................................................................................................................. 30 Read and Load Performance Tests: Objectives and Results ............................................................ 30 RecoverPoint Disaster Recovery Tests: Objectives and Results ....................................................... 31 Conclusion ................................................................................................................................................ 34 References ................................................................................................................................................ 35

© 2011 VCE Company, LLC. All rights reserved.

3

Vblock 700 delivers high-performance. This approach is scalable and supports a dynamic workload in a cost effective model. All rights reserved. operating. managers. LLC. Audience The target audience for this document includes technical engineering staff. and optimized data life cycle management requirements for hosting Greenplum along with other applications. Specifically. The architecture provides a building block approach for hosting Greenplum applications. EMC Symmetrix VMAX. IT planners. and others involved in evaluating. and VMware vSphere 4 and can include flash technology to meet the high-performance demands of mission-critical applications. 4 . The Vblock 700 enables enterprises to meet their mobility. for a total of forty-eight Greenplum Segment Server VMs o Monitor the I/O Workload of the VMs © 2011 VCE Company. security. disaster recovery. managing. administrators. largescale virtualization across data centers of large enterprise customers. or designing Greenplum Vblock platform deployments.Introduction Goal The purpose of this document is to provide architecture for hosting the Greenplum application on Vblock™ Infrastructure Platforms. the Greenplum application is hosted on the Vblock Series 700 model MX shared infrastructure and is part of a multi-applications environment. Scope The project demonstrates the ability to:  Run the Greenplum application on the Vblock platform  Prove that Greenplum Data Warehousing (DW) is a viable solution for use on a Vblock platform Objectives The business objectives of the new Vblock Solution for Greeplum include advantages in the following areas:  Provide a proven performance platform for a Greenplum and Vblock 700 architecture  Establish a building-block scalable model with predictable performance growth  Provide a showcase environment for a Greenplum and Vblock 700  Greenplum Workload Tests using massively parallel processing (MPP) o MPP/Row and MPP/Columnar Load queries o MPP/Row Sequential Read and Random Read queries o MPP/Columnar Sequential Read and Random Read queries o Mixed Workloads  Automated Virtualization Functionality o Scale out to a new ESX server infrastructure with six VMs per ESX host. It includes Cisco Unified Computing system.

query optimizations. Greenplum Database also includes features designed to optimize PostgreSQL for business intelligence (BI) workloads. resource management. 5 . and in most cases is very similar to PostgreSQL with respect to SQL support. MPP (also known as a shared nothing architecture) refers to systems with two or more processors which cooperate to carry out an operation. and is able to use all of a system’s resources in parallel to process a query. each processor with its own memory. query executor. Figure 1.2.com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-011-538. query planner. LLC.pdf © 2011 VCE Company. MPP Shared-nothing Architecture For further Greenplum information.Greenplum Architecture Overview Greenplum Database is a massively parallel processing (MPP) database server based on PostgreSQL opensource technology. The Greenplum interconnect (the networking layer) enables communication between the distinct PostgreSQL instances and allows the system to behave as one logical database. Greenplum Database is essentially several PostgreSQL database instances acting together as one cohesive database management system (DBMS). features. Database users interact with Greenplum Database as they would a regular PostgreSQL DBMS. For example. see the following: http://powerlink. and transaction manager components have been modified and enhanced to be able to execute queries in parallel across all of the PostgreSQL database instances at once. and storage enhancements which are not found in regular PostgreSQL. Greenplum has added parallel data loading (external tables). operating system. Many features and optimizations developed by Greenplum do make their way back into the PostgreSQL community. It is based on PostgreSQL 8. the system catalog. Greenplum leverages this high-performance system architecture to distribute the load of multi-terabyte data warehouses.emc.9. optimizer. The internals of PostgreSQL have been modified or supplemented to support the parallel structure of Greenplum Database. now in standard PostgreSQL. and disks. For example. All rights reserved. and end-user functionality. configuration options.

LLC. © 2011 VCE Company. More throughputs can be achieved by adding additional pairs of VMAX engines and disks. 6 .com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-011-541.Setup This section addresses the Greenplum setup. It will support 4 GB/sec scan rate.0 U2 servers  vCenter connects to another Windows 2008 R2 Enterprise Edition Server running SQLServer 2005 Enterprise Edition. All rights reserved.emc. 16 front-end processors and 192 x Fibre Channel (FC) 15K RPM disks). Table 1 700MX Compute Building block configuration specifications for Greenplum B200 M2 Blades  8 x B200 M2 Blades w/96GB memory per Blade (8 Blades for Greenplum Segment Servers) across two or more chassis  2 Engines  16 x 8Gb/sec FC ports  192 x 300GB FC drives (excluding hot spares)  RAID5 (3+1) 700MX Storage Symmetrix VMAX storage 700MX Virtualization  VMware vSphere ESX 4. See below for additional details. Installation The following link points to the Greenplum installation documentation. http://powerlink. or by adding more building blocks. setup per VMware’s installation guide  VMware Distributed Virtual Switch Greenplum application on Vblock platform Greenplum v4.006 software Greenplum Utilities  GP PerfMon  Psql  GP-Load  Greenplum Perl Greenplum Connectivity  GP/port 5432  GP PerfMon /port 8888 Scan Rate for the given building block  4 GB /sec This scan rate is achieved by using 2 x VMAX engines.pdf Vblock Series 700 model MX Building Block Configuration Specifications for Greenplum The 700MX building block configuration for Greenplum comprises forty-eight Greenplum Segment Server VMs running on eight Blades and two storage engines. More throughputs can be achieved by adding an additional pair of Vmax engines and disks.

illustrate the conceptual design through both the physical topology and the logical topology. 130 are in director A. The SAN directors are fabric A and fabric B. Figures 2 and 3. 7 . which in turn are connected to the Nexus 7000 switches as the aggregation layer. depicts the connectivity from the UCS to the SAN layer and LAN layer. VSANs 130 and 131 are the corresponding front-end VSANs consisting of server HBA ports. before figure 3. In this case. In the SAN layer. All rights reserved. The front-end and backend VSANs are required by Cisco’s SANTap to function as the write-splitter for RecoverPoint Appliances. below. a total of 16 x 8 Gb Fiber Channel (FC) connections were utilized from the USC Fabric Interconnects (A and B) to the MDS SAN directors.Greenplum Design and Configuration Details This section presents design and configuration details. figure 2. a total of 16 x 10 Gb Ethernet port connections are used between the UCS Fabric Interconnects (A and B) and the Nexus 5020 access layer LAN switches. VSANs 30 and 31 are backend VSANs consisting of the VMax storage ports and the RecoverPoint Appliance HBA ports. © 2011 VCE Company. and VSANs 31 and 131 are in director B. The logical topology is explained below. In the LAN layer. the UCS blade servers are used as ESX servers. VSANs 30. LLC. The physical topology.

LLC.Figure 2. Conceptual Design – Physical Topology Diagram © 2011 VCE Company. All rights reserved. 8 .

The logical topology depicts the configuration of the backend disk. below. one active and one standby node. 3 x VMs per ESX server. All rights reserved. 2 x VMs are used as the master servers. RecoverPoint was configured in local replication mode called Continuous Data Protection (CDP). A total of 26 x VMs were created on 10 x ESX servers (the UCS blade servers).See Figure 3. VMs. Conceptual Design – Logical Topology Diagram © 2011 VCE Company. 9 . and Greenplum components. 24 x VMs are used as the data segment servers. The 24 VMs are evenly distributed across 8 x ESX servers. ESX host LUNs. Figure 3. LLC. The master servers (active and standby node) are on the ninth and tenth ESX blade servers.

create one additional NIC in each fabric for the Greenplum segment network. In UCS terminology. 2. For Greenplum. 10 . The template allows users to deploy one or more service profiles as a Service Profile Template. Each vNIC has the following VLAN configuration: VLAN ID vNIC 0 (vmnic0) Service Console/Management vMotion Network Greenplum Segment VLan-1 * Public Network vLAN** A vNIC 1 (vmnic1) Service Console/Management vMotion Network Greenplum Segment VLAN – 2 * Public Network vLAN** B Fabric ID *Greenplum segment VLANs are private VLANs and completely isolated from each other. vNIC Template: This template is used for vNIC configuration for all Service Profiles deployed using a Service Profile Template. In this Greenplum on Vblock 700 configuration. LLC. Service Profile Template A Service Profile Template has the following components: 1. All rights reserved. a server is described as a Service Profile.Compute – Unified Computing System (UCS) UCS Server Blade Provisioning The Cisco Unified Computing System (UCS) allows provisioning of Physical Servers using a Template. a total of eight Service Profiles were used. If the UCS is equipped with Palo Cards. The following Service Profile Template details are used for deploying Greenplum on Vblock 700. **The Public Network VLAN accesses the Master Segment in the Greenplum setup. disable the failover between vNICs. © 2011 VCE Company.

11 . which were named Boot from SAN (BFS). The FC Adapter policy for VMware was used for all the vHBAs. VSAN ID Fabric ID vHBA-1 201 A vHBA-2 202 B 4. For this setup. All rights reserved. a total of eight fiber connections from each fabric Interconnect to the MDS Switches were used. This way. each Service Profile has a dedicated SAN Connection.3. Boot From SAN Policies (BFS) Policy 1: Policy 2: Policy 3: © 2011 VCE Company. Boot Policy: Four Boot policies were created. Eight SAN Pin Groups also were created. vHBA Template: This template is used for a vHBA configuration for all service profiles deployed using a service profile template. LLC. Each Boot policy points to front-end directors of the EMC Symmetrix VMAX storage array.

LLC. MAC pools for each fabric: In this setup. © 2011 VCE Company. The UCS blade allows users to modify the last three octets of the MAC Address pools. one for each fabric. we have two MAC Pools. All rights reserved.Policy 4: 5. 12 .

Service Profiles gpesx101-105. and gpesx201-205.6. The UCS blade allows users to modify the last three octets of the worldwide port and node names. were used to host the Greenplum (segment server) virtual machines. one for each fabric. two WWPN Pools were used. Service Profile Template and Service Profiles: The results achieved at the end of the whole process are shown in the following figure. LLC. 13 . © 2011 VCE Company. Service profiles gpesx107 and gpesx207 formed the management cluster for the Greenplum environment. All rights reserved. The Universal Unique Identifier (UUID) pool for each Service Profile: 7. World Wide Node and Port Name Pool: In this setup.

which has Palo card (M81KR) firmware. 14 . To apply the update to all service profiles: 1. go to Policies. Additionally. In the Service Profile template. 2. All rights reserved. Select the firmware package as shown below: UCS Network Configuration 1. The following shows a firmware package. The following screen captures show the port channel (LAN) configurations on the fabric interconnect. 2. Port-Channels used on the UCS © 2011 VCE Company. LAN Configuration Ports 1/37-40 were selected as uplink ports for 10GB ethernet traffic. it is necessary to attach the firmware package with each service profile. configure the LAN and SAN. LLC. It is also necessary to update the BIOS to the latest version using a firmware package.3 (1c). to enable the Palo interface cards. 1. This can be done in the service profile template. and BIOS updates. was used for this deployment. which then propagates the firmware to each service profile bound to this template. Under networking .UCS Firmware The latest firmware available at the time. Create uplink ports for etherenet traffic.

15 . SAN Pin groups are created to isolate SAN traffic.Port-channel Configuration details for fabric Interconnect A Port-channel Configuration details for fabric Interconnect B SAN Configuration on Fabric Interconnects There are a total of eight fiber channel ports in use on the fabric interconnect. © 2011 VCE Company. All rights reserved. The following screen capture shows the SAN Pin Group configurations on the fabric interconnect. LLC.

All rights reserved. It is used as the Greenplum Standby Server. This server also handles client requests. Table 2 VM/ESX storage configuration details © 2011 VCE Company. The table below provides the VM/ESX storage configuration details.  1 x VM is created on a 10th ESX server. 16 . LLC. These are used as Greenplum segment servers. It is used as the Greenplum Master/metadata Server. Note: XFS-formatted devices on VMware ESX Guest RDMs are recommended for GP data segments for the highest performance.Greenplum Segment Server/ESX Server Provisioning VM servers are provisioned as follows:  3 x VMs are created per ESX server on a total of 8 x ESX servers.  1 x VM is created on a 9th ESX server.

In this setup. All rights reserved. LLC. 17 . © 2011 VCE Company.The following diagram depicts the graphical layout of a single blade. VM/ESX/LUN Layout The following setup shows the ESX Servers in the vCenter server. with an ESX instance and a VM (Greenplum Data Segment server) with two LUNs (FC) per VM. Figure 4. which holds the three VM instances. The ESX environment is located on the 500GB LUN. there are six (only three VMs are used) Greenplum Segment servers (gpssx) on a single ESX server.

SQL Server Database and EMC Control Center. These include the vCenter server. All rights reserved.gp. 18 .vce Other Management Virtual Machines are hosted on the following ESX server.The following shows the Greenplum Segment Servers (gpssx) Virtual Machines distribution on single ESX servers. The VM distribution on the remaining seven ESX servers is identical to gpesx101. LLC. © 2011 VCE Company.

additional VMAX Engines can be added non-disruptively for efficient and dynamic scaling of capacity and performance. messaging. and management operations including internal discovery and initialization. Each director has two connections to the VMAX Matrix Interface Board Enclosure (MIBE) via the System Interface Board (SIB) ports. This design eliminates the need for separate interconnects for data. All rights reserved. and supports all Global Memory references. To meet growth requirements. this is a highly available interconnect with no single point of failure. The Symmetrix VMAX array comprises from one to eight VMAX Engines. control. © 2011 VCE Company. which reduces complexity. Each VMAX Engine contains two integrated directors. Symmetrix VMAX engines connect to and scale-out linearly through the Virtual Matrix Architecture. 19 . The Virtual Matrix is redundant and dual active. load balancing. Figure 5. and system test. memory. while dramatically simplifying and automating operational tasks are critical to addressing the infrastructure requirements and driving down cost in both virtual and physical deployments. messaging. which allows resources to be shared within and across VMAX Engines. environmental. The following figure illustrates the building block approach. path management. and connectivity on two directors for fault tolerance. Symmetrix VMAX Virtual Matrix Interconnect The Symmetrix VMAX design is based on an individual Symmetrix VMAX engine with redundant CPU.Symmetrix VMAX Storage Array The Symmetrix VMAX Architecture Overview At the heart of the Symmetrix VMAX series storage array architecture is the scalable Virtual Matrix interconnect design. failover. and fault isolation within the array. Since every director has two separate physical paths to every other director via the Virtual Matrix. LLC. A single highly-available interconnect suffices for all communications between the directors.

All rights reserved.  Each metavolume is allocated to each of the total 24 Greenplum segment VMs as an RDM disk © 2011 VCE Company. This is to achieve IO isolation on disk level. 20 . Symmetrix VMAX Engine Building Blocks Hardware List  Number of VMAX Engines: 2  Global Memory (GB): 256  Number of Front End (8 Gbps Ports): 32 ports  Number of 300G (15K RPM) FC disks (excluding hot spares): 192 Disk Layout  48 x RAID5(3+1) RAID groups are created out of the total 192 x Fibre Channel (FC) disks o 24 x RAID groups are used as RecoverPoint CDP source o 24 x RAID groups are used as RecoverPoint CDP target  Two 190 GB hyper volumes are created from each RAID group  One concatenated metavolume is created from the above hyper-volumes from each RAID group.Symmetrix VMAX Engine Building Block Front End Back End Easily Add More Symmetrix VMAX Engines Virtual Servers Front End Back End Host & Disk Ports Host & Disk Ports Symmetrix VMAX Engine Core Core Core Core Core Core Core Core Core Core Core Core Core Core Core Core Global Memory CPU Complex B CPU Complex A VMAX Engine Global Memory VMAX Engine VMAX Engine VMAX Engine VMAX Engine VMAX Engine VMAX Engine VMAX Engine Virtual Matrix Interface Virtual Matrix Interface A B Figure 6. LLC.

2 x FC ports on each  2 x FC ports on each processor with total 32 x 8 Gbs FC ports. Each ESX server HBA initiator accesses storage via 2 x Vmax storage ports. 21 .128G Port Dir 10 0 1 FA 0 1 FA 0 1 FA 0 1 FA Processor # H G F E Dir 7 FA FA FA FA Port Port 0 0 0 0 Port 1 1 1 1 Dir 9 FA FA FA FA Port Port 0 0 0 0 Port 1 1 1 1 Figure 8. see the product documentation: http://powerlink.Figure 7.pdf?mtcs=ZXZlbnRUeXBlPUttQ2xpY2tDb250ZW50RXZlbnQsZG9jdW1lbnRJZD0wOTAxNDA2NjgwNTIyMz FkLG5hdmVOb2RlPVNvZndhcmVEb3dubG9hZHMtMg__ © 2011 VCE Company.128G Port Dir 8 0 1 FA 0 1 FA 0 1 FA 0 1 FA Engine 5 .com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-008603. 8 x for the ESX hosts running Greenplum segment VMs. See figure below.emc. All rights reserved. Engine 4 . Greenplum Front-end Storage Port layout VMax Mask View (LUN masking): A total of 10 x Mask Views are configured. For more information about Symmetrix VMAX. 1 x for the ESX host running Greenplum master VM server. 1 x for the ESX host running standby VM server. Port 0 is taken from each processor.  16 FC ports (from each processor) are utilized In this configuration. LLC. only 16 x FC ports are used. Greenplum Backend Disk Layout on VMAX Front-end Storage Port Layout  2 x Engines  4 x Directors  16 x front-end processors. or total of 4 x Vmax storage ports per ESX server (with dual HBAs).

RecoverPoint continuous remote replication (CRR) provides remote asynchronous replication between two sites for LUNs that reside in one or more arrays. Table 3 Splitter Type Splitter details How Deployed Overhead Adds write traffic at the HBA. the Cisco SANTap service is used in a RecoverPoint CDP deployment. block-level replication product for a heterogeneous server and storage environment. All rights reserved. The bi-directional replication and any-point-in-time recovery capability can be enabled simultaneously with RecoverPoint concurrent local and remote (CLR) data protection. below. RecoverPoint continuous data protection (CDP) provides local synchronous replication between LUNs that reside in one or more arrays at the same site. See figure 9.RecoverPoint RecoverPoint is EMC’s leading out-of-band. © 2011 VCE Company. active in both storage processors CLARiiON-based No impact In the Greenplum configuration. Both the GP segment server data and the GP master metadata replicate locally on the VMAX for continuous data protection. no impact Host-based In I/O stack just above the multi-path software Fabric-based In intelligent storage services hardware on a Brocade. Both RecoverPoint CDP and RecoverPoint CRR feature bi-directional replication and an any-point-in-time recovery capability. which allows the target LUNs to be rolled back to a previous point in time and used for read/write operations without affecting the ongoing replication or data protection. LLC. 22 . Cisco 18/5 MSMs (Multi Service Modules) are installed in MDS 9513.or Cisco-based switch In FLARE operating system. no other impact Operates at wire speeds. RecoverPoint supports three types of write-splitting technologies for maximal flexibility.

a zone that contains a member of the ESX server HBA virtual initiators and the corresponding physical storage ports. The I/O is then copied to the RPA. This allows the RPAs to access the storage on the Vmax.  Zone Type 4 .a zone that contains RPA HBA ports and CVTs.1. RecoverPoint Sequence RecoverPoint VSAN Zoning RecoverPoint with Cisco SANTap deployment requires placing different components into two VSANs:  Front-end VSAN  Backend VSAN All I/O activity between the host and the storage is relayed by SANTap from the actual host port via the DVT (Data Virtual Target) in the front-end VSAN to the VI (Virtual Initiator) in the backend VSAN. created by the RecoverPoint Appliance):  Zone Type 1 . LLC. These zones were almost identical to the zones in the front-end VSAN that contain the host HBA port and DVTs. the RecoverPoint Appliance HBA ports. Zones in the backend VSAN The backend VSAN contains the physical storage ports. © 2011 VCE Company. Writes are acknowledged back from the RecoverPoint appliance 4. and AVTs (Appliance Virtual Targets. Write-order-consistent data is distributed to the replica volumes Figure 9.a zone that contains a member of the RPA HBA port and the physical storage ports.  Zone Type 3 .a zone that contains RPA HBA ports and the AVTs. The appliance writes data to the journal volume. All rights reserved. the CVTs (control virtual targets created by SANTap service). This relay mechanism is completely transparent to the hosts.  Zone Type 2 . along with time stamp and application-specific bookmarks /A /B / C rA rB rC Journal volume Production volumes Replica volumes 5. and then to the actual storage port. See explanation below. allowing the RPA to replicate data to the target. This allows the RPA to request the CVT to open a splitting session. The following types of zones are required for each VSAN. Data is split and sent to the RecoverPoint appliance in one of three ways Intelligentfabric splitter 3. 23 .

Note: DVTs are created as the virtual storage port entity during SANTap configuration. Zones in the front-end VSAN A zone that is between the host HBA ports. allowing it to appear as the host. and the SANTap Data Virtual Targets (the DVTs). All rights reserved. VSAN 30 (BE VSAN) Zone gpesx101_HBA1_Vmax gpesx102_HBA1_Vmax gpesx103_HBA1_Vmax gpesx104_HBA1_Vmax gpesx107_HBA1_Vmax gpesx201_HBA1_Vmax gpesx202_HBA1_Vmax gpesx203_HBA1_Vmax gpesx204_HBA1_Vmax gpesx207_HBA1_Vmax RPA1_HBA1_2_Vmax Member gpesx101_HBA1 gpesx102_HBA1 gpesx103_HBA1 gpesx104_HBA1 gpesx107_HBA1 gpesx201_HBA1 gpesx202_HBA1 gpesx203_HBA1 gpesx204_HBA1 gpesx207_HBA1 RPA1_HBA1 RPA1_HBA2 Member Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA Member Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA … RPA8_HBA1_2_Vmax … RPA8_HBA1 RPA8_HBA2 … Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA All SANTap CVTs in Fabric A All RPA AVTs … Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA RPA_CVT_A RPA_AVT_A All above RPA HBA ports All above RPA HBA ports © 2011 VCE Company. This masking is necessary to allow the RPA to overcome SCSI reservation of storage ports by the hosts and to get the same view of the SAN that the hosts have. 24 . Table 4 RecoverPoint VSAN zoning table: Fabric A Fabric A. see the following zoning tables. in this case the UCS blade server HBA ports. Each physical storage port used in the backend VSAN would need a corresponding DVT created. For more information. LLC.AVT are used to mask the identity of the appliance (RPA).

Table 5 RecoverPoint VSAN zoning table: Fabric B Fabric B. VSAN 31 (BE VSAN) Zone gpesx101_HBA2_Vmax gpesx102_HBA2_Vmax gpesx103_HBA2_Vmax gpesx104_HBA2_Vmax gpesx107_HBA2_Vmax gpesx201_HBA2_Vmax gpesx202_HBA2_Vmax gpesx203_HBA2_Vmax gpesx204_HBA2_Vmax gpesx207_HBA2_Vmax RPA1_HBA3_4_Vmax Member gpesx101_HBA2 gpesx102_HBA2 gpesx103_HBA2 gpesx104_HBA2 gpesx107_HBA2 gpesx201_HBA2 gpesx202_HBA2 gpesx203_HBA2 gpesx204_HBA2 gpesx207_HBA2 RPA1_HBA3 RPA1_HBA4 Member Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA Member Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA … RPA8_HBA3_4_Vmax … RPA8_HBA3 RPA8_HBA4 … Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA All SANTap CVTs in Fabric B All RPA AVTs … Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA RPA_CVT_B RPA_AVT_B Table 6 All above RPA HBA ports All above RPA HBA ports RecoverPoint VSAN zoning table: Fabric A’ Fabric A'. 25 . All rights reserved. LLC. VSAN 130 (FE VSAN) Zone gpesx101_HBA1_Vmax gpesx102_HBA1_Vmax gpesx103_HBA1_Vmax gpesx104_HBA1_Vmax gpesx107_HBA1_Vmax gpesx201_HBA1_Vmax gpesx202_HBA1_Vmax gpesx203_HBA1_Vmax gpesx204_HBA1_Vmax gpesx207_HBA1_Vmax Member gpesx101_HBA1 gpesx102_HBA1 gpesx103_HBA1 gpesx104_HBA1 gpesx107_HBA1 gpesx201_HBA1 gpesx202_HBA1 gpesx203_HBA1 gpesx204_HBA1 gpesx207_HBA1 Member (DVT) Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA Vmax_8eA Vmax_8fA Vmax_8gA Vmax_8hA Vmax_8gA Member (DVT) Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA Vmax_10eA Vmax_10fA Vmax_10gA Vmax_10hA Vmax_10gA © 2011 VCE Company.

Table 7 RecoverPoint VSAN zoning table: Fabric B’ Fabric B'. Data consistency is maintained at the Group Set level. A Group Set was created to contain all 8 x CGs to provide replication consistency for the entire Greenplum environment. All rights reserved. point-in-time recovery of the Greenplum environment. Each Consistency Group contains one or many replication sets. © 2011 VCE Company. LLC. Each replication set is a paring between the replication source LUN and target LUN. Below is the Consistency Group configuration table. VSAN 131 (FE VSAN) Zone gpesx101_HBA2_Vmax gpesx102_HBA2_Vmax gpesx103_HBA2_Vmax gpesx104_HBA2_Vmax gpesx107_HBA2_Vmax gpesx201_HBA2_Vmax gpesx202_HBA2_Vmax gpesx203_HBA2_Vmax gpesx204_HBA2_Vmax gpesx207_HBA2_Vmax Member gpesx101_HBA2 gpesx102_HBA2 gpesx103_HBA2 gpesx104_HBA2 gpesx107_HBA2 gpesx201_HBA2 gpesx202_HBA2 gpesx203_HBA2 gpesx204_HBA2 gpesx207_HBA2 Member (DVT) Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA Vmax_9eA Vmax_9fA Vmax_9gA Vmax_9hA Vmax_9gA Member (DVT) Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA Vmax_7eA Vmax_7fA Vmax_7gA Vmax_7hA Vmax_7gA RecoverPoint Consistency Groups RecoverPoint replicates data by using logical groups called Consistency Groups (CGs). a total of 8 x CGs were created with each CG containing 3 x replication sets. Since each Consistency Group can be active on a particular RPA. This allows rapid. 26 . in order to utilize all 8 x RPAs for optimal performance.

27 . see the RecoverPoint documentation set located at: http://powerlink. LLC.com/km/appmanager/km/secureDesktop?_nfpb=true&_pageLabel=freeformlinks2&internalId =0b014066800f517e&_irrt=true&rnavid=PT-2%3A0b0140668037fed7 © 2011 VCE Company. All rights reserved.emc.Table 8 RecoverPoint consistency group configuration table For more information about RecoverPoint.

TimeFinder/Snap provides the following:  Storage-based information replication. and: o Reduces backup windows o Minimizes/eliminates impact on the application o Improves Recovery Point Objectives (RPO) and Recovery Time Objectives(RTO) © 2011 VCE Company. LLC. point-in-time replicas  Auto-discovers the environment  Has intelligence to orchestrate replicas with deep application awareness  Is easy to use with point-and-click controls. Key benefits are that it:  Automates the creation. wizards. TimeFinder/Snap Replication Manager Replication Manager (RM) is EMC’s software that improves access to information by automating and managing disk-based replicas. Replication Manager is used to manage the TimeFinder Snap operations. All rights reserved. TimeFinder helps companies perform backups. no-host cycles  Snapshots create logical point-in-time images of a source volume  Requires only a fraction of the source volume’s capacity (~20–30%)  Multiple snapshots can be created from a source volume and are available immediately  Snapshots support both read and write processing In the Greenplum Vblock platform. load data warehouses. 28 . and use of EMC disk-based. Leveraging the industry leading high-end EMC Symmetrix system. management. SATA disks are configured into a Snap pool for the snaps. and user access  Supports VMware ESX Server Windows and Linux guest operating system environments and Virtual Machine File System (VMFS) containing virtual machines. and easily provide data for application test and development—without downtime. Production view Production volume Save area Snapshot view Cache-based pointer map Figure 10.TimeFinder/Snap TimeFinder provides local storage replication for increased application availability and faster data recovery. TimeFinder offers unmatched deployment flexibility and massive scalability to meet any service level requirement.

All rights reserved. 29 .o Enhances productivity o Offers data-warehouse refreshes o Provides decision support o Provides database-recovery ―checkpoints‖ o Enables application development and testing o Enables fast restore o Enables application restart and business resumption © 2011 VCE Company. LLC.

Run a stored procedure with two variations: One sequential run with no other workload on the system.6 minutes for 15 million records Run #2 result: 11. Test Greenplum load performance on a Vblock 700. Database size: eight TB Test data load rate on Vblock 700 with a specific building block as previously described. 1.1 minutes for 15 million records    Online = 1100 queries completed Production = 50 jobs completed Ad hoc = 15 jobs completed 1. Table 10 Test Scenario Test 1: Read test summary and results Description Read scan query and response times. All rights reserved. Table 9 Read and load performance test objectives Description Test read scan rate and query response time on a Vblock 700 with a specific building block as previously described. Test Result Description Run #1 result: 9. © 2011 VCE Company. Test Greenplum read performance on a Vblock 700 with the specified building block.Test Results This section presents the objectives and results for two different. LLC. 30 . 2. One sequential run while additional workload was being run on the system. 2. 2. One sequential run while additional workload is being run on the system. One sequential run with no other workload on the system. complementary test sets:  Tests of the read and load performance with Greenplum database on the Vblock 700  Tests of disaster recovery success with the RecoverPoint appliance Read and Load Performance Tests: Objectives and Results The following tests were performed with the Greenplum DB on a Vblock 700. Database size: eight TB Test Objective 1.

Load an empty monthly partition with one day of data. Load a half full monthly partition with one day of data. Verify that Snapshot consolidation works correctly. Perform a PIT copy on production. 2. LLC. 2. 4. Switch over the production DB to the secondary side. Test Result Empty Partition Load = 11. 2. All rights reserved.32 minutes Full Partition Load = 11. 3.33 minutes Half Full Partition Load = 11. The following test scenarios to validate the DR testing solutions are within the BI/DW solution stack proposed by EMC. 3. 1. RecoverPoint Disaster Recovery Tests: Objectives and Results The RecoverPoint (RP) appliance replication capability is leveraged to perform site Disaster Recovery (DR) testing. Objectives for DR Site recovery tests using RecoverPoint replication Description Perform a local bookmark function test to verify that users have access to the database. Load a full monthly partition with one day of data. Verify that the Point in Time (PIT) image was the correct image. Enable Snapshot consolidation. Verify the Point in Time (PIT) bookmark for entire dataset. Test 2: Load test summary and results Description One day of data equated to roughly 15 million records. 31 . The objectives and results for four tests are summarized in the following tables. 3.17 minutes Description Met performance metrics. © 2011 VCE Company. Table 12 Test Objective 1.Table 11 Test Scenario 1. Switch over to target side.

DBAs insert 35 million records on production A PIT image is copied. Verify PIT. Table 15 Test Scenario Test 3: Snapshot consolidation summary and results Description DBA must verify the image on Snapshot consolidation. Test Result 1. LLC. 7. while enabling the image on secondary hosts.Table 13 Test Scenario Test 1: Bookmark test summary and results Description Verify that the database is open and accessible. Test Result 1. Copy a PIT image before inserting records to production. DBA restores the Snapshot consolidation bookmark. 3. Take the database out of gp suspend mode. All rights reserved. Table 14 Test Scenario Test 2: PIT image test summary and results Description Verify that the correct PIT image is copied before inserting 35 million records into production. © 2011 VCE Company. DBAs are able to get correct image before the insert. Bookmark an image during Snapshot consolidation. Enable the image on secondary hosts. 32 . Description Database was open and DBA’s were able to access the database. 4. Follow a sequence of steps to place the database in and out of suspend mode. 6. 5. Create RP BM. Start database. 2. 2. Place database in gp suspend mode. Test Result 1. Description DBA successfully completed Snapshot-consolidated image. Description DBAs successfully copied correct image on production. Mount BM image. 3.

© 2011 VCE Company. Enable the latest point in time image and switch over the production DB to the target DB. Test Result 1. 33 . Determine the latest point in time image and switch over from the production database to the target DB. Description Successfully completed. LLC.Table 16 Test Scenario Test 4: Switch production DB summary and results Description Target DB has to be primary. All rights reserved.

or run with additional workload on the system where the results achieved were 11. a half full partition load with results at 11.1 minutes to read 15 million records.6 minutes to read 15 million records. The key results from the RecoverPoint tests show the successful recovery and restoration of the database image and validate the disaster recovery solution included in the system. © 2011 VCE Company. or a full partition load with results at 11.Conclusion Our testing supports the benefits of the building block system approach used for hosting Greenplum applications. 34 .32 minutes. All rights reserved. Key results from the read and load performance tests illustrate the scalability of Greenplum on a Vblock 700 solution:  The Read Test results show that the scan rate and query response time on Vblock 700 have a similar read performance time whether a stored procedure was run alone where results showed 9.33 minutes.17 minutes.  The Load Test results showed a similar load performance time whether the load was an empty partition load with results at 11. LLC.

go to www.emc. © 2011 VCE Company. All other trademarks used herein are the property of their respective owners. through the Vblock platform.com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-011-541. All rights reserved.emc. VCE. and/or its affiliates in the United States or other countries.com/collateral/software/white-papers/h2781-emc-recoverpoint-3-new-features. 35 . All rights reserved.vce. Vblock and the VCE logo are registered trademarks or trademarks of VCE Company. vertical industry offerings.com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-011-538.pdf Introduction to EMC RecoverPoint 3.com/km/live1/en_US/Offering_Technical/Technical_Documentation/300-008603.emc.com/km/appmanager/km/secureDesktop?_nfpb=true&_pageLabel=freeformlinks2&internalId =0b014066800f517e&_irrt=true&rnavid=PT-2%3A0b0140668037fed7 Symmetrix VMAX product guide http://powerlink.References For further Greenplum information. and application development environments. LLC. For more information. accelerates the adoption of converged infrastructure and cloud-based computing models that dramatically reduce the cost of IT while improving time to market for our customers.pdf Greenplum installation DOC link http://powerlink. validating and managing IT infrastructure.emc.pdf?mtcs=ZXZlbnRUeXBlPUttQ2xpY2tDb250ZW50RXZlbnQsZG9jdW1lbnRJZD0wOTAxNDA2NjgwNTIyMz FkLG5hdmVOb2RlPVNvZndhcmVEb3dubG9hZHMtMg ABOUT VCE VCE. the Virtual Computing Environment Company formed by Cisco and EMC with investments from VMware and Intel. VCE prepackaged solutions are available through an extensive partner network. LLC. see the following: http://powerlink.emc. THE INFORMATION IN THIS PUBLICATION IS PROVIDED "AS IS. delivers the industry's first completely integrated IT offering with end-to-end vendor accountability." VCE MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION. Copyright © 2011 VCE Company. AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OR MECHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. and cover horizontal applications. allowing customers to focus on business innovation instead of integrating.pdf RecoverPoint Guides – full set http://powerlink. LLC.3 New Features and Functions Applied Technology white paper http://www.com.