You are on page 1of 2

DELL AND EMC VPLEX

VPLEX is EMC's SAN virtualisation manager. It is deployed as a cluster of engines, though it is possible
to have a single engine cluster. The architecture is in-band which means that all IO flows through the
VPLEX. The cluster itself consists of between one and four pairs of directors, and any director can fail
over to any other director in the cluster if any component fails. The back end storage is open, that is, it
can be EMC and non-EMC.
The VPLEX engine is the core building block of a VPLEX cluster that can be configured as single, dual or
quad engine in a single rack. Engines contain IO directors and a Linux kernel that runs proprietary cluster
software. The VPLEX components use the same standard building blocks that EMC uses in its Symmetrix
products. The two IO directors in the cluster run on a virtualised x86 processing unit and have 16 or 32
Fibre Channel ports, split between front end and back end.
A storage volume is simply a physical storage LUN that exists in one of the 'back end' storage controllers.
The VPLEX carves this volume up into extents, then combines these extents with extents from other
storage volumes to create the virtual volumes that are presented to the servers.
So, externally, the 'front end' of the VPLEX looks like a storage controller, as far as the connecting
servers are concerned. The 'back end' is connected to actual storage controllers and from them it looks
like a host.
A V-Plex can migrate data between different devices or between different extents, without affecting any
running user applications.
Write caching works differently depending on the VPLEX type. With vplex local and metro, while writes
are cached in the VPLEX, a write will not be marked as complete until it is successfully written to both
back end storage arrays. Because VPLEX Geo can potentially have elongated response times from the
remote cluster, a write is acknowledged completed when it is cached.
Several types of VPLEX configuration exist.
1. Vplex local is a single VPLEX cluster within one data center. It provides a high availability active-
active storage configuration and lets you move and share data nondisruptively, within a data
center.
2. Vplex metro is two VPLEX clusters between two data centers. More detail is provided below.
3. Vplex Geo is two VPLEX asynchronous clusters between two sites, with an Round Trip Latency
(RTL) of up to 50ms.
4. VPLEX for All-Flash is an all-inclusive solution with a software license for any capacity on one or
multiple Dell EMC all-flash arrays.
For configuration details see the VPLEX configuration page

VPLEX METRO
The VPLEX Metro system contain two clusters, with each cluster having one, two, or four engines,
although the two clusters need not have the same number of engines. The two clusters can be in the
same datacenter or in two datacenters, as long as they are not too far apart. The important thing is that
the RTL communication latency needs to be in the order of 5 -10ms.
If you run with both clusters in the same datacenter then you get a high availability, active-active local
solution that should survive the loss of one cluster. If you span between datacenters then you stretch that
high availability and your applications should survive the loss of one datacenter.
While the VPLEX data disks are spread inside a datacenter or even between datacenters, you want your
hosts to think they are accessing data from a traditional disk. This is achieved by the VPLEX cache
memory, where the memory in each director is combined to form a distributed VPLEX cache. Every
director within a cluster can service an I/O request for any virtual volume served by that cluster. Every
director within a cluster is connected to the same set of physical storage volumes from the back-end
arrays and has the same virtual-to-physical storage mapping metadata for its volumes. However the
cache algoriths are designed so that each director accesses local disks wherever possible, so minimizing
inter-director messaging. The global cache also ensures that a director always accesses the most recent
consistent data for given volume.
VPLEX Metro uses the write-through cache mode, which means that an application must wait until data is
written to both cache and disk storage before it is considered complete. This has a performance impact
but is much safer than accepting the writing to cache as completion then writing the data to disk later, and
it guarantees that both local and remote copies of data are consistent.
One of the issues facing a active-active cluster design is what to do if one side of the cluster fails, or if the
comms links fail. VLPEX Storage disks can be aggregated together into consistency groups, so that the
I/O activity will be consistent within that group, even after a failure. The cluster sides are marked as
'preferred' and 'non-preferred', so if the non-preferred site goes down, then processing will continue at the
preferred site. However if it's not possible to find out which site has gone down, usually because the links
have failed, then processing should stop at both sites. If it does not stop, then both sites will continue to
update local data independently and data corruption could well happen.
VPLEX Metro resolves this problem with VPLEX WITNESS, a third server running on a separate VM,
which is connected to each cluster side independently. This VM needs to be hosted on a different
hypervisor to any cluster volume, and configured in such a way that it would not be affected by any cluster
failure. We still designate one cluster as the preferred cluster for data availability and now if the comms
links between the VPLEX clusters fail, but the connections to the VPLEX Witness are retained, the
VPLEX Witness will indicate to the clusters that the preferred cluster should continue providing service to
the volumes in the consistency group. However the non-preferred cluster will stop servicing the volumes
to ensure data consistency and this will continue until the link issue is resolved and the data is correctly
mirrored again and consistent in both sites. So now we have three failure scenarios and three safe
actions to resolve them, provided the volumes are in a consistency group.
1. The non-preferred cluster fails, the Witness tells the preferred cluster to continue to handle the
I/O
2. The preferred cluster fails, the Witness tells the non-preferred cluster to continue to handle the
I/O
3. The links fail, the Witness tells the preferred cluster to continue to handle the I/O and the non-
preferred cluster to suspend I/O
VPLEX supports Microsoft Hyper-V server virtualization deployment as well as Oracle RAC cluster
technologies.
Storage M&R for VPLEX is included with VPLEX systems, providing in depth views of all VPLEX
components as well as trending data. It is easy to visualize and analyze the utilization, capacity, health
and performance of the VPLEX system.

VPLEX Virtual Edition


VPLEX Virtual Edition is implemented on a VMware ESXi infrastructure as a VMware Vapp. It extends the
distance that VMware can operate in, to provide failover and resilience between data centers. Tools like
VMware Vmotion can be used to move VMs transparently between data centers.
back to top

You might also like